WO2004100561A1 - Photographing apparatus, device and method for obtaining images to be used for creating a three-dimensional model - Google Patents

Photographing apparatus, device and method for obtaining images to be used for creating a three-dimensional model Download PDF

Info

Publication number
WO2004100561A1
WO2004100561A1 PCT/GB2004/001902 GB2004001902W WO2004100561A1 WO 2004100561 A1 WO2004100561 A1 WO 2004100561A1 GB 2004001902 W GB2004001902 W GB 2004001902W WO 2004100561 A1 WO2004100561 A1 WO 2004100561A1
Authority
WO
WIPO (PCT)
Prior art keywords
photographing
mode
transparent table
lighting
camera
Prior art date
Application number
PCT/GB2004/001902
Other languages
French (fr)
Inventor
Koichi Matsumura
Adam Michael Baumberg
Alexander Ralph Lyons
Kenichi Nagasawa
Takashi Saito
Original Assignee
Canon Europa, Nv
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Europa, Nv filed Critical Canon Europa, Nv
Priority to GB0522704A priority Critical patent/GB2416973A/en
Publication of WO2004100561A1 publication Critical patent/WO2004100561A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/246Calibration of cameras
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C11/00Photogrammetry or videogrammetry, e.g. stereogrammetry; Photographic surveying
    • G01C11/02Picture taking arrangements specially adapted for photogrammetry or photographic surveying, e.g. controlling overlapping of pictures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/243Image signal generators using stereoscopic image cameras using three or more 2D image sensors

Definitions

  • This invention relates to a photographic apparatus, device and method for taking images to be used for creating three-dimensional model. Further, this invention typically relates to those taking images for textures and silhouettes of such images.
  • Japanese-Laid Open Patent 2001-148021 introduces a technology described as follows. This technology is to prepare displays as the background for the object and the table for setting the object on, and to manually vary the displayed colour and brightness depending on the colour and brightness of the object. Thus, the object is easily extracted from the photographed image, which includes the background, the table and the object.
  • Inverting the object would also cause another adjustment between the image of the bottom and other pre-photographed images to be necessary, since the focal length for photographing the bottom is normally slightly different from the one for photographing other images.
  • a user would have to adjust size of these images by using computer programs increasing the workload of the user. Accordingly, though such technologies are very useful and can contribute industrial progress, users may well be restricted to ones who have some photographic skill. In this respect, these technologies have something to be improved.
  • This invention has been made to improve three-dimensional model creating technology.
  • One of the objects of this invention is to provide an apparatus for photographing images to create full three-dimensional models shown by all orientation by taking silhouettes and textures, (hereinafter “silhouette images” and “texture images”) without manual user operations or further workload.
  • this invention provides, in one aspect, an apparatus for creating a three-dimensional object model, comprising photographing means for photographing an object to be modeled, setting means for setting a relative position between the object and said photographing means, said setting means being capable of setting a plurality of different relative positions, said setting means including a transparent table on which the object is to be set and rotating means for rotating the transparent table so that the object is set at a plurality of different longitudinal orientations to said photographing means, said setting means setting the relative positions so that said photographing means photographs the object from a plurality of latitudinal positions which has different latitudinal distances from the transparent table, the plurality of latitudinal positions including a position below the transparent table; lighting means including front lighting means capable of lighting a front side of the object confronting a photographing position of said photographing means and back lighting means capable of lighting a back side of the object hidden from the photographing position of said photographing means, and control means for controlling said photographing means, said lighting means and said setting
  • Another object of this invention is to provide an apparatus for photographing images to create a high quality three-dimensional image by taking silhouette images and texture images with ideal photographing conditions so that resultant quality of the three-dimensional image becomes higher.
  • this invention provides, in one aspect, an apparatus for creating three-dimensional object model, comprising photographing means for photographing an object to be modeled, said photographing means including exposure adjusting means for adjusting an exposing condition under which the object is photographed; setting means for setting a relative position between the object and said photographing means, said setting means being capable of setting a plurality of different relative positions; lighting means including front lighting means capable of lighting a front side of the object confronting a photographing position of said photographing means and back lighting means capable of lighting a back side of the object hidden from the photographing position of said photographing means; and control means for controlling said photographing means, said lighting means and said setting means so that said photographing means photographs the object a plurality of times at a plurality of relative positions set by said setting means for obtaining a plurality of silhouette images in
  • exposure adjusting means included in said photographing means changes the exposing condition between the first mode in which said photographing means photographs for obtaining the silhouette images and in the second mode in which said photographing means photographs for obtaining the texture images.
  • This invention also provides, in another aspect (usable with the other aspects), a method for creating a three-dimensional object model comprising the steps of setting an object to be modeled at a plurality of relative positions to a photographing position, the plurality of relative positions including a plurality of different longitudinal orientations and a plurality of latitudinal positions which has different latitudinal distances from a transparent table on which the object is to be placed, the plurality of latitudinal positions including a position below the transparent table; photographing the object at the plurality of relative positions by using both a front lighting for a front side of the object confronting the photographing position and a back lighting for a back side of the object hidden from the photographing position; and controlling so that the object is photographed a plurality of times at a plurality of relative positions set in said setting step for obtaining a plurality of silhouette images in a first mode and is photographed a plurality of times at a plurality of relative positions set in said setting step for obtaining a plurality of texture
  • Another one of objects of this invention is to provide an apparatus for prevent from being influenced by a transparent table on which the object is to be placed for photographing images, which are necessary for creating three-dimensional object model, which can be observed via a transparent table.
  • this invention introduces, as another aspect of this invention, an apparatus for creating a three-dimensional object model, comprising photographing means for photographing an object to be modeled, said photographing means photographing the object from a plurality of different positions, setting means for setting the object, said setting means including a transparent table, lighting means for illuminating the object through the transparent table, first polarizing control means located between said photographing means and the transparent table, and second polarizing control means located at a position in an optical path of lights illuminating a photographing area of said photographing means by directly reflecting on the transparent table, polarizing angles of said first and second polarizing control means have a predetermined difference.
  • Figure 1 shows a mechanical structure of a three-dimensional modeling system of one embodiment of the invention.
  • Figure 2 shows a cross-section of the three-dimensional modeling system according to this embodiment shown in Figure 1 as a side view.
  • Figure 3 shows a plan view of the system according to this embodiment shown in Figure 1 and Figure 2.
  • Figure 4 shows a calibration mat, which is used with the system shown in Figure 1, Figure 2 and Figure 3.
  • Figure 5 shows a block diagram representing the electrical system of the three- dimensional modeling system described using Figures 1-4.
  • Figure 6 Figure 7 and Figure 8 are flowcharts representing operations of the application program and the three-dimensional object model creating program carried out by the electrical system shown in Figure 5.
  • Figure 9 shows a table of control data stored in CPU shown in Figure 5 and used for positioning the circular glass shown in Figure 1.
  • Figure 10 shows a table of imaging parameters stored in storage within the application program.
  • Figure 11 shows a table of focal lengths of the digital cameras shown in Figure 1.
  • Figure 12 shows control data used for turning on and turning off the flashlights (backlights) and fluorescent lights (front lights) shown in Figure 1.
  • FIG. 13 shows another embodiment of the system according to this invention.
  • Figure 1 shows the mechanical structure of a three-dimensional modeling system of one embodiment of the invention.
  • a frame 1 supports the other elements of the system and is composed of a plurality of pole members, with base plates connecting lower ends of the pole members, and upper plates connecting upper ends of the pole members.
  • a circular glass table 2 is supported on the upper plates of the frame 1.
  • the table is rotated by stepping motors and pinch rollers consisting of rubber. These stepping motors and rollers are numbered 3, 4, and 5.
  • the reference 6 indicates a central axis of rotation of the table 2 and the table rotates around the axis 6.
  • the reference 7 indicates an intersection point of the rotation axis 6 and the table 2.
  • the point 7 is, in this embodiment, for obvious reasons, a central point of the circular table.
  • rotary encoder marks 8 for being detected so as to provide an angle of rotation of the table 2.
  • the encoder marks 8 are composed of evaporated aluminum thin films respectively extending in the radial directions of the table 2.
  • a photo reflector 9 located at a position confronting the thin films detects them and thus detects the angle of rotation of the table 2.
  • 11, 12, 13, 14 and 15 indicate guide rollers guiding the table 2 so as to rotate around the rotation axis 6.
  • 16 and 17 indicate spaces where three-dimensional objects are set. Space 16 is a space for a bigger object and space 17 is a space for a smaller object.
  • a digital camera 18 (digital_cam#0), is located on a parallel plane higher than the plane of the table 2 by a predetermined height A.
  • the reference 19 indicates a horizontal line indicating the plane on which the digital camera 18 is located.
  • the digital camera 18 is aimed at the central point 7 and an optical axis of the camera makes an angle of 10 degrees with the plane of the table 2.
  • Another digital camera 20 (digital_cam#1), is located on a parallel plane higher than the plane of the table 2 by a predetermined height B.
  • the reference 21 indicates a horizontal line indicating the plane on which the digital camera 20 is located.
  • the digital camera 20 is aimed at the central point of the object space 16 and the optical axis of the camera 20 makes an angle of 10 degrees with the plane of the table 2.
  • Another digital camera 22 (digital_cam#2), is aimed at the central point 7 of the table 2, and the optical axis thereof makes an angle of 45 degrees with the plane of the table 2.
  • Another digital camera 23 (digital_cam#3) is aimed at the central point 7 of the table 2, and the optical axis thereof makes at an angle of 80 degrees with the plane of the table 2.
  • Another digital camera 24 (digital_cam#4) below the plane of the table 2, is aimed at the central point 7 of the table 2, and the optical axis thereof makes an angle of 70 degrees with the plane of the table 2.
  • a flashlight 25 for giving backlight is set above the plane of the table 2 opposite the digital cameras 18 and 20.
  • another flashlight 26 gives backlight.
  • the flashlight 26 is set below the plane of the table 2 opposite the digital camera 22 set at an upward angle of 45 degrees towards an object in spaces 16 or 17.
  • a further flashlight 27 for giving backlight is set opposite the digital camera 23 at an upward angle of 80 degrees towards an object in spaces 16 or 17.
  • Another flashlight 28 for giving backlight is set opposite the digital camera 24 at a downward angle of 70 degrees towards an object in spaces 16 or 17.
  • White-light diffuser plates 29, 30, 31 and 32 are located in front of each of the flashlights 25, 26, 27, and 28.
  • the diffuser plate 29 located at a position illuminated by the flashlight 25 diffuses light received from the flashlight 25.
  • the diffuser plate 30 located at a position illuminated by the flashlight 26 diffuses light received from the flashlight 26
  • the diffuser plate 31 located at a position illuminated by the flashlight 27 diffuses light received from the flashlight 27,
  • the diffuser plate 32 located at a position illuminated by the flashlight 28 diffuses light received from the flashlight 28.
  • a polarizing panel 33 having an approximately same size as the diffuser plate 31 is located thereon.
  • Polarizing filter 34 is located in front of the lens of the digital camera 24.
  • the polarizing angle of the polarizing filter 34 is set at a roughly right angle with the polarizing panel 33.
  • a white-light diffuser composed of several plates, is located so as to surround the object 16 or 17.
  • a plurality of fluorescent lights 36 are provided as front light sources and are located behind the light diffuser plates 35 so as to illuminate object spaces 16 and 17 with diffuse light.
  • Figure 2 shows a cross-section of the three-dimensional modeling system according to this embodiment shown in Figure 1 as a side view.
  • Lines extending from each of the digital cameras 18, 20, 22, 23 and 24 represent each of photographing areas thereof in a vertical cross-section.
  • Two solid lines represent an outside border of the photographing area corresponding to a set predetermined focal length, and a central one-dotted-chained line shows a central axis of the photographing area.
  • Figure 3 shows a plan view of the system according to the embodiment shown in Figure 1 and Figure 2. Similar to Figure 2, two solid lines extended from the camera 20 to the table 2 show a photographing area thereof for the predetermined set focal length.
  • FIG. 4 shows a calibration mat 37, which is used with the system according to this embodiment.
  • Calibration dots 38 are prepared on a calibration mat 37 to enable the detection of the position of each digital camera, the orientation of each digital camera, and the focal length of camera lens of each digital camera.
  • there are 32 of the calibration dots 38 four (4) dots being located on each of eight (8) different radii dividing the mat into eight (8) equal angles.
  • These calibration dots may have different sizes, and preferably each set of four dots on a radius has a different pattern of dot sizes compared to the other sets.
  • the calibration mat 37 preferably has the same calibration dots located at the exactly same position on the back as the dots 38 located on the front of the mat.
  • the diameter of the calibration mat 37 is smaller than a diameter of the circular glass table 2 and the size of the mat 37 is designed so as not to interfere with any of stepping motors and rollers 3,4, and 5.
  • Figure 5 shows a block diagram representing the electrical system of the three- dimensional modeling system described using Figures 1-4 before as one embodiment of the invention.
  • a doted line shown with the reference 39 represents a part contained in an ordinary personal computer.
  • this personal computer may be composed of ordinary PC platforms conforming to the well-known PC/AT standard.
  • Central processing unit (CPU) 40 executes an application program. Normally, such an application program is stored in a ROM or a hard disk within the computer 39 as an object code. Then, such a program is read out from the storage and written into a memory within the CPU 40 for execution when the system launched. Since it does not relate to this invention directly, detailed descriptions of data flow, control flow and memory construction within the CPU 40 are omitted in this specification.
  • a video monitor 41 is connected to the computer 39.
  • a video signal to be displayed at the video monitor 41 is output from a video board 42 to which the monitor 41 is connected.
  • the video board 42 is driven by a set of software programs called a video driver 43.
  • a keyboard 44 is provided by which users of this system manually input data and they may also give instructions using a mouse 45. Such input data and instructions are interpreted in a keyboard and mouse driver 46 composed of a set of software programs.
  • All of the digital cameras 18, 19, 20, 21 and 22 are connected to the computer 39 by the well-known Universal Serial Bus (USB).
  • the item referenced 47 represents USB ports to which the digital cameras 18, 19, 20, 21 and 22 are physically connected and their HUB interfaces.
  • a USB device manager 48 manages the USB ports and HUB interfaces 47.
  • Also provided are software programs composing a USB driver 49 for controlling the digital cameras 18, 19, 20, 21 and 22.
  • An interface box 50 controls communications amongst a STM driver 51, a photo reflection detector 53, a lighting control unit 55 and the computer 39. All of elements 51 to 60 are located in the interface box 50 and are described hereafter.
  • the STM driver 51 drives and controls stepping motors 3, 4, and 5 for rotating the glass circular table 2 in accordance with outputs from a digital to analogue converter (DAC) 52 which converts digital data from the computer into an analogue signal to be used in the STM driver 51.
  • the photo reflection detector (PR) 53 detects an output of the photo reflector 9 indicating positions of the encoder marks 8 composed of evaporated aluminum thin films located a circumference of the table 2.
  • the analogue output of the photo reflection detector 53 is converted to digital data at an analogue to digital converter (ADC) 54.
  • the lighting control unit 55 has a register that controls flashlights 25, 26, 27 and 28, which are used as backlights for the object on the table.
  • This register is composed of a 5-bit hardware register, the bits of which control flashlights 25, 26, 27 and 28 and fluorescent lights 36 in accordance with control signals provided via a serial interface 61.
  • Such light-control signals are created in accordance with the application program and are communicated via the serial interface 61 under control of a communication serial port driver (COM port Driver) 62.
  • the control signals for controlling the flashlights are input to two-input AND gates 57, 58, 59 and 60 respectively.
  • a first port, labelled #0, #1 of the register is connected to one input of the AND gate 57.
  • a port #2 of the register is connected to one input of the AND gate 58, a port #3 of the register is connected to one input of the AND gate 59 and a port #4 of the register is connected to one input of the AND gate 60.
  • the register also has a port #FL for front lights and is therefore connected to the plurality of fluorescent lights 36.
  • a five-input OR gate 56 has its inputs respectively connected to X-triggers of digital cameras 18, 20, 22, 23 and 24. These X-triggers are provided for synchronizing flash with photographing and are well known as "Flash Synchronization" connection points.
  • the other input of each of the AND gates 57, 58, 59 and 60 is connected to the output of the OR gate 56.
  • a hard disc unit 63 stores data 64 of texture images and silhouette images.
  • a three-dimensional object model creating program is stored in a ROM or a hard disc within the computer 39 as an object code and is represented as stored at 65. The program is read out from the storage and written into a memory within the CPU 40 for execution when the system is launched.
  • the application program and the model creating program (an object modeling engine) communicate through the communication (COM) interface.
  • a program for displaying a graphical user interface (GUI) for the application program stored in CPU 40 is represented as stored at 66.
  • Figure 6, Figure 7 and Figure 8 are flowcharts of operations of the application program in CPU 40 and the three-dimensional object model creating program in storage 65, as carried out by the CPU 40.
  • Figure 9 shows a table of control data used for positioning the circular glass table 2, which is stored in CPU 40 within the application program.
  • This table shows that the system controls the rotation of the circular table 2 to locate it to each of fifteen (15) rotation positions, each different from a predetermined principal rotation position (#0) by a multiple of 22.5 degrees.
  • Such sixteen (16) positions, including the principal rotation position are determined by data set (rotation_set: #0-#15) and are designated by the application program.
  • Figure 10 shows a table of imaging parameters (exp_param_set) stored in storage within the application program.
  • an exposure value (AV) and a shutter speed value (TV) of the digital cameras 18, 20, 22, 23 and 24 are determined corresponding to a designated mode (#0: front_texture, or #1 : backlight).
  • Figure 11 shows a table of focal lengths of the digital cameras 18, 20, 22, 23 and 24. These parameters representing focal lengths are also stored in the storage 40 within the application program. Such actual focal lengths are determined in response to input data (zoom_pos_set), which represent focal position of zoom lens of the digital cameras.
  • Figure 12 shows control data used for turning on and turning off the flashlights 25, 26, 27 and 28 (backlights) and fluorescent lights 36 (front light). These control data are also stored in the storage 40 within the application program. Operations of the system shown in Figure 1 to Figure 5 are described referring to the flowcharts shown in Figure 6, Figure 7 and Figure 8 as follows. As the subject system in one embodiment of this invention starts to operate, the CPU 40 starts the application program and its operations from a step #101 shown in Figure 6. In the step #101 , all previous settings and data used for previous operations are retrieved and the values are reinitialised. The intialising process is shown in Figure 8. A step #201 is a starting step of the intialising process.
  • step #201 all texture image data and silhouette image data stored in the hard disc unit 63 are cleared.
  • step #202 the CPU 40 resets the USB HUB interfaces 47, the USB device manager 48 and the USB camera driver 49 and confirms communications between the digital cameras 18, 20, 22, 23, 24 and these USB elements 47, 48 and 49.
  • step #203 the CPU 40 initialises the interface box 50, the serial interface 61 and the serial port driver 62 and confirms communication between the interface box to and the interface 61.
  • step #204 of the initializing step #101 the circular glass table 2 is returned to the predetermined principal rotating position in accordance with the application program.
  • the CPU 40 instructs rotation of the table 2 so as to locate it at the principal rotation position. This instruction is transferred to the interface box 50 through the serial interface 61 and the serial port driver 62.
  • Digital data representing the principal rotating position (#0) is compared with digital data representing the actual rotation position of the table 2 and the CPU 40 calculates digital driving data for the stepping motors 3,4, and 5.
  • the digital driving data is converted to an analogue signal at the digital to analogue converter (DAC) 52 and supplied to the STM driver 51.
  • the STM driver 51 drives stepping motors 3, 4, 5 in accordance with the analogue signal.
  • the photo reflection detector (PR) 53 determines the actual rotation position of the table 2 by detecting an output of the photo reflector 9 reflecting the positions of the encoder marks 8.
  • the CPU 40 refers the table shown in Figure 9 to determine a targeted rotation position (angle) of the table 2 and calculates difference between the current actual rotating position of the table 2 and the targeted position. In accordance with the difference, the CPU 40 generates digital driving data necessary to drive the stepping motors 3, 4, and 5. Accordingly, the stepping motors 3, 4, and 5 are driven by the analogue driving signal output from the DAC 52 so that the table 2 can be located at the predetermined principal position. After the above initialising process, operations are returned to a step #102 shown in the Figure 6. In this step #102, a user of this system decides a size of a three-dimensional object to be photographed and modeled. In accordance with the application program and the program for displaying a graphical user interface
  • GUI size-selecting wi ni dow on the display of the video monitor 41 via the video board.
  • the user uti llises the keyboard 44 or the mouse 45 to select the size of the object by referri ni g the displayed page. Since the GUI and the selecting page itself are not important to describe this invention, detailed descriptions thereof are omitted.
  • the user can select the size of the object as “Small” or “Large".
  • the height of the object is larger than the height A but smaller than the height B which corresponds to the horizontal line 21 representing the plane on which the digital camera 20 is located, it is preferable for the user to select "Large”.
  • step #106 the fluorescent lights 36, the front lights, are turned on.
  • the CPU 40 instructs the writing of a flag "1" in the front light control bit of the register in the lighting control unit 55 of the interface box 50 though the serial interface 61 under control of the COM port Driver 62. Accordingly an output of the port of front lights #FL switches to a predetermined level representing "1" and, as shown in the table of Figure 12, the fluorescent lights 36 are turned on.
  • a step #107 the CPU 40 displays a window urging the user to put the calibration mat 37 on the circular glass table 2 via the GUI on the display of the video monitor 41 in accordance with the application program. Further, in a step #108, the CPU 40 displays a window for the user to confirm whether the calibration mat 37 has already been put on the table 2 or not. If the user confirms this, namely ready to calibrate, the process goes to a step #109.
  • a loop variable N is set as one of sixteen integers from "0" to "15”. Initially the variable N is set as "0".
  • a step #110 the CPU 40 determines a targeted rotating position (angle) of the table 2 by outputting one of the data sets (rotation_set: #0-#15).
  • the CPU 40 refers the table shown in Figure 9. Operations of the CPU 40, the stepping motors 3,4, and 5, photo reflection detector (PR) 53, and the DAC 52 to rotate the table 2 to the targeted position are similar to the operations described before for the step #101 (#204).
  • a variable, a software repeat counter C is set as "0", "1", "2", “3”, or "4". Initially the variable C is set as "0".
  • the imaging parameters shown in the table of the Figure 10 are set (in accordance with exp_param_set: #0) for front light, namely the exposure value (AV) is set to F8.0 and the shutter speed (TV) is set 1/15 second. These parameters are transferred to one of the digital cameras 18, 20, 22, 23 and 24. These parameters are transferred through the USB ports HUB interfaces 47 under control of the USB device manager 48 and the USB camera driver 49 to the digital camera (digital_cam_#C). Therefore, in this initial stage, the parameters are transferred to the digital camera 18 (digital_cam_#0).
  • a parameter representing a focal length is also transferred to the digital camera (digital_cam_#C) through the USB ports HUB interfaces 47.
  • the transferred parameter is a set #0 (zoom_pos_set: #0), which represents a wide-end focal length in accordance with the table of the Figure 11.
  • this parameter is also transferred to the digital camera 18 (digital_cam_#0).
  • a step #114 the CPU 40 sends a command to photograph to the digital camera (digital_cam_#C) and the digital camera takes an image.
  • Image data obtained by the digital camera (digital__cam_#C) is transferred to the hard disc unit 63 through the USB ports HUB interfaces 47 after compressing the image data in conformity with well-known JEPG compression scheme in a step #115.
  • the name of such image file is, for example, "img_cam #0. jpg".
  • the JEPG compressed image data obtained in the step #114 and stored in the step #115 is processed and developed and the CPU 40 detects the calibration dots 38 on a calibration mat 37 in the captured image in accordance with the application program and three-dimensional object model creating program 65 in a step #116.
  • the CPU 40 processes and analyses the detected calibration dots and determines a central position of the calibration mat 37 for creating supposed three-dimensional coordinates. In accordance with the supposed three-dimensional coordinates, a position and an orientation of the digital camera, and a focal length of the digital camera, the focal length of the camera can be obtained from the image of the calibration dots 38 by using perspective information.
  • the image of the calibration dots is obtained from the back of the calibration mat 37 and is processed and analysed for obtaining the position and the orientation of the digital camera 24, and the focal length of the digital camera 24.
  • the obtained camera information including the position, the orientation and the focal length of the digital camera (digital_cam_#C_#N) are stored in the hard disc unit 63 as a file named "cal_cam#C" in a step #118.
  • a step #119 the variable C designating the digital camera is incremented.
  • the variable C is changed from “0" to "1".
  • the steps #111 to #119 are repeated for each value of the variable C from “0" to "4".
  • the variable N is incremented and the process goes back to the step #109 from the step #120. (Lines returning from steps #119 to #111 and #120 to #100 are not shown in the Figure 7 and Figure 8.)
  • the CPU 40 determines the targeted rotating position (angle) of the table 2, the table 2 is rotated to the targeted position.
  • the variable N is incremented one by one and finally comes to "15".
  • step #118 the circular glass table has been completely rotated around the object and all camera calibration parameters with each camera position at every rotation angle have been stored by step #118.
  • each of the digital cameras 18(digital_cam#0), 20(digital_cam#1), 22(digital_cam#2), 23(digital_cam#3), and 24(digital_cam#4) have been selected in turn and each digital camera has photographed the image of the calibration dots 38 in order to obtain the camera information (the position, the orientation and the focal length) for each of the digital cameras.
  • These information have been stored in the hard disc unit 63 as files "cal_cam#0", “cal_cam#1", “cal_cam#2", “cal_cam#3", and "cal_cam#4".
  • a value of the variable N is checked in a step #120 to confirm whether the repeated processes in the steps #109 to #120 are completed for all of the digital cameras 18, 20, 22, 23, and 24 or not. If it is confirmed, the process goes to a next step #121.
  • the CPU 40 instructs to write a flag "0" in the front light control bit of the register in the lighting control unit 55. Accordingly an output of the port of front lights #FL switches to another predetermined level representing "0" and the fluorescent lights 36 are turned off in accordance with the table of the Figure 12. All processes for calibration are completed at the step #121 and the process returns to a step #122.
  • a step #122 the CPU 40 displays a window for asking the user if an actual modeling process shall be started, using the GUI, on the display of the video monitor 41. If the user instructs by using keyboard 44 or the mouse 45 to start an actual modeling, the process goes to a step #123.
  • the CPU 40 turns on the fluorescent lights 36, i.e. the front lights. This process is the same as the step #106.
  • the variable N is set as one of sixteen integers from "0" to "15”. Initially the variable N is set as "0" in a step #124.
  • a targeted rotation position (angle) of the table 2 by outputting one of the data sets (rotation_set: #0 ⁇ #15).
  • the CPU 40 refers to the table shown in Figure 9. Operations of the CPU 40, the stepping motors 3,4, and 5, photo reflection detector (PR) 53, and the DAC 52 to rotate the table 2 to the targeted position are similar to the operations described before for the step #101.
  • step #126 if the user selected the object size as "Small”, the process goes to a step #127. If the user selected the object size as "Large”, the process goes to a step #128.
  • numbers to be selectively used as another variable C are set as 0, 2, 3, and 4.
  • numbers to be selectively used as another variable C are set as 1, 2, 3, and 4. This means, the digital camera 18 (digital_cam_#0) is used if the object size is small, but the digital camera 19 (digital_cam_#1) is used if the object size is large. Other digital cameras 20.22, and 24 are commonly used irrespective of the size of the object.
  • this variable C is set either "0" or "1 in accordance with the size of the object.
  • Two steps #129 and #130 both check the vari a; ble C and the variable N. If the variable C is "2" and the variable N is not a multi p
  • the steps #131 to #139 (executing the photographing) are carried out only when the table 2 is at the principal rotation position or rotation positions different from the principal rotation position (#0) by each of multiple angles of 45 degrees. Accordingly, eight (8) images are taken by the camera 20 (digital_cam_#2). Similarly, if the camera 22 (digital_cam_#3) or the camera 24 (digital_cam_#4) is selected, the steps #131 to #139 (executing the photographing) are carried out only when the table 2 is at the principal rotation position or rotation positions different from the principal rotating position (#0) by each of multiple angles of 90 degrees.
  • the CPU 40 instructs to write a flag "1" in one of the backlight control bits of the register in the lighting control unit 55 of the interface box 50 though the serial interface 61 under control of the COM port Driver 62.
  • the corresponding bit of the register is switched to "1", which switches the output of port #0,#1 to a level representing "1".
  • the backlights may be turned on.
  • the imaging parameters are set (in accordance with exp_param_set: #1) for backlight, namely the exposure value (AV) is set to F8.0 and the shutter speed (TV) is set to 1/60 second. Then, these parameters are transferred to one of the digital cameras 18, 20, 22, 23 and 24. These parameters are transferred through the USB ports HUB interfaces 47 under control of the USB device manager 48 and the USB camera driver 49 to the digital camera (digital_cam_#C). Therefore, in this initial stage, the parameters are transferred to the digital camera 18 (digital_cam_#0) or the digital camera 19 (digital_cam_#1 ).
  • a parameter representing a focal length (zoom_pos_set: #X), is also transferred to the digital camera (digital_cam_#C) through the USB ports HUB interfaces 47.
  • the focal length "#X” is manually selected by the user among six focal lengths ("#0" to "#5") shown in the table of the Figure 11 before the photographing process. The user needs to select the focal length so that the whole of the object is in a photographic area of each digital camera.
  • this user's selection may be given by using the GUI.
  • the CPU 40 displays a focal length selecting page showing six focal lengths to be selected, using the GUI, on the display of the video monitor 41 via the video board 42.
  • the user utilises the keyboard 44 or the mouse 45 to select one of the focal lengths by referring the displayed page.
  • a step #134 the CPU 40 sends a command to photograph to the digital camera (digital_cam_#C) and the digital camera takes an image.
  • one of the X-triggers of digital cameras 18, 20, 22, 23 and 24 is closed and becomes a predetermined level representing "1" for a predetermined period of time.
  • the output of the OR gate 56 becomes a predetermined level representing "1” and one of the AND gates 56,58, 59 and 60 corresponding to the digital camera (digital_cam_#C) outputs a predetermined level representing "1".
  • one of the flashlights 25, 26, 27, and 28 is turned on and emits flashlight in synchronization with the X-trigger, namely a photographing operation of the digital camera (digita!__cam_#C).
  • the digital camera 18 digital_cam_#0
  • the digital camera 20 digital_cam_#1
  • a step #135 the CPU 40 instructs to write a flag "0" in one of the backlight control bits of the register in the lighting control unit 55. For example, for port #0,#1 the corresponding bit of the register is switched to "0", which switches the output of port #0,#1 to a level representing "0". Thus the backlight #0,#1 is disabled.
  • the image data obtained by the digital camera (digital_cam_#C) in the step #134 is transferred to the hard disc unit 63 after compressing in conformity with well-known JEPG compression scheme in a step #136.
  • the name of such image file is "sil_cam#C_#N. jpg”.
  • the name of the image file is "sil_cam#0_#0. jpg”.
  • the name of the file includes "sil” indicating a silhouette image, "#C” indicating the camera used for photographing, and "#N” indicating the rotation position of the table 2.
  • a step #137 the imaging parameters shown the Figure 10 are set (in accordance with exp_param_set: #0) for front light. Accordingly the exposure value (AV) is set as F8.0 and the shutter speed (TV) is set as 1/15 second. Similarly to the step #132, these parameters are transferred to the digital cameras (digital_cam_#C). In the initial stage, the parameters are transferred to the digital camera 18 (digital_cam_#0) or the digital camera 19 (digital_cam_#1). The digital camera typically uses the same focal length (zoom_pos_set: #X) for both obtaining the silhouette image and the texture image. Therefore, the same focal length (zoom_pos__set: #X) is used also in the capturing step #133. It shall be noted that the fluorescent lights 36, the front lights, have been continuously turned on since the step #123.
  • a step #138 the CPU 40 sends a command to photograph to the digital camera (digital_cam_#C) and the digital camera takes a texture image.
  • digital_cam_#C digital camera
  • the image data obtained by the digital camera is, in the step #138, transferred to the hard disc unit 63 after compression in conformity with the well-known JEPG compression scheme in a step #139.
  • the name of such image file is, for example, "ima_cam#C_#N. jpg".
  • the name of the image file is "ima_cam#0_#0. jpg”.
  • the name of the file includes "ima” indicating a texture image, "#C” indicating the camera used for photographing, and "#N” indicating the rotating position of the table 2.
  • the variable C is incremented in a step #140. If it was judged that the object is small, the process goes back to the step #127 from the step #139, (the line for that not being shown in Figure 7 and Figure 8). For instance, if the variable C was set as "0" so that the digital camera 18 (digital_cam_#0) is used during the first photographing process, the variable C becomes "2" in the step 127. On the other hand, if it was judged that the object is large, the process goes back to the step #128 from the step #140.
  • variable C For instance, if the variable C was set as "1" so that the digital camera 20 (digital_cam_#1 ) is used during the first photographing, the variable C becomes "2" in the step 128. Thus, whichever the object is small or large, the photographing process using the digital camera 22 (digital_cam_#2) is next executed in the steps #131 to #139. After photographing process using the digital camera 22 (digital_cam_#2) is completed, the variable becomes "3" and the digital camera 23 (digital_cam_#3) is selected to be used in the photographing process. Finally, the variable C comes “4" and the digital camera 24 (digital_cam_#4) is selected to be used in the photographing process.
  • step #141 all of these photographing processes are executed whilst the table 2 is located at the principal rotation position.
  • the variable N is incremented and the process goes back to the step #124 from the step #140, (the line for that not being shown in Figure 7 and Figure 8).
  • the CPU 40 determines the targeted rotation position (angle) of the table 2, the table 2 is rotated to the targeted position.
  • the variable N is incremented by units of one and finally becomes "15".
  • the circular glass table has been completely rotated around the object and all texture images and silhouette images from all angles have been stored in the hard disc unit 63.
  • the CPU 40 instructs to write a flag "0" in the front light control bit of the register in the lighting control unit 55. Accordingly an output of the port of front lights #FL turns to the predetermined level representing "0" and the fluorescent lights 36 are turned off in accordance with the table of the Figure 12.
  • a step #143 executing the three-dimensional object model creating program 65, the CPU 40 creates three-dimensional geometry data of the object by using all silhouette images stored in the hard disc unit 63.
  • the three-dimensional geometry is defined by polygons, including triangles and four-cornered polygons.
  • Detailed methods or processes for obtaining the three-dimensional geometry data by using silhouette images taken from different positions and orientations is disclosed in former patent applications, for example, USP 6317139, USP 4710876, and a UK application number 0114157.1 (CRE235), and these methods or processes can be adopted to this step #143. Therefore, in this specification, detailed descriptions of such methods and processes are omitted.
  • step #143 before creating three-dimensional geometry, parameters including photographing positions, orientations, and focal lengths for each of digital cameras and each photographing are calibrated by using the obtained camera information, including the position, the orientation and the focal length of the digital cameras stored in the hard disc unit 63 as files named "cal_cam#C" in the step #118.
  • the CPU 40 creates three-dimensional texture data to be put on surfaces of each polygons created in the step #143 by using all texture images stored in the hard disc unit 63.
  • a step #145 the resultant three-dimensional object having geometry on which texture images have been put is displayed on the display of the video monitor 41.
  • resultant three-dimensional model can be rotated, magnified or like by the user, using the keyboard 44 or the mouse 45.
  • a step #146 all information of such a resultant three-dimensional object including three-dimensional geometry information and three-dimensional texture information to be put on the geometry is stored in the hard disc unit 63 as an ordinal VRML file (*.wrl), for example.
  • the process is completed after the step #146.
  • the digital camera 24 located beneath the glass table 2 photographs the object
  • the digital camera 24 picks up the light reflected by the glass table 2.
  • This reflected light obviously deteriorates quality of the resultant three- dimensional object.
  • the existence of the white-light diffuser plate 31 causes a significant problem, namely that directly reflected light from the plate 31 illuminates a lower surface on the glass table within a photographing area of the digital camera 24.
  • This reflected light illuminating the lower surface in the main deteriorates the contrast of texture images of a bottom of the object.
  • the polarizing panel 33 which has almost the same size as the white-light diffuser plate 31 , was put thereon and the polarizing filter 34 was set in front of a camera lens of the digital camera 24, which has a photographing area illuminated by the directly reflected light from the white-light diffuser plate 31. Further the polarizing angle of the polarizing filter 34 is set at roughly a right angle with the polarizing panel 33. Accordingly the system of this embodiment can significantly reduce the inappropriate influence caused by the directly reflected light from the white-light diffuser plate 31 on the resultant three-dimensional object model.
  • the fluorescent lights 36 are used as front lights.
  • other types of lights such as tungsten lights can be used.
  • the white balance of the digital cameras has preferably, however, to be set so that colour temperature of the produced image is appropriate.
  • other types of lights such as tungsten lights or fluorescent lights, can be used as backlights instead of the flashlights.
  • the embodiment described above only introduced two kind of image parameters, as shown in Figure 10, for texture images using only front lights and silhouette images using a backlight.
  • such parameters like an exposure value (AV) and a shutter speed (TV) are preferably set more precisely in consideration with luminance value of each of the emitted lights, to effectively obtain high-quality texture images and silhouette images.
  • AV exposure value
  • TV shutter speed
  • one aspect of this embodiment and one advantage of this embodiment is to select the different exposure parameters between texture images and silhouette images. This provides significant improvement in obtaining high-quality texture and silhouette images.
  • the location of cameras and focal lengths of each of camera lens are considered and are decided in accordance with several factors, including the size of the object, distance between the object and backlight, and the size of the backlight, so that the outline of the object is surrounded by the backlight.
  • a plurality of digital cameras are prepared for each vertical (latitudinal) location.
  • FIG. 13 shows another embodiment of the system according to this invention.
  • This embodiment introduces fine nylon fibers roped off on a plane on which the object is put instead of the glass table.
  • the reference 67 shows a rotating table made of metal such as aluminum.
  • the table 67 has a hole where fine nylon fibers 69 are secured.
  • the material of the nylon fibers is selected so that fibers can be made sufficiently fine enough to be unnoticeable or invisible from the digital cameras.
  • the number of fibers is decided in accordance with size and weight of the object.
  • the size of the hole is decided so as to cause no obstacle to the creation of silhouette images taken by all digital cameras.
  • this embodiment shown in Figure 13 it becomes possible to float the object in air substantially without visible materials.
  • this embodiment shown in Figure 13 has an advantage in avoiding the inappropriate influence of the reflected lights without using other elements like a polarizing panel, a polarizing filter 34 and so on.
  • putting the object on a material, which is at least partially transparent, so that the object looks to be substantially floating in air and photographing the object from a plurality of angles including an angle below the table as explained in the embodiments makes it possible to take silhouette images from the bottom of the object and texture images of the bottom of the object without manually changing the direction or orientation of the object. This greatly contributes to the creation of a resultant three-dimensional object that is viewable from all orientations without relying on the user's manual operations.
  • a polarizing control panel is set at a position from which a direct reflected right enter into a digital camera
  • a polarizing control filter is set in front of the digital camera, and the polarizing angles of these are set so as to have a predetermined difference, preferably a right angle difference, it becomes possible to reduce influences of reflected lights on photographed image taken by the digital camera. Accordingly, the three-dimensional object model viewable from all orientations can be automatically obtained without large distortion.
  • an element on which an object to be modeled locates is composed of a glass table or fine transparent fibers secured on the plate.
  • a method or elements for making the object shown in all orientation shall not be limited into the glass table or fine transparent fibers. For example, it is possible to replace these with a transparent fiber hanging the object, a transparent needle-shaped resin piercing the object, or a transparent stool on which the object is set.
  • a plurality of digital cameras are prepared and set at different positions for photographing the object from a plurality of different orientations.
  • This structure may be replaced with a structure, which guides and moves one digital camera, revolving it around the object. If such a structure utilising a single camera is adopted, the camera is moved to locate it at one or more positions for every angle at which the object is rotated and photographs from each position.
  • the present invention can be applied to an apparatus photographing a plurality of silhouette and texture images from different angles having the same height.
  • a polarizing panel is set on a white-light diffuser plate to provide a backlight having the same shape of the white-light diffusible plate.
  • the polarizing panel or filter can be set at different positions within in the optical path of the backlight that illuminates a photographing area of the camera by directly reflecting on the glass table and can be set as a different shape of polarizing control material.
  • longitudinal relative positions namely rotating positions of the table, are commonly used for several latitudinal positions of a camera and photographs for silhouette and texture images.
  • these conditions are not essential for the present invention and, for example, all longitudinal positions can be different for each photographing at different latitudinal positions and can be different between silhouette and texture images.
  • shutter speed is changed.
  • both the exposure value and the shutter speed, or only the exposure value may be changed to obtain ideal exposing condition.
  • the shutter speed can be controlled by controlling a mechanical shutter or by changing charge- accumulating period of imaging device, such as CCD (Charge Coupled Device).
  • CCD Charge Coupled Device
  • automatic photographing is carried out by rotating a table and setting the table at a plurality of predetermined angles and is described as a fully automatic photographing.
  • the photographing conditions for each texture and silhouette images is optimized and quality of resultant texture images and silhouette images is highly improved. Further, photographing the object as shown from all orientations including below the object and from various latitudinal positions, the resultant three-dimensional model is highly improved.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Computer Graphics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Input (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Image Generation (AREA)

Abstract

This invention introduces, as one aspect, a method for creating a three-dimensional object model, wherein an object is set to be modeled at a plurality of relative positions to a photographing position, the plurality of relative positions including a plurality of different longitudinal orientations and a plurality of latitudinal positions which has different latitudinal distances from a transparent table on which the object is to be placed, the plurality of latitudinal positions including a position below the transparent table, the object is photographed at the plurality of relative positions by using both a front lighting for a front side of the object confronting the photographing position and a back lighting for a back side of the object hidden from the photographing position, and control is carried out so that the object is photographed a plurality of times at a plurality of relative positions for obtaining a plurality of silhouette images in a first mode and is photographed a plurality of times at a plurality of relative positions for obtaining a plurality of texture images under a different lighting condition from the first mode in a second mode, the silhouette images and the texture images are to be used for creating the three-dimensional object model.

Description

Photographing Apparatus, Device and Method For Obtaining Images To Be Used For Creating A Three-Dimensional Model
FIELD OF THE INVENTION
This invention relates to a photographic apparatus, device and method for taking images to be used for creating three-dimensional model. Further, this invention typically relates to those taking images for textures and silhouettes of such images.
BACKGROUND OF THE INVENTION
Recently, technologies creating a three-dimensional model of an object from a plurality of images taken from a plurality of positions and/or orientations have been developed. In detail, such technologies generally generate silhouettes from photographed images, create the geometry of the three-dimensional model by using the silhouettes, generate textures from photographed images, and set the generated textures on each polygon of the geometry. Further, technologies to display such three-dimensional models in an internet browser and so on and to make necessary rotations so as to make it possible to observe the models have also been developed. Using these technologies to make three-dimensional models observable through the internet browser, it becomes possible for electronic commerce (E-Commerce) customers to observe merchandise as three-dimensional objects. Thus, it is expected that such three- dimensional object modeling technologies will greatly contribute to the advancement of E-Commerce businesses.
However, to create the three-dimensional model, it is ordinarily necessary to prepare a studio for photographing the objects to be three-dimensionally modeled and make various arrangements for photography. In particular, users have to set lighting conditions and appropriate backgrounds for each of the photographs from various positions and orientations, so as to be able to take textures and silhouettes effectively.
For example, when taking silhouettes by using the choroma-key technique, it is necessary to prepare a single coloured background of a colour that is not significantly used in an object to be photographed. To accomplish this, the
Japanese-Laid Open Patent 2001-148021 introduces a technology described as follows. This technology is to prepare displays as the background for the object and the table for setting the object on, and to manually vary the displayed colour and brightness depending on the colour and brightness of the object. Thus, the object is easily extracted from the photographed image, which includes the background, the table and the object.
Further, another technology is introduced by Japanese-Laid Open Patent H10- 148021. This technology firstly photographs the background of the object and the table (an "object-setting surface") without the object and secondly photographs the image including all of the background, the object-setting surface and the object. After these photographs, a difference between images is determined and a silhouette is obtained.
However, when silhouette information is obtained by using these techniques, a user has to change manually a background and an object-setting surface or has to set and remove the object manually. Therefore, it is impossible to complete all photographs without a manual operation by the user. Accordingly, these techniques cannot make it possible to create a three-dimensional model fully automatically nor to make user operation easy or simple. By not only photographing lateral images but also top and bottom images of the object, it would be possible to create a three-dimensional model observable from all orientations. However, to do so, it would be necessary for a user to invert the object, since the bottom of the object sitting on the object-setting surface is invisible. Inverting the object would also cause another adjustment between the image of the bottom and other pre-photographed images to be necessary, since the focal length for photographing the bottom is normally slightly different from the one for photographing other images. A user would have to adjust size of these images by using computer programs increasing the workload of the user. Accordingly, though such technologies are very useful and can contribute industrial progress, users may well be restricted to ones who have some photographic skill. In this respect, these technologies have something to be improved.
SUMMARY OF THE INVENTION
This invention has been made to improve three-dimensional model creating technology.
One of the objects of this invention is to provide an apparatus for photographing images to create full three-dimensional models shown by all orientation by taking silhouettes and textures, (hereinafter "silhouette images" and "texture images") without manual user operations or further workload.
To accomplish the above object, this invention provides, in one aspect, an apparatus for creating a three-dimensional object model, comprising photographing means for photographing an object to be modeled, setting means for setting a relative position between the object and said photographing means, said setting means being capable of setting a plurality of different relative positions, said setting means including a transparent table on which the object is to be set and rotating means for rotating the transparent table so that the object is set at a plurality of different longitudinal orientations to said photographing means, said setting means setting the relative positions so that said photographing means photographs the object from a plurality of latitudinal positions which has different latitudinal distances from the transparent table, the plurality of latitudinal positions including a position below the transparent table; lighting means including front lighting means capable of lighting a front side of the object confronting a photographing position of said photographing means and back lighting means capable of lighting a back side of the object hidden from the photographing position of said photographing means, and control means for controlling said photographing means, said lighting means and said setting means so that said photographing means photographs the object a plurality of times at a plurality of relative positions set by said setting means for obtaining a plurality of silhouette images in a first mode and photographs the object a plurality of times at a plurality of relative positions set by said setting means for obtaining a plurality of texture images in a second mode, and said lighting means changes the lighting condition between the first and second modes, the silhouette images and the texture images being to be used for creating the three- dimensional object model. By using the above-mentioned apparatus, it becomes possible to create high quality silhouette images and texture images for creating the three-dimensional object model including those photographed from beneath the transparent table without a number of manual user operations. This apparatus also can reduce post processing of such silhouette images and texture images because of their quality.
Another object of this invention is to provide an apparatus for photographing images to create a high quality three-dimensional image by taking silhouette images and texture images with ideal photographing conditions so that resultant quality of the three-dimensional image becomes higher. To accomplish the above object, this invention provides, in one aspect, an apparatus for creating three-dimensional object model, comprising photographing means for photographing an object to be modeled, said photographing means including exposure adjusting means for adjusting an exposing condition under which the object is photographed; setting means for setting a relative position between the object and said photographing means, said setting means being capable of setting a plurality of different relative positions; lighting means including front lighting means capable of lighting a front side of the object confronting a photographing position of said photographing means and back lighting means capable of lighting a back side of the object hidden from the photographing position of said photographing means; and control means for controlling said photographing means, said lighting means and said setting means so that said photographing means photographs the object a plurality of times at a plurality of relative positions set by said setting means for obtaining a plurality of silhouette images in a first mode and photographs the object a plurality of times at a plurality of relative positions set by said setting means for obtaining a plurality of texture images in a second mode, the silhouette images and the texture images being to be used for creating the three-dimensional object model; and wherein said control means controls said exposure adjusting means to change the exposing condition between the first and second modes and controls said lighting means so that said lighting means changes the lighting condition between the first and second modes.
By using the above apparatus, exposure adjusting means included in said photographing means changes the exposing condition between the first mode in which said photographing means photographs for obtaining the silhouette images and in the second mode in which said photographing means photographs for obtaining the texture images. This makes it possible to photograph the object to be modeled in suitable lighting conditions for both obtaining silhouette images and texture images and, therefore, a resultant three-dimensional object model to have better quality.
This invention also provides, in another aspect (usable with the other aspects), a method for creating a three-dimensional object model comprising the steps of setting an object to be modeled at a plurality of relative positions to a photographing position, the plurality of relative positions including a plurality of different longitudinal orientations and a plurality of latitudinal positions which has different latitudinal distances from a transparent table on which the object is to be placed, the plurality of latitudinal positions including a position below the transparent table; photographing the object at the plurality of relative positions by using both a front lighting for a front side of the object confronting the photographing position and a back lighting for a back side of the object hidden from the photographing position; and controlling so that the object is photographed a plurality of times at a plurality of relative positions set in said setting step for obtaining a plurality of silhouette images in a first mode and is photographed a plurality of times at a plurality of relative positions set in said setting step for obtaining a plurality of texture images under a different lighting condition from the first mode in a second mode, the silhouette images and the texture images being to be used for creating the three-dimensional object model. Another one of objects of this invention is to provide an apparatus for prevent from being influenced by a transparent table on which the object is to be placed for photographing images, which are necessary for creating three-dimensional object model, which can be observed via a transparent table. To accomplish the above objects, this invention introduces, as another aspect of this invention, an apparatus for creating a three-dimensional object model, comprising photographing means for photographing an object to be modeled, said photographing means photographing the object from a plurality of different positions, setting means for setting the object, said setting means including a transparent table, lighting means for illuminating the object through the transparent table, first polarizing control means located between said photographing means and the transparent table, and second polarizing control means located at a position in an optical path of lights illuminating a photographing area of said photographing means by directly reflecting on the transparent table, polarizing angles of said first and second polarizing control means have a predetermined difference.
According to the apparatus above, it becomes possible to obtain good quality images, which are necessary for creating three-dimensional object model visible from any orientation, without increasing a number of user's manual operations. Other features or aspects would be clarified by following detailed embodiments with reference with of drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
There will now be described, by way of example only, embodiments of the invention, with reference to the accompanying drawings, of which: Figure 1 shows a mechanical structure of a three-dimensional modeling system of one embodiment of the invention.
Figure 2 shows a cross-section of the three-dimensional modeling system according to this embodiment shown in Figure 1 as a side view. Figure 3 shows a plan view of the system according to this embodiment shown in Figure 1 and Figure 2.
Figure 4 shows a calibration mat, which is used with the system shown in Figure 1, Figure 2 and Figure 3. Figure 5 shows a block diagram representing the electrical system of the three- dimensional modeling system described using Figures 1-4.
Figure 6, Figure 7 and Figure 8 are flowcharts representing operations of the application program and the three-dimensional object model creating program carried out by the electrical system shown in Figure 5. Figure 9 shows a table of control data stored in CPU shown in Figure 5 and used for positioning the circular glass shown in Figure 1.
Figure 10 shows a table of imaging parameters stored in storage within the application program.
Figure 11 shows a table of focal lengths of the digital cameras shown in Figure 1. Figure 12 shows control data used for turning on and turning off the flashlights (backlights) and fluorescent lights (front lights) shown in Figure 1.
Figure 13 shows another embodiment of the system according to this invention.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
Figure 1 shows the mechanical structure of a three-dimensional modeling system of one embodiment of the invention. In the apparatus of Figure 1, a frame 1 supports the other elements of the system and is composed of a plurality of pole members, with base plates connecting lower ends of the pole members, and upper plates connecting upper ends of the pole members. A circular glass table 2 is supported on the upper plates of the frame 1. The table is rotated by stepping motors and pinch rollers consisting of rubber. These stepping motors and rollers are numbered 3, 4, and 5. The reference 6 indicates a central axis of rotation of the table 2 and the table rotates around the axis 6. The reference 7 indicates an intersection point of the rotation axis 6 and the table 2. The point 7 is, in this embodiment, for obvious reasons, a central point of the circular table.
Around a circumference of the table 2 are provided rotary encoder marks 8 for being detected so as to provide an angle of rotation of the table 2. The encoder marks 8 are composed of evaporated aluminum thin films respectively extending in the radial directions of the table 2. A photo reflector 9 located at a position confronting the thin films detects them and thus detects the angle of rotation of the table 2. 11, 12, 13, 14 and 15 indicate guide rollers guiding the table 2 so as to rotate around the rotation axis 6. 16 and 17 indicate spaces where three-dimensional objects are set. Space 16 is a space for a bigger object and space 17 is a space for a smaller object.
A digital camera 18 (digital_cam#0), is located on a parallel plane higher than the plane of the table 2 by a predetermined height A. The reference 19 indicates a horizontal line indicating the plane on which the digital camera 18 is located. The digital camera 18 is aimed at the central point 7 and an optical axis of the camera makes an angle of 10 degrees with the plane of the table 2.
Another digital camera 20 (digital_cam#1), is located on a parallel plane higher than the plane of the table 2 by a predetermined height B. The reference 21 indicates a horizontal line indicating the plane on which the digital camera 20 is located. The digital camera 20 is aimed at the central point of the object space 16 and the optical axis of the camera 20 makes an angle of 10 degrees with the plane of the table 2.
Another digital camera 22 (digital_cam#2), is aimed at the central point 7 of the table 2, and the optical axis thereof makes an angle of 45 degrees with the plane of the table 2. Another digital camera 23 (digital_cam#3) is aimed at the central point 7 of the table 2, and the optical axis thereof makes at an angle of 80 degrees with the plane of the table 2. Another digital camera 24 (digital_cam#4) below the plane of the table 2, is aimed at the central point 7 of the table 2, and the optical axis thereof makes an angle of 70 degrees with the plane of the table 2.
A flashlight 25 (flashlight#0,1) for giving backlight is set above the plane of the table 2 opposite the digital cameras 18 and 20. Similarly, another flashlight 26 (flashlight#2) gives backlight. The flashlight 26 is set below the plane of the table 2 opposite the digital camera 22 set at an upward angle of 45 degrees towards an object in spaces 16 or 17. A further flashlight 27 (flashlight#3) for giving backlight is set opposite the digital camera 23 at an upward angle of 80 degrees towards an object in spaces 16 or 17. Another flashlight 28 (flashlight#4) for giving backlight is set opposite the digital camera 24 at a downward angle of 70 degrees towards an object in spaces 16 or 17. White-light diffuser plates 29, 30, 31 and 32 are located in front of each of the flashlights 25, 26, 27, and 28. For instance, the diffuser plate 29 located at a position illuminated by the flashlight 25 diffuses light received from the flashlight 25. Similarly, the diffuser plate 30 located at a position illuminated by the flashlight 26 diffuses light received from the flashlight 26, the diffuser plate 31 located at a position illuminated by the flashlight 27 diffuses light received from the flashlight 27, and the diffuser plate 32 located at a position illuminated by the flashlight 28 diffuses light received from the flashlight 28.
A polarizing panel 33 having an approximately same size as the diffuser plate 31 is located thereon. Polarizing filter 34 is located in front of the lens of the digital camera 24. The polarizing angle of the polarizing filter 34 is set at a roughly right angle with the polarizing panel 33. A white-light diffuser composed of several plates, is located so as to surround the object 16 or 17. A plurality of fluorescent lights 36 are provided as front light sources and are located behind the light diffuser plates 35 so as to illuminate object spaces 16 and 17 with diffuse light. Figure 2 shows a cross-section of the three-dimensional modeling system according to this embodiment shown in Figure 1 as a side view.
Lines extending from each of the digital cameras 18, 20, 22, 23 and 24 represent each of photographing areas thereof in a vertical cross-section. Two solid lines represent an outside border of the photographing area corresponding to a set predetermined focal length, and a central one-dotted-chained line shows a central axis of the photographing area.
Figure 3 shows a plan view of the system according to the embodiment shown in Figure 1 and Figure 2. Similar to Figure 2, two solid lines extended from the camera 20 to the table 2 show a photographing area thereof for the predetermined set focal length.
Figure 4 shows a calibration mat 37, which is used with the system according to this embodiment. Calibration dots 38 are prepared on a calibration mat 37 to enable the detection of the position of each digital camera, the orientation of each digital camera, and the focal length of camera lens of each digital camera. In this embodiment, there are 32 of the calibration dots 38, four (4) dots being located on each of eight (8) different radii dividing the mat into eight (8) equal angles. These calibration dots may have different sizes, and preferably each set of four dots on a radius has a different pattern of dot sizes compared to the other sets. The calibration mat 37 preferably has the same calibration dots located at the exactly same position on the back as the dots 38 located on the front of the mat. Preferably, of course, the diameter of the calibration mat 37 is smaller than a diameter of the circular glass table 2 and the size of the mat 37 is designed so as not to interfere with any of stepping motors and rollers 3,4, and 5. Figure 5 shows a block diagram representing the electrical system of the three- dimensional modeling system described using Figures 1-4 before as one embodiment of the invention. A doted line shown with the reference 39 represents a part contained in an ordinary personal computer. For example, this personal computer may be composed of ordinary PC platforms conforming to the well-known PC/AT standard. Central processing unit (CPU) 40 executes an application program. Normally, such an application program is stored in a ROM or a hard disk within the computer 39 as an object code. Then, such a program is read out from the storage and written into a memory within the CPU 40 for execution when the system launched. Since it does not relate to this invention directly, detailed descriptions of data flow, control flow and memory construction within the CPU 40 are omitted in this specification.
A video monitor 41 is connected to the computer 39. A video signal to be displayed at the video monitor 41 is output from a video board 42 to which the monitor 41 is connected. The video board 42 is driven by a set of software programs called a video driver 43. A keyboard 44 is provided by which users of this system manually input data and they may also give instructions using a mouse 45. Such input data and instructions are interpreted in a keyboard and mouse driver 46 composed of a set of software programs. All of the digital cameras 18, 19, 20, 21 and 22 are connected to the computer 39 by the well-known Universal Serial Bus (USB). The item referenced 47 represents USB ports to which the digital cameras 18, 19, 20, 21 and 22 are physically connected and their HUB interfaces. A USB device manager 48 manages the USB ports and HUB interfaces 47. Also provided are software programs composing a USB driver 49 for controlling the digital cameras 18, 19, 20, 21 and 22.
An interface box 50 controls communications amongst a STM driver 51, a photo reflection detector 53, a lighting control unit 55 and the computer 39. All of elements 51 to 60 are located in the interface box 50 and are described hereafter.
The STM driver 51 drives and controls stepping motors 3, 4, and 5 for rotating the glass circular table 2 in accordance with outputs from a digital to analogue converter (DAC) 52 which converts digital data from the computer into an analogue signal to be used in the STM driver 51. The photo reflection detector (PR) 53 detects an output of the photo reflector 9 indicating positions of the encoder marks 8 composed of evaporated aluminum thin films located a circumference of the table 2. The analogue output of the photo reflection detector 53 is converted to digital data at an analogue to digital converter (ADC) 54.
The lighting control unit 55 has a register that controls flashlights 25, 26, 27 and 28, which are used as backlights for the object on the table. This register is composed of a 5-bit hardware register, the bits of which control flashlights 25, 26, 27 and 28 and fluorescent lights 36 in accordance with control signals provided via a serial interface 61. Such light-control signals are created in accordance with the application program and are communicated via the serial interface 61 under control of a communication serial port driver (COM port Driver) 62. The control signals for controlling the flashlights are input to two-input AND gates 57, 58, 59 and 60 respectively. A first port, labelled #0, #1 of the register is connected to one input of the AND gate 57. A port #2 of the register is connected to one input of the AND gate 58, a port #3 of the register is connected to one input of the AND gate 59 and a port #4 of the register is connected to one input of the AND gate 60. The register also has a port #FL for front lights and is therefore connected to the plurality of fluorescent lights 36.
A five-input OR gate 56 has its inputs respectively connected to X-triggers of digital cameras 18, 20, 22, 23 and 24. These X-triggers are provided for synchronizing flash with photographing and are well known as "Flash Synchronization" connection points. The other input of each of the AND gates 57, 58, 59 and 60 is connected to the output of the OR gate 56.
A hard disc unit 63 stores data 64 of texture images and silhouette images. A three-dimensional object model creating program is stored in a ROM or a hard disc within the computer 39 as an object code and is represented as stored at 65. The program is read out from the storage and written into a memory within the CPU 40 for execution when the system is launched. The application program and the model creating program (an object modeling engine) communicate through the communication (COM) interface. A program for displaying a graphical user interface (GUI) for the application program stored in CPU 40 is represented as stored at 66. Figure 6, Figure 7 and Figure 8 are flowcharts of operations of the application program in CPU 40 and the three-dimensional object model creating program in storage 65, as carried out by the CPU 40. Figure 9 shows a table of control data used for positioning the circular glass table 2, which is stored in CPU 40 within the application program. This table shows that the system controls the rotation of the circular table 2 to locate it to each of fifteen (15) rotation positions, each different from a predetermined principal rotation position (#0) by a multiple of 22.5 degrees. Such sixteen (16) positions, including the principal rotation position, are determined by data set (rotation_set: #0-#15) and are designated by the application program.
Figure 10 shows a table of imaging parameters (exp_param_set) stored in storage within the application program. In accordance with the table shown in Figure 10, an exposure value (AV) and a shutter speed value (TV) of the digital cameras 18, 20, 22, 23 and 24 are determined corresponding to a designated mode (#0: front_texture, or #1 : backlight).
Figure 11 shows a table of focal lengths of the digital cameras 18, 20, 22, 23 and 24. These parameters representing focal lengths are also stored in the storage 40 within the application program. Such actual focal lengths are determined in response to input data (zoom_pos_set), which represent focal position of zoom lens of the digital cameras.
Figure 12 shows control data used for turning on and turning off the flashlights 25, 26, 27 and 28 (backlights) and fluorescent lights 36 (front light). These control data are also stored in the storage 40 within the application program. Operations of the system shown in Figure 1 to Figure 5 are described referring to the flowcharts shown in Figure 6, Figure 7 and Figure 8 as follows. As the subject system in one embodiment of this invention starts to operate, the CPU 40 starts the application program and its operations from a step #101 shown in Figure 6. In the step #101 , all previous settings and data used for previous operations are retrieved and the values are reinitialised. The intialising process is shown in Figure 8. A step #201 is a starting step of the intialising process. In this step #201, all texture image data and silhouette image data stored in the hard disc unit 63 are cleared. In a following step #202, the CPU 40 resets the USB HUB interfaces 47, the USB device manager 48 and the USB camera driver 49 and confirms communications between the digital cameras 18, 20, 22, 23, 24 and these USB elements 47, 48 and 49.
In a step #203 following the step #202, the CPU 40 initialises the interface box 50, the serial interface 61 and the serial port driver 62 and confirms communication between the interface box to and the interface 61. Finally in a step #204 of the initializing step #101 , the circular glass table 2 is returned to the predetermined principal rotating position in accordance with the application program. In particular, the CPU 40 instructs rotation of the table 2 so as to locate it at the principal rotation position. This instruction is transferred to the interface box 50 through the serial interface 61 and the serial port driver 62. Digital data representing the principal rotating position (#0) is compared with digital data representing the actual rotation position of the table 2 and the CPU 40 calculates digital driving data for the stepping motors 3,4, and 5.
The digital driving data is converted to an analogue signal at the digital to analogue converter (DAC) 52 and supplied to the STM driver 51. The STM driver 51 drives stepping motors 3, 4, 5 in accordance with the analogue signal. The photo reflection detector (PR) 53 determines the actual rotation position of the table 2 by detecting an output of the photo reflector 9 reflecting the positions of the encoder marks 8.
The CPU 40 refers the table shown in Figure 9 to determine a targeted rotation position (angle) of the table 2 and calculates difference between the current actual rotating position of the table 2 and the targeted position. In accordance with the difference, the CPU 40 generates digital driving data necessary to drive the stepping motors 3, 4, and 5. Accordingly, the stepping motors 3, 4, and 5 are driven by the analogue driving signal output from the DAC 52 so that the table 2 can be located at the predetermined principal position. After the above initialising process, operations are returned to a step #102 shown in the Figure 6. In this step #102, a user of this system decides a size of a three-dimensional object to be photographed and modeled. In accordance with the application program and the program for displaying a graphical user interface
(GUI) 66, the CPU 40 displays a size-selecting wi ni dow on the display of the video monitor 41 via the video board. The user uti llises the keyboard 44 or the mouse 45 to select the size of the object by referri ni g the displayed page. Since the GUI and the selecting page itself are not important to describe this invention, detailed descriptions thereof are omitted.
In this embodiment, the user can select the size of the object as "Small" or "Large". In accordance with this embodiment, it is preferable to select "Small" if the height of the object is smaller than the predetermined height A which corresponds to the horizontal line 19 indicating the plane on which the digital camera 18 is located. On the other hand, if the height of the object is larger than the height A but smaller than the height B which corresponds to the horizontal line 21 representing the plane on which the digital camera 20 is located, it is preferable for the user to select "Large".
If the user selects the object size as "Large", this result is registered as "obj_size=large" in a step #103. If the user selects the object size as "Small", this is registered as "obj_size=smaH" in a step #104. Following these steps #103 and #104, the user can select whether calibration is to be carried out or not. Similar to the step #102, the CPU 40 displays a calibration-selecting page via the GUI on the display of the video monitor 41 in accordance with the application program. If the user selects to carry out the calibration, the process goes to a step #106. If the user selects not to carry out the calibration, the process skips steps #106 to #121 and goes to a step #122. The steps #106 to #121 for calibration are now described as follows. In the step #106, the fluorescent lights 36, the front lights, are turned on. In detail, in accordance with the application program, the CPU 40 instructs the writing of a flag "1" in the front light control bit of the register in the lighting control unit 55 of the interface box 50 though the serial interface 61 under control of the COM port Driver 62. Accordingly an output of the port of front lights #FL switches to a predetermined level representing "1" and, as shown in the table of Figure 12, the fluorescent lights 36 are turned on.
In a step #107, the CPU 40 displays a window urging the user to put the calibration mat 37 on the circular glass table 2 via the GUI on the display of the video monitor 41 in accordance with the application program. Further, in a step #108, the CPU 40 displays a window for the user to confirm whether the calibration mat 37 has already been put on the table 2 or not. If the user confirms this, namely ready to calibrate, the process goes to a step #109. In a step #109, a loop variable N is set as one of sixteen integers from "0" to "15". Initially the variable N is set as "0".
In a step #110, the CPU 40 determines a targeted rotating position (angle) of the table 2 by outputting one of the data sets (rotation_set: #0-#15). The CPU 40 refers the table shown in Figure 9. Operations of the CPU 40, the stepping motors 3,4, and 5, photo reflection detector (PR) 53, and the DAC 52 to rotate the table 2 to the targeted position are similar to the operations described before for the step #101 (#204).
In the step #111 , a variable, a software repeat counter C, is set as "0", "1", "2", "3", or "4". Initially the variable C is set as "0". In a following step #112, the imaging parameters shown in the table of the Figure 10 are set (in accordance with exp_param_set: #0) for front light, namely the exposure value (AV) is set to F8.0 and the shutter speed (TV) is set 1/15 second. These parameters are transferred to one of the digital cameras 18, 20, 22, 23 and 24. These parameters are transferred through the USB ports HUB interfaces 47 under control of the USB device manager 48 and the USB camera driver 49 to the digital camera (digital_cam_#C). Therefore, in this initial stage, the parameters are transferred to the digital camera 18 (digital_cam_#0).
In a step #113, a parameter representing a focal length is also transferred to the digital camera (digital_cam_#C) through the USB ports HUB interfaces 47. In this step #113, the transferred parameter is a set #0 (zoom_pos_set: #0), which represents a wide-end focal length in accordance with the table of the Figure 11. Similarly, in this initial stage, this parameter is also transferred to the digital camera 18 (digital_cam_#0).
In a step #114, the CPU 40 sends a command to photograph to the digital camera (digital_cam_#C) and the digital camera takes an image. Image data obtained by the digital camera (digital__cam_#C) is transferred to the hard disc unit 63 through the USB ports HUB interfaces 47 after compressing the image data in conformity with well-known JEPG compression scheme in a step #115. The name of such image file is, for example, "img_cam #0. jpg".
The JEPG compressed image data obtained in the step #114 and stored in the step #115 is processed and developed and the CPU 40 detects the calibration dots 38 on a calibration mat 37 in the captured image in accordance with the application program and three-dimensional object model creating program 65 in a step #116. The CPU 40 processes and analyses the detected calibration dots and determines a central position of the calibration mat 37 for creating supposed three-dimensional coordinates. In accordance with the supposed three-dimensional coordinates, a position and an orientation of the digital camera, and a focal length of the digital camera, the focal length of the camera can be obtained from the image of the calibration dots 38 by using perspective information. Detailed methods or processes for obtaining the central position of the calibration mat 37, the supposed three-dimensional coordinates, the position and the orientation of the digital camera, and the focal length were disclosed in several former patent applications, for example, a Japanese Raid- Open Patents numbered 00-96374, a Japanese Raid-Open Patents numbered 98-170914 and a UK patent application numbered 0012812.4 , and these methods or processes can be adopted for this step #116. Therefore, in this specification, detailed descriptions of such concrete methods and processes.
In the step #116, if the digital camera 24 (digital_cam_#4) located beneath the table is designated by the variable C, the image of the calibration dots is obtained from the back of the calibration mat 37 and is processed and analysed for obtaining the position and the orientation of the digital camera 24, and the focal length of the digital camera 24. After the step #116 and confirming the completion of obtaining information for the digital camera (digital_cam_#C) in a step #117, the obtained camera information, including the position, the orientation and the focal length of the digital camera (digital_cam_#C_#N) are stored in the hard disc unit 63 as a file named "cal_cam#C" in a step #118.
In a step #119, the variable C designating the digital camera is incremented. In the initial stage, the variable C is changed from "0" to "1". As easily understood, the steps #111 to #119 are repeated for each value of the variable C from "0" to "4". In the step #120, the variable N is incremented and the process goes back to the step #109 from the step #120. (Lines returning from steps #119 to #111 and #120 to #100 are not shown in the Figure 7 and Figure 8.) As described before, in step #110, the CPU 40 determines the targeted rotating position (angle) of the table 2, the table 2 is rotated to the targeted position. The variable N is incremented one by one and finally comes to "15". At that point all of the photographic processes have been completed, the circular glass table has been completely rotated around the object and all camera calibration parameters with each camera position at every rotation angle have been stored by step #118. Namely, each of the digital cameras 18(digital_cam#0), 20(digital_cam#1), 22(digital_cam#2), 23(digital_cam#3), and 24(digital_cam#4) have been selected in turn and each digital camera has photographed the image of the calibration dots 38 in order to obtain the camera information (the position, the orientation and the focal length) for each of the digital cameras. These information have been stored in the hard disc unit 63 as files "cal_cam#0", "cal_cam#1", "cal_cam#2", "cal_cam#3", and "cal_cam#4".
After these repeated processes in the steps #109 to #120, a value of the variable N is checked in a step #120 to confirm whether the repeated processes in the steps #109 to #120 are completed for all of the digital cameras 18, 20, 22, 23, and 24 or not. If it is confirmed, the process goes to a next step #121. In the step #121, the CPU 40 instructs to write a flag "0" in the front light control bit of the register in the lighting control unit 55. Accordingly an output of the port of front lights #FL switches to another predetermined level representing "0" and the fluorescent lights 36 are turned off in accordance with the table of the Figure 12. All processes for calibration are completed at the step #121 and the process returns to a step #122. In a step #122, the CPU 40 displays a window for asking the user if an actual modeling process shall be started, using the GUI, on the display of the video monitor 41. If the user instructs by using keyboard 44 or the mouse 45 to start an actual modeling, the process goes to a step #123. In the step #123, the CPU 40 turns on the fluorescent lights 36, i.e. the front lights. This process is the same as the step #106. This time, the variable N is set as one of sixteen integers from "0" to "15". Initially the variable N is set as "0" in a step #124.
In a step #125, the CPU 40 determines a targeted rotation position (angle) of the table 2 by outputting one of the data sets (rotation_set: #0~#15). The CPU 40 refers to the table shown in Figure 9. Operations of the CPU 40, the stepping motors 3,4, and 5, photo reflection detector (PR) 53, and the DAC 52 to rotate the table 2 to the targeted position are similar to the operations described before for the step #101. In a step #126, the CPU 40 consults the selection result of the object size ("Small" or "Large") made by the user by reading out registered information as "obj_size=large" or "obj_size=small".
At step #126, if the user selected the object size as "Small", the process goes to a step #127. If the user selected the object size as "Large", the process goes to a step #128. In the step #127, numbers to be selectively used as another variable C are set as 0, 2, 3, and 4. On the other hand, in the step #128, numbers to be selectively used as another variable C are set as 1, 2, 3, and 4. This means, the digital camera 18 (digital_cam_#0) is used if the object size is small, but the digital camera 19 (digital_cam_#1) is used if the object size is large. Other digital cameras 20.22, and 24 are commonly used irrespective of the size of the object. In an initial stage, this variable C is set either "0" or "1 in accordance with the size of the object. Two steps #129 and #130 both check the vari a; ble C and the variable N. If the variable C is "2" and the variable N is not a multi p| ie of two, the process directly goes from the step #129 to the step #140, skippi ni g steps #131 lo #139. If the variable C is "3" or "4" and the variable N is not a multiple of four, the process directly goes from the step #130 also to the step #140, skipping steps #131 to #139. Otherwise, the process goes to a step #129.
This means, if the camera 18 (digital_cam_#0) or 19 (digital_cam_#1 ) is selected, the steps #131 to #139 (executing the photographing of an object) are always carried out the photographing process irrespective of a photographing angle, namely the rotation position of the table 2. Accordingly, each of sixteen (16) images is taken by the camera (digital_cam_#0) 18 or the camera (digital_cam_#1) 20.
On the other hand, if the camera 20 (digital_cam_#2) is selected, the steps #131 to #139 (executing the photographing) are carried out only when the table 2 is at the principal rotation position or rotation positions different from the principal rotation position (#0) by each of multiple angles of 45 degrees. Accordingly, eight (8) images are taken by the camera 20 (digital_cam_#2). Similarly, if the camera 22 (digital_cam_#3) or the camera 24 (digital_cam_#4) is selected, the steps #131 to #139 (executing the photographing) are carried out only when the table 2 is at the principal rotation position or rotation positions different from the principal rotating position (#0) by each of multiple angles of 90 degrees. Accordingly, only four (4) images are taken by both the camera 22 (digital_cam_#3) and the camera 24 (digital_cam_#4). The photographing process carried out in the steps #131 to #139 is described hereafter. In the step #131, the CPU 40 instructs to write a flag "1" in one of the backlight control bits of the register in the lighting control unit 55 of the interface box 50 though the serial interface 61 under control of the COM port Driver 62. For example, for port #0,#1 the corresponding bit of the register is switched to "1", which switches the output of port #0,#1 to a level representing "1". Thus (in accordance with the table of Figure 12) the backlights may be turned on. In the following step #132, the imaging parameters, shown in the table of the Figure 10, are set (in accordance with exp_param_set: #1) for backlight, namely the exposure value (AV) is set to F8.0 and the shutter speed (TV) is set to 1/60 second. Then, these parameters are transferred to one of the digital cameras 18, 20, 22, 23 and 24. These parameters are transferred through the USB ports HUB interfaces 47 under control of the USB device manager 48 and the USB camera driver 49 to the digital camera (digital_cam_#C). Therefore, in this initial stage, the parameters are transferred to the digital camera 18 (digital_cam_#0) or the digital camera 19 (digital_cam_#1 ). In a step #133, a parameter representing a focal length (zoom_pos_set: #X), is also transferred to the digital camera (digital_cam_#C) through the USB ports HUB interfaces 47. The focal length "#X" is manually selected by the user among six focal lengths ("#0" to "#5") shown in the table of the Figure 11 before the photographing process. The user needs to select the focal length so that the whole of the object is in a photographic area of each digital camera.
Of course, this user's selection may be given by using the GUI. For example, the CPU 40 displays a focal length selecting page showing six focal lengths to be selected, using the GUI, on the display of the video monitor 41 via the video board 42. The user utilises the keyboard 44 or the mouse 45 to select one of the focal lengths by referring the displayed page.
In a step #134, the CPU 40 sends a command to photograph to the digital camera (digital_cam_#C) and the digital camera takes an image. In response to this command to photograph, one of the X-triggers of digital cameras 18, 20, 22, 23 and 24 is closed and becomes a predetermined level representing "1" for a predetermined period of time. Thus, the output of the OR gate 56 becomes a predetermined level representing "1" and one of the AND gates 56,58, 59 and 60 corresponding to the digital camera (digital_cam_#C) outputs a predetermined level representing "1". Accordingly, one of the flashlights 25, 26, 27, and 28 is turned on and emits flashlight in synchronization with the X-trigger, namely a photographing operation of the digital camera (digita!__cam_#C). For example, in the initial stage of this photographing process while the variable C is 0 or 1, the digital camera 18 (digital_cam_#0) or the digital camera 20 (digital_cam_#1 ) photographs the object and the flashlight 25 emits flashlight.
In a step #135, the CPU 40 instructs to write a flag "0" in one of the backlight control bits of the register in the lighting control unit 55. For example, for port #0,#1 the corresponding bit of the register is switched to "0", which switches the output of port #0,#1 to a level representing "0". Thus the backlight #0,#1 is disabled.
The image data obtained by the digital camera (digital_cam_#C) in the step #134 is transferred to the hard disc unit 63 after compressing in conformity with well-known JEPG compression scheme in a step #136. The name of such image file is "sil_cam#C_#N. jpg". For instance, if the digital camera 18 (digital_cam_#0) is used in the initial stage and the photographing is made while the table 2 locates at the principal rotation position, the name of the image file is "sil_cam#0_#0. jpg". As seen, the name of the file includes "sil" indicating a silhouette image, "#C" indicating the camera used for photographing, and "#N" indicating the rotation position of the table 2.
In a step #137, the imaging parameters shown the Figure 10 are set (in accordance with exp_param_set: #0) for front light. Accordingly the exposure value (AV) is set as F8.0 and the shutter speed (TV) is set as 1/15 second. Similarly to the step #132, these parameters are transferred to the digital cameras (digital_cam_#C). In the initial stage, the parameters are transferred to the digital camera 18 (digital_cam_#0) or the digital camera 19 (digital_cam_#1). The digital camera typically uses the same focal length (zoom_pos_set: #X) for both obtaining the silhouette image and the texture image. Therefore, the same focal length (zoom_pos__set: #X) is used also in the capturing step #133. It shall be noted that the fluorescent lights 36, the front lights, have been continuously turned on since the step #123.
In a step #138, the CPU 40 sends a command to photograph to the digital camera (digital_cam_#C) and the digital camera takes a texture image. In this moment, since none of the X-triggers of digital cameras 18, 20, 22, 23 and 24 is closed and each has a predetermined level representing "0", none of the flashlights 25, 26, 27, and 28 is turned on and emits any lights.
The image data obtained by the digital camera (digital_cam_#C) is, in the step #138, transferred to the hard disc unit 63 after compression in conformity with the well-known JEPG compression scheme in a step #139. The name of such image file is, for example, "ima_cam#C_#N. jpg". For instance, if the digital camera 18 (digital_cam_#0) is used in the initial stage and the photographing is made while the table 2 located at the principal rotating position, the name of the image file is "ima_cam#0_#0. jpg". As seen, the name of the file includes "ima" indicating a texture image, "#C" indicating the camera used for photographing, and "#N" indicating the rotating position of the table 2.
After all of the steps #131 to #139 executing photographing are completed, the variable C is incremented in a step #140. If it was judged that the object is small, the process goes back to the step #127 from the step #139, (the line for that not being shown in Figure 7 and Figure 8). For instance, if the variable C was set as "0" so that the digital camera 18 (digital_cam_#0) is used during the first photographing process, the variable C becomes "2" in the step 127. On the other hand, if it was judged that the object is large, the process goes back to the step #128 from the step #140. For instance, if the variable C was set as "1" so that the digital camera 20 (digital_cam_#1 ) is used during the first photographing, the variable C becomes "2" in the step 128. Thus, whichever the object is small or large, the photographing process using the digital camera 22 (digital_cam_#2) is next executed in the steps #131 to #139. After photographing process using the digital camera 22 (digital_cam_#2) is completed, the variable becomes "3" and the digital camera 23 (digital_cam_#3) is selected to be used in the photographing process. Finally, the variable C comes "4" and the digital camera 24 (digital_cam_#4) is selected to be used in the photographing process. Before step #141, all of these photographing processes are executed whilst the table 2 is located at the principal rotation position. In the step #141, the variable N is incremented and the process goes back to the step #124 from the step #140, (the line for that not being shown in Figure 7 and Figure 8). As described before, in the step #125, the CPU 40 determines the targeted rotation position (angle) of the table 2, the table 2 is rotated to the targeted position. The variable N is incremented by units of one and finally becomes "15". At that point all of the photographic processes have been completed, the circular glass table has been completely rotated around the object and all texture images and silhouette images from all angles have been stored in the hard disc unit 63. The final total number of the texture images is 32 (=16+8+4+4) and the final total number of the silhouette images also is 32 (=16+8+4+4). After all photographing process repeatedly executed in the steps #124 to #139 have completed, the process goes to a step #142. In the step #142, the CPU 40 instructs to write a flag "0" in the front light control bit of the register in the lighting control unit 55. Accordingly an output of the port of front lights #FL turns to the predetermined level representing "0" and the fluorescent lights 36 are turned off in accordance with the table of the Figure 12.
In a step #143, executing the three-dimensional object model creating program 65, the CPU 40 creates three-dimensional geometry data of the object by using all silhouette images stored in the hard disc unit 63. The three-dimensional geometry is defined by polygons, including triangles and four-cornered polygons. Detailed methods or processes for obtaining the three-dimensional geometry data by using silhouette images taken from different positions and orientations is disclosed in former patent applications, for example, USP 6317139, USP 4710876, and a UK application number 0114157.1 (CRE235), and these methods or processes can be adopted to this step #143. Therefore, in this specification, detailed descriptions of such methods and processes are omitted. In the step #143, before creating three-dimensional geometry, parameters including photographing positions, orientations, and focal lengths for each of digital cameras and each photographing are calibrated by using the obtained camera information, including the position, the orientation and the focal length of the digital cameras stored in the hard disc unit 63 as files named "cal_cam#C" in the step #118. In a step #144, in accordance with the three-dimensional object model creating program 65, the CPU 40 creates three-dimensional texture data to be put on surfaces of each polygons created in the step #143 by using all texture images stored in the hard disc unit 63. Detailed methods or processes for obtaining the three-dimensional texture data for each polygon by using two- dimensional texture images taken from different positions and orientations is disclosed in former patent applications, for example, UK application number 0022343.4 (CRE223), and such methods or processes can be adopted to this step #144. Therefore, in this specification, detailed descriptions of such concrete methods and processes are omitted. Thus, all information of three-dimensional object including geometry and texture are finally obtained.
In a step #145, the resultant three-dimensional object having geometry on which texture images have been put is displayed on the display of the video monitor 41. As known, such resultant three-dimensional model can be rotated, magnified or like by the user, using the keyboard 44 or the mouse 45.
In a step #146, all information of such a resultant three-dimensional object including three-dimensional geometry information and three-dimensional texture information to be put on the geometry is stored in the hard disc unit 63 as an ordinal VRML file (*.wrl), for example. The process is completed after the step #146.
When the digital camera 24 located beneath the glass table 2 photographs the object, normally the digital camera 24 picks up the light reflected by the glass table 2. This reflected light obviously deteriorates quality of the resultant three- dimensional object. Especially, the existence of the white-light diffuser plate 31 causes a significant problem, namely that directly reflected light from the plate 31 illuminates a lower surface on the glass table within a photographing area of the digital camera 24. This reflected light illuminating the lower surface in the main deteriorates the contrast of texture images of a bottom of the object. To solve this, in this embodiment, the polarizing panel 33, which has almost the same size as the white-light diffuser plate 31 , was put thereon and the polarizing filter 34 was set in front of a camera lens of the digital camera 24, which has a photographing area illuminated by the directly reflected light from the white-light diffuser plate 31. Further the polarizing angle of the polarizing filter 34 is set at roughly a right angle with the polarizing panel 33. Accordingly the system of this embodiment can significantly reduce the inappropriate influence caused by the directly reflected light from the white-light diffuser plate 31 on the resultant three-dimensional object model.
In accordance with the embodiment described above, finally all information of such a resultant three-dimensional object is finally stored as the VRML file. However, it is possible to store by using other types of files. In accordance with the embodiment described above, the fluorescent lights 36 are used as front lights. However, other types of lights such as tungsten lights can be used. The white balance of the digital cameras has preferably, however, to be set so that colour temperature of the produced image is appropriate. Further, other types of lights, such as tungsten lights or fluorescent lights, can be used as backlights instead of the flashlights.
To make description more brief and concise, the embodiment described above only introduced two kind of image parameters, as shown in Figure 10, for texture images using only front lights and silhouette images using a backlight. However, such parameters like an exposure value (AV) and a shutter speed (TV) are preferably set more precisely in consideration with luminance value of each of the emitted lights, to effectively obtain high-quality texture images and silhouette images. It shall be noted that one aspect of this embodiment and one advantage of this embodiment is to select the different exposure parameters between texture images and silhouette images. This provides significant improvement in obtaining high-quality texture and silhouette images.
The location of cameras and focal lengths of each of camera lens are considered and are decided in accordance with several factors, including the size of the object, distance between the object and backlight, and the size of the backlight, so that the outline of the object is surrounded by the backlight. In accordance with the embodiment described above, a plurality of digital cameras are prepared for each vertical (latitudinal) location. However, it would also be possible to move one digital camera so as to locate it at a plurality of latitudinal locations by using a moving mechanism. If the moving mechanism accurately locates the digital camera at each latitudinal location, such a system may give a cheaper solution than this embodiment. This aspect of this embodiment is therefore not in how many cameras are used but rather the existence of a plurality of latitudinal locations where the camera photographs the object. Therefore, it shall be understood that obtaining a plurality of latitudinal locations for photographing by moving the digital camera is within the scope of this invention, Figure 13 shows another embodiment of the system according to this invention. This embodiment introduces fine nylon fibers roped off on a plane on which the object is put instead of the glass table. In Figure 13, the reference 67 shows a rotating table made of metal such as aluminum. The table 67 has a hole where fine nylon fibers 69 are secured. The material of the nylon fibers is selected so that fibers can be made sufficiently fine enough to be unnoticeable or invisible from the digital cameras. The number of fibers is decided in accordance with size and weight of the object. The size of the hole is decided so as to cause no obstacle to the creation of silhouette images taken by all digital cameras. By this embodiment shown in Figure 13, it becomes possible to float the object in air substantially without visible materials. In comparison with the former embodiment, this embodiment shown in Figure 13 has an advantage in avoiding the inappropriate influence of the reflected lights without using other elements like a polarizing panel, a polarizing filter 34 and so on. These embodiments described above have many advantages in comparison with the prior technology for creating three-dimensional object model, as follows. First of all, the above-described embodiments make it possible to minimize the user operation to obtain texture images and silhouette images to be used for creating high quality three-dimensional objects. In accordance with one aspect of these embodiments, it is possible to obtain texture and silhouette images having higher quality without bothering users, by photographing the object a plurality of times under different lighting conditions for silhouette images and texture images, namely by turning on and off the backlight in accordance with the embodiments. Further, optimising photographing conditions by changing exposures between texture images and silhouette images and changing the lighting, quality of texture and silhouette images becomes greatly improved.
Further, putting the object on a material, which is at least partially transparent, so that the object looks to be substantially floating in air and photographing the object from a plurality of angles including an angle below the table as explained in the embodiments, makes it possible to take silhouette images from the bottom of the object and texture images of the bottom of the object without manually changing the direction or orientation of the object. This greatly contributes to the creation of a resultant three-dimensional object that is viewable from all orientations without relying on the user's manual operations. Further, in accordance with the first embodiment described above, since a polarizing control panel is set at a position from which a direct reflected right enter into a digital camera, a polarizing control filter is set in front of the digital camera, and the polarizing angles of these are set so as to have a predetermined difference, preferably a right angle difference, it becomes possible to reduce influences of reflected lights on photographed image taken by the digital camera. Accordingly, the three-dimensional object model viewable from all orientations can be automatically obtained without large distortion.
As described above, in this specification, although only two embodiments have been disclosed, it is possible to replace some of the particular elements with alternative elements and remain within a scope of the present invention. A couple of such alternative elements are disclosed as follows.
In the above-described embodiments, an element on which an object to be modeled locates is composed of a glass table or fine transparent fibers secured on the plate. A method or elements for making the object shown in all orientation shall not be limited into the glass table or fine transparent fibers. For example, it is possible to replace these with a transparent fiber hanging the object, a transparent needle-shaped resin piercing the object, or a transparent stool on which the object is set.
Further, in the above-described embodiments, a plurality of digital cameras are prepared and set at different positions for photographing the object from a plurality of different orientations. This structure may be replaced with a structure, which guides and moves one digital camera, revolving it around the object. If such a structure utilising a single camera is adopted, the camera is moved to locate it at one or more positions for every angle at which the object is rotated and photographs from each position. Further, in the above-described embodiments, although the object is photographed from a plurality of positions having different heights from the object, this is not essential for the present invention. Namely, the present invention can be applied to an apparatus photographing a plurality of silhouette and texture images from different angles having the same height. Even with such a structure, it can be seen, that the advantage of the present invention to obtain high quality texture and silhouette images with minimising user's manual operation similarly exists. Therefore it shall be understood that an apparatus adopting such a structure is with a scope of the present invention. Further, in the above-described embodiments, transparent materials are used for setting an object to be modeled so that the object can be seen from all orientations, namely setting object substantially in the air. However, this setting is not essential for the present invention. For example, a structure to set a backlight very close to the object may be used. If such a structure is utilised, you have the same advantage by turning on the backlight during photographing for silhouette images and turning off the backlight during photographing for texture images.
Further, in the first embodiment using a glass table, a polarizing panel is set on a white-light diffuser plate to provide a backlight having the same shape of the white-light diffusible plate. However, the polarizing panel or filter can be set at different positions within in the optical path of the backlight that illuminates a photographing area of the camera by directly reflecting on the glass table and can be set as a different shape of polarizing control material. Further, in the above-described embodiments, longitudinal relative positions, namely rotating positions of the table, are commonly used for several latitudinal positions of a camera and photographs for silhouette and texture images. However, these conditions are not essential for the present invention and, for example, all longitudinal positions can be different for each photographing at different latitudinal positions and can be different between silhouette and texture images.
Further, in the above-described embodiments, to change an exposing condition under which the object is photographed between texture images and silhouette images, only shutter speed is changed. However, both the exposure value and the shutter speed, or only the exposure value, may be changed to obtain ideal exposing condition. It is also possible that the shutter speed can be controlled by controlling a mechanical shutter or by changing charge- accumulating period of imaging device, such as CCD (Charge Coupled Device). Furthermore, in the above-described embodiments, automatic photographing is carried out by rotating a table and setting the table at a plurality of predetermined angles and is described as a fully automatic photographing. However, even if the table is manually or part-manually rotated, the reduction of other manual user operations are still largely reduced, especially operations of setting the photographing conditions and lighting conditions. Therefore, such a structure in which the table is manually rotated is also within the scope of the present invention.
As described above, according to one aspect of the present invention, it becomes possible to improve quality of silhouette and texture images without additional manual user operations by photographing an object to be three- dimensionally modeled under different lighting conditions. Especially by changing exposing conditions between silhouette images and texture images in addition to the lighting condition, the photographing conditions for each texture and silhouette images is optimized and quality of resultant texture images and silhouette images is highly improved. Further, photographing the object as shown from all orientations including below the object and from various latitudinal positions, the resultant three-dimensional model is highly improved. Further, according to another aspect of the present invention, it becomes possible to obtain silhouette images and texture images that produce high quality three-dimensional object model, which can be observed from any orientation, by reducing influences caused by a reflected right from a transparent plane by setting two polarizing controllers having a different polarizing control angles on a camera side and a lighting side and setting the polarizing angles thereof different from each other.

Claims

CLAIMS:
1. An apparatus for creating a three-dimensional object model, comprising photographing means for photographing an object to be modeled; setting means for setting a relative position between the object and said photographing means, said setting means being capable of setting a plurality of different relative positions, said setting means including a transparent table on which the object is to be set and rotating means for rotating the transparent table so that the object is set at a plurality of different longitudinal orientations to said photographing means, said setting means setting the relative positions so that said photographing means photographs the object from a plurality of latitudinal positions which has different latitudinal distances from the transparent table, the plurality of latitudinal positions including a position below the transparent table; lighting means including front lighting means capable of lighting a front side of the object confronting a photographing position of said photographing means and back lighting means capable of lighting a back side of the object hidden from the photographing position of said photographing means; and control means for controlling said photographing means, said lighting means and said setting means so that said photographing means photographs the object a plurality of times at a plurality of relative positions set by said setting means for obtaining a plurality of silhouette images in a first mode and photographs the object a plurality of times at a plurality of relative positions set by said setting means for obtaining a plurality of texture images in a second mode, and said lighting means changes the lighting condition between the first and second modes, the silhouette images and the texture images being to be used for creating the three-dimensional object model.
2. An apparatus according to Claim 1 , further comprising first polarizing control means located between the glass table and said photographing means as located below the transparent table and second polarizing control means located at a position in an optical path of lights illuminating a photographing area of said photographing means by directly reflecting on the transparent table, and wherein polarizing angles of said first and second polarizing control means have a predetermined difference.
3. An apparatus according to Claim 1 or 2, wherein said control means controls said back lighting means so that said back lighting means operates in the first mode but not in the second mode.
4. An apparatus according to one of Claims 1-3, wherein said photographing means includes exposure adjusting means for adjusting an exposing condition under which the object is photographed and said control means controls said exposure adjusting means to change the exposing condition between the first and second modes.
5. An apparatus according to Claim 4, wherein said exposure adjusting means sets a shutter speed of said photographing means faster in the first mode than that in the second mode.
6. An apparatus according to one of Claims 1-5, wherein at least a part of the relative positions set by said setting means is common between the first and second modes.
7. An apparatus according to one of Claims 1-6, wherein the transparent table is a glass table.
8. An apparatus according to one of Claims 1-6, wherein the transparent table s composed of fine transparent fibers.
9. An apparatus for creating a three-dimensional object model, comprising photographing means for photographing an object to be modeled, said photographing means including exposure adjusting means for adjusting an exposing condition under which the object is photographed; setting means for setting a relative position between the object and said photographing means, said setting means being capable of setting a plurality of different relative positions; lighting means including front lighting means capable of lighting a front side of the object confronting a photographing position of said photographing means and back lighting means capable of lighting a back side of the object hidden from the photographing position of said photographing means; and control means for controlling said photographing means, said lighting means and said setting means so that said photographing means photographs the object a plurality of times at a plurality of relative positions set by said setting means for obtaining a plurality of silhouette images in a first mode and photographs the object a plurality of times at a plurality of relative positions set by said setting means for obtaining a plurality of texture images in a second mode, the silhouette images and the texture images being to be used for creating the three-dimensional object model; wherein said control means controls said exposure adjusting means to change the exposing condition between the first and second modes and controls said lighting means so that said lighting means changes the lighting condition between the first and second modes.
10. An apparatus according to Claim 9, wherein said exposure adjusting means sets a shutter speed of said photographing means faster in the first mode than that in the second mode.
11. An apparatus according to Claim 9 or 10, wherein said control means controls said back lighting means so that said back lighting means operates in the first mode but not in the second mode.
12. An apparatus according to one of Claims 9-11, wherein said setting means includes a transparent table on which the object is to be set and rotating means for rotating the transparent table so that the object is set at a plurality of different longitudinal orientations to said photograph means, said setting the relative positions so that said photographing means photographs the object from a plurality of latitudinal positions which has different latitudinal distances from the transparent table, the plurality of latitudinal positions including a position below the transparent table.
13. An apparatus according to Claim 12, further comprising first polarizing control means located between the transparent table and said photographing means as located below the transparent table and second polarizing control means located at a position in an optical path of lights illuminating a photographing area of said photographing means by directly reflecting on the transparent table, and wherein polarizing angles of said first and second polarizing control means have a predetermined difference.
14. An apparatus according to Claim 12 or 13, wherein the transparent table is a glass table.
15. A method for creating a three-dimensional object model, comprising steps of; setting an object to be modeled at a plurality of relative positions to a photographing position, the plurality of relative positions including a plurality of different longitudinal orientations and a plurality of latitudinal positions which has different latitudinal distances from a transparent table on which the object is to be placed, the plurality of latitudinal positions including a position below the transparent table; photographing the object at the plurality of relative positions by using both a front lighting for a front side of the object confronting the photographing position and a back lighting for a back side of the object hidden from the photographing position; and controlling so that the object is photographed a plurality of times at a plurality of relative positions set in said setting step for obtaining a plurality of silhouette images in a first mode and is photographed a plurality of times at a plurality of relative positions set in said setting step for obtaining a plurality of texture images under a different lighting condition from the first mode in a second mode, the silhouette images and the texture images being to be used for creating the three- dimensional object model.
16. A method according to Claim 15, wherein a first polarizing controller is set between the transparent table and the photographing position below the glass table, a second polarizing controller is set at a position in an optical path of lights illuminating a photographing area by directly reflecting on the transparent table, and polarizing angles of said first and second polarizing controller have a predetermined difference.
17. A method according to Claim 15 or 16, wherein the back lighting is used in the first mode but not in the second mode.
18. A method according to one of Claims 15-17, wherein an exposing condition under which the object is photographed is changed between the first mode and the second mode.
19. A method according to Claim 18, wherein a shutter speed of photographing is set faster in the first mode than the second mode.
20. A method according to one of Claims 15-19, wherein at least a part of the relative positions of the object to be set by said setting means is common between the first and second modes.
21. A method according to Claim 15-20, wherein the transparent table is a glass table.
22. A method according to Claim 15-20, wherein the transparent table is composed of fine transparent fibers.
23. A method for creating a three-dimensional object model, comprising steps of; setting an object to be modeled at a plurality of relative positions to a photographing position; photographing the object at the plurality of relative positions by using both a front lighting for a front side of the object confronting the photographing position and a back lighting for a back side of the object hidden from the photographing position; and controlling so that the object is photographed a plurality of times at a plurality of relative positions set in said setting step for obtaining a plurality of silhouette images in a first mode and is photographed a plurality of times at a plurality of relative positions set in said setting step for obtaining a plurality of texture images under a different lighting condition and different exposing condition from the first mode in a second mode, the silhouette images and the texture images being to be used for creating the three-dimensional object model.
24. A method according to Claim 23, wherein a shutter speed of photographing is set faster in the first mode than the second mode.
25. A method according to Claim 23 or 24, wherein the back lighting is used in the first mode but not in the second mode.
26. A method according to one of Claims 23-25, wherein the object is set at a plurality of different longitudinal orientations to said photograph means and the object is photographed from a plurality of latitudinal positions which has different latitudinal distances from a transparent table on which the object is placed, the plurality of latitudinal positions including a position below the transparent table.
27. A method according to Claim 26, wherein a first polarizing controller is set between the transparent table and the photographing position below the transparent table, a second polarizing controller is set at a position in an optical path of lights illuminating a photographing area by directly reflecting on the transparent table, and polarizing angles of said first and second polarizing controller have a predetermined difference, the object is set at a plurality of different longitudinal orientations to the photographing position by rotating the horizontal plane.
28. A method according to Claim 26 or 27, wherein the transparent table is a glass table.
29. A method according to one of Claims 26-28, wherein a latitudinal photographing position is capable of changing.
30. A device for creating a three-dimensional object model, comprising one or more cameras for photographing an object to be modeled, a position changer for changing a relative position between the camera and the object so as to set the a plurality of different relative positions, the position changer including a transparent table on which the object is to be set and a motor for rotating the transparent table so that the object is set at a plurality of different longitudinal orientations to the camera and at a plurality of different latitudinal orientations positions which has different latitudinal distances from the transparent table, the plurality of latitudinal positions including a position below the transparent table; a front light capable of lighting a front side of the object confronting a photographing position of the camera; a back light capable of lighting a back side of the object hidden from a photographing position of the camera; and a controller for controlling the camera, the position changer and at least one of the front light and the back light so that the camera photographs the object a plurality of limes at a plurality of different relative positions set by said position changer for obtaining a plurality of silhouette images in a first mode and photographs the object a plurality of times at a plurality of relative positions set by said position changer for obtaining the texture image in a second mode, and said front light and the back light changes a lighting condition between the first and second modes, the silhouette images and the texture images being to be used for creating the three-dimensional object model.
31. A device according to Claim 30, further comprising a first polarizing filter located between the transparent table and the camera as located below the table and a second polarizing filter located at a position in an optical path of lights illuminating a photographing area of the camera by directly reflecting on the transparent table, and wherein polarizing angles of the first and second polarizing filters have a predetermined difference.
32. A device according to Claim 30 or 31, wherein the controller controls the back light so that the back light operates in the first mode but not in a second mode.
33. A device according to one of Claims 30-32, wherein the controller also changes an exposing condition of the camera under which the object is photographed between the first mode and the second mode.
34. A device according to Claim 33, wherein the controller sets a shutter speed of the camera faster in the first mode than that in the second mode.
35. A device according to one of Claims 30-34, wherein at least a part of the relative positions set by said position changer is common between the first and second modes.
36. A device according to one of Claims 30-35, wherein the transparent table is a glass table.
37. A device according to one of Claims 30-35, wherein the transparent table is composed of fine transparent fibers.
38. A device for creating a three-dimensional object model, comprising one or more camera for photographing an object to be modeled, a position changer for changing a relative position between the camera and the object so as to set the a plurality of different relative positions; a front light capable of lighting a front side of the object confronting a photographing position of the camera; a back light capable of lighting a back side of the object hidden from a photographing position of the camera; and a controller for controlling the camera, the position changer and at least one of the front light and the back light so that the camera photographs the object a plurality of times at a plurality of different relative positions set by said position changer for obtaining a plurality of silhouette images in a first mode and photographs the object a plurality of times at a plurality of relative positions set by said position changer for obtaining the texture image in a second mode, and said front light and the back light changes a lighting condition between the first and second modes, the silhouette images and the texture images being to be used for creating the three-dimensional object model, the controller also changes an exposing condition of the camera under which the object is photographed between the first mode and the second mode.
39. A device according to Claim 38, wherein the controller sets a shutter speed of the camera faster in the first mode than that in the second mode.
40. A device according to Claim 38or 39, wherein the controller controls the back light so that the back light operates in the first mode but not in a second mode.
41. A device according to one of Claims 38-40, wherein the controller set the object at a plurality of different longitudinal orientations to the camera and set the photographing positions of the camera at a plurality of latitudinal positions which has different latitudinal distances from a transparent table on which the object is placed, the plurality of latitudinal positions including a position below the transparent table.
42. A device according to Claim 41 , further comprising a first polarizing filter locating between the transparent table and the camera as located below the table and a second polarizing filter locating at a position in an optical path of lights illuminating a photographing area of the camera by directly reflecting on the transparent table, and wherein polarizing angles of the first and second polarizing filters have a predetermined difference.
43. A device according to one of Claims 38-41 , wherein the transparent table is a glass table.
44. An apparatus for creating a three-dimensional object model, comprising photographing means for photographing an object to be modeled, said photographing means photographing the object from a plurality of different positions; setting means for setting the object, said setting means including a transparent table; lighting means for illuminating the object through the transparent table; first polarizing control means located between the transparent table and said photographing means as located below the transparent table; and second polarizing control means located at a position in an optical path of lights illuminating a photographing area of said photographing means by directly reflecting on the transparent table, polarizing angles of said first and second polarizing control means have a predetermined difference.
45. A method for creating a three-dimensional object model, setting an object to be modeled on a transparent table, illuminating the object through the transparent table, setting a first polarizing control filter between the trans parent table and a camera located below the transparent table, setting a second polarizing control filter at a position in an optical path of lights illuminating a photographing area by directly reflecting on the transparent table, polarizing angles of the first and second polarizing control filters have a predetermined difference, and photographing the object from a plurality of different positions.
46. A device for creating a three-dimensional object model, comprising a camera for photographing an object to be modeled, the camera photographing the object from a plurality of different positions; a transparent table on which the object is set; a light illuminating the object through the transparent table; a first polarizing control filter located between the transparent table and the camera as located below the transparent table; and a second polarizing control filter located at a position in an optical path of lights illuminating a photographing area of the camera by directly reflecting on the transparent table, polarizing angles of the first and second polarizing control filters have a predetermined difference.
47. A set of computer programmes to be used for creating a three- dimensional object model by photographing an object to be modeled; comprising programme for controlling a camera and a relative position of the object and the camera so that the object is photographed from the plurality of different relative positions, the plurality of relative positions including a plurality of different longitudinal orientations and a plurality of latitudinal positions which has different latitudinal distances from the transparent table, the plurality of latitudinal positions including a position below a transparent table on which the object to be placed; and programme for controlling at least one of a front lighting for a front side of the object confronting a photographing position and a back lighting for a back side of the object hidden from the camera so that the object is photographed a plurality of times at a plurality of relative positions in a first mode for obtaining silhouette images and is photographed a plurality of times at a plurality of relative positions under a different lighting condition from the first mode in a second mode, the silhouette images and the texture images being to be used for creating the three- dimensional object model.
PCT/GB2004/001902 2003-05-07 2004-05-04 Photographing apparatus, device and method for obtaining images to be used for creating a three-dimensional model WO2004100561A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
GB0522704A GB2416973A (en) 2003-05-07 2004-05-04 Photographing apparatus,device and method for obtaining images to be used for creating a three-dimensional model

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB0310502.0 2003-05-07
GBGB0310502.0A GB0310502D0 (en) 2003-05-07 2003-05-07 Photographing apparatus,device and method for obtaining images to be used for creating three-dimensional model

Publications (1)

Publication Number Publication Date
WO2004100561A1 true WO2004100561A1 (en) 2004-11-18

Family

ID=9957612

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/GB2004/001902 WO2004100561A1 (en) 2003-05-07 2004-05-04 Photographing apparatus, device and method for obtaining images to be used for creating a three-dimensional model

Country Status (2)

Country Link
GB (2) GB0310502D0 (en)
WO (1) WO2004100561A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2010049680A2 (en) * 2008-10-27 2010-05-06 Lobster Pot Photography Limited Method and apparatus for generating "3-d photograph" data files
WO2016096161A1 (en) * 2014-12-17 2016-06-23 Martini Studios Ltd. Method and system for spin photography
EP3477254A1 (en) * 2017-10-30 2019-05-01 XYZprinting, Inc. Apparatus for producing 3d point-cloud model of physical object and producing method thereof
US20240142225A1 (en) * 2022-10-27 2024-05-02 Tencent America LLC Three-dimensional object scanning device and method

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1193646A2 (en) * 2000-09-27 2002-04-03 Canon Kabushiki Kaisha Image processing apparatus
US20020159628A1 (en) * 2001-04-26 2002-10-31 Mitsubishi Electric Research Laboratories, Inc Image-based 3D digitizer
US20030035061A1 (en) * 2001-08-13 2003-02-20 Olympus Optical Co., Ltd. Shape extraction system and 3-D (three dimension) information acquisition system using the same
US20030066949A1 (en) * 1996-10-25 2003-04-10 Mueller Frederick E. Method and apparatus for scanning three-dimensional objects

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030066949A1 (en) * 1996-10-25 2003-04-10 Mueller Frederick E. Method and apparatus for scanning three-dimensional objects
EP1193646A2 (en) * 2000-09-27 2002-04-03 Canon Kabushiki Kaisha Image processing apparatus
US20020159628A1 (en) * 2001-04-26 2002-10-31 Mitsubishi Electric Research Laboratories, Inc Image-based 3D digitizer
US20030035061A1 (en) * 2001-08-13 2003-02-20 Olympus Optical Co., Ltd. Shape extraction system and 3-D (three dimension) information acquisition system using the same

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2010049680A2 (en) * 2008-10-27 2010-05-06 Lobster Pot Photography Limited Method and apparatus for generating "3-d photograph" data files
WO2010049680A3 (en) * 2008-10-27 2010-07-22 Lobster Pot Photography Limited Method and apparatus for generating "3-d photograph" data files
WO2016096161A1 (en) * 2014-12-17 2016-06-23 Martini Studios Ltd. Method and system for spin photography
US10260672B2 (en) 2014-12-17 2019-04-16 Martini Studios Ltd. Method and apparatus for spin photography
EP3477254A1 (en) * 2017-10-30 2019-05-01 XYZprinting, Inc. Apparatus for producing 3d point-cloud model of physical object and producing method thereof
US20240142225A1 (en) * 2022-10-27 2024-05-02 Tencent America LLC Three-dimensional object scanning device and method

Also Published As

Publication number Publication date
GB0522704D0 (en) 2005-12-14
GB2416973A (en) 2006-02-08
GB0310502D0 (en) 2003-06-11

Similar Documents

Publication Publication Date Title
US7616886B2 (en) Photographing apparatus, device and method for obtaining images to be used for creating a three-dimensional model
US9531965B2 (en) Controller in a camera for creating a registered video image
US7059729B2 (en) Display apparatus and image pickup apparatus
US7771054B2 (en) Display system and camera system
US20070035539A1 (en) Image acquisition system for generation of three-dimensional models
US9041899B2 (en) Digital, virtual director apparatus and method
US6970166B2 (en) Display apparatus and information processing method
CN105072343A (en) Object appearance panorama shooting method and portable shooting system
JP2003099799A (en) Method for simulating motion of three-dimensional physical object stationary in changeless scene
EP1316211A2 (en) Image projection apparatus
JPH10191155A (en) Interactive desk-top system
KR100780701B1 (en) Apparatus automatically creating three dimension image and method therefore
Jacobs Interactive panoramas: techniques for digital panoramic photography
JP2003015218A (en) Projection display device
JP4665949B2 (en) Display control device, information terminal device, display control program
US20030025649A1 (en) Image projection apparatus
WO2004100561A1 (en) Photographing apparatus, device and method for obtaining images to be used for creating a three-dimensional model
KR102564522B1 (en) Multi-view shooting apparatus and method for creating 3D volume object
JP3601031B2 (en) Image data measuring apparatus and arbitrary viewpoint image creating method including free deformation of target object using image data
JP2005115645A (en) Shading information acquisition device and image processor
JP2003090715A5 (en)
EP1524864A1 (en) Turntable system for photographing three-dimensional objects
JPH05336549A (en) Stereoscopic video device
JP2005115897A (en) Three-dimensional image creation device and method
JP2001197524A (en) Stereoscopic image display device

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 0522704.6

Country of ref document: GB

Ref document number: 0522704

Country of ref document: GB

122 Ep: pct application non-entry in european phase