WO2012029193A1 - 商品撮影装置、商品撮影方法、画像変換装置、画像処理装置、画像処理システム、プログラム、ならびに、情報記録媒体 - Google Patents
商品撮影装置、商品撮影方法、画像変換装置、画像処理装置、画像処理システム、プログラム、ならびに、情報記録媒体 Download PDFInfo
- Publication number
- WO2012029193A1 WO2012029193A1 PCT/JP2010/067003 JP2010067003W WO2012029193A1 WO 2012029193 A1 WO2012029193 A1 WO 2012029193A1 JP 2010067003 W JP2010067003 W JP 2010067003W WO 2012029193 A1 WO2012029193 A1 WO 2012029193A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- unit
- product
- angle
- sensor unit
- Prior art date
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 42
- 238000006243 chemical reaction Methods 0.000 title claims description 41
- 238000000034 method Methods 0.000 claims description 60
- 230000006870 function Effects 0.000 claims description 16
- 230000005540 biological transmission Effects 0.000 claims description 8
- 230000008447 perception Effects 0.000 abstract 1
- 239000011159 matrix material Substances 0.000 description 30
- 239000013598 vector Substances 0.000 description 29
- 230000009466 transformation Effects 0.000 description 28
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 23
- 238000004364 calculation method Methods 0.000 description 10
- 230000005484 gravity Effects 0.000 description 9
- 230000003287 optical effect Effects 0.000 description 7
- 238000000605 extraction Methods 0.000 description 6
- 239000004973 liquid crystal related substance Substances 0.000 description 5
- 230000001133 acceleration Effects 0.000 description 4
- 230000010365 information processing Effects 0.000 description 3
- 241000282414 Homo sapiens Species 0.000 description 2
- 238000007476 Maximum Likelihood Methods 0.000 description 2
- 230000015572 biosynthetic process Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 230000036544 posture Effects 0.000 description 2
- 238000003825 pressing Methods 0.000 description 2
- 238000003786 synthesis reaction Methods 0.000 description 2
- 229930091051 Arenine Natural products 0.000 description 1
- 230000001174 ascending effect Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 210000000744 eyelid Anatomy 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000002945 steepest descent method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T1/00—General purpose image data processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/02—Affine transformations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/207—Image signal generators using stereoscopic image cameras using a single 2D image sensor
- H04N13/211—Image signal generators using stereoscopic image cameras using a single 2D image sensor using temporal multiplexing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/207—Image signal generators using stereoscopic image cameras using a single 2D image sensor
- H04N13/221—Image signal generators using stereoscopic image cameras using a single 2D image sensor using the relative movement between cameras and objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/633—Control of cameras or camera modules by using electronic viewfinders for displaying additional information relating to control or operation of the camera
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/24—Indexing scheme for image data processing or generation, in general involving graphical user interfaces [GUIs]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30244—Camera pose
Definitions
- the present invention relates to a commodity photographing device, a commodity photographing method, an image conversion device, an image processing device, an image processing system, a program, and an information recording medium, and enables a user to easily photograph an entire surrounding image row of commodities. It is a thing.
- An omnidirectional image sequence consisting of a plurality of images obtained by photographing the object sequentially from the periphery of the object generates a three-dimensional shape model of the object (for example, see Non-Patent Document 1) and the appearance of goods. It can be used in various applications, such as making it easy to understand for consumers.
- the omnidirectional image is obtained by sequentially capturing the situation of the real world including the object while moving the imaging direction of the camera toward the object on the circumference centered on the object.
- a lens for photographing collects external light including a subject and forms an image of the external world on an image sensor such as a CCD (Charge Coupled Device).
- CCD Charge Coupled Device
- This electronic file corresponds to a sensitized film or a developed printing paper in a conventional optical camera.
- live preview the image currently detected by the image sensor is displayed as it is on a screen such as a liquid crystal display in real time (see, for example, Non-Patent Document 2). That is, in the live preview, an electronic finder using a liquid crystal display or the like is used.
- the image displayed on the finder screen in the live preview corresponds to "an image that would be obtained if the user has just pressed the shutter at this moment".
- the user can confirm the composition, exposure, etc. of the subject before shooting.
- the present invention solves the problems as described above, and is a commodity photographing apparatus, a commodity photographing method, an image conversion apparatus, and an image processing suitable for enabling a user to easily photograph an all-around image sequence of a commodity.
- An object is to provide an apparatus, an image processing system, a program, and an information recording medium.
- the commodity photographing apparatus is An image sensor unit that senses light projected from the outside where the product is placed and outputs an image representing the result of the sensing;
- An instruction receiving unit that receives an imaging instruction from the user,
- a storage unit configured to store an image sensed by the image sensor unit when the photographing instruction is received;
- a finder display unit is configured to combine the image stored by the storage unit and the image sensed by the image sensor unit and display the combined image on a finder screen.
- the finder display unit may be configured to alternately display the image stored by the storage unit and the image sensed by the image sensor unit on the finder screen.
- the finder display unit may be configured to make a result obtained by applying a predetermined image filter to the image stored by the storage unit to be translucent, and to superimpose it on the image sensed by the image sensor unit and combine them. it can.
- the finder display unit may be configured to combine the image stored first with the image sensed by the image sensor unit when there are a plurality of images stored by the storage unit. .
- the apparatus further comprises a tilt sensor unit that senses the tilt of the product imaging device
- the storage unit stores the image sensed by the image sensor unit in association with the tilt sensed by the tilt sensor unit when the photographing instruction is received
- the display unit may be configured to display, on the finder screen, the inclination stored in association with the image stored by the storage unit and the inclination detected by the inclination sensor unit. it can.
- An image conversion apparatus is An image reception unit that receives a plurality of images captured and stored by the above-mentioned product photographing device from the periphery of the product and an inclination associated with the image; A conversion unit configured to convert the plurality of received images into an all-around image sequence of the product based on the inclination associated with the images is provided.
- An image processing apparatus is An angle reception unit that receives an angle around a predetermined rotation axis, An interpolation unit that interpolates an all-surrounding image sequence of a product obtained by the above image conversion device based on the received angle, and obtains an interpolated image representing the product viewed from the received angle To provide.
- the image processing apparatus is An angle sensor unit that senses an angle around a predetermined axis of the image processing apparatus; It further comprises a monitor display unit for displaying the obtained interpolated image on a monitor screen, The angle receiving unit may be configured to receive the sensed angle.
- An image processing system is The image processing apparatus described above and a terminal device communicably connected to the image processing apparatus are provided.
- the terminal device is An angle sensor unit for detecting an angle around a predetermined axis of the terminal device;
- An angle transmitting unit is provided for transmitting the sensed angle to the image processing apparatus.
- the image processing apparatus An angle reception unit that receives the angle transmitted from the terminal device and causes the angle reception unit to receive the angle; An image transmission unit is provided for transmitting the obtained interpolated image to the terminal device.
- the terminal device An image receiving unit for receiving the interpolated image transmitted from the image processing apparatus;
- the monitor display unit displays the received interpolated image on a monitor screen.
- the commodity photographing method is executed by a commodity photographing device provided with an image sensor unit, an instruction receiving unit, a storage unit, and a display unit, An image sensor process in which the image sensor unit senses light projected from the outside where goods are placed and outputs an image representing the result of the sensing; An instruction accepting step of the instruction accepting unit accepting an imaging instruction from a user; A storage step of storing the image sensed by the image sensor unit when the photographing instruction is received;
- the display unit may be configured to include a finder display step of combining an image stored by the storage unit and an image sensed by the image sensor unit and displaying the synthesized image on a finder screen.
- a program is a computer, An image sensor unit that senses light projected from the outside where the product is placed and outputs an image representing the result of the sensing; An instruction receiving unit that receives an imaging instruction from the user, A storage unit configured to store an image sensed by the image sensor unit when the photographing instruction is received; The image stored by the storage unit and the image sensed by the image sensor unit are combined to function as a finder display unit for displaying on a finder screen.
- a computer readable information storage medium is a computer.
- An image sensor unit that senses light projected from the outside where the product is placed and outputs an image representing the result of the sensing;
- An instruction receiving unit that receives an imaging instruction from the user,
- a storage unit configured to store an image sensed by the image sensor unit when the photographing instruction is received; It is configured to record a program characterized by causing the image stored by the storage unit and the image sensed by the image sensor unit to be combined and displayed on a finder screen.
- the program of the present invention can be recorded on a computer readable information recording medium such as a compact disk, a flexible disk, a hard disk, a magneto-optical disk, a digital video disk, a magnetic tape, a semiconductor memory and the like.
- a computer readable information recording medium such as a compact disk, a flexible disk, a hard disk, a magneto-optical disk, a digital video disk, a magnetic tape, a semiconductor memory and the like.
- the program can be distributed and sold via a computer communication network independently of the computer on which the program is executed. Also, the information recording medium can be distributed and sold independently of the computer.
- a commodity photographing device a commodity photographing method, an image conversion device, an image processing device, an image processing system, a program, and a commodity photographing method suitable for enabling the user to easily photograph the entire surrounding image sequence of a commodity.
- An information recording medium can be provided.
- FIG. 1 is an explanatory view showing a schematic configuration of an image processing system according to an embodiment of the present invention. It is an explanatory view explaining a schematic structure of a goods photographing device concerning one embodiment of the present invention. It is explanatory drawing which shows the relationship between a photography completed image, a live preview image, and the image displayed on a finder screen. It is explanatory drawing which shows the relationship between a photography completed image, a live preview image, and the image displayed on a finder screen. It is a flowchart which shows the flow of control of the goods imaging process performed with the goods imaging device which concerns on one Embodiment of this invention. It is an explanatory view showing a schematic structure of an image conversion device concerning one embodiment of the present invention.
- FIG. 1 is an explanatory view showing a schematic configuration of an image processing system according to the present embodiment. This will be described below with reference to this figure.
- An image processing system 101 includes a product photographing device 121, an image conversion device 141, an image processing device 161, and a terminal device 181.
- the product photographing device 121 is used by a shop owner who wants to introduce a product, an auction seller, etc. when photographing the product.
- the product photographing device 121 is configured by a digital camera, a mobile phone, a smart phone, etc. having an information processing function, but can also be configured by a personal computer etc. to which a USB (Universal Serial Bus) camera is connected. It is.
- USB Universal Serial Bus
- the photographer uses the product photographing device 121 to arrange the product in the middle of the field of view as much as possible, so that the distance between the product and the product photographing device 121 is as constant as possible. Then, the photographer repeatedly captures images while rotating the product or rotating the product itself to obtain a plurality of product images.
- the image conversion device 141 converts a plurality of product images obtained by shooting by the product shooting device 121 into an all-around image sequence for the product.
- the all-surrounding image row is obtained when it is assumed that the product is photographed in the optical axis direction facing the product from a camera moving on a circumference centered on the rotation axis passing through the product. Collection of images.
- the angle from the rotation axis to the camera through the goods is constant, the distance between the rotation axis and the camera is also constant, and the image assumed to be taken in a situation where the distance between the goods and the camera is constant is all around It is an image.
- the photographer using the commodity photographing apparatus 121 is neither skilled at such a special photographing method nor has a special instrument. Therefore, in the present embodiment, when using the product photographing device 121, various assistances are performed so that a row of all-surrounding images can be obtained as easily as possible.
- an electronic finder such as a liquid crystal display is used for live preview.
- the finder screen of the electronic finder displays the current state of the product as viewed from the product photographing device 121, and also displays the state of the product already photographed in the past.
- the photographer can shoot while comparing with the past history what size and at what angle the product is shot. Therefore, it is possible to make the angle from the rotation axis through the product to the product imaging device 121 and the distance between the product and the product imaging device 121 as constant as possible.
- the image conversion device 141 converts a plurality of product images into an all-surrounding image sequence
- the conversion result obtained in the present embodiment is a certain approximation, and complex calculations as disclosed in Non-Patent Document 1
- This embodiment is also characterized in that it does not require the.
- the image conversion device 141 obtains the all-surrounding image sequence from the product image, and the image processing device 161 manages the all-surrounding image sequence of the product.
- the image processing device 161 manages the all-surrounding image sequence of the product.
- both are the products of the product. It is comprised by one server apparatus which performs net sales.
- the shop owner or the like uploads a plurality of product images photographed by the product photographing device 121 such as a digital camera to the server apparatus.
- the server device functions as the image conversion device 141 to convert the product image into an all-around image sequence.
- the server device functions as the image processing device 161 to allow the user who is examining whether to purchase the commodity in the net mail order or the net auction the entire surrounding image sequence.
- appropriate image processing that attracts the user's interest is performed instead of providing the entire surrounding image sequence as it is to the user.
- the terminal device 181 functions as a web terminal to request the image processing device 161 to properly provide the entire surrounding image sequence. That is, the terminal device 181 is configured by a personal computer, a mobile phone, a smartphone, etc. that can be connected to a computer communication network such as a web.
- the all-surrounding image row is a row in which the images appear as they appear as they go around the product.
- the terminal device 181 notifies the image processing apparatus 161 at which angle of this round the user wants to see an image.
- the image processing device 161 appropriately performs interpolation processing on the entire surrounding image to generate an interpolated image representing the appearance of the product according to the angle.
- the interpolation image is transmitted from the image processing device 161 to the terminal device 181, and the interpolation image is displayed on the monitor screen of the terminal device 181.
- the terminal device 181 can also be configured to allow the user to directly input an angle using a keyboard or a touch panel.
- a keyboard or a touch panel there are cellular phones, smart phones, and the like, which have a compass function, a GPS (Global Positioning System) function, and an inclination detection function using an acceleration sensor or the like.
- GPS Global Positioning System
- the present embodiment is also characterized in that the user can observe the appearance of the product all around in a simple operation.
- the image processing apparatus 161 and the terminal apparatus 181 are configured by different devices, but both may be integrally configured by one personal computer, a smartphone, a mobile phone, or the like.
- transmission and reception of information between the terminal device 181 and the image processing device 161 are realized by transmission and reception of electric signals in the electronic circuit of one device.
- an all-around image sequence of the product is downloaded from the web server constituting the image conversion device 141 to the smartphone.
- the smart phone functions as the image processing device 161, and the smart phone generates the interpolation image from the entire surrounding image sequence according to the inclination thereof, and displays it on the monitor screen.
- FIG. 2 is an explanatory view illustrating a schematic configuration of the product photographing device 121 according to the present embodiment. This will be described below with reference to this figure.
- the product photographing device 121 includes an image sensor unit 201, an instruction receiving unit 202, a storage unit 203, and a finder display unit 204. Further, the tilt sensor unit 205 may be provided.
- the image sensor unit 201 senses the light projected from the outside where the product is placed, and outputs an image representing the result of the sensing.
- the image sensor unit 201 is configured of an image sensor such as a CCD.
- the light sensed by the image sensor unit 201 is reached through a lens and various apertures.
- the instruction receiving unit 202 receives an imaging instruction from the user.
- the instruction receiving unit 202 corresponds to a so-called shutter button.
- the storage unit 203 stores the image sensed by the image sensor unit 201 when a photographing instruction is received.
- the storage unit 203 generally records a photographed image file on a non-volatile information storage medium such as a flash memory card or a hard disk configured by an EEPROM (Electrically Erasable Programmable Read Only Memory) or the like.
- a non-volatile information storage medium such as a flash memory card or a hard disk configured by an EEPROM (Electrically Erasable Programmable Read Only Memory) or the like.
- the finder display unit 204 displays the image sensed by the image sensor unit 201 on the finder screen.
- the photographer performs adjustment of the position and orientation of the product photographing device 121, focusing, and adjustment of the aperture while referring to the image displayed on the finder screen.
- the finder display unit 204 is characterized in that when the image sensed by the image sensor unit 201 is displayed on the finder screen, it is combined with the image saved by the storage unit 203.
- an image representing a field of view already captured in the past and an image representing a field of view to be currently captured are displayed superimposed on the finder screen. Therefore, the photographer can adjust the position and orientation of the product photographing device 121 while comparing the two images in real time.
- the simplest method as a composition method for superimposing two images is an image (live preview image) sensed by the image sensor unit 201 with the image (photographed image) saved by the saving unit 203 being translucent. ) Is superimposed on the live preview image and displayed on the finder screen.
- FIG. 3 and FIG. 4 are explanatory diagrams showing the relationship between the photographed image, the live preview image, and the image displayed on the finder screen.
- description will be made with reference to these figures.
- a dotted line shall be used suitably.
- FIG. 3 shows a photographed image 311 in which the product 301 is disposed substantially in the center, and a live preview image 321 in which the product 301 is disposed substantially in the center. In both cases, the angle of the product 301 is slightly different.
- a finder image 331 obtained as a result of superimposing the captured image 311 semi-transparently and superimposing it on the live preview image 321 the images of the two products 301 are displayed with almost the same size so as to overlap. Therefore, when the shutter is released in the current state, the error can be reduced when converting into the entire surrounding image sequence.
- the product 301 is arranged substantially at the center in the photographed image 311, but in the live preview image 321, the position and size of the product 301 are largely different from those of the photographed image 311.
- the finder image 331 obtained as a result of superimposing the photographed image 311 semi-transparent and superimposing it on the live preview image 321 the images of the two products 301 hardly overlap. Therefore, even if the shutter is released in the current state and a new captured image 311 representing the same state as the live preview image 321 is obtained, it is considered that the error becomes large when converted into the entire surrounding image sequence.
- the photographer shoots the product so that the image of the product 301 taken in the past and the image of the product 301 currently viewed are as similar in size as possible and slightly different in angle
- an instruction input for photographing may be given.
- the photographer can easily distinguish which one of the finder images 331 is an image derived from the photographed image 311 and which is an image derived from the live preview image 321.
- Another possible method is to display the photographed image 311 and the live preview image 321 alternately on the finder screen. As described above, whether the image displayed on the finder screen is the live preview image 321 or the photographed image 311 can be distinguished depending on whether the display on the finder screen follows the movement of the photographer or is fixed.
- the time length for displaying both images can be set arbitrarily.
- the display switching time of both is set so that the display time of the live preview image 321 is longer than the display time of the photographed image 311.
- the tilt sensor unit 205 senses the tilt of the product photographing device 121, and when the storage unit 203 receives a shooting instruction, the tilt detected by the sensor unit is an image sensed by the image sensor unit 201. Associate with and save.
- the tilt sensor unit 205 can be configured by an acceleration sensor.
- the inclination may be detected by a function such as GPS or compass.
- the inclination sensor unit is detected by detecting an inclination of the product imaging device 121 with respect to the axis of gravity direction by the acceleration sensor and detecting a rotation amount of the product imaging device 121 about the axis of gravity direction using GPS or a compass.
- all inclinations of the product imaging device 121 in the three axial directions can be detected. This makes it possible to grasp all the postures of the product photographing device 121 in the real world.
- the information on the inclinations in these three axial directions is referred to when the captured image 311 is normalized to obtain an all-surrounding image sequence by the image conversion device 141 described later.
- the angle from the rotation axis through the product 301 to the product imaging device 121 be constant, that is, the inclination with respect to gravity be constant.
- the finder image 331 displays the currently detected tilt of the product photographing device 121, that is, the information on the tilt of the live preview image 321 and the information on the tilt at the time of shooting the shot image 311. If so, the photographer can capture the product 301 at a constant angle as much as possible, and the error in conversion to the entire surrounding image sequence can be reduced.
- a bar graph 322 representing the inclination of the live preview image 321 and a bar graph 312 representing the inclination of the photographed image 311 are displayed in the finder image 331.
- the photographer may try to shoot after the lengths of the bar graphs 312 and 322 match.
- the method of expressing the inclination is not limited to the bar graphs 312 and 322, and any figure that can express the degree of coincidence and non-coincidence can be adopted. Further, the numerical value of the inclination may be displayed by character information.
- FIG. 5 is a flow chart showing the flow of control of the commodity photographing process executed by the commodity photographing apparatus 121 according to the present embodiment. This will be described below with reference to this figure.
- This process is started by the user selecting to start shooting an image to be a material of the all-surrounding image sequence of the product in a digital camera or the like configuring the product shooting device 121.
- the image sensor unit 201 senses light projected from the outside where the commodity is arranged, and outputs an image representing the result of the sensing (step S501).
- the image output here is the live preview image 321.
- the tilt sensor unit 205 acquires the current tilt of the product photographing device 121 (step S502).
- step S503 it is determined whether a photographing instruction has been issued from the user (step S503).
- step S503 If a shooting instruction has been issued from the user (step S503; Yes), the saving unit 203 saves the live preview image 321 and the inclination detected in step S502 in association with each other (step S504). After this, the control returns to step S501.
- step S502 when the user has not issued a shooting instruction (step S502; No), it is checked whether there is an image stored in the storage unit 203 after the start of the process, ie, the photographed image 311 Step S505).
- step S506 If the photographed image 311 exists (step S505; Yes), one of the photographed images 311 is selected (step S506). As described above, this selection may be an image captured first, may be changed as appropriate by the user during processing, or may be an image captured immediately before.
- the selected photographed image 311 is made translucent and is combined with the live preview image 321 (step S507).
- the photographed image 311 may be combined with a color filter, an outline extraction filter, or the like.
- one of the images may be selected by time division, and the selected image may be used as the result of the synthesis.
- step S502 bar graphs 312 and 322 representing the inclination associated with the selected photographed image 311 and the current inclination of the product photographing device 121 acquired in step S502 are drawn in the synthesis result of step S506, A finder image 331 is obtained (step S508).
- the finder display unit 204 displays the finder image 331 on a finder screen configured of a liquid crystal display or the like (step S509), and the process returns to step S501.
- step S505 when the photographed image does not exist (step S505; No), the bar graph 322 representing the current inclination of the product photographing device 121 acquired in step S502 is drawn on the live preview image 321 to obtain the finder image 331 ( Step S510).
- the finder display unit 204 displays the finder image 331 on a finder screen configured of a liquid crystal display or the like (step S511), and returns to step S501.
- the processing relating to the tilt is also appropriately omitted.
- the inclination with respect to the gravity of the product photographing device 121 is drawn in the finder image 311 as bar graphs 312 and 322.
- the image of the product photographed in the past and the image of the live preview of the current product are superimposed and displayed on the finder screen.
- the distance between the image capturing device 121 and the image capturing device 121 can be kept as constant as possible.
- the image capturing device 121 can be moved so as to rotate in a posture as constant as possible with respect to a predetermined rotation axis.
- the user can easily capture an image that can be converted with high quality into the entire surrounding image sequence of the product.
- the image capturing device 121 can be wrapped around the product 301 while keeping the distance to the product 301 and the inclination with respect to the direction of gravity as constant as possible. By moving the product 301, the product 301 can be photographed sequentially.
- the angle from the rotation axis to the camera through the product 301 is kept constant, and the distance between the product 301 and the camera Is a row of images obtained when it is assumed that the product 301 is taken one after another by rotating the camera around the rotation axis while keeping.
- the plurality of photographed images 311 photographed by the image photographing device 121 can be considered as "all-round image sequence with large error".
- the image photographing apparatus 121 is devised to photograph the product 301 while keeping the above angle and distance as constant as possible, but in photographing by human beings, an error occurs at the time of photographing.
- the image conversion device 141 converts a plurality of photographed images 311 to obtain an all-around image sequence with the smallest possible error.
- Non-Patent Document 1 in which a three-dimensional model of a photographing object is constructed, it is possible to obtain an all-surrounding image sequence by a simple calculation. The details will be described below.
- FIG. 6 is an explanatory view showing a schematic configuration of the image conversion device 141 according to the present embodiment. This will be described below with reference to this figure.
- An image conversion apparatus 141 includes an image reception unit 601 and a conversion unit 602.
- the image receiving unit 601 receives a plurality of photographed images 311 stored and photographed by the product photographing device 121 from the periphery of the product.
- the photographer uploads the photographed image 311 photographed by the photographer via the Internet or the like.
- the image receiving unit 601 receives this upload.
- a mode may be adopted in which the product photographing device 121 automatically transmits the photographed image 311 to the image conversion device 141.
- the commodity photographing apparatus 121 automatically transmits the photographed image 311 every time the photographed image 311 is stored for a predetermined number of times or when the photographing is not performed for a predetermined time.
- This aspect is particularly suitable for realizing the commodity imaging apparatus 121 by executing a predetermined program in a portable computer such as a PDA (Personal Data Assistant) or a smartphone.
- the captured image 311 is moved by inserting and removing a memory card between the two.
- the conversion unit 602 converts the received plurality of photographed images 311 into an all-around image sequence of the product 301.
- the conversion unit 602 of the present embodiment further includes an extraction unit 611, an estimation unit 612, an approximation unit 613, and an output unit 614.
- the extraction unit 611 acquires two captured images 311 in order. Then, corresponding feature points are extracted from the two captured images 311 acquired. At this time, various image recognition algorithms can be applied.
- the position and the orientation of the product photographing device 121 when the two photographed images 311 are photographed are estimated from the arrangement of the characteristic points in the two photographed images 311. Then, an affine transformation matrix for normalizing each of the two photographed images 311 is obtained.
- the affine transformation used here includes perspective transformation by one-point perspective in addition to scaling, rotation and skew transformation. Therefore, when expressing a matrix, homogeneous coordinates are appropriately used.
- the approximation unit 613 applies an affine transformation based on the affine transformation matrix obtained for each of the two photographed images 311 to the photographed image 311 to normalize the same.
- the normalized image approximates the image to be included in the entire surrounding image sequence.
- the output unit 614 outputs the two normalized images obtained here, and how much difference angle they were captured around the rotation axis.
- FIG. 7 is a flowchart showing the flow of control of the image conversion process performed by the image conversion device 141 according to the present embodiment. This will be described below with reference to this figure.
- the image receiving unit 601 receives a plurality of photographed images 311 photographed and stored by the product photographing device 121 and inclinations associated with the photographed images 311 (step S701).
- the number of photographed images 311 is N
- the photographed images are denoted as G [0], G [1],..., G [N-1].
- the inclinations when the images G [0], G [1],..., G [N-1] are captured are denoted as R [0], R [1], ..., R [N-1].
- subscripts expressed by square brackets are expressed by the remainder class of N. That is, when the integer at the subscript part is negative or N or more, the remainder obtained by dividing this by N is used as the value of the subscript.
- the slope R [N] means the same as the slope R [0]
- the image G [-1] means the same as the image R [N-1].
- the inclinations R [0], R [1],..., R [N ⁇ 1] are information for determining the direction of each coordinate axis of the camera coordinate system fixed to the product photographing device 112. Therefore, the degrees of freedom of the slopes R [0], R [1],..., R [N-1] are six.
- step S702 the integer k is changed between 0 and N, and the following processing is repeated (step S702).
- the extraction unit 611 extracts corresponding feature points from the image G [k] and the image G [k + 1] (step S703).
- the estimation unit 612 selects any three sets of the extracted sets of feature points (step S704).
- the method of selecting these three sets is arbitrary, in the image G [k], the distance between corresponding points to be selected is large, and the correspondence selected also in the image G [k + 1]. If the points are selected to be separated, the error can be suppressed. Any three of the feature points near the top, bottom, left center, and right center of the product may be selected.
- the three feature points are hereinafter referred to as P, Q and W, respectively.
- the positions where these feature points are drawn in the image G [k] are p [k], q [k], w [k], and these feature points are drawn in the image G [k + 1]. It is assumed that the existing position is expressed as p [k + 1], q [k + 1], w [k + 1].
- the estimation unit 612 (1) Positions p [k], p [k + 1] at the images G [k], G [k + 1] of the feature point P, and (2) Positions q [k], q [k + 1] at positions G [k], G [k + 1] of the feature point Q, and (3) Positions w [k], w [k + 1] at the images G [k], G [k + 1] of the feature point W, and (4) The inclination R [k] of the product photographing device 121 when photographing the image G [k], (5) The inclination R [k + 1] of the product photographing device 121 when photographing the image G [k + 1], From (X) The position T [k] of the product photographing device 121 when photographing the image G [k], (Y) The position T [k + 1] of the product photographing device 121 when photographing the image G [k + 1], Is estimated (step S 705).
- the direction vectors of the three coordinate axes in the camera coordinate system are set as rx, ry, and rz, respectively (equations 1 to 3).
- rx1, rx2, ..., rz3 are all coordinate values of the world coordinate system, and represent vectors rx, ry, rz.
- T the position vector of the origin of the camera coordinate system with respect to the origin of the world coordinate system (Equation 4).
- the numerical values t1, t2 and t3 are also coordinate values of the world coordinate system of the position vector T.
- the matrix R is a matrix that represents the direction of the coordinate axis of the camera coordinate system. That is, this matrix is obtained from the detection results of the gravity sensor of the tilt sensor unit 205, the compass, and the like when shooting is performed by the product shooting device 121.
- M is a 3-by-4 matrix.
- the projection plane is assumed to be an ideal pinhole camera, fixed to the camera coordinate system, and the first axis of the two-dimensional coordinate system is parallel to the x-axis, and the two-dimensional coordinate system The second axis of is assumed to be parallel to the y axis.
- the matrix A can be expressed as follows using appropriate constants k1, k2, uc1, and uc2 (Equation 13).
- the same affine transformation matrix M as described above can be obtained by appropriately determining the matrix A.
- M [k] and M [k + 1] are 3-by-4 matrices
- M [k] -1 and M [k + 1] -1 are M [k] and M [k], respectively. It is assumed that it is a generalized inverse matrix of +1].
- the generalized inverse is sometimes referred to as the "Moore-Penrose pseudoinverse".
- the number of unknowns is nine in total: six of the coordinate values of the imaging positions T [k] and T [k + 1] and three of the proportional coefficient of the left side and the right side of each eyelid.
- a solution may be obtained by mathematical expression processing based on symbol processing, or various approximate solution methods may be used. For example, this is a method of obtaining a solution by giving an appropriate initial value to T [k], T [k + 1] and a proportional constant and using the steepest descent method.
- the number of simultaneous equations can be further increased by increasing the number of feature points.
- these unknowns can be obtained by performing maximum likelihood estimation so as to minimize the sum of squares of errors obtained by subtracting the right side from the left side of each simultaneous equation.
- the imaging positions T [k] and T [k + 1] are determined by selecting three corresponding feature points in the image. Conversely, if the shooting positions T [k] and T [k + 1] are known, affine transformation matrices M [k] and M [k + 1] in perspective projection of a single point perspective can be obtained.
- step S 707 When these processes are repeated (step S 707), by step S 705 (x), Estimated values of T [0], T [1], T [2], ..., T [N-1] are obtained, and in step S705 (y), Estimated values of T [1], T [2],..., T [N-1], T [0] are obtained.
- the position T [0], T [1], T [2],..., T [N-1] of the product photographing device 121 can be obtained by selecting one of them or finding the average of the two. Are identified (step S 707).
- step S 708 the position H of the product 301 in the real world is identified by the following method (step S 708).
- the optical axis direction for imaging is acquired from the inclination R [k]. Then, a straight line extending in the optical axis direction is assumed to pass through the shooting position T [k]. These N straight lines should pass through the position H of the product 301 or in the vicinity thereof.
- the position H of the product 301 can be determined by the method of least squares so that the sum of squares of the distances between the position H of the product 301 and these N straight lines is minimized.
- the positions T [0], T [1], T [2], ..., T [N-1] of the identified photographing are It is considered to be disposed circumferentially around a central axis passing through the position H of the product 301. Also, when a person takes a picture of a product, the picture is not necessarily taken from the side, but often taken from diagonally above.
- the estimated imaging positions T [0], T [1], T [2],..., T [N-1] are arranged on the bottom of the cone having the position H as the apex. It should be. Also, when shooting from the side, the estimated shooting positions T [0], T [1], T [2], ..., T [N-1] are on the circumference centered on the position H. It should be deployed.
- the position U [k] is on a straight line connecting the position H of the product 301 and the photographing position T [k], and the position U [k] and the photographing position T [k] are separated by sc [k] There is.
- the constraint condition is to minimize sc [0], sc [1], sc [2],..., Sc [N-1].
- the position J and the radius r of the center of the circumference to be the base of the cone are calculated.
- the center of gravity of the position U [k] is the simplest.
- the simplest is to average the distance between the position J and each U [k].
- the circumference has a straight line connecting the position H and the position J as the rotation axis.
- the simplest method is to set the position closest to the photographing position T [k] to S [k].
- the vector OP [k] ((HJ) ⁇ (T [k] ⁇ J)) ⁇ (HJ) is calculated from the vector HJ and the vector T [k] ⁇ J.
- the vector (HJ ⁇ T [k] -J) is orthogonal to both the vector HJ and the vector T [k] -J.
- the vector OP [k] ((HJ) ⁇ (T [k] ⁇ J)) ⁇ (HJ) is parallel to the plane spanned by the vector HJ and the vector T [k] -J, and the vector It is orthogonal to HJ.
- the normalized positions S [0], S [1],..., S [N-1] are arranged on a predetermined circumference. Therefore, on this circumference, how much each of S [0], S [1], ..., S [N-1] rotates with respect to the center J, as viewed from S [0] It is also possible to ask for.
- the rotation angles of the normalized positions S [0], S [1],..., S [N-1] are respectively ⁇ [0], ⁇ [1],.
- ⁇ [k] is the angle at which the vector S [k] -J spans the vector S [0] -J.
- the inclination V [k] of virtual imaging at the virtualized position S [k] is calculated (step S 710).
- the vector HS [k] directed from the virtualized position S [k] to the position H of the product 301 is taken as the optical axis direction (z axis) for shooting,
- the tangent direction of the circumference at the virtualized position S [k] be the left and right direction (x axis) of shooting
- a component orthogonal to HS [k] of the vector JS [k] from the virtualized position S [k] to the center position J of the circumference is taken as the vertical direction (y axis) of imaging.
- the proportionality constant of this equation is the z-axis in the camera coordinate system of the point P, that is, the coordinate value pc3 of the optical axis in the photographing direction, as shown in [Equation 10].
- An affine transformation matrix L [k] is calculated which performs transformation from the actual position to the position in the image (step S711).
- the output unit 614 outputs these as information on the entire surrounding image sequence (step S713), and the main processing is performed. Finish.
- step S719 the affine transformation matrix (1 / d) L [0], (1 / d) L [1], (1 / d) performs conversion from the position in the real world to the position in the normalized image. ) It is desirable to output L [2],..., (1 / d) L [N-1] as part of the information of the entire surrounding image sequence.
- the output all-surrounding image sequence, rotation angle, and affine transformation matrix information is stored in a hard disk or the like in the server device, it is downloaded from here to a smartphone, a mobile phone, a personal computer, etc. It is also possible.
- the photographed image is obtained by photographing the product 301 while moving the product photographing device 121 on the substantially circumference of the bottom surface of the substantially conical shape with the product 301 as the vertex.
- the error included in the result of shooting is corrected, and it may be obtained by shooting the product 301 while moving the product photographing device 121 on the positive circumference of the bottom of the regular cone with the product 301 as the vertex. It is supposed to obtain a normalized image.
- the commodity photographing apparatus 121 moves on a substantially circumference centered on the commodity 301, and the normalized image is viewed from a positive circumference centered on the commodity 301. It may be assumed that the In this method, further speeding up of the calculation can be expected because the constraint conditions are greatly simplified as compared with the above description.
- the position of the product 301 and the shape of the cone may be determined using various maximum likelihood estimation techniques.
- the photographing positions T [k], T [k + 1], the inclinations R [k], R [k +] can be obtained by simpler approximation calculation. 1], Affine transformation matrixes M [k], M [k + 1] are calculated with respect to a method for obtaining them.
- the proportional constant pc3 represents the coordinate value of the point to be photographed on the optical axis in the photographing direction, that is, the depth from the camera to the point to be photographed.
- the photographing device 121 when used, the photographing is performed so that the size of the product 301 is substantially constant in the photographed image. Therefore, the depths to the product 301 are considered to be substantially equal regardless of the shooting position.
- the specific calculation can adopt the same method as the above embodiment.
- the number of feature points to be referred to in order to obtain the imaging position increases, but the computational accuracy improves.
- the number of feature points is three or more, the shooting positions T [k], T [k + 1], the inclination even if there is no special condition in the position of the product 301 with respect to each shooting position T [k].
- R [k], R [k + 1], and affine transformation matrices M [k], M [k + 1] can be obtained.
- the calculation load is reduced if the number of feature points is two. While the approximation is used, the photographing positions T [k], T [k + 1], the slopes R [k], R [k + 1], the affine transformation matrices M [k], M [can be obtained with high precision while using approximation. k + 1] can be obtained.
- the normalized images B [0], B [1], B [2],..., B [N-1] included in the all-surrounding image sequence are the rotation angles ⁇ [0], ⁇ around the product 301, respectively. [1], ⁇ [2],..., ⁇ [N ⁇ 1].
- 0 ⁇ [0] ⁇ ⁇ [1] ⁇ ⁇ [2] ⁇ ... ⁇ ⁇ [N ⁇ 1] ⁇ 2 ⁇
- 0 ⁇ [0] ⁇ ⁇ [1] ⁇ ⁇ [2] ⁇ ... ⁇ ⁇ [N ⁇ 1] ⁇ 2 ⁇
- the image processing device 161 When given an arbitrary rotation angle ⁇ from the user, the image processing device 161 has a function of presenting to the user by interpolating the image when viewing the product 301 from the position corresponding to the rotation angle ⁇ from the information. Play.
- the simplest method of interpolation is as follows.
- the affine transformation matrix (1 / d) L ( ⁇ ) is uniquely determined from the position S ( ⁇ ), the slope V ( ⁇ ), and the length d of the hypotenuse of the cone.
- an interpolated image B ( ⁇ ) for the rotation angle ⁇ is obtained from the normalized image B [k].
- the simplest method is a method of using the normalized image B [k] as the interpolation image B ( ⁇ ) as it is. This is effective in the sense of suppressing the amount of calculation if N is sufficiently large (for example, 20 or more).
- an affine transformation matrix (1 / d) L [k] for projecting the real world onto the image and an affine transformation matrix for performing the inverse thereof d L [k] -1 is associated.
- the left side ( ⁇ [k]) / ( ⁇ [k + 1] ⁇ [k]) of the image obtained by applying the affine transformation to B [k] is extracted, and B [k + 1) extract the right side ( ⁇ [k + 1] ⁇ ) / ( ⁇ [k + 1] ⁇ [k]) of the image obtained by applying the affine transformation to the left and right And the interpolation image B ( ⁇ ).
- the direction is determined so that ⁇ and ⁇ increase when the side close to the user rotates from left to right, the same method can be adopted by changing the direction even if the direction is reversed.
- the normalized image is cut in a strip shape perpendicular to the rotation direction and extracted.
- Interpolated image B ( ⁇ ) can be obtained by arranging
- FIG. 8 is an explanatory view showing the relationship between each element of the image processing device 161 according to the present embodiment and each element of the terminal device 181.
- FIG. 9 is a flow chart showing a flow of control of image processing performed in each element of the image processing device 161 and each element of the terminal device 181.
- the image processing apparatus 161 and the terminal apparatus 181 can be configured integrally, or can be configured as separate devices. In addition, it is possible to arbitrarily change which device each part is disposed. Therefore, this figure illustrates the most typical aspect.
- the image processing apparatus 161 includes an angle reception unit 801, an angle reception unit 802, an interpolation unit 803, and an image transmission unit 804.
- the terminal device 181 includes an angle sensor unit 811, an angle transmission unit 812, an image reception unit 813, and a monitor display unit 814.
- server processing is performed in the image processing apparatus 161, and terminal processing is performed in the terminal device 181.
- the angle sensor unit 811 senses the angle ⁇ around the predetermined axis of the terminal device 181 (step S901).
- the product imaging device 121 When shooting a product with the product imaging device 121, it is generally used to detect inclination of three axes, but in the terminal device 181, only one acceleration sensor is used to detect inclination with respect to the direction of gravity. It is enough to do. Of course, inclinations of the three axes may be detected.
- the angle transmission unit 812 transmits the sensed angle ⁇ to the image processing device 161 (step S902).
- the angle reception unit 801 receives the angle ⁇ transmitted from the terminal device 181 and causes the angle reception unit 802 to receive it (step S903).
- the interpolation unit 803 obtains an interpolated image B ( ⁇ ) from the received angle ⁇ based on the above algorithm (step S904).
- the image transmission unit 804 transmits the obtained interpolated image B ( ⁇ ) to the terminal device 161 (step S905), and the process returns to step S903.
- the image receiving unit 813 receives the interpolated image B ( ⁇ ) transmitted from the image processing device 161 (step S906).
- the monitor display unit 814 displays the received interpolated image B ( ⁇ ) on the monitor screen (step S 907), and the process returns to step S 901.
- the product 301 displayed on the monitor screen of the terminal device 181 is also rotated in conjunction with the tilt, and the user can easily change the product 301 You can see the whole surroundings of
- the angle ⁇ detected by the angle sensor unit 811 is given to the interpolation unit 803, and the interpolation image obtained by the interpolation unit 803 B ( ⁇ ) is given to the monitor display unit 814, and the interpolated image B ( ⁇ ) is displayed on the monitor screen.
- the terminal device 181 includes an angle sensor unit 811, an interpolation unit 803, and a monitor display unit 814.
- the terminal device 181 when the terminal device 181 is realized by a general personal computer having no tilt sensor, the user does not use the angle sensor unit 811 to sense the angle ⁇ , but the user inputs an input such as a keyboard or a mouse.
- a mode may be adopted in which the angle ⁇ is input by operating the device. For example, an operation system in which ⁇ is increased by pressing the first button of the keyboard and ⁇ is decreased by pressing the second button, or the value of ⁇ is linked to the movement amount of the mouse. An operation system can be adopted.
- the interpolated image B ( ⁇ ) is displayed by using the sensed angle ⁇ as it is.
- the interpolation image B (k ⁇ ) may be displayed for the sensed angle ⁇ using a constant k larger than one. That is, the angle receiving unit 802 multiplies the received angle by k.
- the user In the method of displaying the interpolated image B ( ⁇ ), in order to observe the entire periphery of the product 301, the user needs to rotate the terminal device 181 itself by 360 degrees.
- the amount of rotation of the terminal device 181 required to observe the entire periphery of the product 301 is only (360 / k) degrees.
- the user can observe the entire circumference of the product 301 by changing the tilt of the terminal device 181 by 90 degrees.
- the process of multiplying the angle by k may be performed, for example, in the terminal device 181.
- the angle transmitting unit 812 multiplies the detected angle ⁇ by k and transmits the angle k ⁇ , and so on.
- the surface of the earth is cut at every 30 degrees of longitude to make 12 wedges.
- the convex polyhedron model of the product 301 is formed by performing the following processing on each polygon arranged on the surface of the convex polyhedron.
- the normalized image B [k] is virtually arranged at the position of the product 301 at the rotation angle ⁇ [k]. Then, with the position S [k] as the viewpoint position and the polygon as the projection plane, the normalized image B [k] is projected onto the polyhedron.
- a commodity photographing device a commodity photographing method, an image conversion device, an image processing device, an image processing system, a program, and a commodity photographing method suitable for enabling the user to easily photograph the entire surrounding image sequence of a commodity.
- An information recording medium can be provided.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Studio Devices (AREA)
- Image Processing (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
商品が配置された外界から投射された光を感知して、当該感知の結果を表す画像を出力する画像センサ部、
ユーザからの撮影指示を受け付ける指示受付部、
前記撮影指示が受け付けられると、前記画像センサ部により感知されている画像を保存する保存部、
前記保存部により保存された画像と、前記画像センサ部により感知されている画像と、を、合成してファインダ画面に表示するファインダ表示部
を備えるように構成する。
前記ファインダ表示部は、前記保存部により保存された画像と、前記画像センサ部により感知されている画像と、を、交互に前記ファインダ画面に表示する
ように構成することができる。
前記ファインダ表示部は、前記保存部により保存された画像に所定の画像フィルタを施した結果を半透明にして、前記画像センサ部により感知されている画像に重ねて合成する
ように構成することができる。
前記ファインダ表示部は、前記保存部により保存された画像が複数存在する場合、最初に保存された画像と、前記画像センサ部により感知されている画像と、を合成する
ように構成することができる。
前記商品撮影装置の傾きを感知する傾きセンサ部
をさらに備え、
前記保存部は、前記撮影指示が受け付けられると、前記画像センサ部により感知されている画像を前記傾きセンサ部により感知されている傾きに対応付けて保存し、
前記表示部は、前記保存部により保存された画像に対応付けられて保存された傾きと、前記傾きセンサ部により感知されている傾きと、を、前記ファインダ画面に表示する
ように構成することができる。
商品をその周囲から上記の商品撮影装置により撮影して保存された複数の画像と、当該画像に対応付けられる傾きと、を受け付ける画像受付部、
前記受け付けられた複数の画像を、当該画像に対応付けられる傾きに基づいて、前記商品の全周囲画像列に変換する変換部
を備えるように構成する。
所定の回転軸周りの角度を受け付ける角度受付部、
前記受け付けられた角度に基づいて、上記の画像変換装置により得られた商品の全周囲画像列を補間して、当該商品を前記受け付けられた角度から見た様子を表す補間済画像を得る補間部
を備えるように構成する。
前記画像処理装置の所定の軸周りの角度を感知する角度センサ部、
前記得られた補間済画像をモニタ画面に表示するモニタ表示部
をさらに備え、
前記角度受付部は、前記感知された角度を受け付ける
ように構成することができる。
上記の画像処理装置と、当該画像処理装置と通信可能に接続される端末装置と、を備える。
前記端末装置の所定の軸周りの角度を感知する角度センサ部、
前記感知された角度を前記画像処理装置に送信する角度送信部
を備える。
前記端末装置から送信された角度を受信して、前記角度受付部に受け付けさせる角度受信部、
前記得られた補間済画像を前記端末装置に送信する画像送信部
を備える。
前記画像処理装置から送信された補間済画像を受信する画像受信部、
前記受信された補間済画像をモニタ画面に表示するモニタ表示部
を備える。
前記画像センサ部が、商品が配置された外界から投射された光を感知して、当該感知の結果を表す画像を出力する画像センサ工程、
前記指示受付部が、ユーザからの撮影指示を受け付ける指示受付工程、
前記撮影指示が受け付けられると、前記保存部が、前記画像センサ部により感知されている画像を保存する保存工程、
前記表示部が、前記保存部により保存された画像と、前記画像センサ部により感知されている画像と、を、合成してファインダ画面に表示するファインダ表示工程
を備えるように構成する。
商品が配置された外界から投射された光を感知して、当該感知の結果を表す画像を出力する画像センサ部、
ユーザからの撮影指示を受け付ける指示受付部、
前記撮影指示が受け付けられると、前記画像センサ部により感知されている画像を保存する保存部、
前記保存部により保存された画像と、前記画像センサ部により感知されている画像と、を、合成してファインダ画面に表示するファインダ表示部
として機能させるように構成する。
商品が配置された外界から投射された光を感知して、当該感知の結果を表す画像を出力する画像センサ部、
ユーザからの撮影指示を受け付ける指示受付部、
前記撮影指示が受け付けられると、前記画像センサ部により感知されている画像を保存する保存部、
前記保存部により保存された画像と、前記画像センサ部により感知されている画像と、を、合成してファインダ画面に表示するファインダ表示部
として機能させることを特徴とするプログラムを記録するように構成する。
図1は、本実施形態に係る画像処理システムの概要構成を示す説明図である。以下、本図を参照して説明する。
図2は、本実施形態に係る商品撮影装置121の概要構成を説明する説明図である。以下、本図を参照して説明する。
上記のように、画像撮影装置121を利用すれば、商品301の回りをぐるりと一周するように、かつ、商品301に対する距離や重力の方向に対する傾きをできるだけ一定に保ちながら、画像撮影装置121を移動させ、商品301を順次撮影することができる。
(1)特徴点Pの画像G[k],G[k+1]における位置p[k],p[k+1]と、
(2)特徴点Qの画像G[k],G[k+1]における位置q[k],q[k+1]と、
(3)特徴点Wの画像G[k],G[k+1]における位置w[k],w[k+1]と、
(4)画像G[k]を撮影したときの商品撮影装置121の傾きR[k]と、
(5)画像G[k+1]を撮影したときの商品撮影装置121の傾きR[k+1]と、
から、
(x)画像G[k]を撮影したときの商品撮影装置121の位置T[k]と、
(y)画像G[k+1]を撮影したときの商品撮影装置121の位置T[k+1]と、
を、推定する(ステップS705)。
R-1 = Rt
が成立する。また、0tは、3次元の横零ベクトルである。
T[0],T[1],T[2],…,T[N-1]の各推定値
が得られ、ステップS705(y)により、
T[1],T[2],…,T[N-1],T[0]の各推定値
が得られることになる。
T[0],T[1],T[2],…,T[N-1]
を同定する(ステップS707)。
U[k] = T[k] + sc[k](T[k]-H)/|T[k]-H|
とを、sc[0],sc[1],sc[2],…,sc[N-1]をできるだけ小さくように、計算する。
D = Σk=0 N-1 |T[k]-H|/N
により求め、
sc[k] = D - |T[k]-H|
とすれば、各|U[k]-H|は、すべて距離Dに等しくなる。
J = Σk=0 N-1 U[k]/N
のように、位置U[k]の重心とするのが最も単純である。次に円周の半径rは、
r = Σk=0 N-1 |U[k]-J|/N
のように、位置Jと各U[k]との距離の平均とするのが最も単純である。
S[k] = J + r OP[k]/|OP[k]|
により求められることになる。
cos(θ[k]) = 《S[0]-J,S[k]-J》/(|S[0]-J|・|S[k]-J|),
sin(θ[k]) = |(S[0]-J)×(S[k]-J)|/(|S[0]-J|・|S[k]-J|)
により、各回転角θ[k]を求めることができる。ここで《x,y》は、ベクトルxとベクトルyの内積である。
0=θ[0]≦θ[1]≦θ[2]≦…≦θ[N-2]<2π
のように昇順に並ぶことと考えても、一般性を失わない。もしあるkにおいて
θ[k]>θ[k+1]
となって、順序が入れかわってしまっている場合には、
(a)画像G[k]、傾きR[k]撮影位置T[k]、正規化位置S[k]、回転角θ[k]と、
(b)画像G[k+1]、傾きR[k+1]撮影位置T[k+1]、正規化位置S[k+1]、回転角θ[k+1]と、
を、まとめて入れ換えれば良いからである。
仮想化位置S[k]から商品301の位置Hに向かうベクトルH-S[k]を、撮影の光軸方向(z軸)とし、
仮想化位置S[k]における円周の接線方向を撮影の左右方向(x軸)とし、
仮想化位置S[k]から円周の中心位置Jに向かうベクトルJ-S[k]の、H-S[k]に直交する成分を、撮影の上下方向(y軸)とする
のが典型的である。この2つの方向を定めることで、カメラ座標系の6つの自由度が決定される。仮想化位置S[k]における撮影の傾きV[k]も自動的に定まることになる。
d[k] = |T[k]-H|
をpc3の近似値とすることができる(数16)。
(1)撮影済画像G[k]の画素の位置を、上記の関係により求め、
(2)求められた位置の撮影済画像G[k]における画素値を取得し、
(3)当該取得された画素値を、正規化画像B[k]の当該位置の画素値として描画する
ことによって、正規化画像B[k]が得られるのである。
(d/d[k])M[k]L[k]-1
による変換を施して、全周囲画像列に含まれる回転角θ[k]に対する正規化画像B[k]を得る(ステップS712)。
全周囲画像列に含まれる正規化画像B[0],B[1],B[2],…,B[N-1]は、それぞれ、商品301の周囲の回転角θ[0],θ[1],θ[2],…,θ[N-1]に対応付けられている。なお、上記のように、
0=θ[0]≦θ[1]≦θ[2]≦…≦θ[N-1]<2π
としても一般性を失わない。
(1/d)L(φ)d L[k]-1 = L(φ)L[k]-1
を施すことにより、補間画像B(φ)が得られることになる。
θ[k]≦φ<θ[k+1]
が満たされる正規化画像B[k],B[k+1]の2つについて、上記のようにアフィン変換を適用する。
(0)正規化画像B[0]から緯度(θ[N-1]+2π)/2~緯度2π、ならびに、緯度0~緯度θ[1]/2の舟形を作り、
(1)正規化画像B[1]から緯度(θ[0]+θ[1])/2~緯度(θ[1]+θ[2])/2の舟形を作り、
(2)正規化画像B[2]から緯度(θ[1]+θ[2])/2~緯度(θ[2]+θ[3])/2の舟形を作り、…、
(k)正規化画像B[k]から緯度(θ[k-1]+θ[k])/2~緯度(θ[k]+θ[k+1])/2の舟形を作り、…、
(N-1)正規化画像B[N-1]から緯度(θ[N-2]+θ[N-1])/2~緯度(θ[N-1]+2π)/2の舟形を作る。
121 商品撮影装置
141 画像変換装置
161 画像処理装置
181 端末装置
201 画像センサ部
202 指示受付部
203 保存部
204 ファインダ表示部
205 傾きセンサ部
301 商品
311 撮影済画像
312 撮影済画像の撮影時の傾きを表す棒グラフ
321 ライブプレビュー画像
322 現在の傾きを表す棒グラフ
331 ファインダ画像
601 画像受付部
602 変換部
611 抽出部
612 推定部
613 近似部
614 出力部
801 角度受信部
802 角度受付部
803 補間部
804 画像送信部
811 角度センサ部
812 角度送信部
813 画像受信部
814 モニタ表示部
Claims (12)
- 商品が配置された外界から投射された光を感知して、当該感知の結果を表す画像を出力する画像センサ部、
ユーザからの撮影指示を受け付ける指示受付部、
前記撮影指示が受け付けられると、前記画像センサ部により感知されている画像を保存する保存部、
前記保存部により保存された画像と、前記画像センサ部により感知されている画像と、を、合成してファインダ画面に表示するファインダ表示部
を備えることを特徴とする商品撮影装置。 - 請求項1に記載の商品撮影装置であって、
前記ファインダ表示部は、前記保存部により保存された画像と、前記画像センサ部により感知されている画像と、を、交互に前記ファインダ画面に表示する
ことを特徴とする商品撮影装置。 - 請求項1に記載の商品撮影装置であって、
前記ファインダ表示部は、前記保存部により保存された画像に所定の画像フィルタを施した結果を半透明にして、前記画像センサ部により感知されている画像に重ねて合成する
ことを特徴とする商品撮影装置。 - 請求項1に記載の商品撮影装置であって、
前記ファインダ表示部は、前記保存部により保存された画像が複数存在する場合、最初に保存された画像と、前記画像センサ部により感知されている画像と、を合成する
ことを特徴とする商品撮影装置。 - 請求項1に記載の商品撮影装置であって、
前記商品撮影装置の傾きを感知する傾きセンサ部
をさらに備え、
前記保存部は、前記撮影指示が受け付けられると、前記画像センサ部により感知されている画像を前記傾きセンサ部により感知されている傾きに対応付けて保存し、
前記表示部は、前記保存部により保存された画像に対応付けられて保存された傾きと、前記傾きセンサ部により感知されている傾きと、を、前記ファインダ画面に表示する
ことを特徴とする商品撮影装置。 - 商品をその周囲から請求項5に記載の商品撮影装置により撮影して保存された複数の画像と、当該画像に対応付けられる傾きと、を受け付ける画像受付部、
前記受け付けられた複数の画像を、当該画像に対応付けられる傾きに基づいて、前記商品の全周囲画像列に変換する変換部
を備えることを特徴とする画像変換装置。 - 所定の回転軸周りの角度を受け付ける角度受付部、
前記受け付けられた角度に基づいて、請求項6に記載の画像変換装置により得られた商品の全周囲画像列を補間して、当該商品を前記受け付けられた角度から見た様子を表す補間済画像を得る補間部
を備えることを特徴とする画像処理装置。 - 請求項7に記載の画像処理装置であって、
前記画像処理装置の所定の軸周りの角度を感知する角度センサ部、
前記得られた補間済画像をモニタ画面に表示するモニタ表示部
をさらに備え、
前記角度受付部は、前記感知された角度を受け付ける
ことを特徴とする画像処理装置。 - 請求項7に記載の画像処理装置と、当該画像処理装置と通信可能に接続される端末装置と、を備える画像処理システムであって、
(a)前記端末装置は、
前記端末装置の所定の軸周りの角度を感知する角度センサ部、
前記感知された角度を前記画像処理装置に送信する角度送信部
を備え、
(b)前記画像処理装置は、
前記端末装置から送信された角度を受信して、前記角度受付部に受け付けさせる角度受信部、
前記得られた補間済画像を前記端末装置に送信する画像送信部
を備え、
(c)前記端末装置は、
前記画像処理装置から送信された補間済画像を受信する画像受信部、
前記受信された補間済画像をモニタ画面に表示するモニタ表示部
を備える
ことを特徴とする画像処理システム。 - 画像センサ部、指示受付部、保存部、表示部を備える商品撮影装置が実行する商品撮影方法であって、
前記画像センサ部が、商品が配置された外界から投射された光を感知して、当該感知の結果を表す画像を出力する画像センサ工程、
前記指示受付部が、ユーザからの撮影指示を受け付ける指示受付工程、
前記撮影指示が受け付けられると、前記保存部が、前記画像センサ部により感知されている画像を保存する保存工程、
前記表示部が、前記保存部により保存された画像と、前記画像センサ部により感知されている画像と、を、合成してファインダ画面に表示するファインダ表示工程
を備えることを特徴とする商品撮影方法。 - コンピュータを、
商品が配置された外界から投射された光を感知して、当該感知の結果を表す画像を出力する画像センサ部、
ユーザからの撮影指示を受け付ける指示受付部、
前記撮影指示が受け付けられると、前記画像センサ部により感知されている画像を保存する保存部、
前記保存部により保存された画像と、前記画像センサ部により感知されている画像と、を、合成してファインダ画面に表示するファインダ表示部
として機能させることを特徴とするプログラム。 - コンピュータを、
商品が配置された外界から投射された光を感知して、当該感知の結果を表す画像を出力する画像センサ部、
ユーザからの撮影指示を受け付ける指示受付部、
前記撮影指示が受け付けられると、前記画像センサ部により感知されている画像を保存する保存部、
前記保存部により保存された画像と、前記画像センサ部により感知されている画像と、を、合成してファインダ画面に表示するファインダ表示部
として機能させることを特徴とするプログラムを記録したコンピュータ読取可能な情報記録媒体。
Priority Applications (9)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
ES10856738T ES2813066T3 (es) | 2010-08-30 | 2010-09-29 | Dispositivo de conversión de imágenes, procedimiento y medio de registro legible por ordenador |
KR1020137005454A KR101477937B1 (ko) | 2010-08-30 | 2010-09-29 | 화상 처리 장치, 화상 처리 방법 및 정보 기록 매체 |
EP10856738.9A EP2592820B1 (en) | 2010-08-30 | 2010-09-29 | Image conversion device, method and computer-readable recording medium |
US13/814,834 US9373170B2 (en) | 2010-08-30 | 2010-09-29 | Product image interpolating device, method, and non-transitory information recording medium |
CN201080068666.0A CN103081454B (zh) | 2010-08-30 | 2010-09-29 | 商品摄影装置、商品摄影方法、图像变换装置、图像处理装置、图像处理系统 |
EP13158249.6A EP2608555B1 (en) | 2010-08-30 | 2010-09-29 | Product imaging device, product imaging method, image coversion device, image processing device, image processing system, program, and information recording medium |
CA2809301A CA2809301C (en) | 2010-08-30 | 2010-09-29 | Product imaging device, product imaging method, image conversion device, image processing device, image processing system, program, and information recording medium |
KR1020137005402A KR101412878B1 (ko) | 2010-08-30 | 2010-09-29 | 상품 촬영 장치, 상품 촬영 방법, 화상 변환 장치, 화상 처리 장치, 화상 처리 시스템, 프로그램, 및, 정보 기록 매체 |
US13/764,283 US9153004B2 (en) | 2010-08-30 | 2013-02-11 | Product image interpolating device, method and non-transitory information recording medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010192927A JP4763847B1 (ja) | 2010-08-30 | 2010-08-30 | 画像変換装置、画像処理装置、ならびに、画像処理システム |
JP2010-192927 | 2010-08-30 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/814,834 A-371-Of-International US9373170B2 (en) | 2010-08-30 | 2010-09-29 | Product image interpolating device, method, and non-transitory information recording medium |
US13/764,283 Division US9153004B2 (en) | 2010-08-30 | 2013-02-11 | Product image interpolating device, method and non-transitory information recording medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012029193A1 true WO2012029193A1 (ja) | 2012-03-08 |
Family
ID=44597234
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2010/067003 WO2012029193A1 (ja) | 2010-08-30 | 2010-09-29 | 商品撮影装置、商品撮影方法、画像変換装置、画像処理装置、画像処理システム、プログラム、ならびに、情報記録媒体 |
Country Status (8)
Country | Link |
---|---|
US (1) | US9373170B2 (ja) |
EP (2) | EP2608555B1 (ja) |
JP (1) | JP4763847B1 (ja) |
KR (2) | KR101477937B1 (ja) |
CN (2) | CN103081454B (ja) |
CA (2) | CA2875392C (ja) |
ES (2) | ES2698059T3 (ja) |
WO (1) | WO2012029193A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2831807A4 (en) * | 2012-03-27 | 2016-06-01 | Amazon Tech Inc | USER-GUIDED OBJECT IDENTIFICATION |
EP2926196A4 (en) * | 2012-11-30 | 2016-08-24 | Thomson Licensing | METHOD AND SYSTEM FOR DETECTING A 3D IMAGE USING A SINGLE CAMERA |
JP2016531281A (ja) * | 2013-07-02 | 2016-10-06 | マーティン ガイ | 撮像装置のモデリングおよび校正のためのシステムおよびその方法 |
Families Citing this family (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013018614A1 (ja) | 2011-07-29 | 2013-02-07 | 日本電気株式会社 | 照合・検索システム、照合・検索サーバ、画像特徴抽出装置、照合・検索方法及びプログラム |
JP5725049B2 (ja) * | 2013-01-24 | 2015-05-27 | カシオ計算機株式会社 | 撮像装置、撮像方法及びプログラム |
US9836875B2 (en) * | 2013-04-26 | 2017-12-05 | Flipboard, Inc. | Viewing angle image manipulation based on device rotation |
US9665895B2 (en) | 2013-08-12 | 2017-05-30 | Mov, Inc. | Technologies for video-based commerce |
JP2016046676A (ja) | 2014-08-22 | 2016-04-04 | 株式会社リコー | 撮像装置および撮像方法 |
KR102173109B1 (ko) * | 2014-09-05 | 2020-11-02 | 삼성전자주식회사 | 디지털 영상 처리 방법, 상기 방법을 기록한 컴퓨터 판독 가능 저장매체 및 디지털 영상 처리 장치 |
CN106034203B (zh) * | 2015-03-11 | 2020-07-28 | 维科技术有限公司 | 拍摄终端的图像处理方法及其装置 |
CN105959551A (zh) * | 2016-05-30 | 2016-09-21 | 努比亚技术有限公司 | 一种拍摄装置及方法、移动终端 |
US10298837B2 (en) * | 2016-12-28 | 2019-05-21 | Facebook, Inc. | Systems and methods for presenting content based on unstructured visual data |
JP6413035B1 (ja) * | 2018-01-18 | 2018-10-24 | 株式会社ほぼ日 | プログラム、情報処理装置及び情報処理方法 |
KR102614026B1 (ko) * | 2019-01-29 | 2023-12-13 | 삼성전자주식회사 | 복수의 렌즈를 구비하는 전자 장치 및 그 제어 방법 |
CN110266952B (zh) * | 2019-06-28 | 2021-03-12 | Oppo广东移动通信有限公司 | 图像处理方法、装置、电子设备以及存储介质 |
JP2021103442A (ja) * | 2019-12-25 | 2021-07-15 | 京セラ株式会社 | 画像データ登録システム、画像データ登録方法 |
US11928758B2 (en) | 2020-03-06 | 2024-03-12 | Christopher Renwick Alston | Technologies for augmented-reality |
EP3937482A4 (en) * | 2020-05-28 | 2022-05-25 | Beijing Xiaomi Mobile Software Co., Ltd. Nanjing Branch | IMAGE CAPTURE METHOD, DEVICE, DEVICE AND STORAGE MEDIA |
US11496674B2 (en) * | 2020-09-18 | 2022-11-08 | Microsoft Technology Licensing, Llc | Camera placement guidance |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH02178646A (ja) * | 1988-12-29 | 1990-07-11 | Canon Inc | パノラマ撮影用表示装置 |
JPH04347979A (ja) * | 1991-05-24 | 1992-12-03 | Olympus Optical Co Ltd | 画像合成撮影装置及び再生装置 |
JPH06105193A (ja) * | 1992-09-24 | 1994-04-15 | Sony Corp | カメラ装置 |
JPH1188731A (ja) * | 1997-09-10 | 1999-03-30 | Nikon Corp | カメラ |
JP2000299804A (ja) * | 1999-04-12 | 2000-10-24 | Casio Comput Co Ltd | パノラマ撮影方法及び撮像装置 |
JP2001285894A (ja) * | 2000-03-31 | 2001-10-12 | Olympus Optical Co Ltd | 3次元画像データの掲載方法 |
JP2002133447A (ja) * | 2000-10-26 | 2002-05-10 | Shinyo Sangyo:Kk | 静止画像表示システム及び電子商品カタログ表示システム |
JP2004072685A (ja) * | 2002-08-09 | 2004-03-04 | Sharp Corp | 画像合成装置、画像合成方法、画像合成プログラム、および画像合成プログラムを記録した記録媒体 |
JP2005049999A (ja) * | 2003-07-30 | 2005-02-24 | Ricoh Co Ltd | 画像入力装置、画像入力方法、この方法を情報処理装置上で実行可能に記述されたプログラム、及びこのプログラムを記憶した記憶媒体 |
JP2006244306A (ja) * | 2005-03-04 | 2006-09-14 | Nippon Telegr & Teleph Corp <Ntt> | 動画像生成システム、動画像生成装置、動画像生成方法、プログラム、及び記録媒体 |
JP2007174301A (ja) * | 2005-12-22 | 2007-07-05 | Toshiba Corp | 画像撮影装置 |
JP2009005254A (ja) * | 2007-06-25 | 2009-01-08 | Toshiba Corp | 映像送信装置及び映像受信装置 |
JP2009183003A (ja) * | 2005-02-23 | 2009-08-13 | Samsung Electronics Co Ltd | メニュー画面を表示する端末機においてメニューナビゲーションを制御する装置及び方法 |
JP2009246937A (ja) * | 2008-03-11 | 2009-10-22 | Ricoh Co Ltd | 撮像装置 |
JP2010192927A (ja) | 1999-02-15 | 2010-09-02 | Fujitsu Optical Components Ltd | 光モジュール |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6549681B1 (en) * | 1995-09-26 | 2003-04-15 | Canon Kabushiki Kaisha | Image synthesization method |
US7015954B1 (en) * | 1999-08-09 | 2006-03-21 | Fuji Xerox Co., Ltd. | Automatic video system using multiple cameras |
JP2002098521A (ja) * | 2000-09-26 | 2002-04-05 | Minolta Co Ltd | 3次元形状データ生成装置 |
CN1375799A (zh) * | 2002-04-15 | 2002-10-23 | 上海罗莱照相器材厂 | 立体照片数码合成方法 |
WO2004015987A1 (ja) | 2002-08-09 | 2004-02-19 | Sharp Kabushiki Kaisha | 画像合成装置、画像合成方法、画像合成プログラム、および画像合成プログラムを記録した記録媒体 |
JP2004128588A (ja) * | 2002-09-30 | 2004-04-22 | Seiko Epson Corp | 撮像装置 |
CN100440933C (zh) * | 2005-12-31 | 2008-12-03 | 北京中星微电子有限公司 | 一种用于手持数码摄像设备的防抖方法及装置 |
JP5084167B2 (ja) * | 2006-03-31 | 2012-11-28 | キヤノン株式会社 | 位置姿勢計測方法及び装置 |
US8331725B2 (en) * | 2007-01-12 | 2012-12-11 | Qualcomm Incorporated | Panoramic imaging techniques |
CN101420525A (zh) * | 2007-10-26 | 2009-04-29 | 鸿富锦精密工业(深圳)有限公司 | 拍照装置及方法 |
JP2010041586A (ja) * | 2008-08-07 | 2010-02-18 | Olympus Corp | 撮像装置 |
JP2010109598A (ja) * | 2008-10-29 | 2010-05-13 | Canon Inc | 画像処理装置および画像処理方法 |
JP2010199971A (ja) * | 2009-02-25 | 2010-09-09 | Casio Computer Co Ltd | 撮像装置、撮像方法、及びプログラム |
-
2010
- 2010-08-30 JP JP2010192927A patent/JP4763847B1/ja active Active
- 2010-09-29 KR KR1020137005454A patent/KR101477937B1/ko active IP Right Grant
- 2010-09-29 ES ES13158249T patent/ES2698059T3/es active Active
- 2010-09-29 CN CN201080068666.0A patent/CN103081454B/zh active Active
- 2010-09-29 EP EP13158249.6A patent/EP2608555B1/en active Active
- 2010-09-29 CN CN201310091710.5A patent/CN103152526B/zh active Active
- 2010-09-29 ES ES10856738T patent/ES2813066T3/es active Active
- 2010-09-29 CA CA2875392A patent/CA2875392C/en active Active
- 2010-09-29 CA CA2809301A patent/CA2809301C/en active Active
- 2010-09-29 KR KR1020137005402A patent/KR101412878B1/ko active IP Right Grant
- 2010-09-29 US US13/814,834 patent/US9373170B2/en active Active
- 2010-09-29 EP EP10856738.9A patent/EP2592820B1/en active Active
- 2010-09-29 WO PCT/JP2010/067003 patent/WO2012029193A1/ja active Application Filing
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH02178646A (ja) * | 1988-12-29 | 1990-07-11 | Canon Inc | パノラマ撮影用表示装置 |
JPH04347979A (ja) * | 1991-05-24 | 1992-12-03 | Olympus Optical Co Ltd | 画像合成撮影装置及び再生装置 |
JPH06105193A (ja) * | 1992-09-24 | 1994-04-15 | Sony Corp | カメラ装置 |
JPH1188731A (ja) * | 1997-09-10 | 1999-03-30 | Nikon Corp | カメラ |
JP2010192927A (ja) | 1999-02-15 | 2010-09-02 | Fujitsu Optical Components Ltd | 光モジュール |
JP2000299804A (ja) * | 1999-04-12 | 2000-10-24 | Casio Comput Co Ltd | パノラマ撮影方法及び撮像装置 |
JP2001285894A (ja) * | 2000-03-31 | 2001-10-12 | Olympus Optical Co Ltd | 3次元画像データの掲載方法 |
JP2002133447A (ja) * | 2000-10-26 | 2002-05-10 | Shinyo Sangyo:Kk | 静止画像表示システム及び電子商品カタログ表示システム |
JP2004072685A (ja) * | 2002-08-09 | 2004-03-04 | Sharp Corp | 画像合成装置、画像合成方法、画像合成プログラム、および画像合成プログラムを記録した記録媒体 |
JP2005049999A (ja) * | 2003-07-30 | 2005-02-24 | Ricoh Co Ltd | 画像入力装置、画像入力方法、この方法を情報処理装置上で実行可能に記述されたプログラム、及びこのプログラムを記憶した記憶媒体 |
JP2009183003A (ja) * | 2005-02-23 | 2009-08-13 | Samsung Electronics Co Ltd | メニュー画面を表示する端末機においてメニューナビゲーションを制御する装置及び方法 |
JP2006244306A (ja) * | 2005-03-04 | 2006-09-14 | Nippon Telegr & Teleph Corp <Ntt> | 動画像生成システム、動画像生成装置、動画像生成方法、プログラム、及び記録媒体 |
JP2007174301A (ja) * | 2005-12-22 | 2007-07-05 | Toshiba Corp | 画像撮影装置 |
JP2009005254A (ja) * | 2007-06-25 | 2009-01-08 | Toshiba Corp | 映像送信装置及び映像受信装置 |
JP2009246937A (ja) * | 2008-03-11 | 2009-10-22 | Ricoh Co Ltd | 撮像装置 |
Non-Patent Citations (3)
Title |
---|
"Wikipedia, The Free Encyclopedia", 15 June 2010 |
KAZUTAKA YASUDA; TAKESHI MIGITA; MASATO AOYAMA; MASAYUKI KUREKI; NAOKI ASADA: "Generation of Dense Three-dimensional Shape Model from Rough Image Sequence of Entire Surroundings", RESEARCH REPORT OF INFORMATION PROCESSING SOCIETY OF JAPAN, CVIM, COMPUTER VISION AND IMAGE MEDIA, vol. CVIM-138, no. 11, 8 May 2003 (2003-05-08), pages 73 - 80, XP009193971, Retrieved from the Internet <URL:http://harp.lib.hiroshima-u.ac.jp/bitstream/harp/6507> |
See also references of EP2592820A4 |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2831807A4 (en) * | 2012-03-27 | 2016-06-01 | Amazon Tech Inc | USER-GUIDED OBJECT IDENTIFICATION |
EP2926196A4 (en) * | 2012-11-30 | 2016-08-24 | Thomson Licensing | METHOD AND SYSTEM FOR DETECTING A 3D IMAGE USING A SINGLE CAMERA |
JP2016531281A (ja) * | 2013-07-02 | 2016-10-06 | マーティン ガイ | 撮像装置のモデリングおよび校正のためのシステムおよびその方法 |
Also Published As
Publication number | Publication date |
---|---|
CA2809301C (en) | 2016-11-01 |
KR20130044345A (ko) | 2013-05-02 |
ES2813066T3 (es) | 2021-03-22 |
CN103152526A (zh) | 2013-06-12 |
CA2809301A1 (en) | 2012-03-08 |
CA2875392C (en) | 2017-04-25 |
US20130202154A1 (en) | 2013-08-08 |
JP2012050041A (ja) | 2012-03-08 |
CN103152526B (zh) | 2016-01-20 |
KR101477937B1 (ko) | 2014-12-30 |
CN103081454A (zh) | 2013-05-01 |
CN103081454B (zh) | 2016-01-27 |
EP2592820A4 (en) | 2017-05-10 |
CA2875392A1 (en) | 2012-03-08 |
EP2608555A3 (en) | 2015-05-06 |
EP2608555B1 (en) | 2018-09-26 |
EP2592820A1 (en) | 2013-05-15 |
EP2608555A2 (en) | 2013-06-26 |
US9373170B2 (en) | 2016-06-21 |
JP4763847B1 (ja) | 2011-08-31 |
EP2592820B1 (en) | 2020-07-15 |
KR20130041297A (ko) | 2013-04-24 |
KR101412878B1 (ko) | 2014-06-26 |
ES2698059T3 (es) | 2019-01-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2012029193A1 (ja) | 商品撮影装置、商品撮影方法、画像変換装置、画像処理装置、画像処理システム、プログラム、ならびに、情報記録媒体 | |
TWI451750B (zh) | 攝影裝置、電腦可讀取記錄媒體及控制方法 | |
US9619933B2 (en) | Model and sizing information from smartphone acquired image sequences | |
EP3101624B1 (en) | Image processing method and image processing device | |
US9686539B1 (en) | Camera pair calibration using non-standard calibration objects | |
JP5110138B2 (ja) | Ar処理装置、ar処理方法及びプログラム | |
EP3189493B1 (en) | Depth map based perspective correction in digital photos | |
JP6405539B2 (ja) | 多視点画像に対するラベル情報の処理装置及びそのラベル情報の処理方法 | |
JP4783859B1 (ja) | 画像変換装置、画像処理システム、画像変換方法、プログラム、ならびに、情報記録媒体 | |
JP4906949B2 (ja) | 画像処理システム、ならびに、画像処理方法 | |
US9153004B2 (en) | Product image interpolating device, method and non-transitory information recording medium | |
JP6812643B2 (ja) | 通信端末、画像通信システム、表示方法、及びプログラム | |
JP4830041B1 (ja) | 画像処理装置、画像処理方法、プログラム、ならびに、情報記録媒体 | |
JP2012248206A (ja) | Ar処理装置、ar処理方法及びプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080068666.0 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10856738 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010856738 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2809301 Country of ref document: CA |
|
ENP | Entry into the national phase |
Ref document number: 20137005402 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13814834 Country of ref document: US |