WO2013133456A1 - Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium - Google Patents

Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium Download PDF

Info

Publication number
WO2013133456A1
WO2013133456A1 PCT/JP2013/057338 JP2013057338W WO2013133456A1 WO 2013133456 A1 WO2013133456 A1 WO 2013133456A1 JP 2013057338 W JP2013057338 W JP 2013057338W WO 2013133456 A1 WO2013133456 A1 WO 2013133456A1
Authority
WO
WIPO (PCT)
Prior art keywords
image capturing
tilt
conversion data
image
capturing apparatus
Prior art date
Application number
PCT/JP2013/057338
Other languages
French (fr)
Inventor
Hirokazu Takenaka
Yoshiaki Irino
Tomonori Tanaka
Nozomi Imae
Hideaki Yamamoto
Kensuke Masuda
Yoichi Ito
Satoshi Sawaguchi
Daisuke Bessho
Hiroyuki Satoh
Makoto Shohara
Noriyuki Terao
Toru Harada
Original Assignee
Ricoh Company, Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to KR1020147025261A priority Critical patent/KR101613387B1/en
Priority to EP17155235.9A priority patent/EP3182696B1/en
Priority to CA2866621A priority patent/CA2866621C/en
Priority to KR1020187031508A priority patent/KR101961364B1/en
Priority to EP13757254.1A priority patent/EP2823637B1/en
Priority to CN201380012472.2A priority patent/CN104160693B/en
Priority to KR1020167036128A priority patent/KR101770088B1/en
Priority to KR1020197007730A priority patent/KR102046032B1/en
Application filed by Ricoh Company, Limited filed Critical Ricoh Company, Limited
Priority to CN201810707970.3A priority patent/CN108876714B/en
Priority to KR1020167008975A priority patent/KR101692194B1/en
Priority to US14/384,027 priority patent/US9607358B2/en
Priority to KR1020177022160A priority patent/KR101916616B1/en
Publication of WO2013133456A1 publication Critical patent/WO2013133456A1/en
Priority to US15/400,610 priority patent/US11049215B2/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/72Combination of two or more compensation controls
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/12Panospheric to cylindrical image transformations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/04Context-preserving transformations, e.g. by using an importance map
    • G06T3/047Fisheye or wide-angle transformations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/69Control of means for changing angle of the field of view, e.g. optical zoom objectives or electronic zooming
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/698Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • H04N23/81Camera processing pipelines; Components thereof for suppressing or minimising disturbance in the image signal generation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2628Alteration of picture size, shape, position or orientation, e.g. zooming, rotation, rolling, perspective, translation

Definitions

  • the present invention relates to an image capturing apparatus, an image capture system, an information
  • hyperboloidal mirror or a fisheye lens for example.
  • These image capturing apparatuses are specific optical systems that need a non-linear image processing such as distortion correction or projective transformation so that the captured image with the apparatus can be visualized to a user.
  • One of known technologies is that a semispherical image is obtained with a wide-angle capture employing a fisheye lens and recorded as a distorted circular image. A part of the distorted circular image is cut out and image processing is performed on the image with a computer, whereby the distorted image is transformed into a plane regular image.
  • Another known technology is that while performing the image processing above, when the center position of the distorted circular image does not correctly correspond to the vertical point direction, a user specifies a parameter of a tilt angle. The load while performing distortio'n correction on the image is therefore reduced.
  • Japanese Patent Application Laid-open No. 2003-223633 and Japanese Patent Application Laid-open No. 2006-059202 disclose a technology of generating an omni-directional spherical image having the correct vertical direction. This is achieved by adding rotational transformation according to the tilt of a camera to the process of nonlinear image processing.
  • Japanese Patent Application Laid-open No. Hll-309137 and Japanese Patent No. 4175832 disclose a technology of generating an omni-directional spherical image having the correct vertical direction at high speed, for the purpose of reducing the time required for generating the omnidirectional spherical image having the correct vertical direction or reducing the cost on arithmetic operations. This technology is achieved by adding rotational
  • Japanese Patent Application Laid-open No. Hll-309137 and Japanese Patent No. 4175832 disclose in particular a technology generating images using a conversion table that includes rotational transformation according to the tilt of the camera.
  • the conversion table is made so as to
  • the conversion table needs to be re-made from the beginning. Also in this case, the processing still requires a relatively long time to generate an omnidirectional spherical image having the correct vertical direction according to any tilt of the image capturing apparatus, therefore the problem is not yet solved.
  • an image capturing apparatus that includes a tilt detection unit configured to detect a tilt in a vertical direction;
  • conversion data used for transforming plane coordinates into spherical coordinates a correction unit configured to correct the conversion data according to the tilt; a plurality of image capturing units; a coordinate
  • transformation unit configured to transform plane
  • a combining unit configured to combine the images including the pixels transformed into spherical coordinates by the coordinate transformation unit
  • FIG. 1 is a schematic block diagram for explaining the overall structure of an image capturing apparatus according to an embodiment of the present invention.
  • FIG. 2 is an external side view of an omni-directional spherical image capturing apparatus according to the embodiment .
  • FIG. 3 is a flowchart for explaining operations of the image capturing apparatus according to the embodiment.
  • FIG. 4 illustrates a projection relation of a fisheye lens employed in the image capturing apparatus according to the embodiment .
  • FIG. 5 illustrates a format for the omni-directional spherical image captured by the image capturing apparatus according to the embodiment.
  • FIG. 6 illustrates the outline of generation
  • FIG. 7 illustrates the conversion table for the omnidirectional spherical image captured by the image capturing apparatus according to the embodiment.
  • FIG. 8 is a flowchart for explaining operations of correction processing through the conversion table for the omni-directional spherical image captured by the image capturing apparatus according to the embodiment.
  • FIG. 9 is a schematic diagram for explaining a tilt of the image capturing apparatus according to the embodiment .
  • FIG. 10 illustrates an example of calculation of vertical correction on the omni-directional spherical image captured by the image capturing apparatus according to the embodiment .
  • FIG. 11 illustrates another example of calculation of vertical correction on the omni-directional spherical image captured by the image capturing apparatus according to the embodiment.
  • FIG. 12 is another flowchart for explaining operations of correction processing through the conversion table for the omni-directional spherical image captured by the image capturing apparatus according to the embodiment.
  • FIG. 13 is a schematic diagram for explaining the overall structure of an image capture system according to another embodiment of the present invention.
  • FIG. 14 is a schematic block diagram for explaining the overall structure of an electronic circuit serving as a destination device of the image capture system according to another embodiment.
  • FIG. 15 is a flowchart for explaining operations of the image capture system according to another embodiment.
  • FIG. 16 is a diagram for explaining the structure of a multi-eye image capturing apparatus that captures images in all directions spherically at once.
  • FIG. 17 is a diagram for explaining an image obtained through a lens in the multi-eye image capturing apparatus illustrated in FIG. 16.
  • FIG. 18 is a diagram for explaining an image obtained by performing distortion correction on images illustrated in FIG. 17 and then combining the images with each other.
  • FIG. 19 is a diagram for explaining an image captured by the multi-eye image capturing apparatus illustrated in FIG. 16, with the image capturing apparatus in a state of being tilted.
  • FIG. 20 is a diagram for explaining an image obtained by performing distortion correction and image combination on the image illustrated in FIG. 19 without consideration of the tilt.
  • FIG. 21 is a schematic diagram for explaining an acceleration sensor that measures a tilt angle.
  • the omni-directional spherical image capturing apparatus detects the vertical direction and corrects a conversion table used for image processing according to the vertical direction.
  • the omni-directional spherical image is generated through the corrected conversion table. Accordingly, the
  • the image capturing apparatus refers to a stand-alone apparatus, such as a digital camera; an image capture system refers to a system in which a plurality of apparatuses such as a digital camera and an information processing apparatus are utilized separately. In the embodiments, unless otherwise specified, the image capturing apparatus conceptually includes an image capture system.
  • a multi-eye image capturing apparatus will now be described that captures images in all directions
  • FIG. 16 is a diagram for explaining the structure of a multi-eye image capturing apparatus that captures images in all directions spherically at once.
  • An image capture system is prepared that employs a plurality of wide-angle lenses as illustrated in FIG. 16 so as to capture images in all directions spherically at once.
  • the multi-eye image capturing apparatus illustrated on the left in FIG. 16 employs a fisheye lens (ultra wide-angle lens) with an angle of view equal to or larger than 180 degrees.
  • the multi-eye image capturing apparatus requires at least the two-eye structure (image capture lenses with two different image capturing
  • the number of image capture lenses may be increased by employing, for example, the four-eye structure for the multi-eye image capturing apparatus as illustrated on the right in FIG. 16. With this structure, only center parts with high image quality are selectively used to reduce the image distortion. In this image capturing apparatus with the four-eye structure, it is recommended that respective lenses have an angle of view equal to or larger than 90 degrees, and preferably about 100 degrees.
  • the image capturing apparatus with two-eye structure will be described hereafter, for convenience of description.
  • the basic idea of the present invention may be applied to the image capturing apparatus with the four-eye structure illustrated on the right in FIG. 16, however.
  • FIG. 17 is a diagram for explaining an image obtained through a lens in a two-eye image capturing apparatus illustrated on the left in FIG. 16.
  • images are captured so that the vertical point and the horizontal line recognized by a user of the image capturing apparatus correspond to the vertical point and the horizontal line of the image capturing apparatus.
  • a lens has an angle of view equal to or larger than 180 degrees, and preferably about 190 degrees so as to provide the superimposed field.
  • a lens has an angle of view of about 100 degrees so as to provide an appropriate superimposed field.
  • the apparatus corresponds to the position of the vertical point of the image capturing apparatus.
  • the above-described technique is widely known for combining captured images to perform distortion correction on the combined image.
  • FIG. 18 is a diagram for explaining an image obtained by performing distortion correction on images illustrated in FIG. 17 and then combining the images with each other.
  • a correct image can be obtained having the straight horizontal line as illustrated in FIG. 18 after distortion correction and combination are simply performed on the captured images.
  • images can be captured in the correct ⁇ vertical direction when securing the image capturing apparatus to a certain retainer, and adjusting the horizontal and vertical lines of the apparatus using a level, for example.
  • FIG. 19 is a diagram for explaining an image ⁇ captured by the multi-eye (two-eye) image capturing apparatus illustrated on the right in FIG. 16, with the image
  • FIG. 19 exemplifies images captured in such a state. As illustrated in FIG. 19 the vertical points of the images do not correspond to each other and the horizontal lines of the images are distorted. As described above, if distortion correction and combination are performed on images captured with the image capturing apparatus in a state of being tilted without consideration of the tilt, an image is obtained in which the distortion exemplified in FIG. 19 is directly reflected, as
  • the horizontal line is curved like a graph of a
  • FIG. 21 is a schematic diagram for explaining an acceleration sensor that measures a tilt angle.
  • An acceleration sensor as illustrated in FIG. 21 is
  • FIG. 21 illustrates the outline of obtaining an angle using a single-axis acceleration sensor. This is employed to simplify the description.
  • the sensor obtains only the tilt of the image capt ring apparatus within the plane including the central axis of the lens surface in the two-eye structure.
  • a three-axis acceleration sensor is employed so as to measure the deviated angle from the plane including the central plane of the lens surface.
  • FIG. 1 is a schematic block diagram for explaining the overall
  • an image capturing element 1 an image capturing apparatus (hereinafter, also referred to as a digital camera) 100, an image capturing element 1
  • controllej 10 (reference numeral: 109), an image capturing element 2 (reference numeral: 110), a synchronous dynamic random access memory (SDRAM) 111, an external storage device 112, and the acceleration sensor 113 are coupled to a controllej 10.
  • SDRAM synchronous dynamic random access memory
  • two capturing elements i.e., the two-eye structure
  • three or more capturing elements may be employed.
  • a lens corresponding to the image capturing element which will be described with reference to FIG. 2
  • Another angle of view of the lens may be employed as appropriate.
  • Wide-angle' lenses including a fisheye lens are typically employed for the lens.
  • the image capturing apparatus is not limited to the omnidirectional image capturing apparatus. .
  • Another image capturing apparatus capable of capturing images covering 360 degrees in the horizontal direction may be used.
  • the controller 10 includes a central processing unit (CPU) 101, a static RAM (SRAM) 102, a read-only memory
  • ROM read only memory
  • image processing block 104 an SDRAM
  • I/F interface
  • processing block 104 performs typical image processing such as distortion correction and pixel defect correction, and the CPU reads a predetermined table or a computer program, thereby performing correction processing in the vertical direction according to the tilt of the digital camera 100. It is to be understood, however, that the image processing block 104 may perform correction processing in the vertical direction.
  • the pieces of image data input are image- processed using the image processing block 104, the CPU 101, the SRAM 102, and the SDRAM 111, for example, and finally stored in the external storage device 112.
  • Examples of an external storage device include CompactFlash (registered trademark) or a Secure Digital (SD) memory.
  • a universal serial bus (USB) connection interface for connecting to an external device or a wired or wireless network I/F for connecting to a network may be provided.
  • a conversion table which will be described later, a correction processing program for the conversion table, and a processing program for calculation of vertical correction are stored in the SRAM 102 or the SDRAM 111.
  • the acceleration sensor 113 is used for detecting a tilt of the digital camera 100 during capture. This enables detection in the tilting direction of the digital camera instantly and readily.
  • the acceleration sensor 113 is a ' three-axis
  • acceleration sensor that detects acceleration of the digital camera 100 in the three directions perpendicular to each other, i.e., the up and down direction, the right and left direction, and the back and front direction.
  • the acceleration sensor 113 detects the gravity acceleration only.
  • the up and down direction of the digital camera 100 corresponds to the top and bottom direction with respect to the ground.
  • the digital camera is held horizontally in the same manner usually operating a digital camera.
  • the acceleration sensor 113 detects acceleration in the right and left direction and the back and front direction according to the actual tilting
  • a tilt angle of the digital camera 100 can be obtained according to the magnitude of acceleration in the up and down direction, the right and left direction, and the back and front direction.
  • FIG. 2 is an external side view of the omni-directional spherical image capturing apparatus according to the embodiment.
  • the present embodiment aims to generate the omnidirectional spherical image using the omni-directional spherical image capturing apparatus capable of capturing images in all directions from a capture point. That is to say, the omni-directional spherical image capturing
  • apparatus can capture images in all directions viewed from the capture point.
  • the digital camera 100 (illustrated in FIG. 1) serving as an omni-directional spherical image capturing apparatus, captures images through the two. capturing element 1
  • elements have a fisheye lens as an example of wide-angle lenses having an angle of view exceeding 180 degrees.
  • the images captured through these two fisheye lenses have a superimposed field with each other.
  • a predetermined image processing such as distortion correction is performed on the images, and the resulting images are converted, and then combined with each other, thereby generating an omnidirectional spherical image.
  • FIG. 3 is a flowchart for explaining operations of the image capturing apparatus according to the embodiment.
  • the acceleration sensor 113 detects a tilt angle of the digital camera 100 at Step' (hereinafter, noted, as "S") 301.
  • the controller 10 reads out a conversion table stored in the SDRAM 111, for example, according to a tilt angle of the digital camera 100 detected at S301.
  • the controller 10 then corrects the conversion table in a predetermined correction method. The correction method for the conversion table will be described later.
  • processing block 104 performs typical image processing such as distortion correction.
  • the controller 10 converts the two captured fisheye images (like the image as illustrated in FIG. 17) using the conversion table
  • the controller 10 At S305, the controller 10 generates a composed omnidirectional spherical image utilizing a superimposed field between the two images converted at S304. At S306, the controller 10 stores the omni-directional spherical image generated at S305 in the external storage device 112 through the external storage device I/F 106.
  • FIG. 4 illustrates a projection relation of a fisheye lens
  • FIG. 4(a) is an external . side view of the fisheye lens;
  • FIG (b) illustrates a projection function f in a plan view of a captured image.
  • FIG. 17 illustrates an image captured through the fisheye lens with an angle of view over 180 degrees, which is a captured image of an almost hemispherical scene viewed from a capture position.
  • the image is generated with an image height h according to an angle of incidence ⁇ .
  • the projection function varies depending on the characteristic of the fisheye lens.
  • Examples of a technique (function) of projective transformation include central projection,, stereographic projection, equidistant projection, equi-solid-angle projection, and orthographic projection.
  • the central projection is employed when capturing images, using a digital camera including a typical angle of view.
  • Other four methods above are employed in a digital camera
  • a wide-angle lens with an ultra wide angle of view such as a fisheye lens.
  • FIG. 5 illustrates a format of the omni-directional spherical image captured by the digital camera according to the . embodiment.
  • FIG. 5(a) illustrates a plane format;
  • FIG. 5(a) illustrates a plane format
  • 5(b) illustrates a spherical format
  • FIG. 5(a) illustrates a format for developing the omni-directional spherical image into plane coordinates.
  • the plane format is an image including pixel values corresponding to angular coordinates with horizontal angles from 0 to 360 degrees and vertical angles from 0 to 180 degrees.
  • the angular coordinates are related to the points in spherical coordinates as
  • FIG. 5(b) which resemble latitude and longitude coordinates on a globe.
  • FIG. 6 illustrates the outline of generation processing of the omni-directional spherical image captured by the digital camera according to the embodiment.
  • FIG. 6(al) and (a2) are images captured with capturing elements through the two fisheye lenses;
  • FIG. 6(bl) and (b2) are images converted through a conversion table (corresponding to FIG. 5(a));
  • FIG. 6(c) is an image generated by combining (compositing) the two converted images (corresponding to FIG. 5(b)).
  • the respective images illustrated in FIG. 6 ⁇ al) and (a2) captured with capturing elements through the two fisheye ⁇ lenses , which are also schematically illustrated in FIG. 17, are converted into the images illustrated in FIG. 6(bl) and (b2).
  • This conversion is performed through the processing at S304 illustrated in FIG. 3, i.e., image conversion processing in which the corrected conversion table is used.
  • the images illustrated in FIG. 6(bl) and (b2) are presented in a corresponding manner to the omni-directional spherical image format, that is, corresponding to the images illustrated in FIG. 5(a).
  • the processing at S305 illustrated in FIG. 3 is performed. Specifically, two converted images are combined, thereby generating the omni-directional spherical image. More specifically, the two converted images are superimposed using a superimposed field of the images illustrated in FIG. 6(bl) and (b2) as a key and composed, whereby an image illustrated in FIG. 6(c), i.e., the omni-directional spherical image is generated.
  • FIG. 7 illustrates the conversion table for the omni-directional spherical image captured by the digital camera according to the embodiment .
  • FIG. 7 (a) is a diagram for explaining a conversion table representing a matrix of coordinates of a pre- conversion image and a post-conversion image.
  • FIG. 7(b) and (c) are diagrams for explaining a correspondence relation of the coordinates of the pre-conversion image (FIG. 7(b)) and the coordinates of post-conversion image (FIG. 7(c)).
  • FIG. 7 (a) illustrates the conversion table used for the image processing at S304 illustrated in FIG. 3.
  • the conversion table includes data sets of coordinates of pixel values of the post-conversion image ( ⁇ , ⁇ ) and the
  • the conversion table here represents a tabular data structure, however, another structure may be used as long as it is conversion data.
  • a post-conversion image is generated from the captured image (pre-conversion image) according to the conversion table illustrated in FIG. 7 (a) . Specifically, as
  • pixels in the post- conversion image are generated by referring to the
  • FIG. 8 is a flowchart for explaining operations of
  • the controller 10 obtains a camera tilt parameter ( ⁇ , ⁇ ) according to a tilt value of the digital camera 100 detected by the acceleration sensor 113 (FIG. 1) a and ⁇ in the camera tilt parameter ( ⁇ , ⁇ ) represent rotation angles, which will be described later.
  • values ( ⁇ 1, ⁇ ) of the conversion table are set.
  • values in a camera coordinate system are presented as ( ⁇ 0, ⁇ ) and values in a global coordinate system are presented as ( ⁇ 1, ⁇ ) so as to be differentiated from ( ⁇ , ⁇ ) , which are parameter values that depend on a coordinate system. Accordingly, values ( ⁇ 1, ⁇ ) in the global coordinate system are set at S802.
  • the input values ( ⁇ 1, ⁇ ) in the global coordinate system are transformed into the values ( ⁇ 0, ⁇ ) in the camera coordinate system through calculation of vertical correction by the controller 10. This calculation of vertical correction will be described later.
  • converted values ( ⁇ 0, ⁇ ) in the camera coordinate system are transformed into the coordinates (x, y) of the pre-conversion image through the conversion table not ⁇ yet corrected (FIG. 7(a)).
  • a conversion table for generating a correct omni-directional spherical image needs to be provided in advance.
  • the correct omni-directional spherical image is generated under the condition of the camera without being tilted during capture.
  • the conversion table for generating the correct omni-directional spherical image needs to be stored in a predetermined storage unit such as an SRAM or an SDRAM.
  • the controller 10 stores input values ( ⁇ 1, ⁇ ) in the global coordinate system and coordinates (x, y) not yet corrected, which are calculated in the end above, as a set of coordinates corresponding to each other in the conversion table already corrected.
  • the controller 10 determines whether any outstanding input values ( ⁇ 1, ⁇ ) remain. In other words, the controller 10 determines whether any outstanding input values ( ⁇ 1, ⁇ ) in the global coordinate system, in which the corresponding coordinates (x, y) not yet corrected is not calculated remain. If the controller 10 determines that outstanding input values ( ⁇ 1, ⁇ ) remain (Yes at S806) the processing returns to S802 so that input values ( ⁇ 1, ⁇ ) in the global coordinate system are set as a subsequent value.
  • controller 10 determines that no outstanding input values ( ⁇ 1, ⁇ ) remain (No at S806) , the processing ends. In this case, the controller 10 completes
  • FIG. 9 is a schematic diagram for explaining the tilt of the image capturing apparatus according to the embodiment.
  • the vertical direction corresponds to the z-axis in a three-dimensional (x, y, z) Cartesian
  • a tilt angle a in the gravity vector and a gradient angle ⁇ in the xy plane are obtained from the equation below using an output of the acceleration sensor.
  • Ax refers to a component value in the xO-axis direction in the camera coordinate system of the
  • acceleration sensor Ay refers to a component value in the yO-axis direction in the camera coordinate system of the .
  • acceleration sensor and Az refers to a component value in the zO-axis direction in the camera coordinate system of the acceleration sensor.
  • FIGS. 10(a) and 10(b) are diagrams for
  • FIG. 10(a) illustrates a camera coordinate system
  • three-dimensional Cartesian coordinates in the global coordinate system are presented as (xl, yl, zl) and spherical coordinates in the global coordinate system are presented as ( ⁇ 1, ⁇ ) .
  • three-dimensional Cartesian coordinates in the camera coordinate system are presented as (xO, yO, zO) and spherical coordinates in the camera coordinate system are presented as ( ⁇ 0, ⁇ ) .
  • FIG. 10 To correct a tilt, three- dimensional Cartesian coordinates are used to perform rotational transformation at first, thus spherical
  • a camera tilt parameter ( ⁇ , ⁇ ) is used to transform coordinates (xl, yl, zl) in the global
  • Equation (4) is the definition of the camera tilt parameter ( ⁇ , ⁇ ) .
  • FIG. 11 illustrates another example of calculation of vertical correction on the omnidirectional spherical image captured by the digital camera according to the embodiment.
  • FIG. 11(a) illustrates a camera coordinate system
  • FIG. 11(b) illustrates a global coordinate system.
  • Equations (1) to (6) described above and illustrated in FIG. 10 according to the present embodiment the calculation of vertical correction is accelerated. Equations (1) to (6) described above and illustrated in FIG. 10 according to the
  • the rotation a and rotation ⁇ around the z-axis is the rotation ⁇ itself in spherical coordinates ( ⁇ , ⁇ ) .
  • Calculation for rotational transformation can be performed with simple addition or subtraction operations, without transforming into Cartesian coordinate system, thereby accelerating the calculation. Accordingly,
  • FIG. 12 is another flowchart for explaining operations of correction processing through the conversion table for the omni-directional spherical image captured by the image capturing apparatus according to the embodiment.
  • the correction processing on the conversion table can be accelerated.
  • the conversion table corresponding to the value of the camera tilt parameter ( ⁇ , ⁇ ) is obtained , then the processing ends. Specifically, the calculation of vertical correction is omitted by storing a plurality of conversion tables so that the conversion tables provide different values depending on the camera tilt parameter ( a ⁇ ) ⁇
  • the camera tilt parameter ( ⁇ , ⁇ ) is a three- dimensional real vector, in principle. If conversion tables are provided only for specific camera tilt
  • FIG 13 is a schematic block diagram for explaining the overall structure of the image capture system according to the embodiment of the present invention.
  • the tilt correction is performed in a spherical image capturing apparatus (digital camera) , however, the tilt correction is not limited to be performed in a spherical image capturing apparatus.
  • a wired or wireless communication function is added on the spherical image capturing apparatus to obtain an information processing apparatus capable of performing image ' processing, such as a computer or a portable information terminal (e.g.,
  • the images before correction are transmitted to the information processing apparatus and processing for correcting the tilt can be performed in ' the destination information processing
  • the tilt correction processing requires tilt information on the spherical image capturing apparatus.
  • the tilt information on the spherical image capturing apparatus can be detected only in the image capturing apparatus itself, thus the tilt information needs to be transmitted along with the images before correction.
  • the conversion table may be stored in the spherical image capturing apparatus and transmitted along with the images. Alternatively, the conversion table may be transmitted to the destination device in advance as preparation for capturing (e.g., when the destination device recognizes the image capturing apparatus, or when the image capturing apparatus recognizes the destination device) .
  • the conversion table does not need to be transmitted along with the images every time.
  • the conversion table may be transmitted when the destination device requests the conversion table or checks whether the conversion table is the latest one.
  • the conversion table may not be transmitted necessarily from the image capturing
  • the conversion table may be uploaded to a website of the manufacturer that sells the image capturing apparatus so as to be downloaded and stored in the destination device.
  • the conversion table may be uploaded to a website of the manufacturer that sells the image capturing apparatus so as to be downloaded and stored in the destination device.
  • transmitted from the image capturing apparatus may be customized so as to suit the image capturing apparatus.
  • FIG. 14 is a schematic block diagram for explaining the overall structure of an electronic circuit of the destination device 200 of the image capture system according to another embodiment.
  • an SDRAM 1411 and an external storage device 1412 are coupled to a controller 140.
  • the controller 140 includes a CPU 1401, an SRAM 1402, a ROM 1403, an image processing block 1404, an SDRAM I/F 1405, and an external storage device I/F 1406.
  • the CPU 1401 includes a CPU 1401, an SRAM 1402, a ROM 1403, an image processing block 1404, an SDRAM I/F 1405, and an external storage device I/F 1406.
  • controller 140 also includes a USB connection interface 1407 for connecting to an external device and a wired or wireless network I/F 1408 for connecting to a network.
  • the image processing block 1404 performs typical image processing such as distortion correction and pixel defect correction, and the CPU 1401 reads a predetermined table or a computer program, thereby performing correction processing in the vertical direction according to the tilt of the digital camera 100. It is to be understood, however, that the image processing block 1404 may perform correction
  • FIG. 15 is a . flowchart for explaining operations of the. image capture system according to another embodiment.
  • FIG. 15 is a . flowchart for explaining operations of the. image capture system according to another embodiment.
  • the image is stored in a storage unit of the destination device.
  • Processing is performed in the digital camera 100 firstly.
  • a tilt angle of the digital camera 100 is detected by the acceleration sensor 113 (FIG. 1).
  • two fisheye images are transmitted to the destination device 200 such as a
  • the fisheye images have been captured with the image capturing element 1 (reference numeral: 109) and the image capturing element 2 (reference numeral: 110) and digitized.
  • the transmission above is performed through the USB connection interface 1407 or the network I/F 1408 as illustrated in FIG. 13. Additionally, tilt information on the digital camera and the conversion table are transmitted to the destination device 200 through the USB connection interface 1407 or the network I/F 1408.
  • the conversion table may be transmitted in advance when the digital camera 100 and the destination device 200 recognize each other.
  • conversion table may be transmitted to the destination device 200 only once, that is, the conversion table does not need to be transmitted every time.
  • the conversion table is stored in the SDRAM 1411 coupled to the digital camera 100, for example, and read out therefrom, then transmitted. This is the end of processing performed in the digital camera 100. The subsequent steps are performed in the destination device 200.
  • the conversion table is corrected in the controller 140 through a predetermined correction method according to the tilt information, which is the transmitted angular information of the digital camera.
  • the correction method for the conversion table is the same in the
  • the transmitted fisheye images captured through the two capturing elements are input to the image processing block 1404 of the
  • controller 140 through the conversion table corrected at S1503.
  • the transformation method is the same as in the embodiment described above.
  • a composed omni-directional spherical image is generated by the controller 140 utilizing the
  • the omni-directional spherical image generated at S1506 is stored by the controller 140 in the external storage device 1412 through the external storage device I/F 1406.
  • the operations relating to the flowcharts according to the embodiments illustrated in FIGS. 3, 8, 12, and 15 can be executed by a computer program.
  • the CPU (control circuit) 101 (FIG. 1) controlling operations of the image capturing apparatus or the.
  • CPU (control circuit) 1401 (FIG. 14) controlling operations of the destination device 200 loads various computer programs stored in a recording medium such as the ROMs 103 and 1403, the SRAMs 102 and 1402, and sequentially executes the programs.
  • the vertical direction is detected, and correction is performed on a conversion table used for image processing, and an omni-directional spherical image is generated

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)
  • Studio Devices (AREA)
  • Stereoscopic And Panoramic Photography (AREA)

Abstract

An image capturing apparatus includes a tilt detection unit that detects a tilt in the vertical direction, conversion data used for transforming plane coordinates into spherical coordinates, a correction unit that corrects the conversion data according to the tilt, a plurality of image capturing units, a coordinate transformation unit that transforms plane coordinates of a plurality of pixels included in images captured by the image capturing units into spherical coordinates according to the conversion data corrected by the correction unit, and a combining unit that combines the images including the pixels transformed into spherical coordinates by the coordinate transformation unit.

Description

DESCRIPTION
IMAGE CAPTURING APPARATUS, IMAGE CAPTURE SYSTEM, IMAGE PROCESSING METHOD, INFORMATION PROCESSING APPARATUS, AND
COMPUTER-READABLE STORAGE MEDIUM
TECHNICAL FIELD
The present invention relates to an image capturing apparatus, an image capture system, an information
processing apparatus, and a computer-readable storage medium.
BACKGROUND ART
Conventional image capturing apparatuses that capture images in all directions or spherically employ a
hyperboloidal mirror or a fisheye lens, for example. These image capturing apparatuses are specific optical systems that need a non-linear image processing such as distortion correction or projective transformation so that the captured image with the apparatus can be visualized to a user.
One of known technologies is that a semispherical image is obtained with a wide-angle capture employing a fisheye lens and recorded as a distorted circular image. A part of the distorted circular image is cut out and image processing is performed on the image with a computer, whereby the distorted image is transformed into a plane regular image.
Another known technology is that while performing the image processing above, when the center position of the distorted circular image does not correctly correspond to the vertical point direction, a user specifies a parameter of a tilt angle. The load while performing distortio'n correction on the image is therefore reduced.
As described above, when images are captured in a state of the image capturing apparatus being tilted, an omni-directional spherical image in the wrong vertical direction may be generated. Some inventions to solve this problem have been already known.
Japanese Patent Application Laid-open No. 2003-223633 and Japanese Patent Application Laid-open No. 2006-059202 disclose a technology of generating an omni-directional spherical image having the correct vertical direction. This is achieved by adding rotational transformation according to the tilt of a camera to the process of nonlinear image processing.
Japanese Patent Application Laid-open No. Hll-309137 and Japanese Patent No. 4175832 disclose a technology of generating an omni-directional spherical image having the correct vertical direction at high speed, for the purpose of reducing the time required for generating the omnidirectional spherical image having the correct vertical direction or reducing the cost on arithmetic operations. This technology is achieved by adding rotational
transformation according to the tilt of the camera, in addition to distortion correction or projective
transformation, preparing a conversion table used for nonlinear transformation in advance, and performing batch transformation through the conversion' table when capturing images.
With the structure of previous spherical image capturing apparatuses, however, users are required to specify a tilt angle of the apparatus during capture when combining a plurality of images captured at once from a plurality of image capturing directions, when transforming the combined image into a plane regular image, and when displaying the resulting image. That is to say, there was a problem in that a necessary parameter for correction could not be obtained during capture for automatically correcting the image.
Additionally, with the previously performed method for capturing images in all directions or spherically through a conversion table according to the tilt, if the amount or orientation of the tilt changes, the conversion table needs to be re-made from the beginning. In this case, processing requires a relatively long time to generate an omnidirectional spherical image having the correct vertical direction according to any tilt of the image capturing apparatus.
Japanese Patent Application Laid-open No. Hll-309137 and Japanese Patent No. 4175832 disclose in particular a technology generating images using a conversion table that includes rotational transformation according to the tilt of the camera. The conversion table is made so as to
correspond to the amount of tilt determined in advance. Accordingly, if the amount or orientation of the tilt changes so as to differ from the predetermined amount or orientation, the conversion table needs to be re-made from the beginning. Also in this case, the processing still requires a relatively long time to generate an omnidirectional spherical image having the correct vertical direction according to any tilt of the image capturing apparatus, therefore the problem is not yet solved.
Therefore, there is a need to provide an image
capturing apparatus capable of generating an omnidirectional spherical image having the correct vertical direction, according to any tilt of the image capturing apparatus . DISCLOSURE OF INVENTION
It is an object of the present invention to at least partially solve the problems in the conventional technology
According to an embodiment, there is provided an image capturing apparatus that includes a tilt detection unit configured to detect a tilt in a vertical direction;
conversion data used for transforming plane coordinates into spherical coordinates; a correction unit configured to correct the conversion data according to the tilt; a plurality of image capturing units; a coordinate
transformation unit configured to transform plane
coordinates of a plurality of pixels included in images captured by the image capturing units into spherical coordinates according to the conversion data corrected by the correction unit; and a combining unit configured to combine the images including the pixels transformed into spherical coordinates by the coordinate transformation unit
The above and other objects, features, advantages and technical and industrial significance of this invention will be better understood by reading the following detailed description of presently preferred embodiments of the invention, when considered in connection with the
accompanying drawings .
BRIEF DESCRIPTION OF DRAWINGS
FIG. 1 is a schematic block diagram for explaining the overall structure of an image capturing apparatus according to an embodiment of the present invention.
FIG. 2 is an external side view of an omni-directional spherical image capturing apparatus according to the embodiment .
FIG. 3 is a flowchart for explaining operations of the image capturing apparatus according to the embodiment. FIG. 4 illustrates a projection relation of a fisheye lens employed in the image capturing apparatus according to the embodiment .
FIG. 5 illustrates a format for the omni-directional spherical image captured by the image capturing apparatus according to the embodiment.
FIG. 6 illustrates the outline of generation
processing of the omni-directional spherical image captured by the image capturing apparatus according to the
embodiment .
FIG. 7 illustrates the conversion table for the omnidirectional spherical image captured by the image capturing apparatus according to the embodiment.
FIG. 8 is a flowchart for explaining operations of correction processing through the conversion table for the omni-directional spherical image captured by the image capturing apparatus according to the embodiment.
FIG. 9 is a schematic diagram for explaining a tilt of the image capturing apparatus according to the embodiment .
FIG. 10 illustrates an example of calculation of vertical correction on the omni-directional spherical image captured by the image capturing apparatus according to the embodiment .
FIG. 11 illustrates another example of calculation of vertical correction on the omni-directional spherical image captured by the image capturing apparatus according to the embodiment.
FIG. 12 is another flowchart for explaining operations of correction processing through the conversion table for the omni-directional spherical image captured by the image capturing apparatus according to the embodiment.
FIG. 13 is a schematic diagram for explaining the overall structure of an image capture system according to another embodiment of the present invention.
FIG. 14 is a schematic block diagram for explaining the overall structure of an electronic circuit serving as a destination device of the image capture system according to another embodiment.
FIG. 15 is a flowchart for explaining operations of the image capture system according to another embodiment.
FIG. 16 is a diagram for explaining the structure of a multi-eye image capturing apparatus that captures images in all directions spherically at once.
FIG. 17 is a diagram for explaining an image obtained through a lens in the multi-eye image capturing apparatus illustrated in FIG. 16.
FIG. 18 is a diagram for explaining an image obtained by performing distortion correction on images illustrated in FIG. 17 and then combining the images with each other.
FIG. 19 is a diagram for explaining an image captured by the multi-eye image capturing apparatus illustrated in FIG. 16, with the image capturing apparatus in a state of being tilted.
FIG. 20 is a diagram for explaining an image obtained by performing distortion correction and image combination on the image illustrated in FIG. 19 without consideration of the tilt.
FIG. 21 is a schematic diagram for explaining an acceleration sensor that measures a tilt angle.
BEST MODE(S) FOR CARRYING OUT THE INVENTION
Exemplary embodiments according to the present
invention are described below with reference to the
accompanying drawings. In the drawings and descriptions below, the same or equivalent components are designated by the same reference numerals, and overlapped explanation thereof will be simplified or omitted as appropriate. In the embodiments, when generating an omni-directional spherical image, the omni-directional spherical image capturing apparatus detects the vertical direction and corrects a conversion table used for image processing according to the vertical direction. In the embodiments, the omni-directional spherical image is generated through the corrected conversion table. Accordingly, the
conversion table does not need to be re-made from the beginning, thereby reducing the processing time. The image capturing apparatus refers to a stand-alone apparatus, such as a digital camera; an image capture system refers to a system in which a plurality of apparatuses such as a digital camera and an information processing apparatus are utilized separately. In the embodiments, unless otherwise specified, the image capturing apparatus conceptually includes an image capture system.
A multi-eye image capturing apparatus will now be described that captures images in all directions
spherically at once. FIG. 16 is a diagram for explaining the structure of a multi-eye image capturing apparatus that captures images in all directions spherically at once. An image capture system is prepared that employs a plurality of wide-angle lenses as illustrated in FIG. 16 so as to capture images in all directions spherically at once.
For example, the multi-eye image capturing apparatus illustrated on the left in FIG. 16 employs a fisheye lens (ultra wide-angle lens) with an angle of view equal to or larger than 180 degrees. The multi-eye image capturing apparatus requires at least the two-eye structure (image capture lenses with two different image capturing
directions) to capture images in all directions spherically at once. If image distortion needs to be reduced as much as possible, the number of image capture lenses may be increased by employing, for example, the four-eye structure for the multi-eye image capturing apparatus as illustrated on the right in FIG. 16. With this structure, only center parts with high image quality are selectively used to reduce the image distortion. In this image capturing apparatus with the four-eye structure, it is recommended that respective lenses have an angle of view equal to or larger than 90 degrees, and preferably about 100 degrees.
The image capturing apparatus with two-eye structure will be described hereafter, for convenience of description. The basic idea of the present invention may be applied to the image capturing apparatus with the four-eye structure illustrated on the right in FIG. 16, however.
In the image capturing apparatus, it is assumed that the central axis of the image capturing apparatus in the vertical direction corresponds to the vertical axis of the sphere. Specifically, an image obtained through each of two fisheye lenses is presented as illustrated in FIG. 17. FIG. 17 is a diagram for explaining an image obtained through a lens in a two-eye image capturing apparatus illustrated on the left in FIG. 16. As described above, it is preferred that images are captured so that the vertical point and the horizontal line recognized by a user of the image capturing apparatus correspond to the vertical point and the horizontal line of the image capturing apparatus.
The borderline of 180 degrees illustrated in FIG. 17 will be described. In an image capture system including two lenses (two image capturing directions), two images are captured so that the images have a superimposed field with each other. The images are combined utilizing the
superimposed field. The outside part of the borderline of 180 degrees is the superimposed field. When two lenses are employed for the image capturing apparatus, accordingly, a lens has an angle of view equal to or larger than 180 degrees, and preferably about 190 degrees so as to provide the superimposed field. When four lenses are employed for the image capturing apparatus, a lens has an angle of view of about 100 degrees so as to provide an appropriate superimposed field.
With the structure employing four lenses illustrated on the right in FIG. 16 and when images are captured with the image capturing apparatus correctly being upright, the horizontal line is captured just horizontally in the same manner described above. The position of the vertical point recognized by the image capturing person using the
apparatus corresponds to the position of the vertical point of the image capturing apparatus. The above-described technique is widely known for combining captured images to perform distortion correction on the combined image.
For example, when distortion correction is performed on a plurality of images and the images are combined with each other in the same manner as in Mercator projection, an image illustrated in FIG. 18 is obtained. FIG. 18 is a diagram for explaining an image obtained by performing distortion correction on images illustrated in FIG. 17 and then combining the images with each other.
When the image capturing apparatus is placed so as to be correctly upright and images are captured without the apparatus being tilted, a correct image can be obtained having the straight horizontal line as illustrated in FIG. 18 after distortion correction and combination are simply performed on the captured images. Specifically, images can be captured in the correct · vertical direction when securing the image capturing apparatus to a certain retainer, and adjusting the horizontal and vertical lines of the apparatus using a level, for example.
FIG. 19 is a diagram for explaining an image captured by the multi-eye (two-eye) image capturing apparatus illustrated on the right in FIG. 16, with the image
capturing apparatus in a state of being tilted. When a person holds the image capturing apparatus in its hand(s), it is 'generally difficult to capture images horizontally and vertically. FIG. 19 exemplifies images captured in such a state. As illustrated in FIG. 19 the vertical points of the images do not correspond to each other and the horizontal lines of the images are distorted. As described above, if distortion correction and combination are performed on images captured with the image capturing apparatus in a state of being tilted without consideration of the tilt, an image is obtained in which the distortion exemplified in FIG. 19 is directly reflected, as
illustrated in FIG. 20.
If distortion correction and combination are performed' on images captured with the image capturing apparatus in a state of being tilted without consideration of the tilt., the horizontal line is curved like a graph of a
trigonometric function as illustrated in FIG. 20. In addition, the orthogonality of the image is broken. To avoid these problems, correction should be performed in consideration of a tilt angle in the vertical direction.
The principles of measurement of a tilt angle will now be described with reference to FIG. 21. FIG. 21 is a schematic diagram for explaining an acceleration sensor that measures a tilt angle.
An acceleration sensor as illustrated in FIG. 21 is
embedded in the image capturing apparatus so as to measure how the image capturing apparatus tilts in the vertical direction using a gravity ' acceleration sensor. FIG. 21 illustrates the outline of obtaining an angle using a single-axis acceleration sensor. This is employed to simplify the description. In FIG. 21, the sensor obtains only the tilt of the image capt ring apparatus within the plane including the central axis of the lens surface in the two-eye structure. When a user actually captures images, however, it is assumed that images may be captured from the deviated angle from the plane as
described above. To solve. this problem a three-axis acceleration sensor is employed so as to measure the deviated angle from the plane including the central plane of the lens surface.
The image capturing apparatus according to the
embodiment will now be described in detail. FIG. 1 is a schematic block diagram for explaining the overall
structure of an electrical circuit of the image capturing apparatus according to the embodiment of the present invention. As illustrated in FIG. 1, in this image
capturing apparatus (hereinafter, also referred to as a digital camera) 100, an image capturing element 1
(reference numeral: 109), an image capturing element 2 (reference numeral: 110), a synchronous dynamic random access memory (SDRAM) 111, an external storage device 112, and the acceleration sensor 113 are coupled to a controllej 10.
In the embodiment, two capturing elements (i.e., the two-eye structure) are employed so as to obtain omnidirectional images, however, three or more capturing elements may be employed. When three capturing elements are employed, a lens corresponding to the image capturing element, which will be described with reference to FIG. 2, does not require an angle of view equal to or larger than 180 degrees. Another angle of view of the lens may be employed as appropriate. Wide-angle' lenses including a fisheye lens are typically employed for the lens. The image capturing apparatus is not limited to the omnidirectional image capturing apparatus.. Another image capturing apparatus capable of capturing images covering 360 degrees in the horizontal direction may be used.
The controller 10 includes a central processing unit (CPU) 101, a static RAM (SRAM) 102, a read-only memory
(ROM) 103, an image processing block 104, an SDRAM
interface (I/F) 105, an external storage device I/F 106, and an external sensor I/F 107.
In the description of the embodiment, the image
processing block 104 performs typical image processing such as distortion correction and pixel defect correction, and the CPU reads a predetermined table or a computer program, thereby performing correction processing in the vertical direction according to the tilt of the digital camera 100. It is to be understood, however, that the image processing block 104 may perform correction processing in the vertical direction.
During capture, pieces of digitized image data are input to the image processing block 104 of the controller 10 through the image capturing element 1 (reference
numeral: 109) and the image capturing element 2 (reference numeral: 110) . The pieces of image data input are image- processed using the image processing block 104, the CPU 101, the SRAM 102, and the SDRAM 111, for example, and finally stored in the external storage device 112. Examples of an external storage device include CompactFlash (registered trademark) or a Secure Digital (SD) memory.
In the controller 10, a universal serial bus (USB) connection interface for connecting to an external device or a wired or wireless network I/F for connecting to a network may be provided. A conversion table, which will be described later, a correction processing program for the conversion table, and a processing program for calculation of vertical correction are stored in the SRAM 102 or the SDRAM 111.
The acceleration sensor 113 is used for detecting a tilt of the digital camera 100 during capture. This enables detection in the tilting direction of the digital camera instantly and readily.
The acceleration sensor 113 is a' three-axis
acceleration sensor that detects acceleration of the digital camera 100 in the three directions perpendicular to each other, i.e., the up and down direction, the right and left direction, and the back and front direction. When a user holds the digital camera 100 in its hand(s) so that the digital camera 100 remains still, the acceleration sensor 113 detects the gravity acceleration only.
When acceleration is detected only in the down
direction in the up and down direction, it is found that the up and down direction of the digital camera 100 corresponds to the top and bottom direction with respect to the ground. In other words, it is known that the digital camera is held horizontally in the same manner usually operating a digital camera.
When the image capturing apparatus tilts in the top and bottom direction, the acceleration sensor 113 detects acceleration in the right and left direction and the back and front direction according to the actual tilting
direction. A tilt angle of the digital camera 100 can be obtained according to the magnitude of acceleration in the up and down direction, the right and left direction, and the back and front direction.
The omni-directional spherical image capturing apparatus will now be described. FIG. 2 is an external side view of the omni-directional spherical image capturing apparatus according to the embodiment.
The present embodiment aims to generate the omnidirectional spherical image using the omni-directional spherical image capturing apparatus capable of capturing images in all directions from a capture point. That is to say, the omni-directional spherical image capturing
apparatus (digital camera) can capture images in all directions viewed from the capture point.
The digital camera 100 (illustrated in FIG. 1) serving as an omni-directional spherical image capturing apparatus, captures images through the two. capturing element 1
(reference numeral: 109) and the image capturing element 2 (reference numeral: 110). The respective capturing
elements have a fisheye lens as an example of wide-angle lenses having an angle of view exceeding 180 degrees. The images captured through these two fisheye lenses have a superimposed field with each other. A predetermined image processing such as distortion correction is performed on the images, and the resulting images are converted, and then combined with each other, thereby generating an omnidirectional spherical image.
Operations of the omni-directional spherical image capturing apparatus according to the embodiment will now be described. FIG. 3 is a flowchart for explaining operations of the image capturing apparatus according to the
embodiment. An explanation will be given on the operations of the omni-directional spherical image capturing apparatus from the time when^ captured images are input to the time when the images are stored in the external storage device 112 (FIG. 1) with reference to FIG. 3.
The acceleration sensor 113 detects a tilt angle of the digital camera 100 at Step' (hereinafter, noted, as "S") 301. At S302, the controller 10 reads out a conversion table stored in the SDRAM 111, for example, according to a tilt angle of the digital camera 100 detected at S301. The controller 10 then corrects the conversion table in a predetermined correction method. The correction method for the conversion table will be described later.
At S303, two digitized fisheye images captured through the image capturing element 1 (reference numeral: 109) and the image capturing element 2 (reference numeral: 110) are input to the image processing block 104. The image
processing block 104 performs typical image processing such as distortion correction. At S304, the controller 10 converts the two captured fisheye images (like the image as illustrated in FIG. 17) using the conversion table
corrected at S302. The conversion method will be described later.
At S305, the controller 10 generates a composed omnidirectional spherical image utilizing a superimposed field between the two images converted at S304. At S306, the controller 10 stores the omni-directional spherical image generated at S305 in the external storage device 112 through the external storage device I/F 106.
A projection relation of the fisheye lens will now be described as an example of wide-angle lenses employed in the digital camera according to the embodiment. FIG. 4 illustrates a projection relation of a fisheye lens
employed in the digital camera according to the embodiment. FIG. 4(a) is an external . side view of the fisheye lens; FIG (b) illustrates a projection function f in a plan view of a captured image.
FIG. 17 illustrates an image captured through the fisheye lens with an angle of view over 180 degrees, which is a captured image of an almost hemispherical scene viewed from a capture position. As illustrated in (a) and (b) of FIG. 4, the image is generated with an image height h according to an angle of incidence Θ. The relation of the angle of incidence Θ and the image height h is defined by a projection function f (h = f (Θ) ) . The projection function varies depending on the characteristic of the fisheye lens.
Examples of a technique (function) of projective transformation include central projection,, stereographic projection, equidistant projection, equi-solid-angle projection, and orthographic projection. The central projection is employed when capturing images, using a digital camera including a typical angle of view. Other four methods above are employed in a digital camera
including a wide-angle lens with an ultra wide angle of view such as a fisheye lens.
A format (form of equation) of the omni-directional spherical image will now be described that is captured by the digital camera according to the embodiment. FIG. 5 illustrates a format of the omni-directional spherical image captured by the digital camera according to the . embodiment. FIG. 5(a) illustrates a plane format; FIG.
5(b) illustrates a spherical format.
FIG. 5(a) illustrates a format for developing the omni-directional spherical image into plane coordinates. As illustrated in FIG. 5(a), the plane format is an image including pixel values corresponding to angular coordinates with horizontal angles from 0 to 360 degrees and vertical angles from 0 to 180 degrees. The angular coordinates are related to the points in spherical coordinates as
illustrated in FIG. 5(b), which resemble latitude and longitude coordinates on a globe. Plane coordinates of images captured through the fisheye lens and spherical coordinates of the omnidirectional spherical image can be related to each other using the projection function f (h = f(0)) illustrated in FIG. 4. This enables conversion of two images captured through the fisheye lens and combination (composition) of the resulting images, thereby generating omni-directional spherical images as illustrated in FIG. 5(a) and (b) .
The process of generating the omni-directional spherical image will now be described with reference to an actual image captured through the fisheye lens. FIG. 6 illustrates the outline of generation processing of the omni-directional spherical image captured by the digital camera according to the embodiment. FIG. 6(al) and (a2) are images captured with capturing elements through the two fisheye lenses; FIG. 6(bl) and (b2) are images converted through a conversion table (corresponding to FIG. 5(a)); FIG. 6(c) is an image generated by combining (compositing) the two converted images (corresponding to FIG. 5(b)).
The respective images illustrated in FIG. 6{al) and (a2) captured with capturing elements through the two fisheye lenses , which are also schematically illustrated in FIG. 17, are converted into the images illustrated in FIG. 6(bl) and (b2). This conversion is performed through the processing at S304 illustrated in FIG. 3, i.e., image conversion processing in which the corrected conversion table is used. At this point, the images illustrated in FIG. 6(bl) and (b2) are presented in a corresponding manner to the omni-directional spherical image format, that is, corresponding to the images illustrated in FIG. 5(a).
Subsequently, the processing at S305 illustrated in FIG. 3 is performed. Specifically, two converted images are combined, thereby generating the omni-directional spherical image. More specifically, the two converted images are superimposed using a superimposed field of the images illustrated in FIG. 6(bl) and (b2) as a key and composed, whereby an image illustrated in FIG. 6(c), i.e., the omni-directional spherical image is generated.
The correction method for the conversion table at S304 illustrated in FIG. 3 will now be described. FIG. 7 illustrates the conversion table for the omni-directional spherical image captured by the digital camera according to the embodiment .
FIG. 7 (a) is a diagram for explaining a conversion table representing a matrix of coordinates of a pre- conversion image and a post-conversion image. FIG. 7(b) and (c) are diagrams for explaining a correspondence relation of the coordinates of the pre-conversion image (FIG. 7(b)) and the coordinates of post-conversion image (FIG. 7(c)).
FIG. 7 (a) illustrates the conversion table used for the image processing at S304 illustrated in FIG. 3. The conversion table includes data sets of coordinates of pixel values of the post-conversion image ( θ , φ ) and the
corresponding coordinates of pixel values of the pre- conversion image (x, y) , for every coordinates of the post- conversion image. The conversion table here represents a tabular data structure, however, another structure may be used as long as it is conversion data.
A post-conversion image is generated from the captured image (pre-conversion image) according to the conversion table illustrated in FIG. 7 (a) . Specifically, as
illustrated in FIG. 7(b) and (c) , pixels in the post- conversion image are generated by referring to the
corresponding pixel values in the coordinates in the pre- conversion image, based on the correspondence relation between the pre-conversion image and the post-conversion image in the conversion table (FIG. 7(a)).
An explanation will now be given on processing of correcting the conversion table for the omni-directional spherical image captured by the digital camera according to the embodiment, according to the tilt of the digital camera FIG. 8 is a flowchart for explaining operations of
correction processing on the conversion table for the omnidirectional spherical image captured by the image capturing apparatus according to the embodiment.
At S801, the controller 10 obtains a camera tilt parameter (α, β) according to a tilt value of the digital camera 100 detected by the acceleration sensor 113 (FIG. 1) a and β in the camera tilt parameter (α, β) represent rotation angles, which will be described later.
At S802, input values (Θ1, φΐ) of the conversion table are set. In FIG. 8, values in a camera coordinate system are presented as (Θ0, φθ) and values in a global coordinate system are presented as (Θ1, φΐ) so as to be differentiated from (θ, φ) , which are parameter values that depend on a coordinate system. Accordingly, values (Θ1, φΐ) in the global coordinate system are set at S802.
At S803, the input values (Θ1, φΐ) in the global coordinate system are transformed into the values (Θ0, φθ) in the camera coordinate system through calculation of vertical correction by the controller 10. This calculation of vertical correction will be described later.
At S804, converted values (Θ0, φθ) in the camera coordinate system are transformed into the coordinates (x, y) of the pre-conversion image through the conversion table not · yet corrected (FIG. 7(a)). It should be noted that a conversion table for generating a correct omni-directional spherical image needs to be provided in advance. The correct omni-directional spherical image is generated under the condition of the camera without being tilted during capture. The conversion table for generating the correct omni-directional spherical image needs to be stored in a predetermined storage unit such as an SRAM or an SDRAM.
At S805, the controller 10 stores input values (Θ1, φΐ) in the global coordinate system and coordinates (x, y) not yet corrected, which are calculated in the end above, as a set of coordinates corresponding to each other in the conversion table already corrected.
At S806, the controller 10 determines whether any outstanding input values (Θ1, φΐ) remain. In other words, the controller 10 determines whether any outstanding input values (Θ1, φΐ) in the global coordinate system, in which the corresponding coordinates (x, y) not yet corrected is not calculated remain. If the controller 10 determines that outstanding input values (Θ1, φΐ) remain (Yes at S806) the processing returns to S802 so that input values (Θ1, φΐ) in the global coordinate system are set as a subsequent value.
If the controller 10 determines that no outstanding input values (Θ1, φΐ) remain (No at S806) , the processing ends. In this case, the controller 10 completes
calculation of coordinates (x, y) not yet corrected
corresponding to the pixels in the format of the omnidirectional spherical ' image that have input values (Θ1, φΐ) in the global coordinate system as coordinates.
The tilt of the digital camera 100 according to the embodiment will now be described. FIG. 9 is a schematic diagram for explaining the tilt of the image capturing apparatus according to the embodiment.
In FIG. 9, the vertical direction corresponds to the z-axis in a three-dimensional (x, y, z) Cartesian
coordinates in the global coordinate system. When this direction corresponds to the vertical direction of the digital camera illustrated in FIG. 9, the camera is in a state of not being tilted. When the direction does not correspond to the^ vertical direction of the digital camera, the digital camera is in a state of being tilted.
A tilt angle a in the gravity vector and a gradient angle β in the xy plane are obtained from the equation below using an output of the acceleration sensor. In the equation, Ax refers to a component value in the xO-axis direction in the camera coordinate system of the
acceleration sensor, Ay refers to a component value in the yO-axis direction in the camera coordinate system of the . acceleration sensor, and Az refers to a component value in the zO-axis direction in the camera coordinate system of the acceleration sensor.
a = Arc tan (Ax / Ay)
Figure imgf000023_0001
The calculation of vertical correction will now be described. FIGS. 10(a) and 10(b) are diagrams for
explaining an example of calculation of vertical correction on the omni-directional spherical image captured by the digital camera according to the embodiment. FIG. 10(a) illustrates a camera coordinate system; FIG. 10 (b)
illustrates a global coordinate system.
In FIG. 10, three-dimensional Cartesian coordinates in the global coordinate system are presented as (xl, yl, zl) and spherical coordinates in the global coordinate system are presented as (Θ1, φΐ) . In addition, three-dimensional Cartesian coordinates in the camera coordinate system are presented as (xO, yO, zO) and spherical coordinates in the camera coordinate system are presented as (Θ0, φθ) .
Spherical coordinates (Θ1, φΐ) are transformed into spherical coordinates (Θ0, φθ) through the equations
illustrated in FIG. 10. To correct a tilt, three- dimensional Cartesian coordinates are used to perform rotational transformation at first, thus spherical
coordinates (Θ1, φΐ) are transformed into three-dimensional Cartesian coordinates (xl, yl, zl) through Equations (1) to (3) illustrated in FIG. 10.
Subsequently, a camera tilt parameter (α, β) is used to transform coordinates (xl, yl, zl) in the global
coordinate system into coordinates (xO, yO, zO) in the camera coordinate system through the rotational coordinate transformation illustrated in FIG. 10 (Equation (4)) . In other words, this equation (Equation (4) illustrated in FIG. 10) is the definition of the camera tilt parameter (α, β) .
This means that the camera coordinate system is
obtained by rotating the global coordinate system by a
around the z-axis, and then rotating the global coordinate system by β around the x-axis. Finally, the three- dimensional Cartesian coordinates (xO, yO, zO) in the camera coordinate system are transformed back to spherical coordinates (Θ0, φθ) in the camera coordinate system
through Equations (5) and (6) illustrated in FIG. 10.
Another example of calculation of vertical correction will now be described. FIG. 11 illustrates another example of calculation of vertical correction on the omnidirectional spherical image captured by the digital camera according to the embodiment. FIG. 11(a) illustrates a camera coordinate system; FIG. 11(b) illustrates a global coordinate system.
In the present embodiment, the calculation of vertical correction is accelerated. Equations (1) to (6) described above and illustrated in FIG. 10 according to the
embodiment may be presented as Equations (7) to (14)
illustrated in FIG. 11.
Specifically, the rotation a and rotation γ around the z-axis is the rotation Θ itself in spherical coordinates (Θ, φ) . Calculation for rotational transformation can be performed with simple addition or subtraction operations, without transforming into Cartesian coordinate system, thereby accelerating the calculation. Accordingly,
transformation with Cartesian coordinate system is required only for the rotational transformation of the rotation β around the x-axis. This speeds up the calculation.
An explanation will now be given on another processing of correcting the conversion table for the omni-directional spherical image captured by the digital camera according to the embodiment, according to the tilt of the digital camera. FIG. 12 is another flowchart for explaining operations of correction processing through the conversion table for the omni-directional spherical image captured by the image capturing apparatus according to the embodiment.
In the present embodiment, the correction processing on the conversion table can be accelerated. In the
embodiment illustrated in FIG. 11, the correction
processing on the conversion table is performed while accelerating the calculation of vertical correction. By contrast, the calculation of vertical correction is omitted to accelerate the calculation in the present embodiment.
With reference to FIG. 12, the correction processing on the conversion table will be described. At S1201, a camera tilt parameter ( α, β ) 'is obtained. This processing is the same as the processing at S801 illustrated in FIG. 8
At S1202, the conversion table corresponding to the value of the camera tilt parameter ( α , β ) is obtained , then the processing ends. Specifically, the calculation of vertical correction is omitted by storing a plurality of conversion tables so that the conversion tables provide different values depending on the camera tilt parameter ( a β ) ·
The camera tilt parameter ( α, β ) is a three- dimensional real vector, in principle. If conversion tables are provided only for specific camera tilt
parameters ( , β ) , and the conversion table closest to the detected camera tilt parameter ( α , β ) is used, all
parameters are covered with the provided tables.
Alternatively, it is also effective that a plurality of tables close to the detected camera tilt parameter ( α , β ) are extracted and an interpolation operation such as weighting or obtaining difference. This enables correction of the conversion table only with an interpolation
operation, which is relatively simple, thereby suppressing the processing on the operation.
The overall structure of the image capture system according to another embodiment will no be described. FIG 13 is a schematic block diagram for explaining the overall structure of the image capture system according to the embodiment of the present invention. In the embodiment described above, the tilt correction is performed in a spherical image capturing apparatus (digital camera) , however, the tilt correction is not limited to be performed in a spherical image capturing apparatus.
- As illustrated" n FIG. 13, a wired or wireless communication function is added on the spherical image capturing apparatus to obtain an information processing apparatus capable of performing image' processing, such as a computer or a portable information terminal (e.g.,
smartphone tablet personal computer) . The images before correction are transmitted to the information processing apparatus and processing for correcting the tilt can be performed in ' the destination information processing
apparatus (destination device) .
The tilt correction processing, however, requires tilt information on the spherical image capturing apparatus. The tilt information on the spherical image capturing apparatus can be detected only in the image capturing apparatus itself, thus the tilt information needs to be transmitted along with the images before correction. The conversion table may be stored in the spherical image capturing apparatus and transmitted along with the images. Alternatively, the conversion table may be transmitted to the destination device in advance as preparation for capturing (e.g., when the destination device recognizes the image capturing apparatus, or when the image capturing apparatus recognizes the destination device) .
The conversion table does not need to be transmitted along with the images every time. The conversion table may be transmitted when the destination device requests the conversion table or checks whether the conversion table is the latest one. In addition, the conversion table may not be transmitted necessarily from the image capturing
apparatus. For example, the conversion table may be uploaded to a website of the manufacturer that sells the image capturing apparatus so as to be downloaded and stored in the destination device. The conversion table
transmitted from the image capturing apparatus may be customized so as to suit the image capturing apparatus.
An explanation will now be given on the destination device 200 in an image capture system according to another embodiment. FIG. 14 is a schematic block diagram for explaining the overall structure of an electronic circuit of the destination device 200 of the image capture system according to another embodiment. In FIG. 14, an SDRAM 1411 and an external storage device 1412 are coupled to a controller 140.
The controller 140 includes a CPU 1401, an SRAM 1402, a ROM 1403, an image processing block 1404, an SDRAM I/F 1405, and an external storage device I/F 1406. The
controller 140 also includes a USB connection interface 1407 for connecting to an external device and a wired or wireless network I/F 1408 for connecting to a network.
In the description of the embodiment above, the image processing block 1404 performs typical image processing such as distortion correction and pixel defect correction, and the CPU 1401 reads a predetermined table or a computer program, thereby performing correction processing in the vertical direction according to the tilt of the digital camera 100. It is to be understood, however, that the image processing block 1404 may perform correction
processing' in the vertical direction.
Operations of the image capture system according to another embodiment will now be described. FIG. 15 is a. flowchart for explaining operations of the. image capture system according to another embodiment. FIG. 15
illustrates that after a captured image is input, the image is stored in a storage unit of the destination device.
Processing is performed in the digital camera 100 firstly. At S1501, a tilt angle of the digital camera 100 is detected by the acceleration sensor 113 (FIG. 1). At S1502, two fisheye images (as illustrated in FIG. 17) are transmitted to the destination device 200 such as a
portable information terminal or a computer by the
controller 10. The fisheye images have been captured with the image capturing element 1 (reference numeral: 109) and the image capturing element 2 (reference numeral: 110) and digitized. The transmission above is performed through the USB connection interface 1407 or the network I/F 1408 as illustrated in FIG. 13. Additionally, tilt information on the digital camera and the conversion table are transmitted to the destination device 200 through the USB connection interface 1407 or the network I/F 1408.
As described above, the conversion table may be transmitted in advance when the digital camera 100 and the destination device 200 recognize each other. The
conversion table may be transmitted to the destination device 200 only once, that is, the conversion table does not need to be transmitted every time.
The conversion table is stored in the SDRAM 1411 coupled to the digital camera 100, for example, and read out therefrom, then transmitted. This is the end of processing performed in the digital camera 100. The subsequent steps are performed in the destination device 200.
At S1503, the conversion table is corrected in the controller 140 through a predetermined correction method according to the tilt information, which is the transmitted angular information of the digital camera. The correction method for the conversion table is the same in the
embodiment described above. At S1504, the transmitted fisheye images captured through the two capturing elements are input to the image processing block 1404 of the
destination device 200. In the image processing block 1404, typical image processing such as distortion correction is performed. At S1505, the two fisheye images are transformed by the
controller 140 through the conversion table corrected at S1503. The transformation method is the same as in the embodiment described above.
At S1506, a composed omni-directional spherical image is generated by the controller 140 utilizing the
superimposed field of the two images transformed at S1505. At S1507, the omni-directional spherical image generated at S1506 is stored by the controller 140 in the external storage device 1412 through the external storage device I/F 1406.
The operations relating to the flowcharts according to the embodiments illustrated in FIGS. 3, 8, 12, and 15 can be executed by a computer program. Specifically, the CPU (control circuit) 101 (FIG. 1) controlling operations of the image capturing apparatus or the. CPU (control circuit) 1401 (FIG. 14) controlling operations of the destination device 200 loads various computer programs stored in a recording medium such as the ROMs 103 and 1403, the SRAMs 102 and 1402, and sequentially executes the programs.
As described above, in the omni-directional spherical image capturing apparatus, or in the image capture system, the vertical direction is detected, and correction is performed on a conversion table used for image processing, and an omni-directional spherical image is generated
through the corrected conversion table. This eliminates the necessity of re-making the conversion table from the beginning, thereby reducing the processing time.
Although the invention has been described with respect to specific embodiments for a complete and clear disclosure, the appended claims are not to be thus limited but are to be construed as embodying all modifications and alternative constructions that may occur to one skilled in the art that fairly fall within the basic teaching herein set forth.

Claims

1. An image capturing apparatus comprising:
a tilt detection unit configured to detect a tilt in a vertical direction;
conversion data used for transforming plane
coordinates into spherical coordinates;
a correction unit configured to correct the conversion data according to the tilt;
a plurality of image capturing units;
a coordinate transformation unit configured to
transform plane coordinates of a plurality of pixels included in images captured by the image capturing units into spherical coordinates according to the conversion data corrected by the correction unit; and
a combining unit configured to combine the images including the pixels transformed into spherical coordinates by the coordinate transformation unit.
2. The image capturing apparatus according to claim 1, wherein
the tilt is a gradient vector in three dimensions from the vertical direction,
conversion data corresponding to the gradient vector is provided in advance, and
the conversion data corresponding to the gradient vector closest to the tilt detected is extracted out of the conversion data provided in advance.
3. The image capturing apparatus according to claim 1, wherein
the tilt is a gradient vector in three dimensions from the vertical direction, conversion data corresponding to the gradient vector is provided in advance,
the conversion data corresponding to the gradient vector closest to the tilt detected is extracted out of the conversion data provided in advance, and
a difference between the conversion data extracted and conversion data corresponding to the tilt detected is interpolated.
4. The image capturing apparatus according to any one of claims 1 to 3, wherein the tilt is detected by an
acceleration sensor.
5. An image capture system comprising:
an image capturing apparatus; and
an information processing apparatus coupled to the image capturing apparatus through a network, wherein the image capture system includes
a tilt detection unit configured to detect a tilt in the vertical direction;
conversion data used for transforming plane coordinates into spherical coordinates;
a correction unit configured to correct the conversion data according to the tilt;
a plurality of image capturing units;
a coordinate transformation unit configured to transform plane coordinates of a plurality of pixels included in images captured by the image capturing units into spherical coordinates according to the conversion data corrected by the correction unit; and
a combining unit configured to combine the images including the pixels transformed into spherical coordinates by the coordinate transformation unit.
6. The image capture system according to claim 5, wherein the image capturing apparatus transmits the tilt, the conversion data, and a plurality of images captured by the image capturing units to the information processing
apparatus, and
the information processing apparatus corrects the conversion data according to the tilt transmitted, performs coordinate transformation on the images transmitted
according to the conversion data corrected, and combines the images obtained through the coordinate transformation.
7 . An image processing method comprising:
detecting a tilt in the vertical direction by a tilt detection unit;
transforming plane coordinates into spherical
coordinates using conversion data;
correcting the conversion data according to the tilt by a correction unit;
transforming plane coordinates of a plurality of pixels included in images captured by a plurality of image capturing units into spherical coordinates by a coordinate transformation unit according to the conversion data corrected at the correcting; and
combining the images including the pixels transformed into spherical coordinates at the transforming, by a combining unit.
8. A computer-readable storage medium with an executable program stored thereon, wherein the program instructs a computer as an image capturing apparatus to perform:
detecting a tilt in the vertical direction by a tilt detection unit; transforming plane coordinates into spherical
coordinates using conversion data;
correcting the conversion data according to the tilt; transforming plane coordinates of a plurality of pixels included in images captured by a plurality of image capturing units into spherical coordinates according to the conversion data corrected at the correcting; and
combining the images including the pixels transformed into spherical coordinates at the transforming.
9. An information processing apparatus coupled to an image capturing apparatus through a network, the
information processing apparatus comprising:
a storage unit configured to store therein conversion data used for transforming plane coordinates into spherical coordinates;
a receiving unit configured to receive a tilt in the vertical direction and a plurality of images captured from the image capturing apparatus;
a correction unit configured to correct the conversion data according to the tilt;
a coordinate transformation unit configured to
transform plane coordinates of a plurality of pixels included in the images captured into spherical coordinates according to the conversion data corrected by the
correction unit; and
a combining unit configured to combine the images including the pixels transformed into spherical coordinates by the coordinate transformation unit.
10. A- computer-readable storage medium with an executable program stored thereon, wherein the program instructs a computer as an information processing apparatus coupled to an image capturing apparatus through a network to perform: storing conversion data used for transforming plane coordinates into spherical coordinates;
receiving a tilt in the vertical direction and a plurality of images captured from the image capturing apparatus ;
transforming plane coordinates into spherical
coordinates using conversion data;
correcting the conversion data according to the tilt; transforming plane coordinates of a plurality of pixels included in images captured by a plurality of image capturing units into spherical coordinates according to th' conversion data corrected at the correcting; and
combining the images including the pixels transformed into spherical coordinates at the transforming.
PCT/JP2013/057338 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium WO2013133456A1 (en)

Priority Applications (13)

Application Number Priority Date Filing Date Title
KR1020167036128A KR101770088B1 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
CA2866621A CA2866621C (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
KR1020187031508A KR101961364B1 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
EP13757254.1A EP2823637B1 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
CN201380012472.2A CN104160693B (en) 2012-03-09 2013-03-11 Image capture apparatus, image capture system, image processing method, information processing unit and computer readable storage medium
KR1020147025261A KR101613387B1 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
KR1020197007730A KR102046032B1 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
EP17155235.9A EP3182696B1 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
CN201810707970.3A CN108876714B (en) 2012-03-09 2013-03-11 Image capturing apparatus, information processing apparatus, and image processing method
KR1020167008975A KR101692194B1 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
US14/384,027 US9607358B2 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
KR1020177022160A KR101916616B1 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
US15/400,610 US11049215B2 (en) 2012-03-09 2017-01-06 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2012053190 2012-03-09
JP2012-053190 2012-03-09
JP2012-263542 2012-11-30
JP2012263542A JP2013214947A (en) 2012-03-09 2012-11-30 Image capturing apparatus, image capturing system, image processing method, information processing apparatus, and program

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US14/384,027 A-371-Of-International US9607358B2 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
US15/400,610 Continuation US11049215B2 (en) 2012-03-09 2017-01-06 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium

Publications (1)

Publication Number Publication Date
WO2013133456A1 true WO2013133456A1 (en) 2013-09-12

Family

ID=49116920

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2013/057338 WO2013133456A1 (en) 2012-03-09 2013-03-11 Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium

Country Status (7)

Country Link
US (2) US9607358B2 (en)
EP (2) EP2823637B1 (en)
JP (1) JP2013214947A (en)
KR (6) KR101916616B1 (en)
CN (2) CN104160693B (en)
CA (1) CA2866621C (en)
WO (1) WO2013133456A1 (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103871038A (en) * 2014-03-06 2014-06-18 中国人民解放军国防科学技术大学 Super-resolution omnidirectional image reconstruction method based on non-uniform measurement matrix
WO2015174547A1 (en) * 2014-05-15 2015-11-19 Ricoh Company, Limited Imaging system, imaging apparatus, computer program, and system
EP3018529A1 (en) * 2014-11-06 2016-05-11 Ricoh Company, Ltd. Image processing apparatus and method for image processing
WO2017002360A1 (en) * 2015-07-01 2017-01-05 Ricoh Company, Ltd. Full-spherical video imaging system and computer-readable recording medium
US9609212B2 (en) 2013-08-28 2017-03-28 Ricoh Company, Ltd. Image processing apparatus, image processing method, and image system
US20180146136A1 (en) * 2015-07-01 2018-05-24 Hideaki Yamamoto Full-spherical video imaging system and computer-readable recording medium
GB2571611A (en) * 2017-12-28 2019-09-04 Canon Kk Electronic apparatus, method for controlling same, and program
US10863083B2 (en) * 2015-03-10 2020-12-08 Ricoh Company, Ltd. Image processing system and image processing method
US10893217B2 (en) 2017-12-28 2021-01-12 Canon Kabushiki Kaisha Electronic apparatus and method for clipping a range out of a wide field view image

Families Citing this family (56)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2013214947A (en) * 2012-03-09 2013-10-17 Ricoh Co Ltd Image capturing apparatus, image capturing system, image processing method, information processing apparatus, and program
JP5910485B2 (en) * 2012-03-16 2016-04-27 株式会社リコー Imaging system
JP2015125002A (en) * 2013-12-25 2015-07-06 株式会社ズームスケープ Photographing method for images for measuring use, and image measuring program
JP2015149600A (en) * 2014-02-06 2015-08-20 ソニー株式会社 image processing apparatus, image processing method, and program
JP5843034B1 (en) * 2014-05-15 2016-01-13 株式会社リコー Movie display device and program
JP5846268B1 (en) 2014-08-12 2016-01-20 株式会社リコー Image processing system, image processing apparatus, program, and imaging system
JP6398457B2 (en) * 2014-08-14 2018-10-03 株式会社リコー Imaging apparatus, imaging method, and program
JP6398472B2 (en) * 2014-08-27 2018-10-03 株式会社リコー Image display system, image display apparatus, image display method, and program
JP2016058886A (en) * 2014-09-09 2016-04-21 パナソニックIpマネジメント株式会社 Imaging device, imaging system, image processing method and program
JP6568374B2 (en) * 2015-03-27 2019-08-28 キヤノン株式会社 Information processing apparatus, information processing method, and program
CN104835118A (en) * 2015-06-04 2015-08-12 浙江得图网络有限公司 Method for acquiring panorama image by using two fish-eye camera lenses
TWI558208B (en) * 2015-07-14 2016-11-11 旺玖科技股份有限公司 Image processing method, apparatus and system
JP2017040687A (en) 2015-08-17 2017-02-23 株式会社リコー Image display system, information processor, image display method, and program
EP3338176A1 (en) 2015-08-17 2018-06-27 Ricoh Company, Ltd. Wide view image display system, information processing apparatus, and image display method
JP6627352B2 (en) * 2015-09-15 2020-01-08 カシオ計算機株式会社 Image display device, image display method, and program
CN108293090B (en) 2015-12-15 2020-10-20 株式会社理光 Image processing apparatus and image processing method
JP6645245B2 (en) * 2016-02-19 2020-02-14 株式会社ザクティ Spherical shooting system
JP6942940B2 (en) * 2016-03-14 2021-09-29 カシオ計算機株式会社 Image processing equipment, image processing methods and programs
EP3435652A4 (en) 2016-03-22 2019-01-30 Ricoh Company, Ltd. Image processing system, image processing method, and program
WO2017179136A1 (en) * 2016-04-12 2017-10-19 株式会社オプティム Entire celestial sphere camera robot altitude adjusting system, entire celestial sphere camera robot altitude adjusting method and program
US20170302714A1 (en) * 2016-04-15 2017-10-19 Diplloid Inc. Methods and systems for conversion, playback and tagging and streaming of spherical images and video
US10958834B2 (en) * 2016-07-22 2021-03-23 Immervision, Inc. Method to capture, store, distribute, share, stream and display panoramic image or video
JP6724659B2 (en) * 2016-08-30 2020-07-15 株式会社リコー Imaging device, method and program
KR20180051288A (en) * 2016-11-08 2018-05-16 삼성전자주식회사 Display apparatus and control method thereof
WO2018211782A1 (en) 2017-05-18 2018-11-22 ソニー株式会社 Image processing device, image processing method, and program
JP6919334B2 (en) * 2017-05-26 2021-08-18 株式会社リコー Image processing device, image processing method, program
JP6981106B2 (en) 2017-08-29 2021-12-15 株式会社リコー Image pickup device, image display system, operation method, program
CN111183334B (en) * 2017-08-31 2022-03-15 株式会社OPTiM Image analysis distance information providing system, method and recording medium
JP6935970B2 (en) * 2017-10-04 2021-09-15 株式会社ザクティ Camera device for motorcycles
KR20190057995A (en) 2017-11-21 2019-05-29 한국자동기술산업 주식회사 Smoking Booth For Personal
JP7031280B2 (en) 2017-12-18 2022-03-08 株式会社リコー Image processing equipment, image processing systems and programs
CN108040242A (en) * 2017-12-18 2018-05-15 横琴峰云视觉技术有限公司 A kind of bearing calibration at angle of inclination of Double-camera device and device
JP6688277B2 (en) * 2017-12-27 2020-04-28 本田技研工業株式会社 Program, learning processing method, learning model, data structure, learning device, and object recognition device
JP7081473B2 (en) 2018-03-02 2022-06-07 株式会社リコー Imaging optical system, imaging system and imaging device
EP3537399B1 (en) 2018-03-05 2021-03-31 Ricoh Company, Ltd. Imaging optical system, imaging system, and imaging apparatus
US10855916B2 (en) 2018-03-15 2020-12-01 Ricoh Company, Ltd. Image processing apparatus, image capturing system, image processing method, and recording medium
CN108462838B (en) * 2018-03-16 2020-10-02 影石创新科技股份有限公司 Panoramic video anti-shake method and device and portable terminal
JP2019164303A (en) 2018-03-20 2019-09-26 株式会社リコー Optical system and imaging apparatus
US10852503B2 (en) 2018-03-20 2020-12-01 Ricoh Company, Ltd. Joint structure
JP7124366B2 (en) 2018-03-20 2022-08-24 株式会社リコー Imaging element fixing structure and imaging device
JP6610741B2 (en) * 2018-09-06 2019-11-27 株式会社リコー Image display system, image display apparatus, image display method, and program
JP6610740B2 (en) * 2018-09-06 2019-11-27 株式会社リコー Imaging apparatus, imaging method, and program
EP3644600B1 (en) 2018-10-22 2022-06-15 Ricoh Company, Ltd. Imaging device, information processing method, system, and carrier means
CN111263037B (en) * 2018-11-30 2023-04-18 株式会社理光 Image processing device, imaging device, video playback system, method, and program
JP2020153796A (en) 2019-03-19 2020-09-24 株式会社リコー Distance measuring device and method for measuring distance
EP3719529A1 (en) 2019-03-20 2020-10-07 Ricoh Company, Ltd. Range finding device and range finding method
JP7350510B2 (en) * 2019-05-14 2023-09-26 キヤノン株式会社 Electronic equipment, electronic equipment control method, program, and storage medium
JP7383911B2 (en) * 2019-06-27 2023-11-21 株式会社リコー Imaging system, image processing device, imaging device and program
JP7379884B2 (en) 2019-06-27 2023-11-15 株式会社リコー Imaging device, image processing system, method and program
JP7388084B2 (en) * 2019-09-20 2023-11-29 株式会社リコー Information processing device, method and recording medium
JP6879350B2 (en) * 2019-10-31 2021-06-02 株式会社リコー Image display systems, image display devices, image display methods, and programs
JP7025042B2 (en) * 2019-12-26 2022-02-24 株式会社ザクティ Spherical image generation method, spherical image generation and display method, spherical image generation system program, and spherical image generation and display system program
US11470250B2 (en) * 2019-12-31 2022-10-11 Gopro, Inc. Methods and apparatus for shear correction in image projections
US11825230B2 (en) * 2020-02-14 2023-11-21 Canon Kabushiki Kaisha Imaging apparatus and storage medium
JP2020174363A (en) * 2020-06-23 2020-10-22 株式会社リコー Photographing system, method, and program
CN112419172B (en) * 2020-09-02 2022-10-21 中国人民解放军93114部队 Remote sensing image processing method for correcting and deblurring inclined image

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005056295A (en) * 2003-08-07 2005-03-03 Iwane Kenkyusho:Kk 360-degree image conversion processing apparatus
JP2005063141A (en) * 2003-08-12 2005-03-10 D Link Corp Image conversion system and image conversion method
JP2008205811A (en) * 2007-02-20 2008-09-04 Alpine Electronics Inc Camera attitude calculation target device and camera attitude calculation method using it, and image display method
JP2009064225A (en) * 2007-09-06 2009-03-26 Canon Inc Image processor and image processing method
JP2009267792A (en) * 2008-04-25 2009-11-12 Panasonic Corp Imaging apparatus
JP2010239221A (en) * 2009-03-30 2010-10-21 Panasonic Corp Image communication system
JP2011040898A (en) * 2009-08-07 2011-02-24 Casio Computer Co Ltd Imaging reproducing device and program
JP2011114496A (en) * 2009-11-25 2011-06-09 Olympus Imaging Corp Imaging apparatus

Family Cites Families (44)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04175832A (en) 1990-11-07 1992-06-23 Nec Eng Ltd Trouble processing system
US7382399B1 (en) * 1991-05-13 2008-06-03 Sony Coporation Omniview motionless camera orientation system
JPH1184482A (en) 1997-09-12 1999-03-26 Nec Corp Digital electronic still camera
US6304284B1 (en) * 1998-03-31 2001-10-16 Intel Corporation Method of and apparatus for creating panoramic or surround images using a motion sensor equipped camera
JPH11309137A (en) 1998-04-27 1999-11-09 Toshiba Corp Tomogram imaging device
US20010015751A1 (en) * 1998-06-16 2001-08-23 Genex Technologies, Inc. Method and apparatus for omnidirectional imaging
AU4221000A (en) * 1999-04-08 2000-10-23 Internet Pictures Corporation Remote controlled platform for camera
JP4469471B2 (en) * 2000-07-13 2010-05-26 亮介 柴崎 Wide viewing angle multidirectional image acquisition device for moving body and wide viewing angle multidirectional image acquisition system for moving body
CN1315098C (en) 2001-05-25 2007-05-09 松下电器产业株式会社 Wide-angle image generating device
JP4175832B2 (en) 2001-05-25 2008-11-05 松下電器産業株式会社 Wide-angle image generator
JP3705162B2 (en) * 2001-06-20 2005-10-12 ソニー株式会社 Imaging device
JP2003223633A (en) 2002-01-29 2003-08-08 Sharp Corp Omnidirectional visual system
JP4378994B2 (en) * 2003-04-30 2009-12-09 ソニー株式会社 Image processing apparatus, image processing method, and imaging apparatus
JP2005175952A (en) 2003-12-11 2005-06-30 Canon Inc Camera
JP2005210651A (en) * 2004-01-26 2005-08-04 Kazuo Iwane Full surround photographing apparatus and lens complex
JP2005244861A (en) * 2004-02-27 2005-09-08 Japan Science & Technology Agency Imaging apparatus and imaging system parameter correction method
JP2005252626A (en) * 2004-03-03 2005-09-15 Canon Inc Image pickup device and image processing method
JP2005275789A (en) * 2004-03-24 2005-10-06 Japan Science & Technology Agency Three-dimensional structure extraction method
JP2006033570A (en) * 2004-07-20 2006-02-02 Olympus Corp Image generating device
EP1771811A4 (en) * 2004-07-26 2010-06-09 Silicon Optix Inc Panoramic vision system and method
JP2006054504A (en) * 2004-08-09 2006-02-23 Olympus Corp Image generating method and apparatus
JP2006059202A (en) * 2004-08-20 2006-03-02 Iwate Univ Imaging device and image correction method
JP4631048B2 (en) * 2005-02-14 2011-02-16 国立大学法人岩手大学 Imaging apparatus and imaging system parameter calibration method
CN101305596B (en) * 2005-11-11 2010-06-16 索尼株式会社 Image processing device, image processing method and image picking device
JP2007311904A (en) 2006-05-16 2007-11-29 Victor Co Of Japan Ltd Drive recorder, video image correction method thereof, drive recorder, and system thereof
JP2008141649A (en) 2006-12-05 2008-06-19 Alpine Electronics Inc Vehicle periphery monitoring apparatus
KR100882011B1 (en) 2007-07-29 2009-02-04 주식회사 나노포토닉스 Methods of obtaining panoramic images using rotationally symmetric wide-angle lenses and devices thereof
JP5067336B2 (en) 2007-12-26 2012-11-07 大日本印刷株式会社 Image conversion apparatus and image conversion method
KR101404527B1 (en) 2007-12-26 2014-06-09 다이니폰 인사츠 가부시키가이샤 Image converter and image converting method
US20100097444A1 (en) * 2008-10-16 2010-04-22 Peter Lablans Camera System for Creating an Image From a Plurality of Images
WO2010123011A1 (en) * 2009-04-20 2010-10-28 京セラ株式会社 Image capturing device and image capturing method
JP5233926B2 (en) 2009-09-10 2013-07-10 大日本印刷株式会社 Fisheye monitoring system
US9110365B2 (en) 2009-11-19 2015-08-18 Olympus Corporation Imaging apparatus
JP5506499B2 (en) 2010-03-30 2014-05-28 キヤノン株式会社 IMAGING DEVICE, ITS CONTROL METHOD, PROGRAM, AND RECORDING MEDIUM
WO2011155161A1 (en) * 2010-06-11 2011-12-15 パナソニック株式会社 Image processing apparatus and image processing method
JP6123274B2 (en) * 2012-03-08 2017-05-10 株式会社リコー Imaging device
JP2013214947A (en) 2012-03-09 2013-10-17 Ricoh Co Ltd Image capturing apparatus, image capturing system, image processing method, information processing apparatus, and program
JP5910485B2 (en) * 2012-03-16 2016-04-27 株式会社リコー Imaging system
US10666860B2 (en) * 2012-09-11 2020-05-26 Ricoh Company, Ltd. Image processor, image processing method and program, and imaging system
JP6044328B2 (en) * 2012-12-26 2016-12-14 株式会社リコー Image processing system, image processing method, and program
EP2860699A1 (en) * 2013-10-11 2015-04-15 Telefonaktiebolaget L M Ericsson (Publ) Technique for view synthesis
JP2015149600A (en) * 2014-02-06 2015-08-20 ソニー株式会社 image processing apparatus, image processing method, and program
US9754159B2 (en) * 2014-03-04 2017-09-05 Gopro, Inc. Automatic generation of video from spherical content using location-based metadata
JP5843033B1 (en) * 2014-05-15 2016-01-13 株式会社リコー Imaging system, imaging apparatus, program, and system

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005056295A (en) * 2003-08-07 2005-03-03 Iwane Kenkyusho:Kk 360-degree image conversion processing apparatus
JP2005063141A (en) * 2003-08-12 2005-03-10 D Link Corp Image conversion system and image conversion method
JP2008205811A (en) * 2007-02-20 2008-09-04 Alpine Electronics Inc Camera attitude calculation target device and camera attitude calculation method using it, and image display method
JP2009064225A (en) * 2007-09-06 2009-03-26 Canon Inc Image processor and image processing method
JP2009267792A (en) * 2008-04-25 2009-11-12 Panasonic Corp Imaging apparatus
JP2010239221A (en) * 2009-03-30 2010-10-21 Panasonic Corp Image communication system
JP2011040898A (en) * 2009-08-07 2011-02-24 Casio Computer Co Ltd Imaging reproducing device and program
JP2011114496A (en) * 2009-11-25 2011-06-09 Olympus Imaging Corp Imaging apparatus

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2823637A4 *

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9609212B2 (en) 2013-08-28 2017-03-28 Ricoh Company, Ltd. Image processing apparatus, image processing method, and image system
US10834316B2 (en) 2013-08-28 2020-11-10 Ricoh Company, Ltd. Image processing apparatus, image processing method, and imaging system
US10063774B2 (en) 2013-08-28 2018-08-28 Ricoh Company, Ltd. Image processing apparatus, image processing method, and imaging system
US10412302B2 (en) 2013-08-28 2019-09-10 Ricoh Company, Ltd. Image processing apparatus, image processing method, and imaging system
CN103871038B (en) * 2014-03-06 2014-11-05 中国人民解放军国防科学技术大学 Super-resolution omnidirectional image reconstruction method based on non-uniform measurement matrix
CN103871038A (en) * 2014-03-06 2014-06-18 中国人民解放军国防科学技术大学 Super-resolution omnidirectional image reconstruction method based on non-uniform measurement matrix
CN110430355A (en) * 2014-05-15 2019-11-08 株式会社理光 Imaging system, imaging device, computer program and system
WO2015174547A1 (en) * 2014-05-15 2015-11-19 Ricoh Company, Limited Imaging system, imaging apparatus, computer program, and system
CN106464804A (en) * 2014-05-15 2017-02-22 株式会社理光 Imaging system, imaging apparatus, computer program, and system
US10681268B2 (en) 2014-05-15 2020-06-09 Ricoh Company, Ltd. Imaging system, imaging apparatus, and system
EP3018529A1 (en) * 2014-11-06 2016-05-11 Ricoh Company, Ltd. Image processing apparatus and method for image processing
US9635286B2 (en) 2014-11-06 2017-04-25 Ricoh Company, Ltd. Image processing apparatus and method for image processing
US10863083B2 (en) * 2015-03-10 2020-12-08 Ricoh Company, Ltd. Image processing system and image processing method
US20210084221A1 (en) * 2015-03-10 2021-03-18 Ricoh Company, Ltd. Image processing system and image processing method
US11659282B2 (en) 2015-03-10 2023-05-23 Ricoh Company, Ltd. Image processing system and image processing method
US20180146136A1 (en) * 2015-07-01 2018-05-24 Hideaki Yamamoto Full-spherical video imaging system and computer-readable recording medium
WO2017002360A1 (en) * 2015-07-01 2017-01-05 Ricoh Company, Ltd. Full-spherical video imaging system and computer-readable recording medium
GB2571611A (en) * 2017-12-28 2019-09-04 Canon Kk Electronic apparatus, method for controlling same, and program
US10893216B2 (en) 2017-12-28 2021-01-12 Canon Kabushiki Kaisha Electronic apparatus and method for controlling same
US10893217B2 (en) 2017-12-28 2021-01-12 Canon Kabushiki Kaisha Electronic apparatus and method for clipping a range out of a wide field view image
GB2571611B (en) * 2017-12-28 2021-04-21 Canon Kk Electronic apparatus, method for controlling same, and program

Also Published As

Publication number Publication date
KR20160043138A (en) 2016-04-20
KR101961364B1 (en) 2019-03-25
EP3182696A1 (en) 2017-06-21
KR101770088B1 (en) 2017-08-21
CN108876714B (en) 2022-08-26
CA2866621A1 (en) 2013-09-12
KR102046032B1 (en) 2019-11-19
EP3182696B1 (en) 2019-07-17
KR20180121691A (en) 2018-11-07
CN104160693B (en) 2018-07-27
KR20140129144A (en) 2014-11-06
JP2013214947A (en) 2013-10-17
KR20190031343A (en) 2019-03-25
US11049215B2 (en) 2021-06-29
KR20170002679A (en) 2017-01-06
CA2866621C (en) 2018-05-01
EP2823637B1 (en) 2017-05-03
KR101613387B1 (en) 2016-04-18
US20170116704A1 (en) 2017-04-27
CN108876714A (en) 2018-11-23
KR20170095396A (en) 2017-08-22
EP2823637A1 (en) 2015-01-14
KR101692194B1 (en) 2017-01-02
CN104160693A (en) 2014-11-19
US20150062363A1 (en) 2015-03-05
KR101916616B1 (en) 2018-11-07
US9607358B2 (en) 2017-03-28
EP2823637A4 (en) 2015-07-29

Similar Documents

Publication Publication Date Title
US11049215B2 (en) Image capturing apparatus, image capture system, image processing method, information processing apparatus, and computer-readable storage medium
US10594941B2 (en) Method and device of image processing and camera
JP5920507B1 (en) Image processing system, image processing method, and program
JP6398472B2 (en) Image display system, image display apparatus, image display method, and program
JP2017017689A (en) Imaging system and program of entire-celestial-sphere moving image
JP7024817B2 (en) Image pickup device, image pickup system, image processing method, information processing device, and program
JP6256513B2 (en) Imaging system, imaging apparatus, method, and program
US20170223281A1 (en) Image processing apparatus, control method, and computer readable storage medium
JP6879350B2 (en) Image display systems, image display devices, image display methods, and programs
JP2018026706A (en) Image processing apparatus, image display device, image processing system, and image processing method
WO2017002360A1 (en) Full-spherical video imaging system and computer-readable recording medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13757254

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 20147025261

Country of ref document: KR

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2866621

Country of ref document: CA

REEP Request for entry into the european phase

Ref document number: 2013757254

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2013757254

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 14384027

Country of ref document: US