US20140292811A1 - Mixed reality image processing apparatus and mixed reality image processing method - Google Patents
Mixed reality image processing apparatus and mixed reality image processing method Download PDFInfo
- Publication number
- US20140292811A1 US20140292811A1 US14/226,726 US201414226726A US2014292811A1 US 20140292811 A1 US20140292811 A1 US 20140292811A1 US 201414226726 A US201414226726 A US 201414226726A US 2014292811 A1 US2014292811 A1 US 2014292811A1
- Authority
- US
- United States
- Prior art keywords
- image data
- mixed reality
- unit
- illumination environment
- reality image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000012545 processing Methods 0.000 title claims abstract description 120
- 238000003672 processing method Methods 0.000 title claims 2
- 238000005286 illumination Methods 0.000 claims abstract description 137
- 238000003384 imaging method Methods 0.000 claims abstract description 55
- 238000006243 chemical reaction Methods 0.000 claims description 85
- 238000000605 extraction Methods 0.000 claims description 33
- 238000000034 method Methods 0.000 claims description 27
- 238000012937 correction Methods 0.000 claims description 14
- 239000000284 extract Substances 0.000 claims description 10
- 230000010365 information processing Effects 0.000 description 22
- 230000006870 function Effects 0.000 description 13
- 238000010586 diagram Methods 0.000 description 8
- 238000005259 measurement Methods 0.000 description 7
- 230000003287 optical effect Effects 0.000 description 7
- 230000008901 benefit Effects 0.000 description 5
- 238000012986 modification Methods 0.000 description 5
- 230000004048 modification Effects 0.000 description 5
- 230000002123 temporal effect Effects 0.000 description 5
- 239000004973 liquid crystal related substance Substances 0.000 description 3
- 239000003550 marker Substances 0.000 description 3
- 230000035807 sensation Effects 0.000 description 3
- 230000035945 sensitivity Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 239000003086 colorant Substances 0.000 description 2
- 238000005401 electroluminescence Methods 0.000 description 2
- 238000010521 absorption reaction Methods 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 239000000428 dust Substances 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 230000001771 impaired effect Effects 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 210000001525 retina Anatomy 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/50—Lighting effects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2215/00—Indexing scheme for image rendering
- G06T2215/16—Using real world measurements to influence rendering
Definitions
- An imaging unit mounted on the video see-through HMD captures observation image data in an external world that substantially coincides with the line-of-sight position of the HMD user.
- the imaging unit includes two pairs of an image sensor and an optical system for a right eye and a left eye for generating stereo image data, and a digital signal processor (DSP) for performing image processing.
- DSP digital signal processor
- a display unit displays the mixed reality image data created by combining CG to the object that substantially coincides with the object observed from the line-of-sight position of the HMD user. Further, the display unit is configured to deal with a pair of images of the right side and the left side like the imaging unit, and includes two pairs of a display device and an optical system for the right eye and the left eye.
- the above-described display unit is configured to display image data input from an external apparatus without making any adjustment thereto. Therefore, even when the external world is at dust, the external world is at noon so that it is extremely bright, or the external world is under a slightly dark environment in the shade of a tree, generally, the mixed reality image data displayed on the display unit is image data unaffected by the environment of the external world around the HMD user, in which a brightness and a color are uniformly adjusted. Therefore, a gap may be generated between brightness and color sensations felt by the HMD user to the ambient environment, and the actual environment of the external world.
- Japanese Patent Application Laid-Open No. 2002-244077 discusses a technique that reduces a shutter speed of an imaging unit and increases a lighting time of an illumination light source of a display unit under a bright environment, while increasing the shutter speed and reducing the lighting time of the illumination light source under a dark environment.
- the technique discussed in Japanese Patent Application Laid-Open No. 2002-244077 aims at maintaining the display brightness of the display unit at a brightness that matches see-through light even under various kinds of environments of an external world.
- the technique discussed in Japanese Patent Application Laid-Open No. 2002-244077 aims at imaging an object while maintaining the brightness within a certain range by changing the shutter speed of the imaging unit according to the brightness of the external world.
- Japanese Patent No. 03423402 discusses a technique that detects a color temperature of an external world by a sensor, and adjusts a color balance of display image data to be displayed on a display unit according to the detected color temperature. With this adjustment, the technique discussed in Japanese Patent No. 03423402 aims at generating display image data less unbalanced with see-through light.
- a mixed reality image processing apparatus includes an extraction unit configured to extract illumination environment information, which indicates an illumination environment of an external world, from image data imaged by an imaging unit, and a conversion unit configured to convert mixed reality image data, which is formed by combining virtual image data to the image data, into image data corresponding to the illumination environment of the external world based on the illumination environment information.
- the present disclosure it is possible to generate the mixed reality image data that matches the illumination environment of the external world.
- FIG. 1 is a block diagram illustrating a functional configuration of a mixed reality system using a video see-through HMD according to a first exemplary embodiment.
- FIG. 2 is a block diagram illustrating a detailed configuration of and around a standard illumination environment processing unit and a local illumination environment processing unit in the video see-through HMD according to the first exemplary embodiment.
- FIG. 3 illustrates a graph indicating a corresponding relationship between Red-Green-Blue (RGB) signal intensities in an RGB image format and a color temperature.
- RGB Red-Green-Blue
- FIG. 6 is a flowchart illustrating timing adjustment processing performed by the local illumination environment processing unit 107 in the video see-through HMD according to the first exemplary embodiment.
- FIGS. 8A , 8 B, 8 C, and 8 D illustrate tables indicating data structures of illumination environment information.
- FIGS. 9A , 9 B, 9 C, and 9 D illustrate what kind of characteristics an imaging device and a display device have.
- FIG. 10 is a block diagram illustrating a functional configuration of a mixed reality system using a video see-through HMD according to a second exemplary embodiment.
- FIG. 11 is a block diagram illustrating a further detailed configuration of the video see-through HMD according to the second exemplary embodiment.
- FIG. 1 illustrates a functional configuration of a mixed reality system using a video see-through head mounted display (HMD) 101 according to the first exemplary embodiment.
- the mixed reality system illustrated in FIG. 1 includes the video see-through HMD 101 according to the first exemplary embodiment.
- the video see-through HMD 101 includes an imaging unit 103 , a standard illumination environment processing unit 104 , an image output unit 105 , an image input unit 106 , a local illumination environment processing unit 107 , and a display unit 108 .
- the mixed reality system is configured to use the video see-through HMD 101 as an example of a mixed reality image processing apparatus.
- the image output unit 105 converts a format of output image data according to an interface between the apparatuses.
- the image input unit 106 converts a format of input image data according to an interface between the apparatuses.
- a method that can meet a requirement of a real-time capability and can transmit a large amount of data is used for each of interfaces of the image output unit 105 and the image input unit 106 . Examples of such a method include a metallic cable such as a universal serial bus (USB) or Institute of Electrical and Electronics Engineers (IEEE) 1394, and an optical fiber such as Gigabit Ethernet (registered trademark).
- the mixed reality system includes an image processing apparatus 102 .
- the image processing apparatus 102 includes the image output unit 105 , the image input unit 106 , a position/orientation measurement unit 109 , a captured image storage unit 110 , a content database (DB) 111 , a CG drawing unit 112 , and an image combining unit 113 .
- the image processing apparatus 102 can be embodied by an information processing apparatus that has a high-performance calculation processing function and graphics generation/display function, such as a personal computer (PC) and a workstation.
- PC personal computer
- the content DB 111 is a database that stores contents of virtual image data.
- the CG drawing unit 112 draws virtual image data based on the position/orientation information measured by the position/orientation measurement unit 109 and the contents stored in the content DB 111 .
- the captured image storage unit 110 stores the captured image data input from the image input unit 106 .
- the image combining unit 113 combines the captured image data and the virtual image data. It is desirable that the captured image data on which the virtual image data is to be superimposed is captured image data from which the position/orientation information for drawing the virtual image data is detected.
- the mixed reality system may be configured to superimpose the virtual image data onto latest updated captured image data at a timing when the virtual image data is generated. Using a predicted value as the position/orientation information for CG drawing at this time can reduce a time lag between the captured image data and the virtual image data.
- the captured image data output from the imaging unit 103 is output to the image processing apparatus 102 and is stored into the captured image storage unit 110 , after being processed by the standard illumination environment processing unit 104 . Then, virtual image data is superimposed on the captured image data stored in the captured image storage unit 110 by the image combining unit 113 . As a result, mixed reality image data is generated.
- the generated mixed reality image data is output to the video see-through HMD 101 and is processed by the local illumination environment processing unit 107 . Then, the mixed reality image data is displayed on the display unit 108 .
- the imaging unit 103 captures captured image data so as to maintain an actual appearance as much as possible. If the external world is dark, a dark image is captured as the captured image data. If the external world is bright, a bright image is captured as the captured image data. Further, if a color temperature is high in the external world, the captured image data also has a high color temperature. If the color temperature is low in the external world, the captured image data also has a low color temperature. In this manner, the imaging unit 103 performs imaging processing capable of outputting captured image data that matches the external world as much as possible. This does not mean that image data formed by performing an analog-digital (AD) conversion on an analog signal photoelectrically converted by the imaging device without any arrangement made thereto is output as the captured image data that matches the external world.
- AD analog-digital
- the imaging unit 103 performs processing that enables the environment of the captured image data to match the external world as the whole unit.
- the imaging unit 103 includes the imaging device characteristic correction unit 201 .
- the imaging device characteristic correction unit 201 corrects a characteristic depending on an individual imaging device.
- the standard illumination environment processing unit 104 includes an illumination environment information extraction unit 202 , an image identification information source 203 , a luminance conversion unit 204 , and a color conversion unit 205 .
- the illumination environment information extraction unit 202 extracts information that indicates an illumination environment of the external world (hereinafter referred to as illumination environment information), such as color temperature information and luminance information, from the captured image data by performing image processing. Then, the illumination environment information extraction unit 202 outputs the extracted illumination environment information together with image identification information contained in the image identification information source 203 to the local illumination environment processing unit 107 .
- the luminance information here means, for example, information that indicates a Y signal of the captured image data in a YUV format.
- the color temperature information means, for example, information that indicates a color temperature calculated from a ratio among RGB signal intensities of the captured image data in an RGB format.
- the illumination environment information extraction unit 202 embeds the image identification information into the captured image data by an electronic watermark technique, and outputs it to the luminance conversion unit 204 .
- the image identification information only has to be added as an attribute of the captured image data. Therefore, the image identification information may be added to a header of image data, and the image identification information may be combined or added by a different method depending on a system configuration.
- how to realize the image identification information and an amount of the information are determined in consideration of factors such as a variation in the time period from a capture of captured image data to superimposition of virtual image data onto this captured image data, and a possibility that the image identification information is also used as information that indicates a date/time when the captured image data is captured.
- the luminance conversion unit 204 converts a luminance of the captured image data.
- the luminance conversion unit 204 adjusts, for example, the Y signal of the captured image data in the YUV format. If the Y signal is expressed by 8 bits, i.e., within a decimal range of 0 to 255, for example, the Y signal is adjusted to around 127, which is a middle value. At that time, the Y signal is adjusted so that each of RGB values is contained in a domain after conversion into the RGB format. If each of RGB signals is expressed by 8 bits, and the following RGB-YUV conversion equation is used, the luminance conversion unit 204 adjusts the Y signal so that each of the RGB signals is contained within the decimal range of 0 to 255.
- a value that makes the captured image data not too bright and not too dark, and visually comfortable is selected as an adjustment value of the Y signal, and this value is determined in consideration of the captured target, the environment under which the captured image data is observed, the application, and the like.
- the format of the image data is converted if necessary, before and after the luminance conversion.
- the luminance conversion processing here means a brightness adjustment performed on a digital signal level.
- the color conversion unit 205 converts a color temperature of the captured image data.
- the color conversion unit 205 makes an adjustment so that, for example, a ratio among RGB signal intensities becomes 1:1:1 in the captured image data in the RGB format.
- the adjusted ratio among RGB signals may be a ratio corresponding to a color temperature defined by a standard such as standard RGB (sRGB) selected in consideration of the environment under which the captured image data is observed, the application, and the like.
- RGB standard RGB
- a method for calculating the color temperature from the ratio among RGB signal intensities will be described below with reference to FIG. 3 .
- the format of the captured image data is converted if necessary, before and after the color conversion.
- the color conversion by the color conversion unit 205 means a white balance adjustment performed on the digital signal level.
- the local illumination environment processing unit 107 includes an illumination environment information processing unit 206 , an image identification information extraction unit 207 , a timing adjustment unit 208 , an image storage unit 209 , a luminance conversion unit 210 , and a color conversion unit 211 .
- the illumination environment information processing unit 206 includes a storage unit (not illustrated) for storing illumination environment information, and sequentially stores therein the illumination environment information input from the illumination environment information extraction unit 202 . Further, image identification information is input from the timing adjustment unit 208 into the illumination environment information processing unit 206 , by which the illumination environment information processing unit 206 reads out illumination environment information corresponding to this image identification information, and outputs a readout success signal (Acknowledgement (ACK)) to the timing adjustment unit 208 . In addition thereto, the illumination environment information processing unit 206 outputs the image identification information and the luminance information contained in the illumination environment information to the luminance conversion unit 210 , and outputs the image identification information and the color temperature information contained in the illumination environment information to the color conversion unit 211 .
- ACK readout success signal
- the image identification information extraction unit 207 extracts the image identification information from the mixed reality image data. This image identification information is the image identification information embedded in the captured image data by the illumination environment information extraction unit 202 . The image identification information extracted from the mixed reality image data is output to the timing adjustment unit 208 . Further, the image identification information extraction unit 207 bundles the mixed reality image data and the extracted image identification information together as one set, and stores them into the image storage unit 209 .
- the timing adjustment unit 208 outputs the image identification information input from the image identification information extraction unit 207 to the illumination environment information processing unit 206 .
- the illumination environment information processing unit 206 outputs a readout success signal (ACK) to the timing adjustment unit 208 , if succeeded in reading out illumination environment information corresponding to this image identification information.
- the timing adjustment unit 208 adjusts a timing of reading out the mixed reality image data from the image storage unit 209 , after confirming the ACK. More specifically, the timing adjustment unit 208 adjusts the timing so that the mixed reality image data is read out from the image storage unit 209 at a timing at which it becomes possible to perform processing using the read illumination environment information by the luminance conversion unit 210 . This adjustment allows the luminance conversion unit 210 and the color conversion unit 211 to sequentially perform the luminance conversion and the color conversion using the illumination environment information with a delay as short as possible.
- the image storage unit 209 includes a line buffer or a frame buffer that stores the mixed reality image data.
- the image storage unit 209 stores the mixed reality image data together with the image identification information extracted by the image identification information extraction unit 207 .
- the luminance conversion unit 210 converts a luminance of the mixed reality image data.
- the content of the processing is similar to the luminance conversion of captured image data by the luminance conversion unit 204 , but the luminance information in the illumination environment information is used as an adjustment value therefor.
- the luminance conversion unit 210 can also check an error by reading out the image identification information together with the mixed reality image data from the image storage unit 209 , and comparing it with the image identification information of the illumination environment information. As a result of this luminance conversion, the luminance of the mixed reality image data becomes substantially equivalent to the luminance of the captured image data before the conversion processing performed by the luminance conversion unit 204 .
- the color conversion unit 211 converts a color temperature of the mixed reality image data.
- the content of the processing is similar to the color conversion of captured image data by the color conversion unit 205 , but the color conversion unit 211 adjusts RGB signals based on the color temperature information in the illumination environment information.
- the color conversion unit 211 can also check an error by inputting the image identification information together with the mixed reality image data from the luminance conversion unit 210 , and comparing it with the image identification information of the illumination environment information. As a result of this color conversion processing, the color temperature of the mixed reality image data becomes substantially equivalent to the color temperature of the captured image data before the conversion processing performed by the color conversion unit 205 .
- the display unit 108 displays the input mixed reality image data while maintaining its brightness and tint. If the input mixed reality image data is dark, dark HMD observation image data is displayed. If the input mixed reality image data is bright, bright HMD observation image data is displayed. Further, if the input mixed reality image data has a high color temperature, displayed HMD observation image data also has a high color temperature. If the input mixed reality image data has a low color temperature, displayed HMD observation image data also has a low color temperature.
- the display unit 108 performs processing so as to restrain a characteristic specific to the display device by a display device characteristic correction unit 212 , which will be described below, and converts the mixed reality image data input into the display unit 108 into HMD observation image data.
- the display device 108 includes the display device characteristic correction unit 212 .
- the display device characteristic correction unit 212 corrects a characteristic depending on an individual display device.
- FIG. 3 illustrates a graph indicating a corresponding relationship between RGB signal intensities in the RGB image format and a color temperature.
- a color temperature 6500K is set as a reference, and an RGB signal gain is adjust so that an RGB signal ratio becomes 1:1:1 when a white plate is imaged under the color temperature 6500K. This is referred to as a white RGB signal gain under the reference color temperature 6500K.
- FIG. 3 illustrates the relationship of RGB signal intensities to each color temperature when the same white plate is imaged under various color temperatures, while maintaining this RGB signal gain. For calculating a color temperature under various kinds of environments of the external world, first, RGB signal intensities are calculated using the white RGB signal gain under the reference color temperature 6500K.
- a color temperature can be acquired from the ratio among these RGB signal intensities and the corresponding relationship illustrated in FIG. 3 .
- the ratio between R/G and B/G is 4:1, the color temperature is approximately 3500K according to the graph in FIG. 3 .
- the ratio is used at this time in consideration of an influence of a bias applied to the whole RGB signal intensities depending on a degree of brightness of image data, but the color temperature can be acquired from RGB signal intensities as long as the luminance is adjusted in advance.
- the method described with reference to FIG. 3 is a method for calculating a color temperature from a ratio among RGB signal intensities, and is merely an example of a method for calculating a color temperature from image data.
- FIG. 4 is a flowchart illustrating processing performed by the standard illumination environment processing unit 104 in the video see-through HMD 101 .
- a central processing unit (CPU) in the video see-through HMD 101 reads out a required program and required data from a recording medium such as a read only memory (ROM) to execute the program, by which the processing illustrated in FIG. 4 , and the processing illustrated in FIGS. 5 and 6 that will be described below are realized.
- CPU central processing unit
- step S 401 the illumination environment information extraction unit 202 inputs captured image data from the imaging device characteristic correction unit 201 of the imaging unit 103 .
- step S 402 the illumination environment information extraction unit 202 extracts illumination environment information that contains color temperature information and luminance information from the input captured image data. The extracted illumination environment information is output to the illumination environment information processing unit 206 together with image identification information from the image identification information source 203 .
- the illumination environment information extraction unit 202 combines or adds the image identification information supplied from the image identification information source 203 to the captured image data.
- the illumination environment information extraction unit 202 may combine the image identification information to the captured image data by the electronic watermark technique, or may add the image identification information to a header of the captured image data.
- the captured image data with the image identification information combined or added thereto is output to the luminance conversion unit 204 .
- step S 404 the luminance conversion unit 204 converts a luminance of the captured image data.
- the captured image data after the luminance conversion is output to the color conversion unit 205 .
- step S 405 the color conversion unit 205 converts a color of the captured image data.
- the captured image data after the color conversion is output to the image output unit 105 in the video see-through HMD 101 .
- FIG. 5 is a flowchart illustrating a flow of image data in the local illumination environment processing unit 107 .
- step S 501 the image identification information extraction unit 207 inputs mixed reality image data from the image input unit 106 .
- step S 502 the image identification information extraction unit 207 extracts image identification information from the mixed reality image data.
- the mixed reality image data, from which the image identification information is extracted, is output to the image storage unit 209 together with the image identification information. Further, the extracted image identification information is also output to the timing adjustment unit 208 .
- step S 503 the image storage unit 209 stores the image identification information and the mixed reality image data in association with each other.
- step S 504 the luminance conversion unit 210 reads out the mixed reality image data from the image storage unit 209 . More specifically, in response to an input of the image identification information and an image readout permission from the timing adjustment unit 208 , the image storage unit 209 makes mixed reality image data corresponding to this image identification information ready for being read out from the image storage unit 209 . The luminance conversion unit 210 reads out this display image data, once the mixed reality image data becomes ready for being read out from the image storage unit 209 .
- step S 505 the luminance conversion unit 210 converts a luminance of the mixed reality image data.
- the luminance conversion unit 210 uses the luminance information contained in the illumination environment information from the illumination environment information processing unit 206 as an adjustment value.
- the display image data after the luminance conversion is output to the color conversion unit 211 .
- step S 506 the color conversion unit 211 converts a color of the display image data. More specifically, the color conversion unit 211 adjusts RGB signals based on the color temperature information contained in the illumination environment information from the illumination environment information processing unit 206 .
- the display image data after the color conversion is output to the display device characteristic correction unit 212 of the display unit 108 .
- FIG. 6 is a flowchart illustrating timing adjustment processing performed by the local illumination environment processing unit 107 of the video see-through HMD 101 .
- step S 601 the image identification information extraction unit 207 extracts the image identification information from the mixed reality image data.
- the extracted image identification information is output to the timing adjustment unit 208 .
- step S 602 the timing adjustment unit 208 outputs the image identification information input from the image identification information extraction unit 207 to the illumination environment information processing unit 206 .
- the illumination environment information processing unit 206 reads out illumination environment information corresponding to the image identification information input from the timing adjustment unit 208 , from the storage unit thereof.
- the illumination environment information processing unit 206 stores the image identification information and the illumination environment information input from the illumination environment information extraction unit 202 in a state bundled as one set. If the illumination environment information processing unit 206 succeeds in reading out the illumination environment information, the illumination environment information processing unit 206 outputs a readout success signal (ACK) to the timing adjustment unit 208 .
- ACK readout success signal
- the illumination environment information processing unit 206 outputs the illumination environment information and the image identification information to each of the luminance conversion unit 210 and the color conversion unit at the same time as the output of the ACK.
- step S 604 the timing adjustment unit 208 determines whether an ACK is input. If an ACK is not input even after a time-out time period has elapsed (NO in step S 604 ), the processing returns to step S 602 . Then, the timing adjustment unit 208 outputs image identification information of mixed reality image data one frame before the current frame to the illumination environment information processing unit 206 . On the other hand, if an ACK is input within the time-out time period (YES in step S 604 ), the processing proceeds to step S 605 .
- step S 605 the luminance conversion unit 210 outputs an image input ready signal to the timing adjustment unit 208 together with the image identification information at a timing at which it becomes possible to perform the luminance conversion processing using the illumination environment information from the illumination environment information processing unit 206 .
- step S 606 by being triggered by an input of the image identification information and the image input ready signal, the timing adjustment unit 208 transmits a permission for reading out mixed reality image data corresponding to this image identification information to the image storage unit 209 .
- the luminance conversion unit 210 reads out the mixed reality image data, once it becomes possible to read out the image from the image storage unit 209 .
- the luminance conversion unit 210 and the color conversion unit 211 can sequentially perform the luminance conversion processing and the color conversion processing using the illumination environment information with a delay as short as possible.
- FIG. 7 is a time chart indicating a temporal relationship between a flow of image data and a flow of information in the local illumination environment processing according to the first exemplary embodiment.
- mixed reality image data is input into the image identification information extraction unit 207 , and image identification information is extracted from this mixed reality image data. Thereafter, the mixed reality image data is stored into the image storage unit 209 . While the mixed reality image data is stored into the image storage unit 209 , the following processing is performed.
- the image identification information of this mixed reality image data is output from the timing adjustment unit 208 . Illumination environment information corresponding to this image identification information is read out by the illumination environment information processing unit 206 .
- an ACK is input into the timing adjustment unit 208 , and an image input ready signal is acquired by the timing adjustment unit 208 . Then, after the mixed reality image data is stored into the image storage unit 209 and the image input ready signal is acquired by the timing adjustment unit 208 , the luminance and the color are converted by the luminance conversion unit 210 and the color conversion unit 211 , respectively.
- FIGS. 8A , 8 B, 8 C, and 8 D illustrate data structures of the illumination environment information.
- the illumination environment information includes the image identification information (date/time information), the color temperature information, and the luminance information.
- FIG. 8B illustrates an example of a data structure of the image identification information (the date/time information).
- the image identification information (the date/time information) is information that indicates when captured image data is captured, and is mainly divided into date information and time information.
- the date information includes year according to the Western calendar, month, and day.
- the time information includes hour, minute, second, and millisecond.
- FIG. 8B illustrates a structure example of 40 bits in total.
- the image identification information (the date/time information) only has to be information that allows identification of image data from the time when captured image data is captured to the time when virtual image data is superimposed onto this captured image data. Therefore, as described above, if this time interval corresponds to 10 frames, the image identification information only has to be information that allows image data pieces of the 10 frames to be identified, respectively. However, actually, the data structure of the image identification information is determined in consideration of factors such as a variation in a time period required to generate virtual image data, and a possibility that the image identification information is also used as attribute information of the captured image data.
- the image identification information has a data structure having a margin, for example, a data structure that allows image data pieces of 240 frames to be identified, respectively. If the image identification information is also used as attribute information of the captured image data, it is desirable that the image identification information has such a data structure that attribute information such as a resolution, an image format, and a shutter speed is added to the data structure illustrated in FIG. 8B .
- FIGS. 8C and 8D illustrate examples of data structures of the color temperature information and the luminance information, respectively.
- the color temperature information is information that indicates a color temperature itself, as illustrated in FIG. 8C .
- the luminance information is information that indicates a Y signal when image data is expressed in the YUV format, as illustrated in FIG. 8D .
- FIGS. 9A , 9 B, 9 C, and 9 D illustrate what kind of characteristics an imaging device and a display device have.
- Representative examples of the imaging device characteristic include a receivable light frequency characteristic and a sensitivity characteristic.
- FIG. 9A illustrates an example of the receivable light frequency characteristic of the imaging device.
- the characteristic of frequency of light receivable by the imaging device varies for each imaging device due to influences of absorption and scattering of a color filter and a peripheral member used in the imaging device. This affects a tint of captured image data.
- FIG. 9B illustrates an example of the sensitivity characteristic of the imaging device.
- the imaging device has a Bayer array, and measures signal outputs (VGr, VGb, VR, and VB) at a center of a screen when a white object is imaged.
- a gain of an electric signal is to be adjusted so that, if a light intensity of each RGB single color is the same, an intensity of an electric signal after photoelectric conversion becomes also the same for each RGB. This affects a luminance of captured image data.
- the imaging device characteristic is not limited to the above-described receivable light frequency characteristic and sensitivity characteristic, and may be another element in the imaging unit 103 such as an infrared (IR) filter and an analog transmission characteristic.
- the imaging device characteristic correction unit 201 corrects a characteristic (a tint, a luminance, and the like) that depends on a component of the imaging unit 103 .
- Representative examples of the display device characteristic include an emittable light frequency characteristic and a luminance characteristic of a light source.
- FIG. 9C illustrates an example of the emittable light frequency characteristic of the light source. This example corresponds to emittable light frequencies of respective RGB colors when light-emitting diodes (LEDs) of RGB three colors are used as the light source. This affects a tint of mixed reality image data.
- LEDs light-emitting diodes
- FIG. 9D illustrates an example of the luminance characteristic of the light source.
- a current supplied to the LED is to be adjusted for each RGB to adjust a light intensity of each RGB single color. This affects a luminance of mixed reality image data.
- the display device characteristic is not limited to the above-described emittable light frequency characteristic and luminance characteristic of the light source, and may be another element in the display unit 108 such as a liquid-crystal filter characteristic of a display panel, and a lens characteristic.
- the display device characteristic correction unit 212 corrects a characteristic (a tint, a luminance, and the like) that depends on a component of the display unit 108 .
- mixed reality image data can be dynamically converted into image data that matches an illumination environment of an external world with use of illumination environment information extracted from captured image data. Therefore, brightness and color sensations felt by an HMD user to an ambient environment can substantially match an actual environment of the external world, which allows the user to be further engrossed into a mixed reality space.
- captured image data output from the video see-through HMD 101 into the image processing apparatus 102 is image data in which the characteristics depending on the imaging device are canceled, and a brightness and a white balance are adjusted. Therefore, the present exemplary embodiment has advantages of compliance with a basic idea of color matching, facilitation of marker extraction from captured image data to detect a position and an orientation, elimination of the necessity of changing a color temperature and a luminance of the light source during drawing of virtual image data for each system, a reduction in loss in image data due to compression, and the like.
- the characteristics depending on the display device are canceled from mixed reality image data in the display unit 108 , which allows formation of image data according to a standard that does not depend on the display device. Further, the present exemplary embodiment does not require an additional sensor different from the components of the video see-through HMD 101 itself, and utilizes a corresponding relationship between captured image data and mixed reality image data, thereby simplifying the configuration.
- FIG. 10 is a block diagram illustrating a functional configuration of a mixed reality system using a video see-through HMD 1001 according to the second exemplary embodiment.
- the mixed reality system includes the video see-through HMD 1001 according to the second exemplary embodiment.
- the video see-through HMD 1001 includes the imaging unit 103 , the standard illumination environment processing unit 104 , the captured image storage unit 110 , the image input unit 106 , a time alignment unit 1004 , the image combining unit 113 , the local illumination environment processing unit 107 , the display unit 108 , the position/orientation measurement unit 109 , an interface (I/F) unit 1003 , and the image output unit 105 .
- the mixed reality system includes an image processing apparatus 1002 .
- the image processing apparatus 1002 includes the I/F unit 1003 , the content DB 111 , the CG drawing unit 112 , and the image output unit 105 .
- the image processing apparatus 1002 can be embodied by an apparatus that has a high-performance calculation processing function and graphics generation/display function, such as a personal computer and a workstation.
- the mixed reality system further includes an external output apparatus 1005 .
- PDP plasma display panel
- EL organic electroluminescence
- FIG. 10 units similar to those in FIG. 1 are labeled in a similar manner to FIG. 1 .
- captured image data is output from the video see-through HMD 101 to the image processing apparatus 102 .
- position/orientation information and image identification information are output from the video see-through HMD 1001 to the image processing apparatus 1002 , and captured image data and the image identification information are bundled together to be stored as one set into the captured image storage unit 110 of the video see-through HMD 1001 .
- the first exemplary embodiment has such an advantage that the video see-through HMD 101 can be simply configured, because the first exemplary embodiment can be realized by using functions of an existing image processing apparatus.
- the second exemplary embodiment has such advantages that a communication amount can be reduced between the video see-through HMD 1001 and the image processing apparatus 1002 , and a delay due to an output of image data can be reduced.
- the image processing apparatus 1002 generates virtual image data from the position/orientation information, and outputs the virtual image data to the video see-through HMD 1001 together with the image identification information.
- the image identification information is embedded in the virtual image data by the electronic watermark technique, or is added to a header of the virtual image data.
- the captured image data and the virtual image data are combined within the video see-through HMD 1001 , and mixed reality image data, which is the combined image data therefrom, is displayed on the display unit 108
- the time alignment unit 1004 is newly provided so as to allow the image combining unit 113 to combine captured image data and virtual image data that coincides with the captured image data on a temporal axis.
- the time alignment unit 1004 performs time alignment processing for the local illumination environment processing unit 107 , in addition to time alignment processing for the image combining unit 113 .
- the video see-through HMD 1001 includes the image combining unit 113 , whereby the second exemplary embodiment is configured in such a manner that mixed reality image data is held only by the video see-through HMD 1001 .
- mixed reality image data is output from the image combining unit 113 to not only the local illumination environment processing unit 107 but also the image output unit 105 , to allow another person than an HMD user to also observe the mixed reality image data.
- the mixed reality image data is output from the image output unit 105 to the external output apparatus 1005 .
- image data that does not depend on an observation environment of the HMD user can be output to the external output apparatus 1005 as mixed reality image data.
- Preparing mixed reality image data as image data according to the sRGB standard or the like allows the external output apparatus 1005 to convert this data into image data suitable for its observation environment.
- the imaging unit 103 has a function similar to the imaging unit 103 illustrated in FIG. 2 .
- the standard illumination environment processing unit 104 includes the illumination environment information extraction unit 202 , the image identification information source 203 , the luminance conversion unit 204 , and the color conversion unit 205 .
- the image identification information source 203 , the luminance conversion unit 204 , and the color conversion unit 205 have functions similar to the image identification information source 203 , the luminance conversion unit 204 , and the color conversion unit 205 illustrated in FIG. 2 , respectively.
- the illumination environment information extraction unit 202 extracts illumination environment information that contains color temperature information and luminance information from captured image data by performing image processing, and outputs the extracted illumination environment information to the local illumination environment processing unit 107 together with image identification information from the image identification information source 203 .
- the illumination environment information extraction unit 202 adds the image identification information to the captured image data, and outputs this captured image data to the luminance conversion unit 204 .
- the present exemplary embodiment is configured so as not to transmit the captured image data from the video see-through HMD 1001 to the image processing apparatus 1002 .
- the image identification information is desirable to handle the image identification information as different information from the captured image data instead of embedding the image identification information into the captured image data, in consideration that the image identification information is stored together with the captured image data as one set into the captured image storage unit 110 , which is a later stage.
- the captured image storage unit 110 stores image identification information and captured image data corresponding to this image identification information as one set.
- the position/orientation measurement unit 109 extracts a marker and/or a natural feature from the captured image data input from the standard illumination environment processing unit 104 , and measures position/orientation information at a line-of-sight position of the HMD user.
- the position/orientation measurement unit 109 bundles the measured position/orientation information and image identification information corresponding to this position/orientation information together as one set, and outputs them to the I/F unit 1003 .
- the I/F unit 1003 transmits the position/orientation information and the image identification information bundled together as one set to the image processing apparatus 1002 .
- the local illumination environment processing unit 107 includes the illumination environment information processing unit 206 , the luminance conversion unit 210 , and the color conversion unit 211 .
- the image identification information extraction unit 207 and the timing adjustment unit 208 are provided as components of the local illumination environment processing unit 107 .
- they are provided in another block as the time alignment unit 1004 .
- the local illumination environment processing unit 107 has a function and a processing content similar to FIG. 2 , only in terms of the local illumination environment processing.
- the time alignment unit 1004 configured as an independent block in FIG. 11 is extended so as to function even in time alignment processing for the image combining processing by the image combining unit 113 .
- the time alignment unit 1004 includes the image identification information extraction unit 207 and the timing adjustment unit 208 .
- the time alignment unit 1004 performs not only the time alignment processing for the image combining unit 113 but also the time alignment processing for the local illumination environment processing unit 107 .
- the image identification information extraction unit 207 extracts image identification information from virtual image data. This image identification information is the image identification information added to the captured image data by the illumination environment information extraction unit 202 . The extracted image identification information is output to the timing adjustment unit 208 . Further, the virtual image data is output to the image combining unit 113 together with the extracted image identification information.
- the image combining unit 113 receives the output virtual image data and reads out captured image data corresponding to the image identification information thereof from the captured image storage unit 110 , and then combines the virtual image data and the captured image data.
- Mixed reality image data after the combining processing is output to the luminance conversion unit 210 and the image output unit 105 .
- FIG. 12 is a time chart indicating a temporal relationship between a flow of image data and a flow of information in the image combination processing and the local illumination environment processing according to the second exemplary embodiment.
- Image identification information of the mixed reality image data is output by the timing adjustment unit 208 .
- Illumination environment information corresponding to this image identification information is read out by the illumination environment information processing unit 206 .
- an ACK is input into the timing adjustment unit 208 , and an image input ready signal is acquired by the timing adjustment unit 208 . This is a difference from the first exemplary embodiment.
- mixed reality image data output from the video see-through HMD 1001 into the external output apparatus 1005 is image data in which the characteristics depending on the device are canceled, and a brightness and a white balance are adjusted. Therefore, mixed reality image data according to a standard that does not depend on the display device can be output.
- the present exemplary embodiment complies with a basic idea of color matching, thereby having an advantage of being able to provide a display suitable for an external environment and a device characteristic of the external output apparatus 1005 .
- FIGS. 13A and 13B are block diagrams illustrating a modification of the mixed reality system according to the second exemplary embodiment. More specifically, FIG. 13A illustrates a configuration of the modification of the mixed reality system according to the second exemplary embodiment, and FIG. 13B illustrates switching processing by a selector 1301 .
- the mixed reality system illustrated in FIG. 13A corresponds to the configuration of the mixed reality system illustrated in FIG. 10 with the selector 1301 added thereto.
- the selector 1301 For displaying only captured image data on the video see-through HMD 1001 instead of displaying mixed reality image data, the selector 1301 is switched to (2) illustrated in FIG. 13B , thereby changing a data flow so as to transmit image data from the imaging unit 103 to the display unit 108 .
- the mixed reality system can be switched to a low power consumption mode with the components other than the imaging unit 103 , the selector 1301 , and the display unit 108 substantially stopping working.
- the imaging unit 103 performs processing that enables an environment of captured image data to match an environment of an external world, and the display unit 108 displays the input mixed reality image data while maintaining its luminance and tint, thereby allowing an HMD user to observe an image that matches the environment of the external world.
- the selector 1301 is switched to (1) illustrated in FIG. 13B .
- the mixed reality system according to the modification has functions and processing similar to the mixed reality system according to the second exemplary embodiment illustrated in FIG. 10 .
- mixed reality image data can be dynamically adjusted to image data that matches an illumination environment of an external world. Further, a delay in the processing of the system, a processing load, and power consumption can be reduced.
- Embodiments of the present invention can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions recorded on a storage medium (e.g., non-transitory computer-readable storage medium) to perform the functions of one or more of the above-described embodiment(s) of the present invention, and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s).
- the computer may comprise one or more of a central processing unit (CPU), micro processing unit (MPU), or other circuitry, and may include a network of separate computers or separate computer processors.
- the computer executable instructions may be provided to the computer, for example, from a network or the storage medium.
- the storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)TM), a flash memory device, a memory card, and the like.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Graphics (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Processing Or Creating Images (AREA)
- Television Receiver Circuits (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
Description
- 1. Field of the Invention
- The present invention relates to a technique for generating mixed reality image data by combining virtual image data to captured image data.
- 2. Description of the Related Art
- In recent years, a mixed reality technique, i.e., an MR technique has been known as a technique for mixing a real world and a virtual world seamlessly in real time. As one type of the MR technique, there is known a technique that uses a video see-through head mounted display (HMD) to allow an HMD user to observe a mixed reality image. The mixed reality image is created by imaging an object that substantially coincides with an object observed from a line-of-sight position of the HMD user with use of a video camera and the like, and displaying this captured image data with computer graphics (CG) superimposed thereon.
- An imaging unit mounted on the video see-through HMD captures observation image data in an external world that substantially coincides with the line-of-sight position of the HMD user. The imaging unit includes two pairs of an image sensor and an optical system for a right eye and a left eye for generating stereo image data, and a digital signal processor (DSP) for performing image processing. A display unit displays the mixed reality image data created by combining CG to the object that substantially coincides with the object observed from the line-of-sight position of the HMD user. Further, the display unit is configured to deal with a pair of images of the right side and the left side like the imaging unit, and includes two pairs of a display device and an optical system for the right eye and the left eye.
- A relationship between the external world and the mixed reality image data will be described now. The above-described display unit is configured to display image data input from an external apparatus without making any adjustment thereto. Therefore, even when the external world is at dust, the external world is at noon so that it is extremely bright, or the external world is under a slightly dark environment in the shade of a tree, generally, the mixed reality image data displayed on the display unit is image data unaffected by the environment of the external world around the HMD user, in which a brightness and a color are uniformly adjusted. Therefore, a gap may be generated between brightness and color sensations felt by the HMD user to the ambient environment, and the actual environment of the external world.
- This influence cannot be also ignored for an optical see-through HMD. When CG data is superimposed on a see-through image of the external world in the optical see-through HMD, the brightness and the color are uniformly adjusted only in the CG data, whereby the balance and the color are unbalanced between the CG data and the see-through image of the external world. This leads to such a problem that the CG data may have an unnatural brightness and color to the see-through image of the external world depending on the environment of the external world, thereby impairing a realistic sensation.
- Under this circumstance, Japanese Patent Application Laid-Open No. 2002-244077 discusses a technique that reduces a shutter speed of an imaging unit and increases a lighting time of an illumination light source of a display unit under a bright environment, while increasing the shutter speed and reducing the lighting time of the illumination light source under a dark environment. With this adjustment, the technique discussed in Japanese Patent Application Laid-Open No. 2002-244077 aims at maintaining the display brightness of the display unit at a brightness that matches see-through light even under various kinds of environments of an external world. Further, the technique discussed in Japanese Patent Application Laid-Open No. 2002-244077 aims at imaging an object while maintaining the brightness within a certain range by changing the shutter speed of the imaging unit according to the brightness of the external world.
- Japanese Patent No. 03423402 discusses a technique that detects a color temperature of an external world by a sensor, and adjusts a color balance of display image data to be displayed on a display unit according to the detected color temperature. With this adjustment, the technique discussed in Japanese Patent No. 03423402 aims at generating display image data less unbalanced with see-through light.
- However, the techniques discussed in Japanese Patent Application Laid-Open No. 2002-244077 and Japanese Patent No. 03423402 are techniques for adjusting the brightness and color of the display image data according to an illumination environment of the external world, and cannot display mixed reality image data that matches the illumination environment of the external world on the video see-through HMD.
- According to an aspect of the present invention, a mixed reality image processing apparatus includes an extraction unit configured to extract illumination environment information, which indicates an illumination environment of an external world, from image data imaged by an imaging unit, and a conversion unit configured to convert mixed reality image data, which is formed by combining virtual image data to the image data, into image data corresponding to the illumination environment of the external world based on the illumination environment information.
- According to the present disclosure, it is possible to generate the mixed reality image data that matches the illumination environment of the external world.
- Further features of the present invention will become apparent from the following description of exemplary embodiments with reference to the attached drawings.
-
FIG. 1 is a block diagram illustrating a functional configuration of a mixed reality system using a video see-through HMD according to a first exemplary embodiment. -
FIG. 2 is a block diagram illustrating a detailed configuration of and around a standard illumination environment processing unit and a local illumination environment processing unit in the video see-through HMD according to the first exemplary embodiment. -
FIG. 3 illustrates a graph indicating a corresponding relationship between Red-Green-Blue (RGB) signal intensities in an RGB image format and a color temperature. -
FIG. 4 is a flowchart illustrating processing performed by the standard illumination environment processing unit in the video see-through HMD according to the first exemplary embodiment. -
FIG. 5 is a flowchart illustrating a flow of image data in the local illuminationenvironment processing unit 107 according to the first exemplary embodiment. -
FIG. 6 is a flowchart illustrating timing adjustment processing performed by the local illuminationenvironment processing unit 107 in the video see-through HMD according to the first exemplary embodiment. -
FIG. 7 is a time chart indicating a temporal relationship between a flow of image data and a flow of information in local illumination environment processing according to the first exemplary embodiment. -
FIGS. 8A , 8B, 8C, and 8D illustrate tables indicating data structures of illumination environment information. -
FIGS. 9A , 9B, 9C, and 9D illustrate what kind of characteristics an imaging device and a display device have. -
FIG. 10 is a block diagram illustrating a functional configuration of a mixed reality system using a video see-through HMD according to a second exemplary embodiment. -
FIG. 11 is a block diagram illustrating a further detailed configuration of the video see-through HMD according to the second exemplary embodiment. -
FIG. 12 is a time chart indicating a temporal relationship between a flow of image data and a flow of information in image combining processing and local illumination environment processing according to the second exemplary embodiment. -
FIGS. 13A and 13B are block diagrams illustrating a modification of the mixed reality system according to the second exemplary embodiment. - Various exemplary embodiments, features, and aspects of the invention will be described in detail below with reference to the drawings.
- First, a first exemplary embodiment of the present invention will be described.
FIG. 1 illustrates a functional configuration of a mixed reality system using a video see-through head mounted display (HMD) 101 according to the first exemplary embodiment. The mixed reality system illustrated inFIG. 1 includes the video see-throughHMD 101 according to the first exemplary embodiment. The video see-throughHMD 101 includes animaging unit 103, a standard illuminationenvironment processing unit 104, animage output unit 105, animage input unit 106, a local illuminationenvironment processing unit 107, and adisplay unit 108. The mixed reality system is configured to use the video see-throughHMD 101 as an example of a mixed reality image processing apparatus. - The
imaging unit 103 captures image data (hereinafter referred to as captured image data) of an external world that substantially coincides with a line-of-sight position of an HMD user. Theimaging unit 103 includes two pairs of image sensor and optical system for a right eye and a left eye for generating stereo image data, and a DSP for performing image processing. A solid-state image sensor, which is represented by a charge coupled device (CCD) image sensor and a complementary metal-oxide semiconductor (CMOS) image sensor, is used for each of the image sensors. - The
display unit 108 displays mixed reality image data (MR image data) formed by combining the captured image data and CG. Thedisplay unit 108 is configured to deal with images of the right side and the left side similar to theimaging unit 103, and includes two pairs of display device and optical system for the right eye and the left eye. A small liquid-crystal display or a retina scan type device using Micro Electro Mechanical Systems (MEMS) is used for each of the display devices. - The
image output unit 105 converts a format of output image data according to an interface between the apparatuses. Theimage input unit 106 converts a format of input image data according to an interface between the apparatuses. A method that can meet a requirement of a real-time capability and can transmit a large amount of data is used for each of interfaces of theimage output unit 105 and theimage input unit 106. Examples of such a method include a metallic cable such as a universal serial bus (USB) or Institute of Electrical and Electronics Engineers (IEEE) 1394, and an optical fiber such as Gigabit Ethernet (registered trademark). - Referring to
FIG. 1 , the mixed reality system includes animage processing apparatus 102. Theimage processing apparatus 102 includes theimage output unit 105, theimage input unit 106, a position/orientation measurement unit 109, a capturedimage storage unit 110, a content database (DB) 111, aCG drawing unit 112, and animage combining unit 113. Theimage processing apparatus 102 can be embodied by an information processing apparatus that has a high-performance calculation processing function and graphics generation/display function, such as a personal computer (PC) and a workstation. - The position/
orientation measurement unit 109 measures position/orientation information that indicates at least any one of a position and an orientation of the HMD user. More specifically, the position/orientation measurement unit 109 extracts a marker and/or a natural feature from the captured image data input from theimage input unit 106, and measures the position/orientation information using them. The measured position/orientation information is output as necessary to theCG drawing unit 112 to be used for a calculation of a shape of virtual image data (CG) to be drawn. - The
content DB 111 is a database that stores contents of virtual image data. TheCG drawing unit 112 draws virtual image data based on the position/orientation information measured by the position/orientation measurement unit 109 and the contents stored in thecontent DB 111. The capturedimage storage unit 110 stores the captured image data input from theimage input unit 106. Theimage combining unit 113 combines the captured image data and the virtual image data. It is desirable that the captured image data on which the virtual image data is to be superimposed is captured image data from which the position/orientation information for drawing the virtual image data is detected. However, if the real-time capability is impaired due to a transmission delay between the systems, a processing time required to draw CG, and the like, the mixed reality system may be configured to superimpose the virtual image data onto latest updated captured image data at a timing when the virtual image data is generated. Using a predicted value as the position/orientation information for CG drawing at this time can reduce a time lag between the captured image data and the virtual image data. - Based on this configuration, a flow of image data will be briefly described. The captured image data output from the
imaging unit 103 is output to theimage processing apparatus 102 and is stored into the capturedimage storage unit 110, after being processed by the standard illuminationenvironment processing unit 104. Then, virtual image data is superimposed on the captured image data stored in the capturedimage storage unit 110 by theimage combining unit 113. As a result, mixed reality image data is generated. The generated mixed reality image data is output to the video see-throughHMD 101 and is processed by the local illuminationenvironment processing unit 107. Then, the mixed reality image data is displayed on thedisplay unit 108. -
FIG. 2 illustrates a detailed configuration of and around the standard illuminationenvironment processing unit 104 and the localillumination environment processing 107 in the video see-throughHMD 101. - The
imaging unit 103 captures captured image data so as to maintain an actual appearance as much as possible. If the external world is dark, a dark image is captured as the captured image data. If the external world is bright, a bright image is captured as the captured image data. Further, if a color temperature is high in the external world, the captured image data also has a high color temperature. If the color temperature is low in the external world, the captured image data also has a low color temperature. In this manner, theimaging unit 103 performs imaging processing capable of outputting captured image data that matches the external world as much as possible. This does not mean that image data formed by performing an analog-digital (AD) conversion on an analog signal photoelectrically converted by the imaging device without any arrangement made thereto is output as the captured image data that matches the external world. It is desirable to reduce an analog noise by guiding a large amount of light to the imaging device, and correct a luminance by an imaging devicecharacteristic correction unit 201, which will be described below. Further, it is desirable to also correct a color due to a color filter and the like. Theimaging unit 103 performs processing that enables the environment of the captured image data to match the external world as the whole unit. Theimaging unit 103 includes the imaging devicecharacteristic correction unit 201. The imaging devicecharacteristic correction unit 201 corrects a characteristic depending on an individual imaging device. - As illustrated in
FIG. 2 , the standard illuminationenvironment processing unit 104 includes an illumination environmentinformation extraction unit 202, an imageidentification information source 203, aluminance conversion unit 204, and acolor conversion unit 205. The illumination environmentinformation extraction unit 202 extracts information that indicates an illumination environment of the external world (hereinafter referred to as illumination environment information), such as color temperature information and luminance information, from the captured image data by performing image processing. Then, the illumination environmentinformation extraction unit 202 outputs the extracted illumination environment information together with image identification information contained in the imageidentification information source 203 to the local illuminationenvironment processing unit 107. The luminance information here means, for example, information that indicates a Y signal of the captured image data in a YUV format. The color temperature information means, for example, information that indicates a color temperature calculated from a ratio among RGB signal intensities of the captured image data in an RGB format. In addition to the above-described processing, the illumination environmentinformation extraction unit 202 embeds the image identification information into the captured image data by an electronic watermark technique, and outputs it to theluminance conversion unit 204. The image identification information only has to be added as an attribute of the captured image data. Therefore, the image identification information may be added to a header of image data, and the image identification information may be combined or added by a different method depending on a system configuration. - The image
identification information source 203 is a collection of image identification information. The image identification information only has to be at least information that allows identification of image data from the time when captured image data is captured to the time when virtual image data is superimposed onto this captured image data. Therefore, if a time period corresponding to 10 frames is required from a capture of captured image data to superimposition of virtual image data onto this captured image data, only image identification information corresponding to the 10 frames should be prepared, so that the image identification information can be expressed by 4 bits. In this manner, the image identification information only has to be information that allows image data to be uniquely identified. For example, as simple methods, the image identification information can be realized by merely numbering captured image data, or can be also realized as a time stamp of a time when captured image data is captured. Actually, how to realize the image identification information and an amount of the information are determined in consideration of factors such as a variation in the time period from a capture of captured image data to superimposition of virtual image data onto this captured image data, and a possibility that the image identification information is also used as information that indicates a date/time when the captured image data is captured. - The
luminance conversion unit 204 converts a luminance of the captured image data. Theluminance conversion unit 204 adjusts, for example, the Y signal of the captured image data in the YUV format. If the Y signal is expressed by 8 bits, i.e., within a decimal range of 0 to 255, for example, the Y signal is adjusted to around 127, which is a middle value. At that time, the Y signal is adjusted so that each of RGB values is contained in a domain after conversion into the RGB format. If each of RGB signals is expressed by 8 bits, and the following RGB-YUV conversion equation is used, theluminance conversion unit 204 adjusts the Y signal so that each of the RGB signals is contained within the decimal range of 0 to 255. -
R=Y+1.40200V -
G=Y−0.344114U−0.71414V -
B=Y+1.77200U - Desirably, a value that makes the captured image data not too bright and not too dark, and visually comfortable is selected as an adjustment value of the Y signal, and this value is determined in consideration of the captured target, the environment under which the captured image data is observed, the application, and the like. The format of the image data is converted if necessary, before and after the luminance conversion. The luminance conversion processing here means a brightness adjustment performed on a digital signal level.
- The
color conversion unit 205 converts a color temperature of the captured image data. Thecolor conversion unit 205 makes an adjustment so that, for example, a ratio among RGB signal intensities becomes 1:1:1 in the captured image data in the RGB format. The adjusted ratio among RGB signals may be a ratio corresponding to a color temperature defined by a standard such as standard RGB (sRGB) selected in consideration of the environment under which the captured image data is observed, the application, and the like. A method for calculating the color temperature from the ratio among RGB signal intensities will be described below with reference toFIG. 3 . The format of the captured image data is converted if necessary, before and after the color conversion. The color conversion by thecolor conversion unit 205 means a white balance adjustment performed on the digital signal level. - As illustrated in
FIG. 2 , the local illuminationenvironment processing unit 107 includes an illumination environmentinformation processing unit 206, an image identificationinformation extraction unit 207, atiming adjustment unit 208, animage storage unit 209, aluminance conversion unit 210, and acolor conversion unit 211. - The illumination environment
information processing unit 206 includes a storage unit (not illustrated) for storing illumination environment information, and sequentially stores therein the illumination environment information input from the illumination environmentinformation extraction unit 202. Further, image identification information is input from thetiming adjustment unit 208 into the illumination environmentinformation processing unit 206, by which the illumination environmentinformation processing unit 206 reads out illumination environment information corresponding to this image identification information, and outputs a readout success signal (Acknowledgement (ACK)) to thetiming adjustment unit 208. In addition thereto, the illumination environmentinformation processing unit 206 outputs the image identification information and the luminance information contained in the illumination environment information to theluminance conversion unit 210, and outputs the image identification information and the color temperature information contained in the illumination environment information to thecolor conversion unit 211. - The image identification
information extraction unit 207 extracts the image identification information from the mixed reality image data. This image identification information is the image identification information embedded in the captured image data by the illumination environmentinformation extraction unit 202. The image identification information extracted from the mixed reality image data is output to thetiming adjustment unit 208. Further, the image identificationinformation extraction unit 207 bundles the mixed reality image data and the extracted image identification information together as one set, and stores them into theimage storage unit 209. - The
timing adjustment unit 208 outputs the image identification information input from the image identificationinformation extraction unit 207 to the illumination environmentinformation processing unit 206. The illumination environmentinformation processing unit 206 outputs a readout success signal (ACK) to thetiming adjustment unit 208, if succeeded in reading out illumination environment information corresponding to this image identification information. Thetiming adjustment unit 208 adjusts a timing of reading out the mixed reality image data from theimage storage unit 209, after confirming the ACK. More specifically, thetiming adjustment unit 208 adjusts the timing so that the mixed reality image data is read out from theimage storage unit 209 at a timing at which it becomes possible to perform processing using the read illumination environment information by theluminance conversion unit 210. This adjustment allows theluminance conversion unit 210 and thecolor conversion unit 211 to sequentially perform the luminance conversion and the color conversion using the illumination environment information with a delay as short as possible. - The
image storage unit 209 includes a line buffer or a frame buffer that stores the mixed reality image data. Theimage storage unit 209 stores the mixed reality image data together with the image identification information extracted by the image identificationinformation extraction unit 207. Theluminance conversion unit 210 converts a luminance of the mixed reality image data. The content of the processing is similar to the luminance conversion of captured image data by theluminance conversion unit 204, but the luminance information in the illumination environment information is used as an adjustment value therefor. Theluminance conversion unit 210 can also check an error by reading out the image identification information together with the mixed reality image data from theimage storage unit 209, and comparing it with the image identification information of the illumination environment information. As a result of this luminance conversion, the luminance of the mixed reality image data becomes substantially equivalent to the luminance of the captured image data before the conversion processing performed by theluminance conversion unit 204. - The
color conversion unit 211 converts a color temperature of the mixed reality image data. The content of the processing is similar to the color conversion of captured image data by thecolor conversion unit 205, but thecolor conversion unit 211 adjusts RGB signals based on the color temperature information in the illumination environment information. Thecolor conversion unit 211 can also check an error by inputting the image identification information together with the mixed reality image data from theluminance conversion unit 210, and comparing it with the image identification information of the illumination environment information. As a result of this color conversion processing, the color temperature of the mixed reality image data becomes substantially equivalent to the color temperature of the captured image data before the conversion processing performed by thecolor conversion unit 205. - The
display unit 108 displays the input mixed reality image data while maintaining its brightness and tint. If the input mixed reality image data is dark, dark HMD observation image data is displayed. If the input mixed reality image data is bright, bright HMD observation image data is displayed. Further, if the input mixed reality image data has a high color temperature, displayed HMD observation image data also has a high color temperature. If the input mixed reality image data has a low color temperature, displayed HMD observation image data also has a low color temperature. Thedisplay unit 108 performs processing so as to restrain a characteristic specific to the display device by a display devicecharacteristic correction unit 212, which will be described below, and converts the mixed reality image data input into thedisplay unit 108 into HMD observation image data. Thedisplay device 108 includes the display devicecharacteristic correction unit 212. The display devicecharacteristic correction unit 212 corrects a characteristic depending on an individual display device. -
FIG. 3 illustrates a graph indicating a corresponding relationship between RGB signal intensities in the RGB image format and a color temperature. Acolor temperature 6500K is set as a reference, and an RGB signal gain is adjust so that an RGB signal ratio becomes 1:1:1 when a white plate is imaged under thecolor temperature 6500K. This is referred to as a white RGB signal gain under thereference color temperature 6500K.FIG. 3 illustrates the relationship of RGB signal intensities to each color temperature when the same white plate is imaged under various color temperatures, while maintaining this RGB signal gain. For calculating a color temperature under various kinds of environments of the external world, first, RGB signal intensities are calculated using the white RGB signal gain under thereference color temperature 6500K. Next, a color temperature can be acquired from the ratio among these RGB signal intensities and the corresponding relationship illustrated inFIG. 3 . If the ratio between R/G and B/G is 4:1, the color temperature is approximately 3500K according to the graph inFIG. 3 . The ratio is used at this time in consideration of an influence of a bias applied to the whole RGB signal intensities depending on a degree of brightness of image data, but the color temperature can be acquired from RGB signal intensities as long as the luminance is adjusted in advance. The method described with reference toFIG. 3 is a method for calculating a color temperature from a ratio among RGB signal intensities, and is merely an example of a method for calculating a color temperature from image data. -
FIG. 4 is a flowchart illustrating processing performed by the standard illuminationenvironment processing unit 104 in the video see-throughHMD 101. In the following description, the processing by the standard illuminationenvironment processing unit 104 will be described with reference toFIG. 4 . A central processing unit (CPU) in the video see-throughHMD 101 reads out a required program and required data from a recording medium such as a read only memory (ROM) to execute the program, by which the processing illustrated inFIG. 4 , and the processing illustrated inFIGS. 5 and 6 that will be described below are realized. - In step S401, the illumination environment
information extraction unit 202 inputs captured image data from the imaging devicecharacteristic correction unit 201 of theimaging unit 103. In step S402, the illumination environmentinformation extraction unit 202 extracts illumination environment information that contains color temperature information and luminance information from the input captured image data. The extracted illumination environment information is output to the illumination environmentinformation processing unit 206 together with image identification information from the imageidentification information source 203. - In step S403, the illumination environment
information extraction unit 202 combines or adds the image identification information supplied from the imageidentification information source 203 to the captured image data. The illumination environmentinformation extraction unit 202 may combine the image identification information to the captured image data by the electronic watermark technique, or may add the image identification information to a header of the captured image data. The captured image data with the image identification information combined or added thereto is output to theluminance conversion unit 204. - In step S404, the
luminance conversion unit 204 converts a luminance of the captured image data. The captured image data after the luminance conversion is output to thecolor conversion unit 205. In step S405, thecolor conversion unit 205 converts a color of the captured image data. The captured image data after the color conversion is output to theimage output unit 105 in the video see-throughHMD 101. - Next, processing performed by the local illumination
environment processing unit 107 in the video see-throughHMD 101 will be described with reference toFIGS. 5 and 6 .FIG. 5 is a flowchart illustrating a flow of image data in the local illuminationenvironment processing unit 107. - In step S501, the image identification
information extraction unit 207 inputs mixed reality image data from theimage input unit 106. In step S502, the image identificationinformation extraction unit 207 extracts image identification information from the mixed reality image data. The mixed reality image data, from which the image identification information is extracted, is output to theimage storage unit 209 together with the image identification information. Further, the extracted image identification information is also output to thetiming adjustment unit 208. - In step S503, the
image storage unit 209 stores the image identification information and the mixed reality image data in association with each other. In step S504, theluminance conversion unit 210 reads out the mixed reality image data from theimage storage unit 209. More specifically, in response to an input of the image identification information and an image readout permission from thetiming adjustment unit 208, theimage storage unit 209 makes mixed reality image data corresponding to this image identification information ready for being read out from theimage storage unit 209. Theluminance conversion unit 210 reads out this display image data, once the mixed reality image data becomes ready for being read out from theimage storage unit 209. - In step S505, the
luminance conversion unit 210 converts a luminance of the mixed reality image data. Theluminance conversion unit 210 uses the luminance information contained in the illumination environment information from the illumination environmentinformation processing unit 206 as an adjustment value. The display image data after the luminance conversion is output to thecolor conversion unit 211. In step S506, thecolor conversion unit 211 converts a color of the display image data. More specifically, thecolor conversion unit 211 adjusts RGB signals based on the color temperature information contained in the illumination environment information from the illumination environmentinformation processing unit 206. The display image data after the color conversion is output to the display devicecharacteristic correction unit 212 of thedisplay unit 108. -
FIG. 6 is a flowchart illustrating timing adjustment processing performed by the local illuminationenvironment processing unit 107 of the video see-throughHMD 101. - In step S601, the image identification
information extraction unit 207 extracts the image identification information from the mixed reality image data. The extracted image identification information is output to thetiming adjustment unit 208. In step S602, thetiming adjustment unit 208 outputs the image identification information input from the image identificationinformation extraction unit 207 to the illumination environmentinformation processing unit 206. - In step S603, the illumination environment
information processing unit 206 reads out illumination environment information corresponding to the image identification information input from thetiming adjustment unit 208, from the storage unit thereof. The illumination environmentinformation processing unit 206 stores the image identification information and the illumination environment information input from the illumination environmentinformation extraction unit 202 in a state bundled as one set. If the illumination environmentinformation processing unit 206 succeeds in reading out the illumination environment information, the illumination environmentinformation processing unit 206 outputs a readout success signal (ACK) to thetiming adjustment unit 208. The illumination environmentinformation processing unit 206 outputs the illumination environment information and the image identification information to each of theluminance conversion unit 210 and the color conversion unit at the same time as the output of the ACK. - In step S604, the
timing adjustment unit 208 determines whether an ACK is input. If an ACK is not input even after a time-out time period has elapsed (NO in step S604), the processing returns to step S602. Then, thetiming adjustment unit 208 outputs image identification information of mixed reality image data one frame before the current frame to the illumination environmentinformation processing unit 206. On the other hand, if an ACK is input within the time-out time period (YES in step S604), the processing proceeds to step S605. - In step S605, the
luminance conversion unit 210 outputs an image input ready signal to thetiming adjustment unit 208 together with the image identification information at a timing at which it becomes possible to perform the luminance conversion processing using the illumination environment information from the illumination environmentinformation processing unit 206. - In step S606, by being triggered by an input of the image identification information and the image input ready signal, the
timing adjustment unit 208 transmits a permission for reading out mixed reality image data corresponding to this image identification information to theimage storage unit 209. Theluminance conversion unit 210 reads out the mixed reality image data, once it becomes possible to read out the image from theimage storage unit 209. - By the above-described processing from step S601 to step S606, the
luminance conversion unit 210 and thecolor conversion unit 211 can sequentially perform the luminance conversion processing and the color conversion processing using the illumination environment information with a delay as short as possible. -
FIG. 7 is a time chart indicating a temporal relationship between a flow of image data and a flow of information in the local illumination environment processing according to the first exemplary embodiment. As illustrated inFIG. 7 , according to the present exemplary embodiment, mixed reality image data is input into the image identificationinformation extraction unit 207, and image identification information is extracted from this mixed reality image data. Thereafter, the mixed reality image data is stored into theimage storage unit 209. While the mixed reality image data is stored into theimage storage unit 209, the following processing is performed. The image identification information of this mixed reality image data is output from thetiming adjustment unit 208. Illumination environment information corresponding to this image identification information is read out by the illumination environmentinformation processing unit 206. Further, an ACK is input into thetiming adjustment unit 208, and an image input ready signal is acquired by thetiming adjustment unit 208. Then, after the mixed reality image data is stored into theimage storage unit 209 and the image input ready signal is acquired by thetiming adjustment unit 208, the luminance and the color are converted by theluminance conversion unit 210 and thecolor conversion unit 211, respectively. -
FIGS. 8A , 8B, 8C, and 8D illustrate data structures of the illumination environment information. As illustrated inFIG. 8A , the illumination environment information includes the image identification information (date/time information), the color temperature information, and the luminance information. -
FIG. 8B illustrates an example of a data structure of the image identification information (the date/time information). As illustrated inFIG. 8B , the image identification information (the date/time information) is information that indicates when captured image data is captured, and is mainly divided into date information and time information. The date information includes year according to the Western calendar, month, and day. The time information includes hour, minute, second, and millisecond.FIG. 8B illustrates a structure example of 40 bits in total. - The image identification information (the date/time information) only has to be information that allows identification of image data from the time when captured image data is captured to the time when virtual image data is superimposed onto this captured image data. Therefore, as described above, if this time interval corresponds to 10 frames, the image identification information only has to be information that allows image data pieces of the 10 frames to be identified, respectively. However, actually, the data structure of the image identification information is determined in consideration of factors such as a variation in a time period required to generate virtual image data, and a possibility that the image identification information is also used as attribute information of the captured image data. For example, if the above-described time interval during which image data should be identified is not a constant interval from about 120 frames to about 180 frames due to a variation in the time period required to generate virtual image data, it is desirable that the image identification information has a data structure having a margin, for example, a data structure that allows image data pieces of 240 frames to be identified, respectively. If the image identification information is also used as attribute information of the captured image data, it is desirable that the image identification information has such a data structure that attribute information such as a resolution, an image format, and a shutter speed is added to the data structure illustrated in
FIG. 8B .FIGS. 8C and 8D illustrate examples of data structures of the color temperature information and the luminance information, respectively. The color temperature information is information that indicates a color temperature itself, as illustrated inFIG. 8C . The luminance information is information that indicates a Y signal when image data is expressed in the YUV format, as illustrated inFIG. 8D . -
FIGS. 9A , 9B, 9C, and 9D illustrate what kind of characteristics an imaging device and a display device have. Representative examples of the imaging device characteristic include a receivable light frequency characteristic and a sensitivity characteristic. -
FIG. 9A illustrates an example of the receivable light frequency characteristic of the imaging device. The characteristic of frequency of light receivable by the imaging device varies for each imaging device due to influences of absorption and scattering of a color filter and a peripheral member used in the imaging device. This affects a tint of captured image data. -
FIG. 9B illustrates an example of the sensitivity characteristic of the imaging device. Suppose that the imaging device has a Bayer array, and measures signal outputs (VGr, VGb, VR, and VB) at a center of a screen when a white object is imaged. A gain of an electric signal is to be adjusted so that, if a light intensity of each RGB single color is the same, an intensity of an electric signal after photoelectric conversion becomes also the same for each RGB. This affects a luminance of captured image data. - The imaging device characteristic is not limited to the above-described receivable light frequency characteristic and sensitivity characteristic, and may be another element in the
imaging unit 103 such as an infrared (IR) filter and an analog transmission characteristic. The imaging devicecharacteristic correction unit 201 corrects a characteristic (a tint, a luminance, and the like) that depends on a component of theimaging unit 103. - Representative examples of the display device characteristic include an emittable light frequency characteristic and a luminance characteristic of a light source.
FIG. 9C illustrates an example of the emittable light frequency characteristic of the light source. This example corresponds to emittable light frequencies of respective RGB colors when light-emitting diodes (LEDs) of RGB three colors are used as the light source. This affects a tint of mixed reality image data. -
FIG. 9D illustrates an example of the luminance characteristic of the light source. A current supplied to the LED is to be adjusted for each RGB to adjust a light intensity of each RGB single color. This affects a luminance of mixed reality image data. - The display device characteristic is not limited to the above-described emittable light frequency characteristic and luminance characteristic of the light source, and may be another element in the
display unit 108 such as a liquid-crystal filter characteristic of a display panel, and a lens characteristic. The display devicecharacteristic correction unit 212 corrects a characteristic (a tint, a luminance, and the like) that depends on a component of thedisplay unit 108. - In this manner, according to the present exemplary embodiment, mixed reality image data can be dynamically converted into image data that matches an illumination environment of an external world with use of illumination environment information extracted from captured image data. Therefore, brightness and color sensations felt by an HMD user to an ambient environment can substantially match an actual environment of the external world, which allows the user to be further engrossed into a mixed reality space.
- Further, captured image data output from the video see-through
HMD 101 into theimage processing apparatus 102 is image data in which the characteristics depending on the imaging device are canceled, and a brightness and a white balance are adjusted. Therefore, the present exemplary embodiment has advantages of compliance with a basic idea of color matching, facilitation of marker extraction from captured image data to detect a position and an orientation, elimination of the necessity of changing a color temperature and a luminance of the light source during drawing of virtual image data for each system, a reduction in loss in image data due to compression, and the like. Further, the characteristics depending on the display device are canceled from mixed reality image data in thedisplay unit 108, which allows formation of image data according to a standard that does not depend on the display device. Further, the present exemplary embodiment does not require an additional sensor different from the components of the video see-throughHMD 101 itself, and utilizes a corresponding relationship between captured image data and mixed reality image data, thereby simplifying the configuration. - Next, a second exemplary embodiment of the present invention will be described.
FIG. 10 is a block diagram illustrating a functional configuration of a mixed reality system using a video see-throughHMD 1001 according to the second exemplary embodiment. Referring toFIG. 10 , the mixed reality system includes the video see-throughHMD 1001 according to the second exemplary embodiment. The video see-throughHMD 1001 includes theimaging unit 103, the standard illuminationenvironment processing unit 104, the capturedimage storage unit 110, theimage input unit 106, atime alignment unit 1004, theimage combining unit 113, the local illuminationenvironment processing unit 107, thedisplay unit 108, the position/orientation measurement unit 109, an interface (I/F)unit 1003, and theimage output unit 105. - Further, Referring to
FIG. 10 , the mixed reality system includes animage processing apparatus 1002. Theimage processing apparatus 1002 includes the I/F unit 1003, thecontent DB 111, theCG drawing unit 112, and theimage output unit 105. Theimage processing apparatus 1002 can be embodied by an apparatus that has a high-performance calculation processing function and graphics generation/display function, such as a personal computer and a workstation. The mixed reality system further includes anexternal output apparatus 1005. This is a display apparatus such as a plasma display panel (PDP), an organic electroluminescence (EL) display, and a liquid-crystal display. InFIG. 10 , units similar to those inFIG. 1 are labeled in a similar manner toFIG. 1 . - According to the first exemplary embodiment, captured image data is output from the video see-through
HMD 101 to theimage processing apparatus 102. On the other hand, according to the second exemplary embodiment, position/orientation information and image identification information are output from the video see-throughHMD 1001 to theimage processing apparatus 1002, and captured image data and the image identification information are bundled together to be stored as one set into the capturedimage storage unit 110 of the video see-throughHMD 1001. The first exemplary embodiment has such an advantage that the video see-throughHMD 101 can be simply configured, because the first exemplary embodiment can be realized by using functions of an existing image processing apparatus. On the other hand, the second exemplary embodiment has such advantages that a communication amount can be reduced between the video see-throughHMD 1001 and theimage processing apparatus 1002, and a delay due to an output of image data can be reduced. - The
image processing apparatus 1002 generates virtual image data from the position/orientation information, and outputs the virtual image data to the video see-throughHMD 1001 together with the image identification information. The image identification information is embedded in the virtual image data by the electronic watermark technique, or is added to a header of the virtual image data. The captured image data and the virtual image data are combined within the video see-throughHMD 1001, and mixed reality image data, which is the combined image data therefrom, is displayed on thedisplay unit 108 - Because captured image data and virtual image data are combined within the video see-through
HMD 1001, time alignment processing for eliminating a time lag between the captured image data and the virtual image data has to be performed in the video see-throughHMD 1001. Therefore, according to the second exemplary embodiment, thetime alignment unit 1004 is newly provided so as to allow theimage combining unit 113 to combine captured image data and virtual image data that coincides with the captured image data on a temporal axis. Thetime alignment unit 1004 performs time alignment processing for the local illuminationenvironment processing unit 107, in addition to time alignment processing for theimage combining unit 113. - According to the second exemplary embodiment, the video see-through
HMD 1001 includes theimage combining unit 113, whereby the second exemplary embodiment is configured in such a manner that mixed reality image data is held only by the video see-throughHMD 1001. Further, mixed reality image data is output from theimage combining unit 113 to not only the local illuminationenvironment processing unit 107 but also theimage output unit 105, to allow another person than an HMD user to also observe the mixed reality image data. The mixed reality image data is output from theimage output unit 105 to theexternal output apparatus 1005. With this configuration, image data that does not depend on an observation environment of the HMD user can be output to theexternal output apparatus 1005 as mixed reality image data. Preparing mixed reality image data as image data according to the sRGB standard or the like allows theexternal output apparatus 1005 to convert this data into image data suitable for its observation environment. -
FIG. 11 is a block diagram illustrating a further detailed configuration of the video see-throughHMD 1001 according to the second exemplary embodiment. InFIG. 11 , units similar to those inFIG. 2 are labeled in a similar manner toFIG. 2 . - Referring to
FIG. 11 , theimaging unit 103 has a function similar to theimaging unit 103 illustrated inFIG. 2 . The standard illuminationenvironment processing unit 104 includes the illumination environmentinformation extraction unit 202, the imageidentification information source 203, theluminance conversion unit 204, and thecolor conversion unit 205. The imageidentification information source 203, theluminance conversion unit 204, and thecolor conversion unit 205 have functions similar to the imageidentification information source 203, theluminance conversion unit 204, and thecolor conversion unit 205 illustrated inFIG. 2 , respectively. - The illumination environment
information extraction unit 202 extracts illumination environment information that contains color temperature information and luminance information from captured image data by performing image processing, and outputs the extracted illumination environment information to the local illuminationenvironment processing unit 107 together with image identification information from the imageidentification information source 203. In addition thereto, the illumination environmentinformation extraction unit 202 adds the image identification information to the captured image data, and outputs this captured image data to theluminance conversion unit 204. The present exemplary embodiment is configured so as not to transmit the captured image data from the video see-throughHMD 1001 to theimage processing apparatus 1002. Therefore, it is desirable to handle the image identification information as different information from the captured image data instead of embedding the image identification information into the captured image data, in consideration that the image identification information is stored together with the captured image data as one set into the capturedimage storage unit 110, which is a later stage. - The captured
image storage unit 110 stores image identification information and captured image data corresponding to this image identification information as one set. The position/orientation measurement unit 109 extracts a marker and/or a natural feature from the captured image data input from the standard illuminationenvironment processing unit 104, and measures position/orientation information at a line-of-sight position of the HMD user. The position/orientation measurement unit 109 bundles the measured position/orientation information and image identification information corresponding to this position/orientation information together as one set, and outputs them to the I/F unit 1003. The I/F unit 1003 transmits the position/orientation information and the image identification information bundled together as one set to theimage processing apparatus 1002. - The local illumination
environment processing unit 107 includes the illumination environmentinformation processing unit 206, theluminance conversion unit 210, and thecolor conversion unit 211. InFIG. 2 , the image identificationinformation extraction unit 207 and thetiming adjustment unit 208 are provided as components of the local illuminationenvironment processing unit 107. InFIG. 11 , they are provided in another block as thetime alignment unit 1004. The local illuminationenvironment processing unit 107 has a function and a processing content similar toFIG. 2 , only in terms of the local illumination environment processing. However, thetime alignment unit 1004 configured as an independent block inFIG. 11 is extended so as to function even in time alignment processing for the image combining processing by theimage combining unit 113. - The
time alignment unit 1004 includes the image identificationinformation extraction unit 207 and thetiming adjustment unit 208. Thetime alignment unit 1004 performs not only the time alignment processing for theimage combining unit 113 but also the time alignment processing for the local illuminationenvironment processing unit 107. - The image identification
information extraction unit 207 extracts image identification information from virtual image data. This image identification information is the image identification information added to the captured image data by the illumination environmentinformation extraction unit 202. The extracted image identification information is output to thetiming adjustment unit 208. Further, the virtual image data is output to theimage combining unit 113 together with the extracted image identification information. - The
image combining unit 113 receives the output virtual image data and reads out captured image data corresponding to the image identification information thereof from the capturedimage storage unit 110, and then combines the virtual image data and the captured image data. Mixed reality image data after the combining processing is output to theluminance conversion unit 210 and theimage output unit 105. -
FIG. 12 is a time chart indicating a temporal relationship between a flow of image data and a flow of information in the image combination processing and the local illumination environment processing according to the second exemplary embodiment. As illustrated inFIG. 12 , at a timing when virtual image data and captured image data are combined by theimage combining unit 113, the following processing is performed. Image identification information of the mixed reality image data is output by thetiming adjustment unit 208. Illumination environment information corresponding to this image identification information is read out by the illumination environmentinformation processing unit 206. Further, an ACK is input into thetiming adjustment unit 208, and an image input ready signal is acquired by thetiming adjustment unit 208. This is a difference from the first exemplary embodiment. - According to the present exemplary embodiment, mixed reality image data output from the video see-through
HMD 1001 into theexternal output apparatus 1005 is image data in which the characteristics depending on the device are canceled, and a brightness and a white balance are adjusted. Therefore, mixed reality image data according to a standard that does not depend on the display device can be output. The present exemplary embodiment complies with a basic idea of color matching, thereby having an advantage of being able to provide a display suitable for an external environment and a device characteristic of theexternal output apparatus 1005. -
FIGS. 13A and 13B are block diagrams illustrating a modification of the mixed reality system according to the second exemplary embodiment. More specifically,FIG. 13A illustrates a configuration of the modification of the mixed reality system according to the second exemplary embodiment, andFIG. 13B illustrates switching processing by aselector 1301. The mixed reality system illustrated inFIG. 13A corresponds to the configuration of the mixed reality system illustrated inFIG. 10 with theselector 1301 added thereto. - For displaying only captured image data on the video see-through
HMD 1001 instead of displaying mixed reality image data, theselector 1301 is switched to (2) illustrated inFIG. 13B , thereby changing a data flow so as to transmit image data from theimaging unit 103 to thedisplay unit 108. At that time, the mixed reality system can be switched to a low power consumption mode with the components other than theimaging unit 103, theselector 1301, and thedisplay unit 108 substantially stopping working. Theimaging unit 103 performs processing that enables an environment of captured image data to match an environment of an external world, and thedisplay unit 108 displays the input mixed reality image data while maintaining its luminance and tint, thereby allowing an HMD user to observe an image that matches the environment of the external world. Further, for superimposing virtual image data onto captured image data, theselector 1301 is switched to (1) illustrated inFIG. 13B . In this case, the mixed reality system according to the modification has functions and processing similar to the mixed reality system according to the second exemplary embodiment illustrated inFIG. 10 . - In this manner, if only captured image data is displayed on the video see-through
HMD 101 without superimposing virtual image data thereon, mixed reality image data can be dynamically adjusted to image data that matches an illumination environment of an external world. Further, a delay in the processing of the system, a processing load, and power consumption can be reduced. - Embodiments of the present invention can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions recorded on a storage medium (e.g., non-transitory computer-readable storage medium) to perform the functions of one or more of the above-described embodiment(s) of the present invention, and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s). The computer may comprise one or more of a central processing unit (CPU), micro processing unit (MPU), or other circuitry, and may include a network of separate computers or separate computer processors. The computer executable instructions may be provided to the computer, for example, from a network or the storage medium. The storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)™), a flash memory device, a memory card, and the like.
- While the present invention has been described with reference to exemplary embodiments, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
- This application claims the benefit of Japanese Patent Application No. 2013-074267 filed Mar. 29, 2013, which is hereby incorporated by reference herein in its entirety.
Claims (11)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013074267A JP5769751B2 (en) | 2013-03-29 | 2013-03-29 | Image processing apparatus, image processing method, and program |
JP2013-074267 | 2013-03-29 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20140292811A1 true US20140292811A1 (en) | 2014-10-02 |
US9501870B2 US9501870B2 (en) | 2016-11-22 |
Family
ID=51620351
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/226,726 Active 2034-10-02 US9501870B2 (en) | 2013-03-29 | 2014-03-26 | Mixed reality image processing apparatus and mixed reality image processing method |
Country Status (2)
Country | Link |
---|---|
US (1) | US9501870B2 (en) |
JP (1) | JP5769751B2 (en) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2533573A (en) * | 2014-12-22 | 2016-06-29 | Nokia Technologies Oy | Image processing method and device |
WO2016122571A1 (en) * | 2015-01-30 | 2016-08-04 | Hewlett Packard Enterprise Development Lp | Dynamic modulation for near eye display |
WO2016139882A1 (en) * | 2015-03-03 | 2016-09-09 | Canon Kabushiki Kaisha | Transfer control apparatus, control method, and mixed-reality presentation apparatus |
US20170161949A1 (en) * | 2015-12-08 | 2017-06-08 | GM Global Technology Operations LLC | Holographic waveguide hud side view display |
CN106855656A (en) * | 2015-12-08 | 2017-06-16 | 通用汽车环球科技运作有限责任公司 | The image procossing of augmented reality system and shielded object |
US20170178565A1 (en) * | 2015-12-21 | 2017-06-22 | Seiko Epson Corporation | Display device, method of controlling display device, and program |
US20180007312A1 (en) * | 2015-03-03 | 2018-01-04 | Canon Kabushiki Kaisha | Transfer control apparatus, control method, and mixed-reality presentation apparatus |
US10267960B1 (en) | 2018-02-05 | 2019-04-23 | GM Global Technology Operations LLC | Cloaking device and apparatus |
US10728531B2 (en) * | 2017-05-01 | 2020-07-28 | Canon Kabushiki Kaisha | Image display system, image display apparatus, image display method, and storage medium |
US20210004996A1 (en) * | 2019-07-01 | 2021-01-07 | Microsoft Technology Licensing, Llc | Adaptive user interface palette for augmented reality |
CN114895471A (en) * | 2016-02-18 | 2022-08-12 | 苹果公司 | Head mounted display for virtual reality and mixed reality with inside-outside position tracking, user body tracking, and environment tracking |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6143242B1 (en) * | 2016-02-15 | 2017-06-07 | 株式会社菊池製作所 | Image display device |
US10872582B2 (en) | 2018-02-27 | 2020-12-22 | Vid Scale, Inc. | Method and apparatus for increased color accuracy of display by compensating for observer's color vision properties |
WO2021146035A1 (en) | 2020-01-15 | 2021-07-22 | Magic Leap, Inc. | Compensation for deformation in head mounted display systems |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3423402B2 (en) * | 1994-03-14 | 2003-07-07 | キヤノン株式会社 | Video display device |
US6885382B1 (en) * | 1999-08-18 | 2005-04-26 | Fuji Xerox Co., Ltd. | Image processing device, image processing system, output device, computer readable recording medium and image processing method |
US20120147163A1 (en) * | 2010-11-08 | 2012-06-14 | DAN KAMINSKY HOLDINGS LLC, a corporation of the State of Delaware | Methods and systems for creating augmented reality for color blindness |
US20130002698A1 (en) * | 2011-06-30 | 2013-01-03 | Disney Enterprises, Inc. | Virtual lens-rendering for augmented reality lens |
US20140002475A1 (en) * | 2012-06-27 | 2014-01-02 | Samsung Electronics Co., Ltd. | Image distortion compensation device, medical imaging device comprising the same and method for compensating image distortion |
US20140168243A1 (en) * | 2012-12-19 | 2014-06-19 | Jeffrey Huang | System and Method for Synchronizing, Merging, and Utilizing Multiple Data Sets for Augmented Reality Application |
US20140198104A1 (en) * | 2011-09-02 | 2014-07-17 | Sharp Kabushiki Kaisha | Stereoscopic image generating method, stereoscopic image generating device, and display device having same |
US20140307007A1 (en) * | 2013-04-11 | 2014-10-16 | Lg Display Co., Ltd. | Electronic device, display controlling apparatus and method thereof |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3658330B2 (en) | 2001-02-21 | 2005-06-08 | キヤノン株式会社 | Composite display device and head mounted display device using the same |
JP2008216343A (en) * | 2007-02-28 | 2008-09-18 | Canon Inc | Image display apparatus, program and storage medium |
-
2013
- 2013-03-29 JP JP2013074267A patent/JP5769751B2/en not_active Expired - Fee Related
-
2014
- 2014-03-26 US US14/226,726 patent/US9501870B2/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3423402B2 (en) * | 1994-03-14 | 2003-07-07 | キヤノン株式会社 | Video display device |
US6885382B1 (en) * | 1999-08-18 | 2005-04-26 | Fuji Xerox Co., Ltd. | Image processing device, image processing system, output device, computer readable recording medium and image processing method |
US20120147163A1 (en) * | 2010-11-08 | 2012-06-14 | DAN KAMINSKY HOLDINGS LLC, a corporation of the State of Delaware | Methods and systems for creating augmented reality for color blindness |
US20130002698A1 (en) * | 2011-06-30 | 2013-01-03 | Disney Enterprises, Inc. | Virtual lens-rendering for augmented reality lens |
US20140198104A1 (en) * | 2011-09-02 | 2014-07-17 | Sharp Kabushiki Kaisha | Stereoscopic image generating method, stereoscopic image generating device, and display device having same |
US20140002475A1 (en) * | 2012-06-27 | 2014-01-02 | Samsung Electronics Co., Ltd. | Image distortion compensation device, medical imaging device comprising the same and method for compensating image distortion |
US20140168243A1 (en) * | 2012-12-19 | 2014-06-19 | Jeffrey Huang | System and Method for Synchronizing, Merging, and Utilizing Multiple Data Sets for Augmented Reality Application |
US20140307007A1 (en) * | 2013-04-11 | 2014-10-16 | Lg Display Co., Ltd. | Electronic device, display controlling apparatus and method thereof |
Non-Patent Citations (1)
Title |
---|
Takemura, Masayuki, Itaru Kitahara, and Yuichi Ohta. "Photometric inconsistency on a mixed-reality face." Proceedings of the 5th IEEE and ACM International Symposium on Mixed and Augmented Reality. IEEE Computer Society, 2006. * |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2533573A (en) * | 2014-12-22 | 2016-06-29 | Nokia Technologies Oy | Image processing method and device |
WO2016122571A1 (en) * | 2015-01-30 | 2016-08-04 | Hewlett Packard Enterprise Development Lp | Dynamic modulation for near eye display |
US10638088B2 (en) * | 2015-03-03 | 2020-04-28 | Canon Kabushiki Kaisha | Transfer control apparatus, control method, and mixed-reality presentation apparatus |
WO2016139882A1 (en) * | 2015-03-03 | 2016-09-09 | Canon Kabushiki Kaisha | Transfer control apparatus, control method, and mixed-reality presentation apparatus |
US20180007312A1 (en) * | 2015-03-03 | 2018-01-04 | Canon Kabushiki Kaisha | Transfer control apparatus, control method, and mixed-reality presentation apparatus |
US20170161949A1 (en) * | 2015-12-08 | 2017-06-08 | GM Global Technology Operations LLC | Holographic waveguide hud side view display |
CN106855656A (en) * | 2015-12-08 | 2017-06-16 | 通用汽车环球科技运作有限责任公司 | The image procossing of augmented reality system and shielded object |
CN106853799A (en) * | 2015-12-08 | 2017-06-16 | 通用汽车环球科技运作有限责任公司 | Holographical wave guide head-up display side view shows |
US20170178565A1 (en) * | 2015-12-21 | 2017-06-22 | Seiko Epson Corporation | Display device, method of controlling display device, and program |
US10121409B2 (en) * | 2015-12-21 | 2018-11-06 | Seiko Epson Corporation | Display device, method of controlling display device, and program |
CN114895471A (en) * | 2016-02-18 | 2022-08-12 | 苹果公司 | Head mounted display for virtual reality and mixed reality with inside-outside position tracking, user body tracking, and environment tracking |
US10728531B2 (en) * | 2017-05-01 | 2020-07-28 | Canon Kabushiki Kaisha | Image display system, image display apparatus, image display method, and storage medium |
US10267960B1 (en) | 2018-02-05 | 2019-04-23 | GM Global Technology Operations LLC | Cloaking device and apparatus |
US20210004996A1 (en) * | 2019-07-01 | 2021-01-07 | Microsoft Technology Licensing, Llc | Adaptive user interface palette for augmented reality |
US11494953B2 (en) * | 2019-07-01 | 2022-11-08 | Microsoft Technology Licensing, Llc | Adaptive user interface palette for augmented reality |
Also Published As
Publication number | Publication date |
---|---|
JP2014199532A (en) | 2014-10-23 |
US9501870B2 (en) | 2016-11-22 |
JP5769751B2 (en) | 2015-08-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9501870B2 (en) | Mixed reality image processing apparatus and mixed reality image processing method | |
US10699473B2 (en) | System and method for generating a virtual viewpoint apparatus | |
US20200294214A1 (en) | Image processing apparatus, method for controlling image processing apparatus, and non-transitory computer-readable storage medium | |
US11494960B2 (en) | Display that uses a light sensor to generate environmentally matched artificial reality content | |
US9117410B2 (en) | Image display device and method | |
JP7254461B2 (en) | IMAGING DEVICE, CONTROL METHOD, RECORDING MEDIUM, AND INFORMATION PROCESSING DEVICE | |
KR20190041586A (en) | Electronic device composing a plurality of images and method | |
EP3402179B1 (en) | Image-capturing system, image-capturing method, and program | |
US20180332210A1 (en) | Video system, video processing method, program, camera system, and video converter | |
CN110458826B (en) | Ambient brightness detection method and device | |
EP3609175A1 (en) | Apparatus and method for generating moving image data including multiple section images in electronic device | |
EP3675477A1 (en) | Electronic device for providing function by using rgb image and ir image acquired through one image sensor | |
JP2006295506A (en) | Image display device and controlling method thereof | |
JP2016142988A (en) | Display device and display control program | |
JP6904684B2 (en) | Image processing equipment, image processing methods, and programs | |
US20200236299A1 (en) | Image capturing apparatus, image compositing method, and recording medium having recorded therein image compositing program to be executed by computer of image capturing apparatus | |
JP2015194567A (en) | display device | |
JP6137910B2 (en) | Information processing apparatus, information processing method, and program | |
JP2017138927A (en) | Image processing device, imaging apparatus, control method and program thereof | |
JP5725271B2 (en) | Color correction system | |
JP2015171038A (en) | Image processing system and method of the same | |
US11212500B2 (en) | Image capture apparatus, electronic apparatus, and recording medium suppressing chroma in white balance correction performed based on color temperature | |
KR20210151613A (en) | Image Processing Device and Image Processing Method | |
US10257444B2 (en) | Image processing apparatus, image processing method, and storage medium storing program | |
KR101488641B1 (en) | Image processing apparatus and Image processing method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: CANON KABUSHIKI KAISHA, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TAKAYAMA, TOMOHIKO;REEL/FRAME:033071/0276 Effective date: 20140303 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
FEPP | Fee payment procedure |
Free format text: 7.5 YR SURCHARGE - LATE PMT W/IN 6 MO, LARGE ENTITY (ORIGINAL EVENT CODE: M1555); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |