JP4755565B2 - Stereoscopic image processing device - Google Patents

Stereoscopic image processing device Download PDF

Info

Publication number
JP4755565B2
JP4755565B2 JP2006282543A JP2006282543A JP4755565B2 JP 4755565 B2 JP4755565 B2 JP 4755565B2 JP 2006282543 A JP2006282543 A JP 2006282543A JP 2006282543 A JP2006282543 A JP 2006282543A JP 4755565 B2 JP4755565 B2 JP 4755565B2
Authority
JP
Japan
Prior art keywords
parallax
image data
image
information
image processing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
JP2006282543A
Other languages
Japanese (ja)
Other versions
JP2008103820A (en
Inventor
竜二 北浦
保孝 若林
Original Assignee
シャープ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by シャープ株式会社 filed Critical シャープ株式会社
Priority to JP2006282543A priority Critical patent/JP4755565B2/en
Publication of JP2008103820A publication Critical patent/JP2008103820A/en
Application granted granted Critical
Publication of JP4755565B2 publication Critical patent/JP4755565B2/en
Application status is Expired - Fee Related legal-status Critical
Anticipated expiration legal-status Critical

Links

Images

Description

  The present invention relates to a stereoscopic image processing apparatus, and more particularly, to a stereoscopic image processing apparatus that makes it possible to visually recognize parallax of a three-dimensional image data portion when reproducing moving image data including at least three-dimensional image data.

  A technique using binocular parallax is known as one of the techniques for humans to obtain a stereoscopic effect. Binocular parallax is a shift in video obtained by viewing an object from two different viewpoints, and the distance in the depth direction is perceived based on this. Research for displaying stereoscopic images using this binocular parallax has been conducted.

  In recent years, various types of stereoscopic display devices have been proposed, and two-view images that are stereoscopically viewed using two-viewpoint images, and more than a plurality of positions using images with a larger number of viewpoints. Multi-view systems that enable natural stereoscopic viewing, super multi-view systems that increase the number of viewpoints and allow two or more viewpoint images to enter the pupil, etc. have been proposed. A stereoscopic effect is obtained by using parallax between viewpoint images.

  Image data for realizing stereoscopic viewing in such a display device includes a method of providing an image corresponding to a plurality of viewpoints, a reference image that is an image from one viewpoint, and an image of a reference image that is different from the reference image. There are various formats such as a method of providing a parallax map having information indicating the distance to the corresponding point, that is, parallax, and creating an image of another viewpoint based on the parallax map and the reference image.

  As an example of image data for stereoscopic display, there is a “stereoscopic video compression encoding device and stereoscopic video decoding / playback device” described in Patent Document 1 below. According to this technique, the encoding device generates a reference image from left-eye image data and right-eye image data input by a camera or the like, and then generates a reference image, a left-eye image, and a right-eye image. Differences are obtained and set as difference data for the left eye and difference data for the right eye, respectively. Then, the reference image, the difference data for the left eye, and the difference data for the right eye are compression encoded and multiplexed, thereby generating three-dimensional image encoded data. On the decoding / playback apparatus side, after decoding the encoded data, a left-eye image and right-eye image data are created from the reference image and each difference data, and a stereoscopic image is displayed.

  In such a stereoscopic view, it is known that the living body is affected by various factors. For example, as the parallax increases between the viewpoint images, the distance in the depth direction increases, and the distance between the viewpoint images greatly protrudes on the display screen or appears to be retracted. However, if the parallax is too large, there is a possibility of causing physiological effects such as giving a feeling of fatigue or feeling uncomfortable to the observer of the stereoscopic image due to a contradiction between convergence and adjustment. Here, convergence is a type of eye movement, which is a movement that attempts to cross the line of sight with the object to be observed, and adjustment is focusing, which is a movement that changes the thickness of the eye lens and changes the refractive index. That is. As shown in FIG. 14, when the distance between the eyes 1407 and 1409 and the display screen 1403 is an adjustment distance A and the distance between the eyes 1407 and 1409 and the in-focus position 1401 is a convergence distance B, the convergence distance B is While the adjustment distance A fits on the display screen in contrast to the image, a contradiction occurs. Even when an image with a large parallax is suddenly displayed, the eyes may not be able to deal with immediately, and stereoscopic viewing may not be achieved, and fatigue or discomfort may be felt.

  In addition to fatigue caused by “distortion” and “displacement” between the left and right images, the observation position may also have physiological effects such as fatigue. Even if the distance to the display device is the same, the larger the screen size, the larger the parallax and the greater the influence.In addition, the closer the distance is, the closer the stereoscopic image is to the image, but if it is too close The image does not fuse well and cannot be recognized as a three-dimensional image, causing fatigue. A method for reducing the influence on the living body has been proposed.

  According to the “stereoscopic image processing method and apparatus” described in Patent Document 2 below, the amount of parallax generated when displaying a stereoscopic image from a plurality of viewpoint images is detected, and the image falls within the allowable parallax amount of the user. A method for changing the amount of parallax between them is described. This suppresses the physiological effects as described above and adjusts the stereoscopic effect so that an appropriate stereoscopic image can be observed even if the display device changes.

  In addition, according to the “stereoscopic image processing method and apparatus” described in Patent Document 3 below, a graphical user interface (GUI) for adjusting the stereoscopic effect of a stereoscopic display image is provided, and an image is displayed based on a user instruction. A technique that enables the amount of parallax to be changed is described.

JP 7-327242 A JP 2003-284093 A JP 2003-284095 A

  As described above, the stereoscopic image may affect the living body depending on the size of the parallax. However, in the conventional 3D moving image reproduction technology, there is no method for determining which scene has a large parallax in a series of moving image data, that is, whether a pop-out or a sense of depth is strong. Therefore, when a stereoscopic moving image is reproduced from the middle, there is a problem that an image with a high stereoscopic effect is suddenly displayed, which may have a strong influence on the living body.

  Conventionally, a method for adjusting the parallax of a stereoscopic image to a parallax within an allowable range of a user has been proposed. However, since this method is a method of editing the image itself, it is not suitable for the purpose of preventing sudden reproduction from an image with strong parallax when a moving image is reproduced from the middle. There is a problem that playback cannot be performed with the amount of popping out as intended.

  The present invention has been made to solve the above-described problems. When reproducing moving image data including at least stereoscopic image data, the stereoscopic effect of any scene in a series of moving images is provided. It is intended to reduce the influence on the living body by presenting to the observer whether or not it is strong. It is another object of the present invention to provide a stereoscopic image processing apparatus that enables various viewing methods such as a user viewing only a scene with a strong stereoscopic effect.

  According to an aspect of the present invention, a stereoscopic image processing device for reproducing moving image data including three-dimensional image data, wherein each position of the moving image data is based on disparity information of the three-dimensional image data. There is provided a stereoscopic image processing apparatus having disparity information presenting means for generating a disparity information presenting image that makes it possible to visually recognize the magnitude of the disparity.

  It is preferable to have image generation means for generating an image to be displayed by synchronizing the parallax information presentation image and the image data. The parallax information presentation image is preferably displayed as an image that can be visually identified according to the magnitude of the parallax at each position of the moving image data, so that the magnitude of the parallax can be visually recognized. The visually identifiable image is, for example, color or brightness.

  The moving image data is characterized in that parallax information regarding three-dimensional image data included in the moving image data is held as metadata. In stereoscopic moving image data in a format that includes disparity information in the metadata, the disparity information presentation image can be changed and the disparity information can be presented by analyzing the disparity information in the metadata.

  It is preferable that the three-dimensional image data includes two-dimensional image data and a parallax map, and that the parallax information of the three-dimensional image data is obtained by detecting from the parallax map. Preferably, the three-dimensional image data includes a plurality of images corresponding to a plurality of viewpoints, and the parallax information of the three-dimensional data is obtained by detecting a parallax between the plurality of images corresponding to the plurality of viewpoints.

  The three-dimensional image data includes two-dimensional image data and a depth map that holds distance information in a three-dimensional space, and the distance information can be used as the parallax information. As the parallax information of the three-dimensional image data, the maximum parallax in the three-dimensional image may be used. The maximum parallax in the three-dimensional image may be divided into a plurality of sections based on the magnitude of the parallax, and the parallax information presentation image may be displayed as an image that can be visually identified according to each section.

  The maximum parallax in the three-dimensional image can be expressed in the parallax information presentation image by dividing the moving image data into fixed units and detecting the maximum parallax for each unit. The parallax information presentation image may include a reproduction position instruction pointer displayed superimposed on the displayed moving image, and the reproduction position of the current moving image may be visually recognized by the reproduction position instruction pointer.

  The apparatus further includes an allowable parallax storage unit that holds an allowable parallax of the user, and when the moving image data is reproduced from the middle, if the parallax at the reproduction position is larger than the allowable parallax, the parallax is calculated from before and after the moving image data. It is preferable to search for a position that falls within the allowable parallax, and to control playback from the searched position.

According to another aspect of the present invention, a step of demultiplexing input moving image data, 3D information necessary for performing 3D display is analyzed to obtain disparity information, and encoded Decoding moving image data; generating a parallax information presentation image that enables visual recognition of the magnitude of parallax at each position of the moving image data based on the parallax information;
Determining the display of the parallax information presentation image, generating an image according to the stereoscopic display method, and displaying the moving image data superimposed on the stereoscopic image when the parallax information presentation image is displayed. A featured stereoscopic image processing method is provided. The present invention may be a program for causing a computer to execute the above steps.

  According to the stereoscopic image processing apparatus of the present invention, there is an advantage that it is possible to easily determine which part of a series of moving images is composed of a three-dimensional image and where a scene with a large parallax is. .

  Hereinafter, a stereoscopic image processing apparatus according to an embodiment of the present invention will be described with reference to the drawings. First, the stereoscopic image processing apparatus according to the first embodiment of the present invention will be described. In the present specification, “3D” is used as a term meaning three-dimensional or three-dimensional, and “2D” is used as two-dimensional. Also, “3D image data” is a moving image data composed of a stereoscopic image, “2D image data” is a moving image data composed of a normal two-dimensional image, and “2D3D is a moving image data in which a stereoscopic image and a two-dimensional image are mixed. This is referred to as “mixed image data”.

  First, the stereoscopic image processing apparatus according to the first embodiment of the present invention will be described. FIG. 1 is a diagram illustrating a configuration example of moving image data input to the stereoscopic image processing apparatus according to the present embodiment. The moving image data is assumed to be 3D image data or 2D3D mixed image data. Hereinafter, an example in the case of 3D image data will be described.

  As shown in FIG. 1A, the 3D image data 100 includes at least four components, ie, header control information 101, 3D identification information 102, 3D display control information 103, and image data 104. . Here, audio data may be recorded together with the 3D image data 100, but it is omitted here for simplicity. Further, other additional information may be included.

  Each component is called an object and has a format shown in FIG. That is, an object ID 106 and an object size 107 for identifying each object 105 are recorded in one object 105, and then object data 108 having a size defined by the object size 107 follows. The object ID 106 and the object size 107 are combined to form an object header. This object can have a hierarchical structure.

  The header control information 101 is control information necessary for reproducing the image data 104 such as an image size. The 3D identification information 102 is information for identifying whether a 3D image is included or whether a 3D image is included. The 3D display control information 103 is information related to the obtained 3D image, and includes parallax information and control information necessary for conversion into a desired stereoscopic format. The image data 104 is image data itself. The image data 104 may be compressed or uncompressed.

  FIG. 2 is a functional block diagram illustrating a configuration example of the stereoscopic image processing apparatus according to the present embodiment. The stereoscopic image processing apparatus 200 includes a data analysis unit 201 that demultiplexes input moving image data, a 3D identification information analysis unit 202 that analyzes 3D identification information, and 3D information necessary for 3D display. 3D display control information analyzing means 203 for analyzing, decoding means 204 for decoding the encoded image data, display control means 205 for controlling the display method and parallax information presenting means for 3D images, and display control means 205 An image generation unit 206 that generates an image according to information on a display method, a parallax information presentation unit 207 that displays a parallax information presentation image as a graphical user interface (GUI) that indicates parallax of moving image data, and a liquid crystal panel that displays a 3D image And image display means 208.

  The data analysis means 201 includes a semiconductor medium represented by an SD memory card (trade name), a disk medium represented by a CD and a DVD, a magnetic medium such as a hard disk drive, or a tape represented by a DV cassette tape. Moving image data recorded on a medium or the like is input. In the data analysis unit 201, each object is separated. First, header control information 101 of moving image data is analyzed, and an object type following the header control information 101 is determined. Next, the object is identified by referring to the ID of each object, and sent to each means in FIG.

  FIG. 3 is a diagram illustrating details of each object. FIG. 3A shows a detailed configuration example of the 3D identification information 102. The ID unique to the object indicating the object relating to the 3D identification information, the size of the entire object (Size), and the 3D identification information (Data) ) Is stored.

  FIG. 3B shows a detailed configuration example of the 3D display control information 103. The ID unique to the object indicating the 3D display control information, the size (Size) of the entire object, the image arrangement information and the view Various information (Data) related to 3D display such as points and parallax information is stored.

  FIG. 3C shows a detailed configuration example of image data, in which an object-specific ID indicating the image data portion, the size of the entire object (Size), and encoded data (Data) are stored. ing.

  Here, when the unique ID of the 3D identification information is 0x11, the unique ID of the 3D display control information is 0x12, and the unique ID of the image data is 0x13, the analysis data unit 201 follows the unique ID and the unique ID is 0x11. If it is 0x12, it is determined to be 3D display control information analysis unit 203, and if it is 0x13, it is determined as image data, and the object is sent to decoding unit 204 separately.

  The 3D identification information analysis unit 202 analyzes the 3D identification information 102 sent from the data analysis unit 201 and sends it to the display control unit 205. The 3D identification information is defined as 1 when the moving image data is 3D image data, 0 when the moving image data is 2D image data, and 3 when the moving image data is 2D3D mixed data.

  The 3D display control information analysis unit 203 analyzes the 3D display control information 103 sent from the data analysis unit 201 and sends it to the display control unit 205. Here, details of information described in the 3D display control information 103 illustrated in FIG. In addition to the above, various information such as information about the camera at the time of shooting and the screen size of the display device assumed at the time of shooting may be included.

  The image arrangement information is so-called arrangement format information indicating in what arrangement the images of different viewpoints included in the 3D image data are stored. FIG. 4 shows an example at this time. FIG. 4A shows an example in which a left-eye image L viewed from the left-eye viewpoint and a right-eye image R viewed from the right-eye viewpoint are juxtaposed side by side. FIG. 4B shows an example of a format in which the left-eye image L is placed on the upper side and the right-eye image R is placed on the lower side. Such arrangement of the 3D image is indicated by image arrangement information. For example, the arrangement of FIG. 4A is defined as “1” and the arrangement of FIG. 4B is defined as “2”. The format of the image data can be identified by the information. In addition to those shown here, there are various formats such as an arrangement in which the left and right arrangements are reversed in FIG. 4A and an arrangement in which the upper and lower positions are reversed in FIG. 4B. In addition, such a stream may be an independent stream instead of a left and right form. In FIGS. 4A and 4B, the number of viewpoints is two. Here, the following description will be given on the assumption that the input 3D image has the format shown in FIG.

  The disparity information is, for example, the maximum disparity for every certain time unit. The maximum parallax indicates the parallax of the most parallax portion in the image when there are two viewpoint images. An example for explaining the maximum parallax is shown in FIG. FIG. 5A shows an image for the left eye, and FIG. 5B shows an image for the right eye. Point AL and point AR, BL and point BR are points corresponding to the left and right images, respectively, AL and AR are the points that appear to protrude most on the image when viewed stereoscopically, and BL and BR are the deepest points. The point that is seen away from is shown. The parallax is calculated from the difference between the corresponding points of the left and right images by calculating the distance (number of pixels) from the left end to the corresponding point in each image.

  When the distance (number of pixels) dL1 from the left end of the left eye image to the point AL is compared with the distance dR1 from the left end of the right eye image to the point AR, dL1> dR1, and dL1−dR1 is a positive value. It appears to pop out from the display screen. AL and AR are the points that appear to protrude most when viewed stereoscopically. Therefore, if the maximum parallax in the direction of protruding from the display screen is d1, d1 = | dL1-dR1 |. Further, when the distances dL2 and dR2 are compared, dL2 <dR2, dL2−dR2 is a negative value, and it appears to be separated from the display screen in the depth direction. If the maximum parallax from the display screen in the depth direction is d2, d2 = | dL2-dR2 |.

  It is assumed that the parallax is calculated in advance when an image is created and stored as parallax information in the 3D display control information. A detailed example of this parallax information is shown in FIG. Here, it is assumed that the 3D image data is a 30-minute stream as a whole, and the leftmost 1 to 6 sections in FIG. 6 represent 5 minutes in the 3D image data. Section 1 shows the maximum parallax of 3D image data at intervals of 5 minutes in order, such as 0 to 5 minutes and Section 2 from 5 to 10 minutes. In the data from 0 to 5 minutes which is Category 1, the maximum parallax d1 in the pop-out direction is 128, and the maximum parallax d2 in the direction from the display screen to the depth is 40, which indicates that the image has a strong pop-out feeling.

  When the image data 104 sent from the data analysis unit 201 has been encoded, the decoding unit 204 decodes the encoded image data and sends it to the image generation unit 206.

  The display control unit 205 recognizes the format of whether the input moving image data is 3D image data or 2D image data based on information from the 3D identification information analysis unit 202. Then, according to the information from the 3D display control information analysis unit 203, information such as the number of viewpoints and arrangement position information is obtained. Also, instruction information from the user is input to the display control means 205. The instruction information includes a stereoscopic display method of the image display means 208 for displaying an image and an image reproduction mode. Hereinafter, the instruction information from the user will be described.

  There are various stereoscopic display methods. Here, the image display means 208 will be described as a parallax barrier liquid crystal display. It is assumed that the barrier is realized by liquid crystal and can be switched between 2D and 3D.

  FIG. 7 is a conceptual diagram for explaining the parallax barrier method. FIG. 7A is a diagram illustrating the principle regarding stereoscopic viewing. On the other hand, FIG.7 (b) is a figure which shows the screen displayed by a parallax barrier system. As shown in FIG. 7A, an image in which the left-eye image and the right-eye image shown in FIG. 7B are alternately arranged every other pixel in the horizontal direction is displayed on the image display panel 701. A parallax barrier 702 is displayed on the front surface of the image display panel 701. The parallax barrier 702 has slits that are narrower than the interval between pixels of the same viewpoint. At this time, the image displayed on the image display panel 701 in FIG. 7A is blocked by the barrier 702, and only the left eye image portion from the left eye 703 and only the right eye image portion from the right eye 704 are displayed. Observation will be possible, and stereoscopic viewing will be possible.

  For this reason, the display control unit 205 refers to the 3D display control information for the image of FIG. 4A that is the input image in the present embodiment in the case of stereoscopic display in accordance with the stereoscopic display method of the image display unit 208. By referring to the number of viewpoints and the format of the stereoscopic image, the image generation unit 206 is notified of image generation information so as to generate a converted image so that the left eye image and the right eye image are arranged in every other pixel row in the horizontal direction. To do.

  Next, the playback mode will be described. FIG. 8 is a diagram showing a list of the relationship between the playback mode, which is instruction information from the user, and the display method. The playback modes are “normal playback”, “reverse playback”, “fast playback”, and “fast forward playback”. , “Fast reverse playback”, “Slow playback”, or “Frame advance playback” is selected. The display method is switched according to the playback mode. Note that the screen switches at high speed during fast forward playback and fast reverse playback. At this time, since the parallax varies depending on the scene, if a high-speed and large-parallax image is displayed in 3D, a large burden is given to the user, and the high-speed makes it impossible to perform accurate stereoscopic viewing. Since it is difficult to determine whether it is a scene, 2D display is performed during fast forward playback and fast reverse playback. There are various 2D display methods. As an example, parallax can be achieved by selecting either the left eye image or the right eye image so that the same image enters the left and right eyes, and presenting the selected image to both the left and right eyes. 2D display becomes possible. Alternatively, only the left eye image may be presented and the right eye image may not be presented.

  As described above, the display control unit 205 sends image generation information necessary for 3D display to the image generation unit 206 in accordance with the instruction information from the user, 3D identification information, and 3D display control information.

  In addition, the display control unit 205 sends the parallax information of the 3D display control information to the parallax information presentation unit 207. Also, the instruction information from the user is sent to the parallax information presenting means 207 every time there is an input. When the playback mode is fast forward or fast reverse in the instruction information from the user, the parallax information presentation unit 207 generates a parallax information presentation image based on the parallax information and sends it to the image generation unit 206.

  The image generation unit 206 generates an image to be displayed on the image display unit 208 based on the 3D image data sent from the decoding unit 204 and the image generation information sent from the display control unit 205. Furthermore, when a parallax information presentation image is sent from the parallax information presentation unit 207, an image in which the parallax information presentation image is superimposed on the generated image is generated and sent to the image display unit 208. The image display unit 208 Display an image. At this time, real-time parallax information in the displayed image can be presented to the user by synchronizing the decoded image data and the parallax information presentation image.

  An example of the parallax information presentation image is shown in FIG. FIG. 9A is a diagram when the 3D image from the image generation unit 206 and the parallax information presentation bar 900 as the parallax information presentation image sent from the parallax information presentation unit 207 are displayed on the image display unit 208. FIG. 9B is a detailed view of the disparity information presentation bar 900 shown in FIG. Since the playback mode is fast forward or fast reverse, the 3D image is displayed in 2D. Similarly, the parallax information presentation bar 900 is also presented in 2D display, that is, parallax is zero. In the parallax information presentation bar 900, the horizontal position indicates the position of the moving image data. In the disparity information presentation bar 900, the leftmost portion 901 corresponds to the section 1 in FIG. 6 and represents the first 0 to 5 minutes of the moving image data. , 5 to 10 minutes of section 2, and 902 on the right side corresponds to 10 to 15 minutes of section 3. It is preferable that the disparity information presentation bar 900 changes the color displayed according to the magnitude of the disparity.

  Here, when the parallax is 0, the signal value of the parallax information presentation bar at that position is displayed in white with red R = 255, green G = 255, blue B = 255, and when the parallax is 1 or more and less than 25, R = 255. G = 200, B = 255, when 25 or more and less than 50, R = 255, G = 100, B = 255, and when 50 or more and less than 75, R = 255, G = 0, B = 200 When 75 or more and less than 100, R = 255, G = 0, B = 200, and when 100 or more, R = 255, G = 0, B = 0 and red display. Assuming that the disparity information presentation bar 900 indicates the maximum disparity d1 in the pop-out direction, referring to the disparity information shown in FIG. , G = 0, B = 0, and in section 2 from parallax 80, R = 255, G = 0, B = 200, and sections 3 and 4 have parallax of 64 and 50, so R = 255, G = 0, Since B = 200 and section 6 has a parallax of 40, it is displayed in colors of R = 255, G = 100, and B = 255.

  FIG. 9B is a detailed view of the disparity information presentation bar 900 at this time, where the area 901 is R = 255, G = 0, B = 0, the numeral 902 is R = 255, G = 0, B = 200, reference numeral 903 indicates colors of R = 255, G = 0, B = 200, and reference numeral 904 indicates colors of R = 255, G = 100, and B = 255. By displaying such a parallax information presentation bar 900, it is possible to confirm at a glance which area of the 3D image data has a large amount of pop-up.

  The maximum parallax may be maintained in all the frames that make up the image data, but in this case, in scenes where the parallax changes frequently, the color changes finely, so the colors look mixed and how much the actual parallax is It may be difficult to determine whether or not the operation is selected, and when selecting a portion with a small error parallax, the user may select a large portion. There is an advantage that it is easy to prevent a bad influence on the living body by displaying from a portion with a large amount of popping out by wrong selection.

  In addition, when the information is held for all frames, the amount of information increases, but as shown here, it is possible to reduce the data amount of the parallax information by holding the maximum parallax in a certain time unit. It becomes possible. In particular, when there is a limit to the amount of information that can be transmitted through a so-called transmission path such as broadcast or streaming, the image quality is improved by reducing this parallax information and increasing the amount of image data to be transmitted by the reduced amount. There is an advantage that the number of frames can be increased.

  In addition, the amount of reception buffer allocated to one frame can be reduced on the receiving device side by the amount that the disparity information can be reduced as described above, and the processing amount when reproducing one frame is reduced. It is also possible to do.

  The moving image data can be received and reproduced by ignoring the parallax information as invalid data even when a conventional receiving apparatus that is not conscious of the parallax information being transmitted is used. There are advantages. In the conventional method, if there is too much disparity information to be transmitted, there is a possibility that the reception buffer of the conventional receiving apparatus will be used up, resulting in data loss and display delays. On the other hand, as described above, it is possible to suppress the occurrence of such a problem by obtaining parallax information in a certain section of moving image data and transmitting it.

  Regarding the magnitude of the parallax, by changing the color for each predetermined range, the approximate size of the parallax can be understood at a glance than when the color is changed finely.

  In FIG. 9B, reference numeral 905 denotes a reproduction position instruction pointer. With this reproduction position instruction pointer 905, the current reproduction position can be displayed. When 2D display is performed in fast forward playback or rewind playback, it is possible to check how much parallax the current position has by referring to the playback position indication pointer 905 and the parallax information presentation bar 900. At the same time, it is possible to confirm how much the amount of pop-up will be in the case of 3D display. When normal playback is started from a part with a large parallax, it depends on the viewing position and screen size, but the amount of popping out is too strong and may adversely affect the living body such as the eyes. Since the parallax at the position to be reproduced can be confirmed at a glance, the user can take measures such as placing a distance from the display means in advance or starting display from a position with a small parallax. There is an advantage that adverse effects on the living body due to the display can be prevented. Also, if you only want to watch a scene with a large pop-up amount, you can determine at a glance which parallax in which scene is large, so the user can select the desired scene immediately and select a scene with the desired pop-out amount. Can be selectively viewed.

  In this embodiment, an example has been described in which data having separate 3D identification information and 3D display control information is used. However, 3D identification information and 3D display control information are held together. Alternatively, it may be held as a hierarchical structure in the header information. The values such as the unique ID of each object are not limited to those listed here. Further, even when a plurality of objects are not recorded as a single piece of data, each object may be recorded as separate data, and a plurality of objects are recorded in the image data area at once. Also good. As a method for storing objects, data with a small amount of information such as a 3D identification information object and a 3D display control information object is stored in a storage area different from an image data area such as an extended area of an image data area or a subcode area. It may be recorded.

  The stereoscopic display method of the image display unit 208 has been described by taking the parallax barrier method as an example, but other stereoscopic display methods such as a time-division method and a polarization method may be used, and the display device is also limited to devices using liquid crystal elements. However, a time-division projector or the like may be used.

  Further, the parallax value itself shown in FIG. 6 is also an example, and is not limited to this, and takes various values depending on the image.

  In addition, for simplicity, one segment is set to 5 minutes and the maximum parallax is used. However, one segment may be set to a shorter time or a longer time. In addition, it is possible to grasp the distribution of parallax in the moving image data. In the case of DVD playback image data or the like, a format in which the maximum parallax is maintained for each chapter attached to the DVD instead of time may be used. In addition, as shown here, it may be held as one section until a certain amount of change in the maximum parallax is changed instead of a certain section. For example, if the maximum parallax of the first frame is 50 and the threshold is determined to be ± 10, the maximum parallax value is set as one section until the maximum parallax is smaller than 40 or larger than 60. It will be in the form of holding.

  The maximum parallax may be used in either the pop-out direction or the depth direction, and may be switched according to the moving image, such as an image with a strong sense of depth or an image with a strong sense of pop-up, or the user may be able to switch . Here, the parallax in the depth direction is set to a negative value, but the reverse calculation method may be used, and the absolute value of the parallax calculation is taken to be a positive value. It may be held as it is to distinguish between the pop-out direction and the depth direction based on the sign.

  In addition, here, as the disparity information presentation bar as shown in FIG. 9, the bar indicating the position of the entire moving image data and the playback position is displayed in synchronization with the decoded image data. An image showing only parallax may be displayed. In this case, the display control unit 205 shown in FIG. 2 sends the disparity information synchronized with the image data to be presented among the 3D display control information from the 3D display control information analysis unit 203 to the disparity information presentation unit 207, and the disparity information The presenting unit 207 generates a red circular image as a parallax information presentation image when the maximum parallax of the segment according to the image data to be presented is large, and a blue circle when it is small, and the image generation unit 206 displays the parallax information presentation image as an image. The parallax information presentation image is generated so as to be superimposed on the data and displayed on the right end of the screen. Of course, the position, shape, color, etc. of the parallax information presentation image are not limited to those exemplified here.

  In addition, the parallax information presentation bar illustrated in FIG. 9 has been described as being presented during fast-forward playback and fast-rewind playback. However, in accordance with an instruction from the user, normal parallax information display bar or normal fast-forward without image playback is performed. It may also be displayed in modes other than those shown here, such as during fast-rewinding. The display method in each playback mode in FIG. 8 may be arbitrarily set by the user as display settings.

  In addition, when changing the color of the parallax information presentation bar, the color is changed by dividing the parallax into 25 units. However, the color is not limited to this value, and the color may be changed one by one according to the parallax. , May be separated by a larger parallax. The color of the parallax information presentation bar may not be a color based on red, but may be a blue color or another color, and the transparency may be changed to change the brightness. Here, the maximum parallax is displayed according to the parallax in the pop-out direction, but the parallax d2 in the depth direction may be displayed, or both may be displayed simultaneously.

  In addition, it has been described that the moving image data may be any of 3D image data, 2D image data, and 2D3D mixed data. However, in the 2D portion when 2D or 2D3D is mixed, the moving image data is displayed in a color corresponding to parallax 0 or in 2D. Set a separate color to indicate that it is present. When the 2D color is set separately, only the 2D portion and the 3D portion can be viewed.

  Next, a stereoscopic image processing technique according to the second embodiment of the present invention will be described with reference to the drawings. FIG. 10 is a functional block diagram illustrating a configuration example of the stereoscopic image processing apparatus according to the present embodiment. In FIG. 10, the same reference numerals are given to the functional blocks similar to those in FIG.

  As illustrated in FIG. 10, the stereoscopic image processing apparatus 1000 includes a data analysis unit 1001 that demultiplexes moving image data including a parallax map and image data, a parallax map analysis unit 1002, and encoded image data. Decoding means 204 for decoding the image, display control means 1003 for controlling the display method and parallax information presentation means for the 3D image, image generation means 206 for generating an image in accordance with information relating to the display method from the display control means 1003, and a moving image It includes a parallax information presentation unit 207 that displays a parallax information presentation image as a graphical user interface (GUI) indicating data parallax, and an image display unit 1004 that displays a stereoscopic image.

  The input 3D image data includes a 2D image and a parallax map. Here, assuming that the number of viewpoints is 4 as a multi-viewpoint image, there are three disparity maps that hold information on the disparity of each viewpoint image with respect to the reference image, in addition to the 2D image that is the reference image. The parallax map is an image of another viewpoint represented by parallax from the reference image to the corresponding point from the different viewpoint image. There are various methods for creating a parallax map, such as a method of detecting by stereo matching from a plurality of viewpoint images. Here, one example of a method of detecting parallax information by stereo matching is given as an example.

  The parallax is calculated by comparing the image for the left eye and the image for the right eye and obtaining the corresponding point of the subject. However, when the input image is viewed as a two-dimensional array having each pixel value as a value, and the corresponding relationship between the points in the same row is obtained, the result becomes very unstable only by comparing the pixels at that point. End up. Therefore, the area correlation method is used to compare the surfaces around the pixel of interest, and the difference between the left and right images is calculated for each point on the surface, and the smallest sum is taken as the corresponding point.

  As shown in FIG. 11, corresponding points are searched for in units of blocks of a certain size. A comparison is made based on the difference in RGB (red, green, blue) components of each pixel. 11A is an image for the left eye, and FIG. 11B is an image for the right eye. In each image, the x-th pixel in the horizontal direction and the y-th pixel in the vertical direction are represented by L (x, y ), R (x, y). Since they match in the vertical direction, the parallax is calculated by comparing only in the horizontal direction. The parallax is represented by d, and the difference between the RGB components is compared for the pixel values of L (x, y) and R (x-d, y). This is performed for each block. For example, if the block has a 4 × 4 pixel configuration, the left and right 16 pixel values are compared, their differences and their sum are obtained, and the similarity of each block is determined by the sum of these differences. to see. When the sum of the differences becomes the smallest, each block corresponds to each other, and d at that time is parallax.

  In FIG. 11, when 1101L in FIG. 11A and 1102L, 1102m, and 1102n blocks in FIG. 11B are respectively calculated as differences in RGB components as described above, the difference becomes the smallest at 1102m. It can be seen that this is a corresponding block, and the parallax d can be calculated. By dividing the entire image into blocks and finding corresponding points, the parallax of the entire image can be calculated in units of blocks.

  The data analysis unit 1001 separates the parallax map and the 2D image. The parallax map is sent to the parallax map analysis unit 1002, and the 2D image is sent to the decoding unit 204.

  The parallax map analysis unit 1002 obtains the maximum parallax from the parallax map. Since there is a parallax map for each image constituting the moving image data, this is analyzed to obtain the maximum parallax in each frame, and the parallax map and the maximum parallax are sent to the display control means 1003. Here, as the maximum parallax in each frame, one reference image and one representative viewpoint image are determined, and the maximum parallax between the two images is calculated as the maximum parallax of the frame.

  The display control unit 1003 sends the transmitted parallax information to the parallax information presenting unit 207, and based on the parallax map, displays image generation information for generating a display image according to the stereoscopic system of the image display unit 1004. The data is sent to the generation means 206. Here, description will be made assuming that the stereoscopic method of the image display unit 1004 is a lenticular method. The lenticular method will be described with reference to the drawings.

  FIG. 12 is a diagram illustrating the principle of the lenticular method. Here, it is assumed that a multi-view system is used and the number of viewpoints is four. FIG. 12 is a diagram of the display device viewed from directly above, and the direction of the arrow in the figure is the horizontal direction when the display device is viewed from the front. An image viewed from four viewpoints 1205-1 to 1205-4 is arranged in order in the horizontal direction on the image display panel 1201 of the display device. Each numerical value from 1 to 4 attached to the image display panel 1201 in FIG. 12 is an image of the viewpoint 1 (1205-1) in the 1 portion and the viewpoint 2 (1205-2) in the 2 portion. As shown, the images corresponding to the viewpoints 1 to 4 are displayed. Then, a lenticular lens 1202 is disposed in front of the lens so that light from a plurality of pixels enters one convex portion of the lenticular lens 1202 as shown in the figure. As a result, light from each viewpoint image gathers at a position corresponding to each viewpoint image, and stereoscopic viewing is possible even when viewed from a plurality of positions.

  If the image display means 1004 has four viewpoints in the lenticular method, the display control means 1003 sends the information of the parallax map to the image generation means 206, and the 2D image of one viewpoint sent from the decoding means 204 and Based on the information of the parallax map, the image generating unit 206 generates images of a plurality of viewpoints. Further, it is matched with the lenticular method to generate an image in which the viewpoint images are arranged in the horizontal direction.

  The disparity information presenting unit 207 generates a disparity information presenting bar that is a disparity information presenting image based on the disparity information transmitted from the display control unit 1003, as in the first embodiment, and the image generating unit 206. Send to. In the first embodiment, the parallax information presentation bar is set to parallax 0. However, the parallax information presentation bar may be displayed in 3D with parallax. Suppose you want to display it.

  The image generation unit 206 generates a stereoscopic image as described above according to the image generation information from the display control unit 1003, and when the parallax information presentation bar is sent from the parallax information presentation unit 207, the image superimposed thereon is displayed. The image is generated and sent to the image display unit 1004, and the image display unit 1004 displays the image.

  At this time, as described above, the parallax information presentation bar is displayed when the playback mode is “fast-forward playback” or the like, but the user determines the position where normal playback starts again while checking the parallax information presentation bar. Even then, in some cases, reproduction may start from a portion with a large parallax due to an operation error or the like. In order to prevent this, if a certain parallax of a certain level or more is set in advance as an allowable parallax and playback is selected from a position where the parallax is larger than the set allowable parallax, the parallax information before and after that is selected. A search is performed to find a position within the closest allowable parallax, and reproduction is performed from that position. For example, as shown in FIG. 13, it is assumed that the maximum parallax d1 in the pop-out direction every 10 seconds is held as parallax information.

  In FIG. 13, the column of 0:03:10 indicates the maximum parallax d1 at a position from the position of 3 minutes to 3 minutes and 10 seconds from the start of the moving image. Assuming that the allowable parallax is set to 100 in advance, and playback from the position 3 minutes 45 seconds from the start is instructed by the user by fast-forwarding, refer to the part of 0:03:50 from FIG. The parallax is 122, which exceeds the allowable parallax. Before and after this, when searching for the portion within the allowable parallax, the closest is the portion of 0:04:00 after 3 minutes 50 seconds, and if playback is started from this portion, the user suddenly plays back from the image with large parallax It is possible to prevent the user from starting, and it is possible to prevent the user from being adversely affected and causing a bad mood.

  Further, when the above method is used, it is not necessary to perform image editing, so that the display is not different from the intention of the image producer by editing, and the procedure is simple. As described above, when the display control unit 1003 receives instruction information from the user and performs playback, the display control unit 1003 determines the playback position according to the parallax information as described above, and generates an image to be played back. Image generation information is sent to the means 206.

  Here, the case where the allowable parallax is given in advance has been described as an example, but the user may be able to freely set by his / her own judgment. Thereby, in addition to being able to set according to each environment such as the screen size of the display device, it is possible to cope with the situation by setting a small allowable parallax in advance when a child or the like sees. In addition, parallax information when a user views a stereoscopic image is stored in a storage device, and the parallax information is analyzed to determine how much parallax the user sees, and depending on the frequency, A function of determining the parallax to be automatically set, such as determining a high allowable parallax, may be provided.

  Further, in the stereo matching in FIG. 11, the differences between the RGB components are compared, but different weights may be applied to the differences between R, G, and B. Furthermore, RGB components may be converted into YUV components and the like, and different weights may be applied to the differences between Y, U, and V. For example, the sum of differences may be obtained using only the luminance component. Here, the block has a 4 × 4 pixel configuration. However, the block may have any number of pixel configurations as long as the block includes at least one pixel in both the horizontal and vertical directions. The stereo matching method described here is an example, and other means may be used.

  In the above description, a plurality of parallax maps corresponding to multiple viewpoints are held, but a depth map that holds a distance in a three-dimensional space is held instead of the parallax map, An intermediate image may be generated from the viewpoint position. As a method of generating a depth map, there is a method of using an infrared camera and an infrared LED to obtain a distance from the time until the light from the LED returns and the speed of light. When the depth map is used, the distance corresponding to the direction perpendicular to the display screen, that is, the pop-out direction and the depth direction is known, and the magnitude and distance of the parallax correspond to each other, so the distance may be used as the parallax information. . From the information of the shooting position and the depth map, the pop-out is stronger as the distance is closer to the shooting position in the direction perpendicular to the display screen, and the distance is further away as the distance is longer. Therefore, by using the closest point or the farthest point farthest from the shooting position as the maximum parallax, it is possible to handle the parallax similarly to the case of using parallax.

  Further, the maximum parallax in the case of multiple viewpoints may be calculated as the parallax between the images of two representative viewpoints, as described here, or the parallax may be calculated for each image of adjacent viewpoints, and the maximum May be the maximum parallax.

  Here, the parallax information presentation bar, which is a parallax information presentation image, is displayed using the maximum parallax between the reference image and another representative viewpoint image, but the parallax is calculated for each adjacent viewpoint image. In this case, with respect to the parallax information presentation bar, a plurality of parallax information presentation bars corresponding to each viewpoint are prepared, and the maximum parallax between the two parallax images observed from the viewpoint position corresponding to each viewpoint position is set as the parallax information. You may make it display with a presentation bar. In the case of multiple viewpoints, the viewpoint image to be viewed differs depending on the observation position, so the maximum parallax also varies depending on the observation position, but all parallaxes are calculated for each adjacent viewpoint image and displayed on each parallax information presentation bar By doing so, an appropriate parallax distribution corresponding to the observation position can be confirmed.

  In addition, here, the 3D image is composed of a 2D image and a parallax map, and stereo matching is cited as a method of creating a parallax map at the time of creating a 3D image of such a format. If the 3D image is in a format that holds a plurality of images as they are, a moving image in a format that holds a multi-viewpoint image by detecting parallax information on the stereoscopic image processing apparatus side by stereo matching as described here It is also possible to deal with image data.

  Further, here, as in the first embodiment, the example in which only the magnitude of the maximum parallax is simply determined and the parallax information presentation bar is changed is shown, but many scenes in which the maximum parallax changes rapidly are included. If so, it may be a different color to call attention. Specifically, the maximum parallax is detected from the parallax map for each frame. First, an amount of parallax that is a threshold for determining that the change is large is defined. This amount of parallax is ds, and now ds = 100. The maximum parallax d1 (t) in the pop-out direction and the maximum parallax d2 (t) in the depth direction in a certain frame are compared, and the larger one is held. Next, d1 (t + 1) and d2 (t + 1) in the next frame are compared, and the larger one is held in the same manner. Assume that d1 (t) = 60 is held in the first frame and d2 (t + 1) = 100 is held in the next frame. The values held between two frames are compared, and when the two values have different directions, such as the pop-out direction and the depth direction, and the sum of the values exceeds the defined parallax amount ds, the parallax is sharp Judge that it will change. Here, since d1 (t) + d2 (t) = 160> ds = 100, it is determined that the change is abrupt. Here, the sum of d1 and d2 is defined because each value is defined to be a positive value.

  Whether there are many scenes where the parallax changes abruptly in this way is determined how many times there are such scenes per minute, for example, and a threshold of a predetermined number of times, for example, 3 times if 3 times. If it exceeds, it is determined that the section is a scene with a large change in parallax. When the change in parallax is large, the physiological burden on the user is large, and thus the user's attention can be urged.

  In addition, here, the display unit is included in the stereoscopic image processing apparatus, and the expression is integrated. However, integration with the display apparatus is not an essential requirement, and may be realized as an external device. Needless to say.

  As described above, according to the stereoscopic image processing device according to the embodiment of the present invention, the stereoscopic image processing apparatus includes disparity information presenting means for displaying the disparity information presenting image indicating the disparity information of the three-dimensional image data. It is characterized by making it possible to visually recognize the magnitude of parallax at each position of image data, and it is easy to determine which part of a series of moving images is composed of 3D images and where a scene with large parallax is located Can be judged.

  In the parallax information presentation image, the position in the length direction indicates each position of the moving image data, and the color or brightness is set to the position constituted by the two-dimensional image data at the position constituted by the three-dimensional image data. By making it different, it is possible to explicitly present a change in parallax depending on a difference in color or brightness.

  Further, the moving image data holds disparity information regarding the three-dimensional image data included in the moving image data as metadata, and in the stereoscopic moving image data in a format including the disparity information in the metadata, the disparity of the metadata By analyzing the information, the parallax information presentation image can be changed and the parallax information can be presented.

  The 3D image data includes 2D image data and a parallax map, and the parallax information of the 3D data is detected from the parallax map. By analyzing the parallax at the position of each moving image from the map and changing the parallax information presentation image, the temporal disparity distribution of the moving image data can be made visible.

  The three-dimensional image data includes two-dimensional image data and a depth map. By using the distance information of the depth map as the parallax information, the three-dimensional image data has a depth map in a format that holds the depth map. The disparity information is treated as disparity information, and the disparity information presentation image is changed, whereby the temporal disparity distribution of the moving image data can be visually recognized.

  The three-dimensional image data includes a plurality of images corresponding to a plurality of viewpoints, and the parallax information of the three-dimensional data is obtained by detecting parallax between the plurality of images corresponding to the plurality of viewpoints. In a stereoscopic moving image data format that holds images, parallax is detected from the difference between viewpoint images, and the disparity information presentation image is changed, whereby the temporal disparity distribution of moving image data can be visually recognized.

  In addition, since the parallax information of the 3D image data is the maximum parallax amount in the 3D image, it is possible to confirm in advance whether or not the parallax at each position of the moving image to be viewed is large. In order to prevent the influence on the living body, it is possible to take a response such as viewing from a portion with less parallax.

  Further, the maximum parallax in the three-dimensional image is divided into a plurality of sections according to the magnitude of the parallax, and at least one of the color or the brightness of the parallax information presentation image is different depending on each section. This makes it possible to see the approximate size of the parallax at a glance compared to the case where the color is changed finely, and the visibility is improved.

  The maximum parallax in the three-dimensional image is finely colored in a scene where the parallax frequently changes by dividing the moving image data into fixed units, detecting the maximum parallax for each unit, and expressing it in the parallax information presentation image. It is difficult to discriminate the magnitude of parallax due to changes in brightness, but it becomes easier for the user to judge by handling every fixed unit, and unintentionally displaying from images with large parallax due to incorrect operations etc. will affect the living body Giving can be prevented.

  In addition, a playback position instruction pointer that is displayed superimposed on the parallax information presentation image is further provided, and the current playback position can be visually recognized by the playback position instruction pointer, so that the observer can perform fast-forwarding or the like in the middle of the moving image. It is possible to determine the degree of parallax at the playback position when viewing from the beginning.

  Further, the apparatus further includes an allowable parallax storage unit that holds the user's allowable parallax. When the moving image data is reproduced from the middle, if the parallax at the reproduction position is larger than the allowable parallax, the parallax is allowed from the front and the back of the moving image data. By searching for a position that falls within the parallax and playing from the searched position, even if you try to play from a part with a large parallax by mistake without editing the image itself, play from the part within the allowable parallax. Therefore, it is possible to prevent the user from having an adverse effect such as a worsening mood.

  The present invention is not construed as being limited to the embodiments described above, and various modifications can be made within the scope of the matters described in the claims, and different embodiments can be applied. Embodiments obtained by appropriately combining the technical means disclosed in the above are also included in the technical scope of the present invention.

  The present invention can be used in a stereoscopic image processing apparatus.

It is a figure which shows an example of the moving image data by the 1st Embodiment of this invention. It is a block diagram which shows the stereo image processing apparatus by this Embodiment. It is a figure which shows the detail of each object in moving image data. It is a figure which shows the example of the 3D image in moving image data. It is a figure explaining the maximum parallax in a 3D image. It is a figure which shows an example regarding distribution of the largest parallax in 3D image. It is a figure for demonstrating a parallax barrier system. It is a figure which shows the example of user instruction information. It is a figure which shows an example of a parallax information presentation image. It is a block diagram which shows the stereo image processing apparatus by the 2nd Embodiment of this invention. It is a figure explaining the stereo matching for obtaining parallax information. It is a figure for demonstrating a lenticular system. It is a figure which shows an example regarding distribution of the largest parallax in 3D image. It is a figure which shows the relationship between the convergence and adjustment in a stereoscopic vision.

Explanation of symbols

DESCRIPTION OF SYMBOLS 100 Moving image data 101 Header control information 102 3D identification information 103 3D display control information 104 3D image data 105 Object 106 Object ID
DESCRIPTION OF SYMBOLS 107 Object size 108 Object data 200 Stereoscopic image processing apparatus 201 Data analysis means 202 3D identification information analysis means 203 3D display control information analysis means 204 Decoding means 205 Display control means 206 Image generation means 207 Parallax information presentation means 208 Image display means 208
701 Image display panel 702 Paralux barrier 703 Left eye 704 Right eye 900, 901, 902, 903, 904 Parallax information presentation bar 905 Playback position indication pointer 602 Right eye 603 Left eye 1000 Stereoscopic image processing apparatus 1001 Data analysis means 1002 Parallax map analysis Means 1003 Display control means 1004 Image display means 1101, 1102 Block 1201 Image display panel 1202 Lenticular lens

Claims (14)

  1. A stereoscopic image processing apparatus for reproducing moving image data including three-dimensional image data,
    An allowable parallax storage means for holding the allowable parallax of the user ;
    Means for re-determining the playback position according to the parallax information when the playback position of the moving image data is halfway and the parallax at the playback position is greater than the allowable parallax;
    Three-dimensional image processing apparatus characterized by having a.
  2. A stereoscopic image processing apparatus for reproducing moving image data including three-dimensional image data,
    An allowable parallax storage means for holding the allowable parallax of the user ;
    When the moving image data is reproduced from the middle, if the parallax at the reproduction position is larger than the allowable parallax, the position where the parallax falls within the allowable parallax is searched from before and after the moving image data, and the searched position Means to control to play from
    Three-dimensional image processing apparatus characterized by having a.
  3. further,
    2. The apparatus according to claim 1, further comprising: a disparity information presenting unit configured to generate a disparity information presenting image that allows a user to visually recognize a magnitude of the disparity at each position of the moving image data based on the disparity information of the three-dimensional image data. Or the three-dimensional image processing apparatus of 2.
  4.   The parallax information presentation image is displayed as an image that can be visually identified according to the magnitude of parallax at each position of the moving image data, whereby the magnitude of the parallax can be visually recognized. The stereoscopic image processing apparatus according to claim 3.
  5.   The stereoscopic image processing according to any one of claims 1 to 4, wherein the moving image data holds disparity information regarding three-dimensional image data included in the moving image data as metadata. apparatus.
  6.   6. The three-dimensional image data includes two-dimensional image data and a parallax map, and parallax information of the three-dimensional image data is obtained by detecting from the parallax map. The three-dimensional image processing apparatus according to any one of the above.
  7.   The three-dimensional image data includes a plurality of images corresponding to a plurality of viewpoints, and the parallax information of the three-dimensional data is obtained by detecting a parallax between the plurality of images corresponding to the plurality of viewpoints. The stereoscopic image processing apparatus according to any one of claims 1 to 6.
  8.   8. The three-dimensional image data includes two-dimensional image data and a depth map that holds distance information in a three-dimensional space, and the distance information is used as the parallax information. A stereoscopic image processing apparatus according to claim 1.
  9.   The stereoscopic image processing apparatus according to any one of claims 1 to 8, wherein a maximum parallax in a three-dimensional image is used as the parallax information of the three-dimensional image data.
  10.   The maximum parallax in the three-dimensional image is obtained by dividing the moving image data into fixed units, detecting the maximum parallax for each unit, and determining whether the parallax falls within the allowable parallax. The stereoscopic image processing apparatus according to any one of claims 9 to 9.
  11.   The parallax information presentation image includes a reproduction position instruction pointer displayed superimposed on a displayed moving image, and the reproduction position of the current moving image can be visually recognized by the reproduction position instruction pointer. Item 13. The stereoscopic image processing apparatus according to any one of Items 1 to 10.
  12. A stereoscopic image processing method for reproducing moving image data including three-dimensional image data,
    A step of reproducing the moving image data from the middle,
    Comparing the parallax of the playback position with the user's allowable parallax held in the allowable parallax storage means;
    Parallax of the playback position, wherein when the allowable disparity user greater tolerance parallax held in the storage means comprises the steps of re-determining a reproduction position according to the parallax information
    Three-dimensional image processing method characterized by having a.
  13. A stereoscopic image processing method for reproducing moving image data including three-dimensional image data,
    A step of reproducing the moving image data from the middle,
    Comparing the parallax of the playback position with the user's allowable parallax held in the allowable parallax storage means;
    A step parallax of the playback position, wherein when the allowable disparity user greater tolerance parallax held in the storage means, to find the position where the parallax from the front and rear of the moving image data fits within said allowable parallax,
    Three-dimensional image processing method characterized by having a <br/> and controlling to play from the search position.
  14. A program for causing a computer to execute each step of the stereoscopic image processing method according to claim 12 or 13.
JP2006282543A 2006-10-17 2006-10-17 Stereoscopic image processing device Expired - Fee Related JP4755565B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2006282543A JP4755565B2 (en) 2006-10-17 2006-10-17 Stereoscopic image processing device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2006282543A JP4755565B2 (en) 2006-10-17 2006-10-17 Stereoscopic image processing device

Publications (2)

Publication Number Publication Date
JP2008103820A JP2008103820A (en) 2008-05-01
JP4755565B2 true JP4755565B2 (en) 2011-08-24

Family

ID=39437840

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2006282543A Expired - Fee Related JP4755565B2 (en) 2006-10-17 2006-10-17 Stereoscopic image processing device

Country Status (1)

Country Link
JP (1) JP4755565B2 (en)

Families Citing this family (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101539935B1 (en) * 2008-06-24 2015-07-28 삼성전자주식회사 Method and apparatus for processing 3D video image
JP5792064B2 (en) * 2008-07-25 2015-10-07 コーニンクレッカ フィリップス エヌ ヴェ Subtitle 3D display processing
US8629898B2 (en) * 2008-09-03 2014-01-14 Sony Corporation Stereoscopic video delivery
JP4998422B2 (en) * 2008-09-25 2012-08-15 Kddi株式会社 Image generating apparatus, method, communication system, and program
WO2010058547A1 (en) 2008-11-18 2010-05-27 パナソニック株式会社 Reproduction device, integrated circuit, and reproduction method considering specialized reproduction
JP5510700B2 (en) * 2009-04-03 2014-06-04 ソニー株式会社 Information processing apparatus, information processing method, and program
JP5400467B2 (en) * 2009-05-01 2014-01-29 キヤノン株式会社 Video output device, its control method, and program
CN102144395A (en) * 2009-06-10 2011-08-03 Lg电子株式会社 Stereoscopic image reproduction method in quick search mode and stereoscopic image reproduction apparatus using same
KR20110025123A (en) * 2009-09-02 2011-03-09 삼성전자주식회사 Method and apparatus for multiple-speed reproduction of video image
JP4823349B2 (en) * 2009-11-11 2011-11-24 パナソニック株式会社 3D video decoding apparatus and 3D video decoding method
JP2011114863A (en) * 2009-11-23 2011-06-09 Samsung Electronics Co Ltd Method for providing 3d image, method for converting 3d image, gui providing method, 3d display apparatus using the same, and system for providing 3d image
KR101615695B1 (en) * 2009-11-23 2016-04-27 삼성전자주식회사 GUI providing method related to 3D image and display apparatus usign the same
KR101639310B1 (en) * 2009-11-23 2016-07-14 삼성전자주식회사 GUI providing method related to 3D image, and display apparatus using the same
JPWO2011080878A1 (en) * 2009-12-28 2013-05-09 パナソニック株式会社 Image reproduction device and display device
JP2011142586A (en) * 2010-01-08 2011-07-21 Sony Corp Image processing device, information recording medium, image processing method, and program
JP2011142585A (en) * 2010-01-08 2011-07-21 Sony Corp Image processing device, information recording medium, image processing method, and program
JP5477128B2 (en) * 2010-04-07 2014-04-23 ソニー株式会社 Signal processing apparatus, signal processing method, display apparatus, and program
JP5545036B2 (en) * 2010-05-28 2014-07-09 三菱電機株式会社 Image processing apparatus, image processing method, and image display apparatus
JP5489223B2 (en) * 2010-06-09 2014-05-14 Necカシオモバイルコミュニケーションズ株式会社 Image display apparatus and program
JP5817091B2 (en) * 2010-07-14 2015-11-18 株式会社ニコン Display device, image processing device, and program
JP5610933B2 (en) * 2010-09-03 2014-10-22 キヤノン株式会社 Reproducing apparatus and control method thereof
KR101640404B1 (en) * 2010-09-20 2016-07-18 엘지전자 주식회사 Mobile terminal and operation control method thereof
JP5764895B2 (en) * 2010-09-28 2015-08-19 ソニー株式会社 Electronic device, playback system, playback method, program
JP5723721B2 (en) * 2010-09-28 2015-05-27 富士フイルム株式会社 Stereoscopic image editing apparatus and stereoscopic image editing method
JP5453552B2 (en) 2010-12-24 2014-03-26 富士フイルム株式会社 Imaging apparatus, method and program
EP2659680A4 (en) * 2010-12-29 2014-05-21 Thomson Licensing Method and apparatus for providing mono-vision in multi-view system
WO2012096065A1 (en) * 2011-01-13 2012-07-19 富士フイルム株式会社 Parallax image display device and parallax image display method
WO2012101916A1 (en) * 2011-01-25 2012-08-02 富士フイルム株式会社 Stereoscopic video processor, stereoscopic video processing program and recording medium therefor, stereoscopic imaging device and stereoscopic video processing method
JP5466773B2 (en) * 2011-02-03 2014-04-09 富士フイルム株式会社 Stereoscopic video playback device, stereoscopic video playback program and recording medium thereof, stereoscopic display device, stereoscopic imaging device, and stereoscopic video playback method
CN103339947B (en) * 2011-02-03 2014-07-30 富士胶片株式会社 3D video playing device, 3D video playing program and recording medium for same, 3D display device, 3D imaging device, and 3D video playing method
JP5198615B2 (en) * 2011-03-28 2013-05-15 株式会社東芝 Image processing apparatus and image processing method
JP5341942B2 (en) * 2011-04-05 2013-11-13 住友電気工業株式会社 Video playback device
GB2489930A (en) * 2011-04-08 2012-10-17 Sony Corp Analysis of Three-dimensional Video to Produce a Time-Varying Graphical Representation of Displacements
JP5914838B2 (en) * 2011-05-23 2016-05-11 パナソニックIpマネジメント株式会社 Image display device
JP5166571B2 (en) 2011-05-30 2013-03-21 株式会社東芝 Electronic device, display control method, and display control program
JP2012253643A (en) 2011-06-06 2012-12-20 Sony Corp Image processing apparatus and method, and program
JP5720475B2 (en) * 2011-07-29 2015-05-20 株式会社Jvcケンウッド 3D image control apparatus, 3D image control method, and 3D image control program
JP5181083B1 (en) 2012-01-19 2013-04-10 パナソニック株式会社 Stereoscopic image display control device, stereoscopic image display control method, and program
WO2013108298A1 (en) * 2012-01-19 2013-07-25 パナソニック株式会社 Device for controlling stereoscopic image display, method for controlling stereoscopic image display, and program
EP2688304B1 (en) 2012-03-01 2019-09-25 Sony Corporation Transmitter, transmission method and receiver
JP5963483B2 (en) * 2012-03-12 2016-08-03 キヤノン株式会社 Imaging device and imaging device control method
JP6377225B2 (en) * 2012-08-13 2018-08-22 キヤノン株式会社 Image processing apparatus, control method therefor, and program

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003346188A (en) * 2002-05-27 2003-12-05 Canon Inc Stereoscopic imaging system
JP4469159B2 (en) * 2003-11-06 2010-05-26 学校法人早稲田大学 3D image evaluation apparatus and 3D image tuner
JP4305269B2 (en) * 2004-04-30 2009-07-29 ソニー株式会社 Signal processing apparatus and method
JP2006041811A (en) * 2004-07-26 2006-02-09 Kddi Corp Free visual point picture streaming method

Also Published As

Publication number Publication date
JP2008103820A (en) 2008-05-01

Similar Documents

Publication Publication Date Title
US7254265B2 (en) Methods and systems for 2D/3D image conversion and optimization
JP4693900B2 (en) Image processing device
JP5529870B2 (en) 2D / 3D playback mode determination method, 2D / 3D playback mode determination device, and storage medium
CA2680696C (en) Recording medium on which 3d video is recorded, recording medium for recording 3d video, and reproducing device and method for reproducing 3d video
TWI477149B (en) Multi-view display apparatus, methods, system and media
ES2537073T3 (en) Reproduction device, reproduction method and program for stereoscopic reproduction
US8013873B2 (en) Depth perception
JP5547725B2 (en) Method and apparatus for 3D video image processing
JP4174001B2 (en) Stereoscopic image display apparatus, recording method, and transmission method
JP2010062695A (en) Image processing apparatus, image processing method, and program
US9124870B2 (en) Three-dimensional video apparatus and method providing on screen display applied thereto
JPWO2003092304A1 (en) Image data generating device, image data reproducing device, and image data recording medium
JP2013118701A (en) Method and device for stereoscopic video data stream generation for reproduction of partial data section of stereoscopic video, and method and device for reproduction of partial data section of stereoscopic video
KR20110097690A (en) Method and apparatus for providing service of multiview still image, method and apparatus for receiving service of multiview still image
WO2010032399A1 (en) Stereoscopic video reproduction device and stereoscopic video reproduction device
JP5551654B2 (en) Semiconductor integrated circuit
US10021377B2 (en) Combining 3D video and auxiliary data that is provided when not reveived
KR20110140088A (en) Stereoscopic display device and control method of stereoscopic display device
US7508485B2 (en) System and method for controlling 3D viewing spectacles
US20100091012A1 (en) 3 menu display
ES2578022T3 (en) Combination of 3D image data and graphics
US20020191841A1 (en) Image processing method and apparatus
KR20110106367A (en) Image based 3d video format
JP2012227924A (en) Image analysis apparatus, image analysis method and program
US8885020B2 (en) Video reproduction apparatus and video reproduction method

Legal Events

Date Code Title Description
A621 Written request for application examination

Free format text: JAPANESE INTERMEDIATE CODE: A621

Effective date: 20090218

A977 Report on retrieval

Free format text: JAPANESE INTERMEDIATE CODE: A971007

Effective date: 20101227

A131 Notification of reasons for refusal

Free format text: JAPANESE INTERMEDIATE CODE: A131

Effective date: 20110111

A521 Written amendment

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20110307

A131 Notification of reasons for refusal

Free format text: JAPANESE INTERMEDIATE CODE: A131

Effective date: 20110329

A521 Written amendment

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20110509

TRDD Decision of grant or rejection written
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 20110524

A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20110527

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20140603

Year of fee payment: 3

R150 Certificate of patent or registration of utility model

Free format text: JAPANESE INTERMEDIATE CODE: R150

LAPS Cancellation because of no payment of annual fees