WO2010151044A2 - 3차원 컨텐츠를 출력하는 디스플레이 기기의 영상 처리 방법 및 그 방법을 채용한 디스플레이 기기 - Google Patents
3차원 컨텐츠를 출력하는 디스플레이 기기의 영상 처리 방법 및 그 방법을 채용한 디스플레이 기기 Download PDFInfo
- Publication number
- WO2010151044A2 WO2010151044A2 PCT/KR2010/004073 KR2010004073W WO2010151044A2 WO 2010151044 A2 WO2010151044 A2 WO 2010151044A2 KR 2010004073 W KR2010004073 W KR 2010004073W WO 2010151044 A2 WO2010151044 A2 WO 2010151044A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image data
- enlargement
- display device
- enlarged
- reduction
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 91
- 238000003672 processing method Methods 0.000 title claims abstract description 22
- 230000009467 reduction Effects 0.000 claims description 96
- 230000008859 change Effects 0.000 claims description 37
- 230000008569 process Effects 0.000 abstract description 29
- 239000011521 glass Substances 0.000 description 38
- 238000010586 diagram Methods 0.000 description 28
- 239000004973 liquid crystal related substance Substances 0.000 description 13
- 230000011664 signaling Effects 0.000 description 13
- 230000000694 effects Effects 0.000 description 8
- 230000005540 biological transmission Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 3
- 230000004888 barrier function Effects 0.000 description 2
- 210000004556 brain Anatomy 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000001360 synchronised effect Effects 0.000 description 2
- 230000006866 deterioration Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000000737 periodic effect Effects 0.000 description 1
- 230000010287 polarization Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/139—Format conversion, e.g. of frame-rate or size
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/398—Synchronisation thereof; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N2013/40—Privacy aspects, i.e. devices showing different images to different viewers, the images not being viewpoints of the same scene
- H04N2013/405—Privacy aspects, i.e. devices showing different images to different viewers, the images not being viewpoints of the same scene the images being stereoscopic or three dimensional
Definitions
- the present invention relates to an image processing method of a display device for outputting three-dimensional content and a display device employing the method. More specifically, the left image data and the right image data of the three-dimensional image data in the display device for outputting the three-dimensional content.
- the present invention relates to an image processing method of a display device that performs image processing of each image data and outputs a 3D format, and a display device employing the method.
- the image processing method of the display device to provide a high-quality image data by processing the three-dimensional image data, the user can conveniently view and use the three-dimensional image data and There is a need for developing a display device employing the method.
- the present invention solves the problems of the prior art as described above, the display device for providing high-quality three-dimensional image data by processing the three-dimensional image data, the user can conveniently view and use the three-dimensional image data It is an object of the present invention to provide a video processing method and a display device employing the method.
- the left image data and the right image data of the 3D image data are respectively enlarged or reduced in proportion to an enlargement command or a reduction command for the 3D image data. Zooming in or out; And outputting the enlarged or reduced processed left image data and right image data in a 3D format.
- the image processing method of the 3D display device comprises the steps of image processing each of the left image data and the right image data of the three-dimensional image data; And outputting the image processed left image data and right image data in a 3D format at a predetermined depth value.
- the 3D display device is to display the left image data and the right image data of the three-dimensional image data at an enlargement ratio or reduction ratio corresponding to the magnification command or reduction command for the three-dimensional image data, respectively.
- an output formatter configured to output the enlarged or reduced processed left image data and right image data in a 3D format.
- the 3D display device includes a scaler for image processing each of the left image data and the right image data of the three-dimensional image data; And an output formatter configured to output the image processed left image data and right image data in a 3D format at a predetermined depth value.
- the present invention has the effect of allowing the user to select the depth value with the enlargement or reduction of the 3D image data so that the user can conveniently use the 3D image data.
- the present invention adjusts the depth value of the 3D image data together to make the image processed region appear more prominently, thereby allowing the user to display the 3D image data. It has the effect of making it more convenient to use.
- the present invention has the effect of providing a more dynamic enlargement reduction function by determining the area to include the enlargement or reduction and the enlargement or reduction ratio in accordance with the user's position change.
- the present invention performs the overscan of the left image data and the right image data, respectively, and then outputs the 3D format, so that the alignment of the left image data and the right image data is performed accurately, and the 3D image data is overscanned and then 3D.
- the present invention performs the overscan of the left image data and the right image data, respectively, and then outputs the 3D format, so that the alignment of the left image data and the right image data is performed accurately, and the 3D image data is overscanned and then 3D.
- FIG. 1 is a view showing an embodiment of a display device for providing three-dimensional content according to the present invention.
- FIG. 2 is a diagram for explaining perspective according to an interval or parallax (hereinafter, referred to as interval) between left image data and right image data;
- interval an interval or parallax
- FIG. 3 is a diagram for explaining an example of a method for implementing a 3D image in a display device in connection with the present invention
- FIG. 4 is a diagram illustrating formats of a 3D video signal including the left image data and the right image data as described above.
- FIG. 5 is a flowchart illustrating a process of image processing 3D image data in a display device according to an embodiment of the present invention.
- FIG. 6 is a flowchart illustrating a process of enlarging or reducing the three-dimensional image data according to an embodiment of the present invention.
- FIG. 7 illustrates a first user interface for receiving an enlargement or reduction command and a second user interface for receiving a depth adjustment command according to an embodiment of the present invention.
- FIG. 8 shows an embodiment of storage means for storing a depth value corresponding to an enlargement ratio in accordance with an embodiment of the present invention.
- FIG 9 illustrates a process of enlarging or reducing three-dimensional image data according to an embodiment of the present invention.
- FIG. 10 is a diagram illustrating an embodiment of three-dimensional image data which has been enlarged or reduced in accordance with one embodiment of the present invention.
- FIG. 11 is a diagram illustrating a process of enlarging or reducing three-dimensional image data according to a user's position change according to another embodiment of the present invention.
- FIG. 12 is a diagram illustrating an embodiment of determining a user position change value according to the present invention.
- FIG. 13 is a diagram illustrating an embodiment in which an enlarged or reduced area and a depth value are determined according to a position change value of a user in a display device according to the present invention
- FIG. 14 is a diagram for one embodiment of storing an enlargement or reduction ratio and a depth value corresponding to a position change value of a user according to one embodiment of the present invention.
- 15 is a flowchart illustrating a process of performing image processing of 3D image data in a display device according to another embodiment of the present invention.
- 16 is a diagram illustrating a process of overscanning 3D image data according to an embodiment of the present invention.
- FIG. 17 illustrates an embodiment of outputting overscanned left image data and right image data in a 3D format according to the present invention
- FIG. 18 is a diagram illustrating a result of overscanning each of left image data and right image data and outputting them in 3D format according to one embodiment of the present invention.
- FIG. 19 is a block diagram illustrating a configuration of a display device according to an embodiment of the present invention.
- FIG. 20 is a block diagram illustrating a display device according to another embodiment of the present invention.
- 21 is a view showing the configuration of shutter glasses according to an embodiment of the present invention.
- the present invention relates to an image processing method of a display device for outputting 3D content and a display device employing the method, which will be described below in more detail with reference to the accompanying drawings.
- FIG. 1 is a view showing an embodiment of a display device for providing three-dimensional content according to the present invention.
- the method of wearing glasses is divided into a passive method and an active method.
- the passive method distinguishes between a left image and a right image by using a polarization filter.
- wearing both blue and red sunglasses on both eyes is also passive.
- the active method is a method of dividing left and right eyes using a liquid crystal shutter, and distinguishing left and right images by sequentially covering the left eye (left eye) and right eye (right eye) in time.
- the active method is a method of periodically wearing a timed screen and wearing glasses that have electronic shutters synchronized with the period, and may be referred to as a time split type or shuttered glass method.
- Representatives known as a glasses-free method without wearing glasses include a lenticular method in which a lenticular lens plate in which a cylindrical lens array is vertically arranged in front of the image panel, and a periodic slit in the upper part of the image panel.
- FIG. 1 illustrates an exemplary embodiment of a stereoscopic type among 3D display methods and an active type among the stereoscopic methods.
- the shutter glasses are described as an example of the active type media, the present invention is not limited thereto, and it is apparent that the present invention can be applied to other media as described below.
- a display device configured to output 3D image data on a display unit and to synchronize the 3D image data output when watching the shutter glasses 200.
- a synchronization signal Vsync related to 3D image data is generated and output to an IR emitter (not shown) in the shutter glasses, so that the shutter glasses 200 can watch in synchronization with the display.
- the shutter glasses 200 adjusts the open period of the left-eye or right-eye shutter liquid crystal panel according to a synchronization signal received through an IR emitter (not shown), thereby outputting the three-dimensional image data 300 output from the display unit 100. Can be motivated.
- the display device processes the three-dimensional image data using the principle of the stereoscopic method. That is, when one object is photographed by two cameras of different positions, the left image data and the right image data are generated, and each generated image data is inputted separately so as to be orthogonal to each other in the left and right eyes of the person. It is a principle that the 3D image is generated by combining the image data inputted into the left and right eyes of the brain. In the above, the fact that the image data is arranged to be orthogonal to each other means that each image data does not interfere with each other.
- FIG. 2 is a diagram illustrating perspective according to a gap or parallax (hereinafter, referred to as a gap) between the left image data and the right image data.
- FIG. 2 (a) illustrates an upper position 203 where both of the above data are combined when the distance between the right image data 201 and the left image data 202 is narrow
- FIG. 2 (b) illustrates the right image data ( 211 and the left image data 212, the position of the upper image 213 is described.
- 2 (a) to 2 (b) show the degree of perspective at which different images are formed at different positions according to the interval between the left image data and the right image data in the image signal processing apparatus.
- the upper side draws extension lines R1 and R2 facing one side and the other side of the right image data 201 to the right eye, and the one side and the other side of the left image data 202 to the left eye.
- the extension line R1 for the right image data and the extension line L1 for the left image data intersect each other at a distance d1 from the right eye and the left eye at a distance 203. Bears.
- the image is based on the foregoing description in FIG. 2A, and the extension line R3 for the right image data and the extension line L3 for the left image data have a predetermined distance from the right eye and the left eye. at point 213 crossing each other at d2).
- the distance d1 is greater from the left eye to the right eye than d2. far. That is, the image in Fig. 2 (a) is formed at a distance farther from the left and right eyes than the image in Fig. 3 (b).
- the interval between the right image data 201 and the left image data 202 in FIG. 2A is the interval between the right image data 203 and the left image data 204 in FIG. 2B. Relatively narrow compared with
- the 3D image data may be implemented in 3D by various methods such as giving a tilt to a depth or giving a 3D effect.
- a method of giving depth among the above methods Briefly describe.
- FIG. 3 is a diagram illustrating an example of a method of implementing a 3D image in a display device in relation to the present invention.
- the distance between the left image data 301 and the right image data 302 constituting the three-dimensional image is narrow.
- the left image constituting the three-dimensional image This is the case where the interval between the data 304 and the right image data 305 is wide.
- the 3D image implemented according to the interval of each image data in FIGS. 3 (a) and 3 (b) is based on the principle of FIG. 2 described above, and the 3D image 303 implemented in FIG. 3D image 306 implemented in FIG. 3 (b) is closer to the human eye, that is, the image is more prominent near the human eye. That is, by adjusting the distance between the left image data and the right image data constituting the three-dimensional image, it is possible to give an appropriate depth to the three-dimensional image.
- FIG. 4 is a diagram illustrating formats of a 3D video signal including the left image data and the right image data as described above.
- 3D content or 3D video signals may be classified into various types.
- Left and right image data are generated by photographing one object with two cameras at different positions.
- Side by side format (401) for inputting the generated left image data and the right image data separately so that they are orthogonal to each other in the left eye and the right eye, respectively,
- FIG. 5 is a flowchart illustrating a process of image processing 3D image data in a display device according to an embodiment of the present invention.
- the display device determines the format of 3D image data to be output.
- format information of the 3D image data may be received from the external input source, and the module includes a module for determining the format of the 3D image data inside the display device.
- the format of the 3D image data to be output in the corresponding module may be determined.
- a format of 3D image data may be selected by a user.
- the format of the 3D image data determined according to the embodiment may be determined as one of a side by side format, a checker board format, and a frame sequential format.
- step S502 the display device classifies the left image data and the right image data of the 3D image data according to the format information of the 3D image data determined in step S501.
- the left region may be determined as the left image data
- the right region may be determined as the right image data
- the display apparatus processes the left image data and the right image data of the 3D image data, respectively.
- various processes related to the output of the 3D image data may be applied to the image processing.
- overscan is applied to the 3D image data to be output, the left image data is overscanned and the right image data is processed.
- Can overscan is applied to the 3D image data to be output.
- the display device may enlarge or reduce the left image data, and enlarge or reduce the right image data.
- the display apparatus outputs the image processed left image data and the right image data in a 3D format with a predetermined depth value.
- the depth value at which the left image data and the right image data are output may be a previously stored value, a value determined in an image processing process, or a value input by a user.
- the display device receives a depth adjustment command, and then shifts the image processed left image data and right image data by pixel shifting to the depth adjustment command. It can output with the corresponding depth value.
- FIG. 6 is a flowchart illustrating a process of enlarging or reducing the three-dimensional image data according to an embodiment of the present invention.
- the display device determines whether an enlargement command or a reduction command for 3D image data is received.
- the magnification command or the reduction command for the 3D image data may be input from the user through the first user interface, or may be input through the remote control apparatus.
- the user when the user's location changes, the user may sense the location change and set the sensed location change value as an enlargement or reduction command.
- step S601 when an enlargement command or a reduction command for the 3D image data is received, the display device determines an enlargement ratio or a reduction ratio corresponding to the enlargement command or the reduction command in step S602.
- the display device determines an enlarged or reduced area of the 3D image data.
- the area to be enlarged or reduced in the 3D image data may be designated by the user. If there is no user specification, the preset area may be determined as the enlarged or reduced area. In addition, according to an embodiment, it may be designated according to a user's location change value.
- step S604 the display device enlarges or reduces each of the enlarged or reduced areas in the left image data and the right image data of the 3D image data at a determined enlargement ratio or reduction ratio.
- step S605 the display device determines whether a depth adjustment command has been received.
- the depth adjustment command for the 3D image data may be input from the user through the second user interface, or may be input through the remote control apparatus.
- a first user interface for receiving an enlargement command or a reduction command for 3D image data and a second user interface for receiving a depth adjustment command for the 3D image data may be output on one display screen.
- the user may select an enlargement or reduction ratio in the first user interface and select a depth value to be output in the second user interface.
- step S607 the display device determines a depth value corresponding to the enlargement ratio or the reduction ratio.
- depth values corresponding to each of the plurality of enlargement ratios or reduction ratios may be preset and stored in storage means included in the display device.
- depth values corresponding to each enlargement ratio or reduction ratio may be set to a constant value, or may be set to different values.
- the depth value at which the enlarged region is output may also be set to a value closer to the user.
- step S608 the display device outputs left image data and right image data that have been enlarged or reduced to the depth value determined in step S607 in a 3D format.
- step S605 If it is determined in step S605 that the depth adjustment command has been received, the display device outputs the enlarged or reduced left image data and the right image data as a depth value corresponding to the depth adjustment command in step S606. do.
- FIG. 7 illustrates a first user interface for receiving an enlargement or reduction command and a second user interface for receiving a depth adjustment command according to an embodiment of the present invention.
- a display device may provide a first user interface 701 that receives an enlargement command or a reduction command for 3D image data and a depth adjustment command for the 3D image data.
- the input second user interface 702 may be output on the display screen.
- only the first interface 701 may be output on the display screen or only the second user interface 702 may be output on the display screen.
- the user may select an enlargement or reduction ratio in the first user interface 701 and select a depth value to be output from the second user interface 702. have.
- the area 703 to be enlarged or reduced in the 3D image data may be designated in various ways.
- the area 703 may be designated by a predetermined pointer through a remote control device, or may be detected by detecting a change in the position of a user to be described later. It is also possible to designate an enlarged or reduced area 703 corresponding to the change.
- a predetermined area for example, a center area of the 3D image
- it may be designated according to a user's location change value.
- the display device When the enlargement or reduction ratio is selected in the first user interface 701, the left image data and the right image data of the 3D image data are enlarged and reduced, respectively, as described above, and the depth adjustment command is executed by the user selecting the depth value. If it is determined that the image is received, the display device outputs the left image data and the right image data, which are enlarged or reduced in accordance with the enlargement or reduction ratio, as a depth value corresponding to the depth adjustment command.
- the present invention has an effect of allowing the user to select the depth value with the enlargement or reduction of the 3D image data so that the user can conveniently use the 3D image data.
- the display device may further output a third user interface 703 that can set the transparency of the 3D image data, along with the reduction, enlargement, and depth adjustment.
- a third user interface 703 can set the transparency of the 3D image data, along with the reduction, enlargement, and depth adjustment.
- the selected transparency may be applied to the left image data and the right image data which are enlarged or reduced.
- FIG. 8 is a view showing an embodiment of a storage means for storing a depth value corresponding to an enlargement ratio according to an embodiment of the present invention.
- a display device may set a depth value corresponding to an enlargement ratio or a reduction ratio.
- the depth value 802 corresponding to each of the plurality of enlargement ratios or reduction ratios 801 may be preset and stored in storage means included in the display device.
- depth values corresponding to each enlargement ratio or reduction ratio 801 may be set to a constant value, or may be set to different values. For example, as the enlargement ratio increases, the depth value at which the enlarged region is output may also be set to a value closer to the user.
- the display device may also store pixel number information 803 in which left image data and right image data should be shifted in order to adjust the depth value.
- the transparency information 804 corresponding to the enlargement or reduction ratio 801 may be stored.
- the display device determines the enlargement ratio or reduction ratio 801 and applies the enlargement ratio or reduction ratio 801 to the left image data and the right image data.
- the 3D image data may be output as the depth value 802 corresponding to the enlargement ratio or the reduction ratio by moving the left image data and the right image data to the pixel shift value corresponding to the determined enlargement ratio or reduction ratio 801. have.
- FIG. 9 is a diagram illustrating a process of enlarging or reducing three-dimensional image data according to an embodiment of the present invention.
- an enlargement process will be described as an example, and a process of reducing may be processed in the following manner.
- the display device may include an enlarged area 901 of the left image data of the 3D image data and an enlarged area of the right image data ( 902 is enlarged at an enlargement ratio.
- the display apparatus pixel shifts the enlarged left image data 903 and the right image data 904.
- the adjusted depth value may be received from the second user interface, and it may be determined according to the enlargement ratio as described above.
- the left image data 903 may be pixel shifted by the pixel d1 in the left direction
- the right image data 904 may be pixel shifted by the pixel d1 in the right direction.
- pixel shifted left image data 905 and pixel shifted right image data 906 are output as three-dimensional image data.
- the display device may output the 3D image data in at least one of line by line, frame sequential, and checker board methods using the determined format information of the 3D image data.
- the format of the 3D image data is converted according to the output method of the display device, and the 3D image data of the converted format is output.
- the 3D image data of line by line, top and bottom, and side by side methods may be converted into a frame sequential method and then output. Can be.
- FIG. 10 is a diagram illustrating an embodiment of three-dimensional image data that has been enlarged or reduced in accordance with one embodiment of the present invention.
- an enlarged or reduced selected area of the 3D image data may be enlarged or reduced, and may be output after being adjusted for depth.
- the area 1001 enlarged and selected from the original three-dimensional image data may be enlarged and each of the corresponding areas of the left image data and the right image data, and may be output as shown by reference numeral 1002.
- the original 3D image data 1001 may also be output as it is before being enlarged or reduced in some embodiments.
- the enlarged 3D image data 1002 may be the original 3D image data 1001. Transparency can be adjusted and output so that can be seen together.
- the present invention when performing image processing on the 3D image data, by adjusting the depth value for the 3D image data together to make the image processed area appear more prominent, the user can see the 3D image data It has the effect of making it more convenient to use.
- FIG. 11 is a diagram illustrating a process of enlarging or reducing 3D image data according to a change in a user's position according to another exemplary embodiment of the present invention.
- the display apparatus determines whether the user selects a predetermined mode (for example, a dynamic zoom function) in which the enlargement or reduction function is adjusted according to the user position. To judge.
- a predetermined mode for example, a dynamic zoom function
- step S1101 when the user selects the corresponding function, the display device determines the location of the user in step S1102.
- the method for determining the user location according to the present invention can be implemented in various ways, when the display device of the present invention is a glasses-free method, to detect the user location in the sensor included in the display device to generate the location information
- the display device of the present invention is an eyeglass type
- the position of the shutter eyeglasses may be sensed or the position information of the shutter eyeglasses may be obtained by receiving the position information from the shutter eyeglasses.
- the shutter glasses sense the information for detecting the user's position by the position sensor, and then transmit the sensed sensing information to the display device, and the display device receives the sensing information sensed by the shutter glasses and receives the received information.
- the position of the shutter glasses, that is, the user position may be determined using the sensed information.
- the position of the shutter glasses may be determined by detecting an IR signal transmitted from the shutter glasses to calculate a distance from the x, y, and z axes.
- the location of the user may be determined.
- the position of the shutter glasses may be determined by analyzing image data of the IR transmission module photographed by the IR camera.
- the positions of the shutter glasses may be determined by analyzing images of the plurality of IR transmitting modules from the images photographed by the shutter glasses, and the positions of the shutter glasses may be used as the user position. .
- step S1103 If the location of the user is changed as a result of the determination in step S1103, the display device determines the location change value of the user in step S1104.
- the display device determines an enlargement ratio or a reduction ratio of the 3D image data according to the determined position change value, and determines an enlargement or reduction region in step S1106.
- the display device senses a user's position at predetermined time intervals, and when the sensed user's position is changed, generates a vector value corresponding to the position change value, and enlarges in response to the vector value. You can determine the ratio or reduction ratio and the area to be enlarged or reduced.
- the display apparatus determines the depth value corresponding to the enlargement or reduction ratio in step S1107.
- the depth value corresponding to the enlargement or reduction ratio may be stored in advance in the storage means as described with reference to FIG. 8.
- the display apparatus enlarges or reduces the determined enlarged or reduced area of the left image data and the right image data of the 3D image data according to the determined enlargement ratio or reduction ratio, respectively, and corresponds to the enlargement or reduction ratio.
- the 3D format is output at the depth value.
- 12 is a diagram illustrating an embodiment of determining a user position change value according to the present invention. 12 illustrates an example of 3D image data 1210 outputted in a spectacles method.
- the display device 1200 includes a predetermined position sensor 1201 to determine whether the position of the shutter glasses 1220 is changed.
- the shutter glasses 1220 and 1230 may include a predetermined IR output unit or IR sensors 1202 and 1203 so that the display apparatus 1200 may determine the position thereof.
- the display apparatus 1200 may generate a vector value 1204 corresponding to the position change value.
- FIG. 13 is a diagram illustrating an embodiment of determining an enlarged or reduced area and a depth value according to a change in position of a user in a display device according to the present invention.
- the display device determines the size d2 and the direction of the vector value 1204 corresponding to the user's position change value, and determines the size and the direction of the vector value 1204.
- the depth value of the area to be enlarged or reduced and the area to be enlarged or reduced is determined accordingly.
- the predetermined area 1310 of the 3D image data 1210 corresponding to the direction of the vector value is determined, and the area is determined as an area to be enlarged or reduced. In addition, it may be determined whether to enlarge or reduce according to the direction of the vector value.
- the 3D image data may be determined to be enlarged, and when the vector value is a direction away from the display device, the 3D image data may be reduced.
- the enlargement or reduction ratio can be determined according to the magnitude d2 of the vector value, and the enlargement or reduction ratio corresponding to each vector value magnitude can be stored in advance in the storage means.
- FIG. 14 is a diagram illustrating an embodiment of storing an enlargement or reduction ratio and a depth value corresponding to a position change value of a user according to an embodiment of the present invention.
- a display device may correspond to an enlargement or reduction ratio 1402 and a position change value corresponding to a position change value of a user (for example, a changed distance, 1401).
- the depth value 1403 may be stored in advance.
- the pixel shift value 1404 to be shifted and the transparency value 1405 corresponding to the enlargement or reduction ratio may be additionally stored to output the area to be enlarged or reduced in the 3D image data as the depth value 1403. .
- the present invention has an effect of providing a more dynamic enlargement reduction function by determining an area to include an enlargement or reduction and an enlargement or reduction ratio according to a change in the position of the user. For example, in consideration of the direction and distance that the user is approaching, the corresponding area is enlarged and the depth is closer to the user so that the user can feel more realistic 3D image including 3D image data. It works.
- 15 is a flowchart illustrating a process of image processing 3D image data in a display device according to another embodiment of the present invention.
- the display device determines whether overscan is set when outputting 3D image data in step S1501.
- the overscan means to scale and output an edge region of the video signal after the edge region of the video signal is removed in order to prevent a phenomenon in which image quality is filmed when noise is present at the edge of the video signal.
- the overscan setting may be preset in the display device according to the type of 3D image data or the type of source providing the 3D image data, and whether or not the user directly overscans the 3D image data to be output from a predetermined user interface. You can also set
- the display device determines a format of the 3D image data.
- the process of determining the format of the 3D image data has been described above with reference to FIGS. 5 and 6.
- format information of the 3D image data may be received from the external input source, and includes a module that determines the format of the 3D image data inside the display device.
- the format of the 3D image data output from the corresponding module may be determined.
- a format of 3D image data may be selected by a user.
- the format of the 3D image data may be determined as one of a side by side format, a checker board format, and a frame sequential format.
- step S1503 the display device classifies the left image data and the right image data of the 3D image data according to the format information of the 3D image data determined in step S1501.
- the left region may be determined as the left image data
- the right region may be determined as the right image data
- the display apparatus overscans the left image data and the right image data of the 3D image data, and then outputs the overscanned left image data and the right image data in the 3D format.
- the depth value at which the left image data and the right image data are output may be a previously stored value, a value determined in an image processing process, or a value input by a user.
- the display device receives a depth adjustment command, and then, after receiving the depth adjustment command, a depth corresponding to the depth adjustment command of the image processed left image data and right image data. Can be output as a value.
- the display device just scans and outputs the 3D image data in step S1506.
- the just scan refers to minimizing and outputting an operation process on an image signal without performing overscan.
- 16 is a diagram illustrating a process of overscanning 3D image data according to an embodiment of the present invention.
- the display device after determining the format of the 3D image data 1601 and 1602, the display device according to an embodiment of the present disclosure classifies the left image data and the right image data according to the format determination result. Each of the left image data and the right image data is overscanned.
- the left region may be determined as the left image data
- the right region may be determined as the right image data
- the left image data is overscanned, the right image data is overscanned, and the overscanned left image data 1602 and the right image data 1603 are output in 3D format.
- the format of the 3D image data 1604 is a top-and-bottom method
- the overscanned left image data 1605 and the right image data 1606 are output in 3D format.
- the overscan area is determined using the determination result, and the overscan process is performed.
- the 3D image data 1608 may be output in a 3D format.
- the overscan area 1608 may determine that the order of the left image data and the right image data is not changed so that an error does not occur in outputting the 3D image data by overscan.
- the left image data and the right image data which are sequentially input may be determined, and the 3D format may be output after overscanning the 1616 and 1611 respectively. Can be.
- 17 is a diagram illustrating an embodiment of outputting overscanned left image data and right image data in a 3D format according to the present invention.
- a display device outputs overscanned left image data 1701 and right image data 1702 as 3D image data 1703.
- the display device may output the 3D image data in at least one of line by line, frame sequential, and checker board methods using the determined format information of the 3D image data.
- the format of the 3D image data is converted according to the output method of the display device, and the 3D image data of the converted format is output.
- the 3D image data of line by line, top and bottom, and side by side methods may be converted into a frame sequential method and then output. Can be.
- FIG. 18 is a diagram illustrating a result of overscanning each of left image data and right image data and outputting them in a 3D format according to an embodiment of the present invention.
- the 3D image data 1801 in which the 3D image data 1800 itself is overscanned by the prior art the alignment of the left image data 1803 and the right image data 1804 is not performed correctly.
- Deterioration of the dimensional image data occurs or there is a problem that is not output in 3D at all, according to the present invention, after performing the left overscan of the left image data and the right image data, respectively, by outputting the 3D format, left image data And the right image data is correctly aligned, and after overscanning the 3D image data 1802, the 3D image data can be output in a 3D format as well as the noise can be output in high quality. have.
- a display device outputs 3D image data in accordance with a format and an image processor 1501 for processing 3D image data according to a panel of a display unit and a user setting.
- the display device outputs a scaler 1503 for image-processing each of the left image data and the right image data of the 3D image data, and outputs the processed left image data and the right image data in a 3D format at a predetermined depth value. It may be configured to include a formatter 1505, a user input unit 1506 for receiving a depth adjustment command for the 3D image data. In some embodiments, the image processing may include overscan processing.
- the output formatter 1505 outputs the image processed left image data and right image data in a 3D format with a depth value corresponding to the depth adjustment command.
- the scaler 1503 enlarges or reduces each of the left image data and the right image data of the 3D image data at an enlargement ratio or a reduction ratio corresponding to an enlargement command or a reduction command for the 3D image data. can do.
- the application controller 1507 may display a first user interface for receiving an enlargement command or a reduction command for 3D image data and a second user interface for receiving a depth adjustment command for the 3D image data. ), And the user input unit 1506 may receive an enlargement command, a reduction command, and a depth adjustment command. In addition, the user input unit 1506 may be designated as an enlarged or reduced area of the 3D image data.
- the FRC 1504 adjusts the frame rate of three-dimensional image data to the output frame rate of the display device.
- the scaler 1503 enlarges or reduces the designated enlarged or reduced area of the left image data and the right image data of the 3D image data according to the enlargement ratio or the reduction ratio, respectively.
- the output formatter 1505 outputs the enlarged or reduced processed left image data and right image data in a 3D format.
- the output formatter 1505 may output the enlarged or reduced processed left image data and the right image data as a depth value corresponding to the enlarged ratio or reduced ratio, or the user input unit 1506 may output three-dimensional image data.
- the output formatter 1505 may output the enlarged or reduced left image data and the right image data as a depth value corresponding to the depth adjustment command.
- the display device further includes a position determination module 1508 for determining a position change value of the user
- the scaler 1503 further includes an enlargement ratio or a reduction ratio for the 3D image data according to the determined position change value.
- the enlarged or reduced area may be determined, and the enlarged or reduced area of the left image data and the right image data of the 3D image data may be enlarged or reduced, respectively, according to the determined enlargement ratio or reduction ratio.
- the position determination module 1508 senses the position of the user at predetermined time intervals, and when the position of the sensed user is changed, generates a vector value corresponding to the position change value, and the scaler 1503 is performed. An enlargement ratio or a reduction ratio and an enlargement or reduction region may be determined corresponding to the vector value.
- 20 is a block diagram illustrating a display device according to another embodiment of the present invention. 20 is a block diagram showing the configuration when the display device of the present invention is a digital broadcast receiver.
- the digital broadcast receiver includes a tuner 101, a demodulator 102, a demultiplexer 103, a signaling information processor 104, an application controller 105, a storage 108, The external input receiver 109, the decoder / scaler 110, the controller 115, the mixer 118, the output formatter 119, and the display 120 may be configured.
- the digital broadcast receiver may further include other necessary components in addition to the configuration shown in FIG. 20.
- the tuner 101 tunes a specific channel to receive a broadcast signal including content.
- the demodulator 102 demodulates the broadcast signal input from the tuner 101.
- the demultiplexer 103 demultiplexes an audio signal, a video signal, and signaling information from the demodulated broadcast signal.
- the demultiplexing may be performed through PID (Packet IDentifier) filtering.
- PID Packet IDentifier
- the signaling information will be described below using System Information (SI) information such as PSI / PSIP (Program Specific Information / Program and System Information Protocol) as an example for convenience of description.
- SI System Information
- the demultiplexer 103 outputs the demultiplexed audio signal / video signal to the decoder / scaler 110 and the signaling information to the signaling information processor 104.
- the signaling information processor 104 processes the demultiplexed signaling information and outputs the demultiplexed signaling information to the application controller 105, the controller 115, and the mixer 118.
- the signaling processor 104 may include a database (not shown) for temporarily storing the processed signaling information.
- the application control unit 105 includes a channel manager 106 and a channel map 107.
- the channel manager 106 may form and manage the channel map 107 based on the signaling information, and control the channel map 107 to be switched based on the channel map 107 according to a user input.
- the decoder / scaler 110 includes a video decoder 111, an audio decoder 112, a scaler 113, and a video processor 114.
- the video decoder / audio decoders 111 and 112 receive and process demultiplexed audio signals and video signals.
- the scaler 113 scales the signal processed by the decoders 111/112 into a signal of an appropriate magnitude for output.
- the user input unit 123 may include a user input unit (not shown) that receives a key input of a user or the like through a remote controller.
- the application controller 105 may further include an OSD data generator (not shown) for configuring the UI.
- the OSD data generator may generate OSD data for UI configuration under the control of the application controller 105.
- the display 120 outputs content, a UI, and the like.
- the mixer 118 mixes and outputs inputs of the signaling processor 104, the decoder / scaler 110, and the application control unit 105.
- the output formatter 119 configures the output of the mixer 118 according to the output format of the display unit.
- the output formatter 119 bypasses, for example, 2D content, and processes the 3D content in 3D format according to the format and the output frequency of the display unit 120 under the control of the controller 115. Can work as a 3D formatter.
- the output formatter 119 outputs 3D image data to the display unit 120 and synchronizes the 3D image data with the configured 3D image data so that the 3D image data is synchronized with the shutter glasses 121.
- (Vsync) is generated and output to an IR emitter (not shown) in the shutter glasses so that the shutter glasses 121 can watch in synchronization with the display.
- the digital broadcast receiver further includes a scaler (not shown) for image processing the left image data and the right image data of the 3D image data, and the output formatter 119 includes the processed left image data and the right image.
- Image data is output in 3D format at a predetermined depth value.
- the user input unit 123 receives a depth adjustment command for 3D image data.
- the output formatter 119 outputs the image processed left image data and right image data as a depth value corresponding to the depth adjustment command.
- the scaler (not shown) enlarges or enlarges each of the left image data and the right image data of the 3D image data at an enlargement ratio or reduction ratio corresponding to an enlargement command or a reduction command for the 3D image data. Can be reduced.
- the application controller 105 may display a first user interface for receiving an enlargement command or a reduction command for 3D image data and a second user interface for receiving a depth adjustment command for the 3D image data. ), And the user input unit 123 may receive an enlargement command, a reduction command, or a depth adjustment command. In addition, the user input unit 1506 may be designated as an enlarged or reduced area of the 3D image data.
- a scaler (not shown) enlarges or reduces the designated enlarged or reduced area of the left image data and the right image data of the 3D image data according to the enlargement ratio or the reduction ratio, respectively.
- the output formatter 119 outputs the enlarged or reduced processed left image data and right image data in a 3D format.
- the output formatter 119 may output the enlarged or reduced processed left image data and the right image data as a depth value corresponding to the enlarged ratio or reduced ratio, or the user input unit 123 may output three-dimensional image data.
- the output formatter 119 may output the enlarged or reduced left image data and the right image data as a depth value corresponding to the depth adjustment command.
- the display apparatus further includes a position determination module 122 for determining a position change value of the user, and a scaler (not shown) may enlarge or reduce the enlargement ratio of the 3D image data according to the determined position change value.
- the ratio and the enlarged or reduced area may be determined, and the enlarged or reduced area of the left image data and the right image data of the 3D image data may be enlarged or reduced according to the determined enlarged ratio or reduced ratio, respectively.
- the position determination module 122 senses the position of the user at predetermined time intervals, and if the position of the sensed user is changed, generates a vector value corresponding to the position change value, and scales (not shown) ) May determine an enlargement ratio or a reduction ratio and an enlargement or reduction region corresponding to the vector value.
- the IR emitter receives the sync signal generated by the output formatter 119 and outputs the sync signal to the light receiving unit (not shown) in the shutter glasses 121, and the shutter glasses 150 pass through the IR emitter (not shown) through the light receiving unit.
- 21 is a diagram illustrating a configuration of shutter glasses according to an embodiment of the present invention.
- the shutter glasses include a left eye shutter liquid crystal panel 1100 and a right eye shutter liquid crystal panel 1130.
- the shutter liquid crystal panels 1100 and 1130 may simply pass or block light according to the source driving voltage.
- the left eye shutter liquid crystal panel 1100 passes light and the right eye shutter liquid crystal panel 1130 blocks light transmission, so that the left image data is transmitted only to the left eye of the user of the shutter glasses. do.
- the right image data is displayed on the display device, the left eye shutter liquid crystal panel 1100 blocks light transmission, and the right eye shutter liquid crystal panel 1130 passes light so that the right image data is transmitted only to the right eye of the user.
- the left eye shutter liquid crystal panel 1100 blocks light transmission
- the right eye shutter liquid crystal panel 1130 passes light so that the right image data is transmitted only to the right eye of the user.
- the infrared receiver 1160 of the shutter glasses converts the infrared signal received from the display device into an electrical signal and provides the converted signal to the controller 1170.
- the controller 1170 controls the left eye shutter liquid crystal panel 1100 and the right eye shutter liquid crystal panel 1130 to be alternately turned on and off according to the synchronization reference signal.
- the shutter glasses may pass or block light to the left eye shutter liquid crystal panel 1100 or the right eye shutter liquid crystal panel 1130 according to a control signal received from the display device.
- the present invention allows the user to select the depth value with the enlargement or reduction of the 3D image data so that the user can conveniently use the 3D image data.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
Description
Claims (20)
- 3D 디스플레이 기기의 영상 처리 방법에 있어서,3차원 이미지 데이터에 대한 확대 명령 또는 축소 명령에 대응하는 확대 비율 또는 축소 비율로 상기 3차원 이미지 데이터의 좌이미지 데이터 및 우 이미지 데이터 각각을 확대 또는 축소 처리하는 단계; 및상기 확대 또는 축소 처리된 좌이미지 데이터 및 우 이미지 데이터를 3D 포맷으로 출력하는 단계를 포함하는 것을 특징으로 하는 영상 처리 방법.
- 제1항에 있어서,상기 확대 또는 축소 처리된 좌이미지 데이터 및 우 이미지 데이터를 3D 포맷으로 출력하는 단계는,상기 확대 또는 축소 처리된 좌이미지 데이터 및 우 이미지 데이터를 상기 확대 비율 또는 축소 비율에 대응하는 뎁스값으로 출력하는 단계를 포함하는 것을 특징으로 하는 영상 처리 방법.
- 제1항에 있어서,상기 3차원 이미지 데이터에 대한 뎁스조정 명령을 입력받는 단계를 더 포함하고,상기 확대 또는 축소 처리된 좌이미지 데이터 및 우 이미지 데이터를 3D 포맷으로 출력하는 단계는,상기 확대 또는 축소 처리된 좌이미지 데이터 및 우 이미지 데이터를 상기 뎁스조정 명령에 대응하는 뎁스값으로 출력하는 단계를 포함하는 것을 특징으로 하는 영상 처리 방법.
- 제3항에 있어서,상기 3차원 이미지 데이터에 대한 확대 명령 또는 축소 명령을 입력받는 제1 사용자 인터페이스 및 상기 3차원 이미지 데이터에 대한 뎁스조정 명령을 입력받는 제2 사용자 인터페이스를 디스플레이 화면에 출력하는 단계를 더 포함하는 것을 특징으로 하는 영상 처리 방법.
- 제1항에 있어서,상기 3차원 이미지 데이터의 확대 또는 축소 영역을 지정받는 단계를 더 포함하고,상기 3차원 이미지 데이터에 대한 확대 명령 또는 축소 명령에 대응하는 확대 비율 또는 축소 비율로 상기 3차원 이미지 데이터의 좌이미지 데이터 및 우 이미지 데이터 각각을 확대 또는 축소 처리하는 단계는,상기 3차원 이미지 데이터의 좌이미지 데이터 및 우이미지 데이터 중 상기 지정받은 확대 또는 축소 영역을 각각 확대 또는 축소 처리하는 단계를 포함하는 것을 특징으로 하는 영상 처리 방법.
- 제1항에 있어서,사용자의 위치 변화값을 판단하고, 상기 판단된 위치 변화값에 따라 상기 3차원 이미지 데이터에 대한 확대 비율 또는 축소 비율 및 확대 또는 축소 영역을 결정하는 단계를 더 포함하고,상기 3차원 이미지 데이터에 대한 확대 명령 또는 축소 명령에 대응하는 확대 비율 또는 축소 비율로 상기 3차원 이미지 데이터의 좌이미지 데이터 및 우 이미지 데이터 각각을 확대 또는 축소 처리하는 단계는,상기 3차원 이미지 데이터의 좌이미지 데이터 및 우이미지 데이터 중 상기 결정된 확대 또는 축소 영역을 상기 결정된 확대 비율 또는 축소 비율에 따라 각각 확대 또는 축소 처리하는 단계를 포함하는 것을 특징으로 하는 영상 처리 방법.
- 제1항에 있어서,사용자의 위치 변화값을 판단하고, 상기 판단된 위치 변화값에 따라 상기 3차원 이미지 데이터에 대한 확대 비율 또는 축소 비율 및 확대 또는 축소 영역을 결정하는 단계는,상기 사용자의 위치를 소정 시간 간격으로 센싱하고, 상기 센싱된 사용자의 위치가 변경되면, 위치 변화값에 대응하는 벡터값을 생성하고, 상기 벡터값에 대응하여 확대 비율 또는 축소 비율 및 확대 또는 축소 영역을 결정하는 단계를 포함하는 것을 특징으로 하는 영상 처리 방법.
- 3D 디스플레이 기기의 영상 처리 방법에 있어서,3차원 이미지 데이터의 좌이미지 데이터 및 우 이미지 데이터를 판단하는 단계;상기 좌이미지 데이터 및 우이미지 데이터 각각을 영상 처리하는 단계; 및상기 영상 처리된 좌이미지 데이터 및 우 이미지 데이터를 소정의 뎁스값으로 3D 포맷 출력하는 단계를 포함하는 것을 특징으로 하는 영상 처리 방법.
- 제8항에 있어서,상기 3차원 이미지 데이터에 대한 뎁스조정 명령을 입력받는 단계를 더 포함하고,상기 영상 처리된 좌이미지 데이터 및 우 이미지 데이터를 소정의 뎁스값으로 3D 포맷 출력하는 단계는,상기 영상 처리된 좌이미지 데이터 및 우 이미지 데이터를 상기 뎁스조정 명령에 대응하는 뎁스값으로 출력하는 단계를 포함하는 것을 특징으로 하는 영상 처리 방법.
- 제8항에 있어서,상기 영상 처리는 오버 스캔 처리를 포함하는 것을 특징으로 하는 영상 처리 방법.
- 3차원 이미지 데이터에 대한 확대 명령 또는 축소 명령에 대응하는 확대 비율 또는 축소 비율로 상기 3차원 이미지 데이터의 좌이미지 데이터 및 우 이미지 데이터 각각을 확대 또는 축소 처리하는 스케일러; 및상기 확대 또는 축소 처리된 좌이미지 데이터 및 우 이미지 데이터를 3D 포맷으로 출력하는 출력 포맷터를 포함하는 것을 특징으로 하는 3D 디스플레이 기기.
- 제11항에 있어서,상기 출력 포맷터는,상기 확대 또는 축소 처리된 좌이미지 데이터 및 우 이미지 데이터를 상기 확대 비율 또는 축소 비율에 대응하는 뎁스값으로 출력하는 것을 특징으로 하는 3D디스플레이 기기.
- 제11항에 있어서,상기 3차원 이미지 데이터에 대한 뎁스조정 명령을 입력받는 사용자 입력부를 더 포함하고,상기 출력 포맷터는,상기 확대 또는 축소 처리된 좌이미지 데이터 및 우 이미지 데이터를 상기 뎁스조정 명령에 대응하는 뎁스값으로 출력하는 것을 특징으로 하는 3D 디스플레이 기기.
- 제13항에 있어서,상기 3차원 이미지 데이터에 대한 확대 명령 또는 축소 명령을 입력받는 제1 사용자 인터페이스 및 상기 3차원 이미지 데이터에 대한 뎁스조정 명령을 입력받는 제2 사용자 인터페이스를 디스플레이 화면에 출력하는 어플리케이션 제어부를 더 포함하는 것을 특징으로 하는 3D 디스플레이 기기.
- 제11항에 있어서,상기 3차원 이미지 데이터의 확대 또는 축소 영역을 지정받는 사용자 입력부를 더 포함하고,상기 스케일러는,상기 3차원 이미지 데이터의 좌이미지 데이터 및 우이미지 데이터 중 상기 지정받은 확대 또는 축소 영역을 각각 확대 또는 축소 처리하는 것을 특징으로 하는 3D 디스플레이 기기.
- 제11항에 있어서,사용자의 위치 변화값을 판단하는 위치 판단 모듈을 더 포함하고상기 스케일러는 상기 판단된 위치 변화값에 따라 상기 3차원 이미지 데이터에 대한 확대 비율 또는 축소 비율 및 확대 또는 축소 영역을 결정하고, 상기 3차원 이미지 데이터의 좌이미지 데이터 및 우이미지 데이터 중 상기 결정된 확대 또는 축소 영역을 상기 결정된 확대 비율 또는 축소 비율에 따라 각각 확대 또는 축소 처리하는 것을 특징으로 하는 3D 디스플레이 기기.
- 제16항에 있어서,상기 위치 판단 모듈은, 상기 사용자의 위치를 소정 시간 간격으로 센싱하고, 상기 센싱된 사용자의 위치가 변경되면, 위치 변화값에 대응하는 벡터값을 생성하고,상기 스케일러는 상기 벡터값에 대응하여 확대 비율 또는 축소 비율 및 확대 또는 축소 영역을 결정하는 것을 특징으로 하는 3D 디스플레이 기기.
- 3D 디스플레이 기기에 있어서,3차원 이미지 데이터의 좌이미지 데이터 및 우 이미지 데이터 각각을 영상 처리하는 스케일러; 및상기 영상 처리된 좌이미지 데이터 및 우 이미지 데이터를 소정의 뎁스값으로 3D 포맷 출력하는 출력 포맷터를 포함하는 것을 특징으로 하는 3D 디스플레이 기기.
- 제18항에 있어서,상기 3차원 이미지 데이터에 대한 뎁스조정 명령을 입력받는 사용자 입력부를 더 포함하고,상기 출력 포맷터는 상기 영상 처리된 좌이미지 데이터 및 우 이미지 데이터를 상기 뎁스조정 명령에 대응하는 뎁스값으로 출력하는 것을 특징으로 하는 3D 디스플레이 기기.
- 제18항에 있어서,상기 영상 처리는 오버 스캔 처리를 포함하는 것을 특징으로 하는 3D 디스플레이 기기.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/265,117 US20120050502A1 (en) | 2009-06-23 | 2010-06-23 | Image-processing method for a display device which outputs three-dimensional content, and display device adopting the method |
EP10792326.0A EP2410753B1 (en) | 2009-06-23 | 2010-06-23 | Image-processing method for a display device which outputs three-dimensional content, and display device adopting the method |
CN201080022817.9A CN102450022B (zh) | 2009-06-23 | 2010-06-23 | 输出三维内容的显示装置的图像处理方法以及采用该方法的显示装置 |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US21973309P | 2009-06-23 | 2009-06-23 | |
US61/219,733 | 2009-06-23 | ||
KR1020100059216A KR101719980B1 (ko) | 2010-06-22 | 2010-06-22 | 3차원 컨텐츠를 출력하는 디스플레이 기기의 영상 처리 방법 및 그 방법을 채용한 디스플레이 기기 |
KR10-2010-0059216 | 2010-06-22 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2010151044A2 true WO2010151044A2 (ko) | 2010-12-29 |
WO2010151044A3 WO2010151044A3 (ko) | 2011-04-07 |
Family
ID=43387042
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2010/004073 WO2010151044A2 (ko) | 2009-06-23 | 2010-06-23 | 3차원 컨텐츠를 출력하는 디스플레이 기기의 영상 처리 방법 및 그 방법을 채용한 디스플레이 기기 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20120050502A1 (ko) |
EP (1) | EP2410753B1 (ko) |
KR (1) | KR101719980B1 (ko) |
CN (1) | CN102450022B (ko) |
WO (1) | WO2010151044A2 (ko) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102215420A (zh) * | 2011-06-20 | 2011-10-12 | 深圳创维-Rgb电子有限公司 | 一种电视机3d格式切换的方法、系统及电视机 |
CN103024410A (zh) * | 2011-09-23 | 2013-04-03 | Lg电子株式会社 | 图像显示设备及其操作方法 |
CN103609104A (zh) * | 2011-05-23 | 2014-02-26 | 高通股份有限公司 | 用于立体效应调整的交互式用户接口 |
EP2574068A3 (en) * | 2011-09-20 | 2014-05-07 | LG Electronics | Image display apparatus and method for operating the same |
US9161016B2 (en) | 2011-01-26 | 2015-10-13 | Fujifilm Corporation | Image processing apparatus, imaging apparatus, reproduction apparatus, and image processing method for processing zoom in 3D imaging |
EP2509325A3 (en) * | 2011-04-08 | 2015-10-21 | Sony Corporation | Image processing apparatus, image processing method, and program |
US10366536B2 (en) | 2016-06-28 | 2019-07-30 | Microsoft Technology Licensing, Llc | Infinite far-field depth perception for near-field objects in virtual environments |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2352118B1 (en) * | 2008-11-25 | 2017-08-16 | Sony Interactive Entertainment Inc. | Image display device, method for displaying image and information storing medium |
EP2355526A3 (en) | 2010-01-14 | 2012-10-31 | Nintendo Co., Ltd. | Computer-readable storage medium having stored therein display control program, display control apparatus, display control system, and display control method |
US20120242807A1 (en) * | 2010-05-27 | 2012-09-27 | Nintendo Co. Ltd | Hand-held electronic device |
US9693039B2 (en) | 2010-05-27 | 2017-06-27 | Nintendo Co., Ltd. | Hand-held electronic device |
CN102340678B (zh) * | 2010-07-21 | 2014-07-23 | 深圳Tcl新技术有限公司 | 一种景深可调的立体显示装置及其景深调整方法 |
KR101850723B1 (ko) * | 2010-08-17 | 2018-04-20 | 엘지전자 주식회사 | 디지털 방송 신호 수신 장치 및 방법 |
JP5059922B2 (ja) * | 2010-09-15 | 2012-10-31 | シャープ株式会社 | 立体画像生成装置、立体画像表示装置、立体画像調整方法、立体画像調整方法をコンピュータに実行させるためのプログラム、及びそのプログラムを記録した記録媒体 |
KR101212170B1 (ko) * | 2010-12-10 | 2012-12-13 | 엘지디스플레이 주식회사 | 패턴드 리타더를 이용한 입체 영상 표시 장치 및 그 구동 방법 |
EP2487915A1 (en) * | 2011-02-14 | 2012-08-15 | HTC Corporation | 3d format conversion systems and methods |
US9628843B2 (en) * | 2011-11-21 | 2017-04-18 | Microsoft Technology Licensing, Llc | Methods for controlling electronic devices using gestures |
US9674499B2 (en) * | 2012-08-15 | 2017-06-06 | Qualcomm Incorporated | Compatible three-dimensional video communications |
CN102843570B (zh) * | 2012-09-11 | 2016-02-03 | 深圳Tcl新技术有限公司 | 基于原始视频选择3d模式的方法和装置 |
KR102329814B1 (ko) * | 2014-12-01 | 2021-11-22 | 삼성전자주식회사 | 3d 디스플레이를 위한 양안 거리 인식 장치 |
KR102132406B1 (ko) * | 2015-04-29 | 2020-07-09 | 삼성전자주식회사 | 디스플레이 장치 및 그의 제어 방법 |
CN106273470B (zh) * | 2015-06-10 | 2018-11-30 | 三纬国际立体列印科技股份有限公司 | 立体打印装置及其打印数据存储方法 |
JP6755809B2 (ja) * | 2017-01-23 | 2020-09-16 | 株式会社ジャパンディスプレイ | 表示装置 |
EP3528496A1 (en) * | 2018-02-16 | 2019-08-21 | Ultra-D Coöperatief U.A. | Overscan for 3d display |
US11224801B2 (en) * | 2019-11-22 | 2022-01-18 | International Business Machines Corporation | Enhanced split-screen display via augmented reality |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AUPN732395A0 (en) * | 1995-12-22 | 1996-01-25 | Xenotech Research Pty Ltd | Image conversion and encoding techniques |
US20050146521A1 (en) * | 1998-05-27 | 2005-07-07 | Kaye Michael C. | Method for creating and presenting an accurate reproduction of three-dimensional images converted from two-dimensional images |
GB2354389A (en) * | 1999-09-15 | 2001-03-21 | Sharp Kk | Stereo images with comfortable perceived depth |
US7639838B2 (en) * | 2002-08-30 | 2009-12-29 | Jerry C Nims | Multi-dimensional images system for digital image input and output |
WO2004066137A2 (en) * | 2002-11-29 | 2004-08-05 | Bracco Imaging, S.P.A. | System and method for managing a plurality of locations of interest in 3d data displays |
JP4118146B2 (ja) | 2003-01-09 | 2008-07-16 | 三洋電機株式会社 | 立体画像処理装置 |
JP4490074B2 (ja) * | 2003-04-17 | 2010-06-23 | ソニー株式会社 | 立体視画像処理装置、立体視画像表示装置、立体視画像提供方法、および立体視画像処理システム |
JP2005073049A (ja) * | 2003-08-26 | 2005-03-17 | Sharp Corp | 立体映像再生装置および立体映像再生方法 |
KR100657275B1 (ko) * | 2004-08-26 | 2006-12-14 | 삼성전자주식회사 | 입체 영상 신호 발생 방법 및 이에 적합한 스케일링 방법 |
DE102006014902B4 (de) * | 2006-03-30 | 2009-07-23 | Siemens Ag | Bildverarbeitungseinrichtung zur erweiterten Darstellung von dreidimensionalen Bilddatensätzen |
US8330801B2 (en) * | 2006-12-22 | 2012-12-11 | Qualcomm Incorporated | Complexity-adaptive 2D-to-3D video sequence conversion |
JP2009135686A (ja) * | 2007-11-29 | 2009-06-18 | Mitsubishi Electric Corp | 立体映像記録方法、立体映像記録媒体、立体映像再生方法、立体映像記録装置、立体映像再生装置 |
KR100942765B1 (ko) * | 2008-03-18 | 2010-02-18 | (주)블루비스 | 입체 영상 편집을 위한 장치 및 방법 |
GB0806183D0 (en) * | 2008-04-04 | 2008-05-14 | Picsel Res Ltd | Presentation of objects in 3D displays |
CN101271678A (zh) * | 2008-04-30 | 2008-09-24 | 深圳华为通信技术有限公司 | 屏幕字体缩放的方法及终端设备 |
KR101667723B1 (ko) * | 2008-12-02 | 2016-10-19 | 엘지전자 주식회사 | 3차원 영상신호 전송 방법과, 3차원 영상표시 장치 및 그에 있어서의 신호 처리 방법 |
US8199186B2 (en) * | 2009-03-05 | 2012-06-12 | Microsoft Corporation | Three-dimensional (3D) imaging based on motionparallax |
US8314832B2 (en) * | 2009-04-01 | 2012-11-20 | Microsoft Corporation | Systems and methods for generating stereoscopic images |
WO2011060525A1 (en) * | 2009-11-19 | 2011-05-26 | Esight Corporation | Image magnification on a head mounted display |
-
2010
- 2010-06-22 KR KR1020100059216A patent/KR101719980B1/ko active IP Right Grant
- 2010-06-23 WO PCT/KR2010/004073 patent/WO2010151044A2/ko active Application Filing
- 2010-06-23 US US13/265,117 patent/US20120050502A1/en not_active Abandoned
- 2010-06-23 CN CN201080022817.9A patent/CN102450022B/zh not_active Expired - Fee Related
- 2010-06-23 EP EP10792326.0A patent/EP2410753B1/en not_active Not-in-force
Non-Patent Citations (2)
Title |
---|
None |
See also references of EP2410753A4 |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9161016B2 (en) | 2011-01-26 | 2015-10-13 | Fujifilm Corporation | Image processing apparatus, imaging apparatus, reproduction apparatus, and image processing method for processing zoom in 3D imaging |
EP2574065B1 (en) * | 2011-01-26 | 2016-09-07 | FUJIFILM Corporation | Image processing device, image-capturing device, reproduction device, and image processing method |
US9516300B2 (en) | 2011-04-08 | 2016-12-06 | Sony Corporation | Image processing apparatus, image processing method, and program |
EP2509325A3 (en) * | 2011-04-08 | 2015-10-21 | Sony Corporation | Image processing apparatus, image processing method, and program |
EP2716052A1 (en) * | 2011-05-23 | 2014-04-09 | Qualcomm Incorporated | Interactive user interface for stereoscopic effect adjustment |
JP2014517619A (ja) * | 2011-05-23 | 2014-07-17 | クゥアルコム・インコーポレイテッド | 立体視効果の調整用の対話型ユーザインターフェース |
CN103609104A (zh) * | 2011-05-23 | 2014-02-26 | 高通股份有限公司 | 用于立体效应调整的交互式用户接口 |
JP2016192773A (ja) * | 2011-05-23 | 2016-11-10 | クゥアルコム・インコーポレイテッドQualcomm Incorporated | 立体視効果の調整用の対話型ユーザインターフェース |
CN102215420A (zh) * | 2011-06-20 | 2011-10-12 | 深圳创维-Rgb电子有限公司 | 一种电视机3d格式切换的方法、系统及电视机 |
EP2574068A3 (en) * | 2011-09-20 | 2014-05-07 | LG Electronics | Image display apparatus and method for operating the same |
EP2574062A3 (en) * | 2011-09-23 | 2014-05-07 | LG Electronics | Image display apparatus and method for operating the same |
US9024875B2 (en) | 2011-09-23 | 2015-05-05 | Lg Electronics Inc. | Image display apparatus and method for operating the same |
CN103024410B (zh) * | 2011-09-23 | 2015-11-18 | Lg电子株式会社 | 图像显示设备及其操作方法 |
CN103024410A (zh) * | 2011-09-23 | 2013-04-03 | Lg电子株式会社 | 图像显示设备及其操作方法 |
US10366536B2 (en) | 2016-06-28 | 2019-07-30 | Microsoft Technology Licensing, Llc | Infinite far-field depth perception for near-field objects in virtual environments |
Also Published As
Publication number | Publication date |
---|---|
EP2410753A2 (en) | 2012-01-25 |
US20120050502A1 (en) | 2012-03-01 |
EP2410753B1 (en) | 2016-10-19 |
CN102450022B (zh) | 2015-03-25 |
KR20110138995A (ko) | 2011-12-28 |
WO2010151044A3 (ko) | 2011-04-07 |
EP2410753A4 (en) | 2014-01-01 |
KR101719980B1 (ko) | 2017-03-27 |
CN102450022A (zh) | 2012-05-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2010151044A2 (ko) | 3차원 컨텐츠를 출력하는 디스플레이 기기의 영상 처리 방법 및 그 방법을 채용한 디스플레이 기기 | |
WO2010107227A2 (en) | A method of displaying three-dimensional image data and an apparatus of processing three-dimensional image data | |
WO2011046279A1 (en) | Method for indicating a 3d contents and apparatus for processing a signal | |
WO2012044130A2 (ko) | 배리어를 이용하는 3d 디스플레이 장치 및 그 구동 방법 | |
WO2010151028A4 (en) | Image display apparatus, 3d glasses, and method for operating the image display apparatus | |
WO2011059261A2 (en) | Image display apparatus and operating method thereof | |
WO2013081429A1 (en) | Image processing apparatus and method for subpixel rendering | |
WO2012074328A2 (ko) | 다시점 3차원 방송 신호를 수신하기 위한 수신 장치 및 방법 | |
WO2011155766A2 (ko) | 영상 처리 방법 및 그에 따른 영상 표시 장치 | |
WO2014010871A1 (en) | Image display apparatus and menu display method | |
WO2012002690A2 (ko) | 디지털 수신기 및 디지털 수신기에서의 캡션 데이터 처리 방법 | |
WO2011102699A2 (ko) | 전자 장치 및 입체영상 재생 방법 | |
WO2011129566A2 (ko) | 이미지 디스플레이 방법 및 장치 | |
WO2012177049A2 (en) | Method and apparatus for processing broadcast signal for 3-dimensional broadcast service | |
WO2011152633A2 (en) | Method and apparatus for processing and receiving digital broadcast signal for 3-dimensional subtitle | |
WO2010151027A4 (ko) | 영상표시장치 및 그 동작방법 | |
WO2011004963A2 (en) | Method for displaying three-dimensional user interface | |
WO2011021894A2 (en) | Image display apparatus and method for operating the same | |
WO2013100376A1 (en) | Apparatus and method for displaying | |
WO2012057434A1 (en) | Stereoscopic image processing system and device and glasses | |
WO2012050366A2 (en) | 3d image display apparatus and display method thereof | |
WO2013100722A1 (en) | Glasses apparatus, display apparatus, content providing method using the same and method for converting mode of display apparatus | |
WO2023210887A1 (ko) | 펜타일 방식 입체표시장치 및 시스템 | |
WO2014109479A1 (en) | Display apparatus, shutter glasses, display method, and method for operating glasses apparatus | |
WO2015002442A1 (ko) | 다시점 영상이 제공되는 시스템에서 부가 오브젝트를 포함하는 3차원 영상 처리 방법 및 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080022817.9 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10792326 Country of ref document: EP Kind code of ref document: A2 |
|
REEP | Request for entry into the european phase |
Ref document number: 2010792326 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13265117 Country of ref document: US Ref document number: 2010792326 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |