MXPA00002201A - Image processing method and apparatus - Google Patents

Image processing method and apparatus

Info

Publication number
MXPA00002201A
MXPA00002201A MXPA/A/2000/002201A MXPA00002201A MXPA00002201A MX PA00002201 A MXPA00002201 A MX PA00002201A MX PA00002201 A MXPA00002201 A MX PA00002201A MX PA00002201 A MXPA00002201 A MX PA00002201A
Authority
MX
Mexico
Prior art keywords
image
images
monoscopic
movement
processing
Prior art date
Application number
MXPA/A/2000/002201A
Other languages
Spanish (es)
Inventor
Victor Harman Philip
Original Assignee
Victor Harman Philip
Xenotech Research Pty Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Victor Harman Philip, Xenotech Research Pty Ltd filed Critical Victor Harman Philip
Publication of MXPA00002201A publication Critical patent/MXPA00002201A/en

Links

Abstract

An image conversion system for converting monoscopic images for viewing in three dimensions including:an input means (1) adapted to receive the monoscopic images;a preliminary analysis means to determine if thereis any continuity between a first image and a second image of the monoscopic image sequence;a secondary analysis means (2) for receiving monoscopic images which have a continuity, and analysing the images to determine the speed and direction of motion, and the depth, size and position of objects;a first processing means (3) for processing the monoscopic images based on data received from the preliminary analysis means or the secondary analysis means;a second processing means capable of further processing images received from the first processing means;a transmission means (4) capable of tranferring the processed images to a stereoscopic display system (5).

Description

METHOD AND APPARATUS FOR THE PROCESSING OF IMAGES FIELD OF THE INVENTION The present invention relates in general to stereoscopic image systems, and in particular to the synthesis of pairs of stereoscopic images from images from monoscopic images for stereoscopic display. The present invention is also directed to a method of five modules, to produce stereoscopic images, which digitizes a monoscopic source, analyzes it for movement, generates pairs of stereoscopic images, optimizes the stereoscopic effect, transmits or stores them and then allows they are displayed on a stereoscopic display device.
BACKGROUND OF THE INVENTION The advent of stereoscopic or three-dimensional (3D) display systems that create a more realistic image for the viewer from conventional monoscopic or 2D (2D) display systems, requires that stereoscopic images be available to be views in the systems REF .: 32967 3D exhibition. In this regard, there are many sources of monoscopic images, for example, existing 2D films or videos, which can be manipulated to produce stereoscopic images to be viewed on a stereoscopic display device. The preexisting methods to convert these monoscopic images for stereoscopic vision do not produce acceptable results. Other attempts in film and video have used techniques to duplicate the reference point of the stereoscopic depth of the "motion parallax". These comprise the production of a delay for the images presented to the trailing eye when the images are presented lateral, left or right movement. • Other attempts have used "lateral shift" of the images to the left and right eyes to provide depth perception. However, these two techniques are limited, and in general, only satisfy specific applications. For example, the motion parallax technique is only good for scenes with left or right movement and is of limited value for the stereoscopic enhancement of fixed scenes. The lateral change technique will only give a full depth effect to a scene and will not allow different objects to be perceived at varying depths in the depths where they occur. Even the combination of these two techniques will only give a limited stereoscopic effect for most 2D movies or videos. Some existing approaches demonstrate the limitations of these techniques. When an image has vertical movement and some lateral movement and a delay is provided to the image presented to the trailing eye, then the result is often a vertical disparity between the left and right visions such that the images are uncomfortable to look at. Scenes with counter-movement, such as objects that move left and right in the same scene are also uncomfortable to look at. Certain modalities of these methods define that when objects of varying depths are presented in an image, there is a "cardboard box cut" appearance different from objects with different depth modules instead of a smooth transition of objects with different modules. of depth instead of a smooth transition of objects from the foreground to the background. In all these approaches, a successful attempt has not been made to develop a system or method to satisfy all the sequences of images or to solve the problem of discomfort to the viewer to optimize the stereoscopic effect for each spectator or display device.
OBJECTS OF THE INVENTION Therefore, there is a need for a system with improved methods for converting monoscopic images into pairs of stereoscopic images and a system for providing enhanced stereoscopic images to a viewer. An object of the present invention is to provide this system with improved methods.
BRIEF DESCRIPTION OF THE INVENTION In order to face the aforementioned problems, the present invention provides in one aspect, a method for converting monoscopic images to be seen in three dimensions, which includes the steps of: receiving the monoscopic images; analyze the monoscopic images to determine the characteristics of the images; process the monoscopic images based on the determined characteristics of the image; transfer the processed images to suitable stereoscopic and / or storage display systems; wherein the analysis of the monoscopic images to determine the movement, includes the steps of: dividing each image into a plurality of blocks, wherein the corresponding blocks in an adjacent image are misaligned horizontally and / or vertically; and compare each block with the corresponding blocks to find the minimum mean square error and thus the movement of the block. An image conversion system for converting monoscopic images to be viewed in three dimensions, including: an input means adapted to receive monoscopic images; a means of preliminary analysis to determine if there is any continuity between a first image and a second image of the sequence of monoscopic images; a means of secondary analysis to receive monoscopic images that have a continuity, and analyze the images to determine at least one of the speed and direction of movement, or the depth, size and position of the objects, where the analysis of the monoscopic images to determine the movement includes the steps of: dividing each image into a plurality of blocks, wherein the corresponding blocks in an adjacent image are misaligned horizontally and / or vertically, and comparing each block with the corresponding blocks to find the minimum square root mean error and in this way the movement of the block; a first processing means for processing the monoscopic images based on the data received from the preliminary analysis means and / or the secondary analysis means. Ideally, the input means also includes a means for capturing and digitizing the monoscopic images.
Preferably, the image analysis means is capable of determining the speed and direction of movement, the depth, size and position of the objects and the second plane within an image. In a further aspect, the present invention provides a method for optimizing the stereoscopic image to further improve the stereoscopic effect and this process is generally applied before transmission, storage and display. In yet a further aspect, the present invention provides a method for improving the pairs of stereoscopic images by adding a reference point of the viewer to the image. In still a further aspect, the present invention provides a method for analyzing monoscopic images for the conversion to pairs of stereoscopic images that includes the steps of: scaling each image in a plurality of regions; comparing each region of a first image with the corresponding and adjacent regions of a second image to determine the nature of the movement between the first image and the second image.
Preferably, a motion vector is defined for each image based on a comparison of the nature of motion detected with predefined movement categories that vary from non-motion to a complete scene change. In yet a further aspect, the present invention provides a system for converting monoscopic images to be viewed in three dimensions, including: a first module adapted to receive a monoscopic image; a second module adapted to receive the monoscopic image and analyze the monoscopic image to create the date of the image, wherein the analysis of the monoscopic image to determine the movement includes the steps of: dividing each image into a plurality of blocks, wherein the corresponding blocks in an adjacent image are misaligned horizontally and / or vertically, and compares each block with the corresponding blocks to find the minimum mean square error and thus the movement of the block; a third module adapted to create pairs of stereoscopic images from the monoscopic image using at least one predetermined technique selected as a function of the image data; a fourth module adapted to transfer the pairs of stereoscopic images to a stereoscopic display medium; a fifth module consisting of a stereoscopic display medium. Preferably, the first module is further adapted to convert any analog image into a digital image. Also, the second module is preferably adapted to detect any object in a scene and make a determination as to the speed and direction of this movement. Conveniently, the image can be compressed before any analysis. Preferably, the third module further includes an Optimization step to further enhance the pairs of stereoscopic images before transmitting the pairs of stereoscopic images to the stereoscopic display means. Conveniently, the fourth module. it may also include a storage medium for storing pairs of stereoscopic images for display in the stereoscopic display medium at a later time.
ADVANTAGES It will be appreciated that the process of the present invention can be suspended at any stage and stored for continuation at a later time or transmitted for continuation at another location if required. The present invention provides a conversion technology with a number of unique advantages including: 1) Conversion in real time or not in real time The ability to convert monoscopic images to pairs of stereoscopic images can be done in real time or not in real time. The intervention of the operator can be applied to manipulate the images manually. An example of this is in the conversion of films or videos where each sequence can be tested and optimized for its stereoscopic effect by an operator. 2) Techniques including erascopic treatment The present invention uses a plurality of techniques to further improve the basic techniques of parallax of movement and lateral displacement (forced parallax) to generate pairs of stereoscopic images. These techniques include, but are not limited to, the use of object analysis, labeling, tracking and training, parallax zones, reference points, motion synthesis, and parallax modulation techniques. 3) Detection and correction of inverted 3D Inverted 3D is ideally detected as part of the 3D generation process when analyzing the movement characteristics of an image. Then correction techniques can be used to minimize inverted 3D to minimize discomfort to the viewer. 4) Use in all applications-includes transmission and storage The present invention describes a technique applicable to a wide range of applications and describes a complete process for applying the stereoscopic conversion process to monoscopic applications. The present invention describes on the one hand techniques for the generation of 3D where both the image processing equipment and the stereoscopic display equipment are located substantially in the same location. On the other hand, techniques are defined for the generation of pairs of stereoscopic images in a location and their transmission, storage and subsequent exhibition in a remote location.
) Can be used with any stereoscopic display device The present invention fits any stereoscopic display device and ideally includes adjustment facilities. The 3D generation process can also take into account the type of display device in order to optimize the stereoscopic effect.
BRIEF DESCRIPTION OF THE FIGURES The invention will be more fully understood from the following detailed description of a preferred embodiment of the conversion method and the integrated system and as illustrated in the appended figures. However, it will be appreciated that the present invention is not limited to the described modality. Figure 1 shows the decomposition of modules a complete system using the present invention. Figure 2 shows a possible use of multiple processors with a complete system which uses the present invention. Figure 3 shows a flow diagram of Module 1 (Video Digitalization) and the first part of Module 2 (Image Analysis). Figure 4 shows the second part of a flow chart of Module 2. Figure 5 shows the third part of a flow chart of Module 2. Figure 6 shows the fourth part of a flow chart of Module 2. Figure 7 shows a flowchart of the first part of Module 3 (3D d Generation). Figure 8 shows the second part of a flow chart of Module 3 and Module 4 (3D Media - Transmission and storage) and Module 5 (3D Display).
DETAILED DESCRIPTION The invention is intended to provide a viewer with a stereoscopic image that uses the visual, complete perception capabilities of an individual. Therefore, it is necessary to provide depth reference points that the brain requires to interpret these images.
INTRODUCTION Humans see a complex combination of physiological and psychological processes involving the eyes and the brain. Visual perception involves the use of short and long term memory to be able to interpret visual information with known and experienced reality as defined by our senses. For example, according to the Cartesian laws in space and perspective, the more an object moves away from the smaller viewer sees it. In other words, the brain expects that if an object is large it is close to the viewer and if it is small it is at some distance. This is a learned process based on the knowledge of the size of the object in the first place. Other monoscopic or minor depth reference points that can be represented in the visual information are for example shadows, defocus, texture, light, atmosphere. These depth reference points are used to take great advantage in the production of video games in "3D perspective" and computer graphics. However, the problem with these techniques in achieving a stereoscopic effect is that you can not quantify the perceived depth; It is an illusion of 2D display objects in a 2D environment. These exhibitions do not seem real since they do not show a stereoscopic image because the visions of both eyes are identical.
DEPTH OF REFERENCE POINTS Stereoscopic images are an attempt to recreate visual images of the real world, and require much more visual information than images "in 3D perspective" so that depth can be quantified. The main or stereoscopic depth reference point provides these data so that a person's visual perception can be stimulated in three dimensions. These main depth reference points are described as follows: 1) Retinal disparity refers to the fact that both eyes see a slightly different view. This can easily be demonstrated by holding an object in front of a person's face and focusing in the background. Once the eyes have focused on the background, it will appear that there are actually two objects in the front of the face. The disparity is the horizontal distance between the left and right image points, corresponding to the overlapping retinal images. While the parallax is the real spatial displacement between the images seen. 2) Motion parallax .- Those objects that are closer to the viewer will seem to move faster even if they are traveling at the same speed as distant objects. Therefore, the relative motion is a - depth reference point, less. But the stereoscopic, principal reference point of lateral movement is the creation of motion parallax. With the movement in an image that moves from right to left, the right eye is the guiding eye while the left eye becomes the trailing eye with its image that is delayed. This delay is a normal function of our visual perception mechanism. For movement from left to right, the right eye becomes the drag eye. The effect of this delay will create retinal disparity (two different views to the eyes), which is perceived as the binocular parallax, thus providing the reference point is ereoscopic known as motion parallax. 3) Accommodation.- The eye puts an object t in acute focus by compressing either the lens of the eye (more convex form for distant object) through motor activity. The amount and type of neuromotor activity is a stereoscopic reference point for depth in an image. 4) Convergence.- It is the response of the neuromotor system of the eye that puts the images and an object in alignment with the central visual area of the eyes such that only an object is seen. For exampleWhen a finger is kept at a prudent distance, it is seen by both eyes and slowly withdraws from the face, the eyes get inwards (converge), indicating that the eye is getting closer. That is, the depth of the finger is decreasing. The convergence response of the eyes is physiologically linked to the mechanism of accommodation in normal vision. In stereoscopic vision, when the spectators are not adjusted to the "fixation plane" (to which the eyes converge, they may experience discomfort.) The "fixation plane" is usually the plane of the screen.
GENERAL VISION - 5-MODEL APPROACH The present invention describes a system that is capable of taking any monoscopic input and converting it to an improved stereoscopic output. For ease of description, this complete system can be decomposed into a number of independent modules or processes, specifically: MODULE 1.- Monoscopic Image Input (typically video input) MODULE 2.- Image Analysis MODULE 3.- 3D GENERATION MODULE 4. • 3D Medium (Storage Transmission) MODULE 5.- 3D display Figure 1 shows this top-down approach to the stereoscopic conversion process, where "video or some other monoscopic image source is introduced, the images are analyzed, the pairs of stereoscopic images are generated, transmitted and / or stored and then They are displayed in a stereoscopic display Each module describes a process independent of the entire system from the input of the monoscopic image to the stereoscopic display, however, it will be appreciated that several modules can be operated independently.
APPLICATIONS In general, the five modules are used, from the monoscopic image input to the display for a particular application. For example, this system can be used in theaters or cinemas. In this application, the 2D video input can take the form of analog or digital video sources. These sources will then be analyzed to determine the direction and speed of any movement. The processes will then work either in real time or not in real time, in order to create the 3D images. This can be further optimized through the use of edges, parallax modification, inverted 3D analysis, hue and / or texturing. The 3D images can then be stored or transmitted to a 3D display, including obturator lenses, polarization lenses or an autostereoscopic display. This system can also be adapted for use with cable or pay television systems. In this application, the 2D video input could be video from a VTR, a laser disk, or some other digital source. Again, the generation and / or optimization of 3D can proceed in either real time or not real time. The 3D medium module will conveniently take the form of cable or satellite transmission to allow 3D display on television, projected on video, or a self-stereoscopic display. The system can also be used with video gallery fires, in multimedia, or on terrestrial or network TV, depending on the application, and the 2D video input module can obtain monoscopic source images from a digital processor. games, video from a laser disk, video from a VTR, video from a network, or some other digital storage device or digital source or telecine process. The generation of 3D can take place in real time or not in real time, and is generated by computer in a central conversion site, in a user's computer, in a central processor, or some other image processor. The stereoscopic images can then be stored in the video or digital storage device, before distribution to the cinemas or transmission over a local network. These stereoscopic images can then be transmitted to the video projectors via a local transmission, or alternatively via VHF / UHF facilities, or satellite. The 3D display is dependent on the required application, and can take the form of an auto-stereoscopic display device, a video projector with polarization lenses, a local monitor with shutter lenses, or a decoder with suitable vision lenses Individual and Multi Processors The complete system can be operated in an individual processor with the five modules that are processed together, or individually in real time or not in real time (Modules 2, 3 and 4). Modules 2 and 3 can be further segmented to fit in a multitasking or multiprocessor environment, as shown in Figure 2, by way of example. The use of multiple processors can also be configured to the application at hand. For example, Modules 1 and 2 can be operated by a first processor, and Modules 3 through 5 by a second processor. If desired, the first processor of this array could be used as a forward search processor, and the second processor could generate the stereoscopic images after a delay. Alternatively, a first processor could be used to receive video or real-time, digitize the video and transfer the digitized video to a suitable digital storage device. A second processor, either on site or remotely, can then analyze the scanned image and perform the necessary tasks to display a stereoscopic image on an • appropriate display device. Forward-looking processing techniques can be used to predict trends in the sequences of the film or video, so that the modes of image processing can be selected more efficiently to optimize the full, stereoscopic effect. The present invention relates mainly to the analysis of monoscopic images and the conversion of the monoscopic images into pairs of stereoscopic images together with the optimization of the stereoscopic effect. In this regard, the present invention can be applied to a wide range of monoscopic inputs, transmission means and viewing means. However, for fullness, the five defined modules will be described here: MODULE 1 - IMAGE OR VIDEO ENTRY Module 1 requires that a monoscopic image source or video input be provided. This source can be provided as either a digital image source or an analog image source which can then be digitized. These image sources may include: 1) Analog Source aJBassed on tape.- VCR / VTR or Filme. b) Based on dicso.- Laser disk. c) Video camera or other device to capture images in real time. d) Images or graphics generated by computer. 2) Digital source a) Tape based.- Typical examples are DAT, AMPEX DCT, SONY Digital Betacam, Panasonic digital video formats or the new Digital Video (DVC) cartridge format using a 6.5 mm tape . b) Disk-based storage. -Data storage devices based on Magneto-Optical (MO) disk, hard disk (HD), compact disc (CD), Laser Disc, CD-ROM, DAT, Digital Video Cartridge (DVC) or digital video disc (DVD) ) - use JPEG, MPEG or other digital formats. c) Video Camera or other device for capturing images, in real time. d) Images or graphics generated by computer. What is important for the conversion process of the present invention is that a source of monoscopic images is provided. It is noted that a source of stereoscopic images can be provided, which will in general avoid the need for modules 1 to 3, however, any stereoscopic image can be passed through an Optimization stage before the display.
MODULE 2. - IMAGE ANALYSIS Referring now to Figures 3 to 8, which show flow charts demonstrating a preferred arrangement of the present invention. After the reception of the 2D images, the digitized video or digital image data is processed on a field basis per field of image by image, in real time or not in real time by the physical equipment, computer program or by a combination of both. First, the image analysis process occurs, including the steps of: 1) Image compression. 2) Motion detection. 3) Object detection. 4) Motion analysis. 1) . Image Compression Image compression is not essential, however, for many processes and applications, compression is a practical option in a particular way, where the processor is not powerful enough to process a full resolution image in the required time . Preferably, the images are scaled to smaller dimensions. The scale factor is dependent on the digital video resolution used __for each image, and is usually defined by the type of image capture facility used in the digitization process. . 2) Motion Detection In a preferred embodiment, each image can be analyzed in blocks of pixels. A motion vector is calculated for each block by comparing the blocks of an image with the corresponding blocks of an adjacent image that are misaligned horizontally and / or vertically by up to a predetermined number of pixels, for example ± 9 and record the position it gives the minimum root mean square error. For each block, the vector and the maximum minimum squared mean error are recorded for further processing. To save processing time, vectors do not need to be calculated if there is no detail in the block, for example, when the block is of homogeneous color. Other methods can be used to calculate the movement, for example, image subtraction. The present method uses the Quadratic Mean Error method. 3) _ Object detection An object is defined as a group of pixels or image elements that identify a part of an image that has common characteristics. These characteristics can be related to regions of similar illuminance value (similar brightness), chrominance value (similar color), motion vector (similar speed and direction of movement) or similar image detail (pattern or similar edge). For example, a car that is driven beyond a house. The car is a region of pixels or block of pixels that is moving at a speed different from the second plane. If the car stops in front of the house, then the car will be difficult to detect, and other methods can be used. A connectivity algorithm can be used to combine motion vectors into regions of similar motion vectors. An object may be comprised of one or more of these regions. Other image processing algorithms, such as edge detection, etc., can be used in the detection of objects. Once the objects in an image are identified, they are preferably marked or given an identification number. These objects and their relevant details (eg, position, size, motion vector, type, depth) are then stored in a database so that they can be presented for further processing. If an object is followed over a sequence of images, then this is known - as an object tracking. By tracking objects and analyzing their characteristics, they can be identified as being foreground or background objects and therefore enhanced to emphasize their depth position in an image. 4) Motion Analysis Once the objects have been detected, the objects can be analyzed to determine the speed and complete friction of the movement in the image. In the preferred embodiment, this step determines the type of image movement, and also provides a complete vector. By using object detection information and comparing data to various models of image movement, a primary determination can be made as to the best method for converting monoscopic images into pairs of stereoscopic images. The image movement models as used in the preferred embodiment of the present invention are: a) Scene Change. b) Simple Panoramic Effect. c) Complex Panoramic Effect. d) Moving object on stationary second plane. e) Close-up object on a moving background. f) No Movement. Other motion models can be used as required. a) Scene Change A scene change as the name suggests is when an image has little or no mediocrity to a previous scene image. It can be detected as a very large absolute difference in the luminance between the two images, or a large difference in the colors of the two images. In a preferred arrangement, a scene change may be determined when the average of the differences in luminance values (0-225) between the previous and current images is typically above 30. This value may vary with application, but The error test determined that this value is appropriate to determine most of the scene changes. A secondary test to determine a scene change can be when there are too many regions of motion vectors, which appears as random noise type in the image and is probably due to a scene change. This can happen if there is a very large amount of image movement. A third technique to detect a scene change is to analyze the few upper lines of each image to detect a scene change. The upper part of each image changes the least. Alternatively, when most of the blocks in the motion vector have large error values, the difference between two images is too large and will therefore be considered as a scene change.
Scene Change and Field Delay In the preferred embodiment, when there is a lateral movement detected in a scene, the image to the trailing eye is delayed by an amount of time that is inversely proportional to the speed of the movement. For an image that moves from right to left, the trailing eye is the left eye and for an image that moves from left to right, the trailing eye is the right eye. The delay of the image sequence (or field delay) to the trailing eye can be created by temporarily delaying the sequence of the video fields to the trailing eye by storing them digitally in memory. The current video field is shown to the guiding eye and the image delayed to the trailing eye is selected from the stored video fields that depend on the speed of lateral movement. More than a number of fields exhibited, a history regarding the change in movement and the change in field delays to the trailing eye can be maintained. This helps in smoothing the stereoscopic effect by allowing the image processor to anticipate any movement and reactions tendencies accordingly by modifying the delay so that there are no sudden changes. If a scene change is detected, the field delay for the preferred embodiment of the present invention is set to zero to prevent the image from breaking and the history of field delay is also readjusted. The field delay history is readjusted preferentially at each scene change. b) Simple Panoramic Effect A simple panoramic effect describes a trend in lateral movement over a series of images, which is why most of the analyzed movement is in one direction. This will also preferentially cover the situation where the majority of the scene has consistent motion, and no stationary objects are detected in the foreground. A simple panoramic example can be detected as the main object that has a non-zero motion vector. The results of a non-simple panoramic effect is that a positive motion vector is generated, if the scene is moving to the right (or panoramic effect on the left). In this case, the image to the right eye will be delayed. Similarly, a negative motion vector is generated if the scene is moving to the left (or right panoramic effect). In this case, the image to the left eye will be delayed. c) Complex Panoramic Effect • A complex panoramic effect differs from a simple panoramic effect in that there is significant vertical movement in the image. Therefore, in the preferred embodiment, to minimize the vertical disparity between the sequences of the stereoscopic image pairs, field delay is not applied and only object processing is used to create a stereoscopic effect. The history of the field delay is stored to maintain continuity with the new lateral movement. d) Moving Object On Stationary Second Plane An object moving on a stationary second plane is simply the situation by which most of a scene has no movement, and one or more medium-sized objects and movements are staged. This situation also results in a positive motion vector if most objects are moving to the right, and a negative motion vector if most objects are moving to the left. A vector of positive movement produces a delay to the right eye, and a vector of negative movement produces a delay to the left eye.
In the case where the motion vectors of the objects in the scene are not consistent, for example, the objects that move to the left and right in the same scene, then there is the counter-movement and correction techniques can be applied. 3D inverted. e) Close-up Object on Second Plane in Motion A close-up object on a second plane in motion refers to the situation where the majority of the scene has movement, an object that has a different movement is in the scene, for example, a camera that follows a person walking. A background object is detected as a main object of the non-zero motion vector (that is, a second panorama plane) behind a zero-sized object with zero or opposite motion vectors to the main object, or a vector main object zero in front of the non-zero vector minor objects that extend over the entire field (that is, a large stationary object that fills most of the field, but a panoramic object is still visible behind it).
A decision must be made as to whether the foreground object should be of priority given in the generation of motion parallax, or if the second plane is given priority. If the second plane contains a large variation in depth (for example, trees), then the motion vectors are assigned as if a simple panoramic effect were occurring. If the second plane contains little variation in depth (eg, a wall) then a motion vector that is negative anti-parallel is assigned. When the second plane contains a large variation in depth, and a motion vector is assigned to the scene as per the simple panoramic effect methods, then the foreground object will be in inverted 3D, and appropriate correction methods must be applied . f) No Movement If no movement is detected such that the motion vectors are completely zero, or alternatively the largest motion object is considered too small, then the field delay will be set to zero. This situation can occur when only random motion vectors are determined over noise, or where no movement information is available, for example, during a panoramic effect through the 'blue sky'.
MODULE 3 ^ 3D GENERATION Once the images are analyzed, then they can be processed to create the pairs of stereoscopic images. When you see a real world scene, both eyes see a slightly different image. This is called retinal disparity. This in turn produces stereopsis or depth perception. In other words, it looks stereoscopically by making each eye see a slightly different image of the same scene. On the other hand, parallax is defined as the amount of horizontal or lateral displacement between the images that are perceived by the viewer as retinal disparity. When a pair of stereoscopic images is created, a three-dimensional scene is observed from two points of view displaced horizontally. The present invention uses a number of image and object processing techniques to generate pairs of stereoscopic images from monoscopic images. These techniques include: 1) Motion Parallax. 2) Forced Parallax (Lateral Displacement). 3) Parallax areas. 4) Image Rotation Around the Y-axis. 5) Processing of objects. 1) Movement Parallax When a scene is moving from right to left, the right eye will observe the scene first while the left eye will receive a delayed image and vice versa for a scene moving in the opposite direction. The faster the movement, the smaller the delay between the images for both eyes. This is known as motion parallax and is a main depth reference point. Therefore, if there is lateral movement in a scene, when creating a delay between the images in the eyes a stereoscopic effect will be perceived. a) Field Delay Calculation Once the nature of the motion of an image has been analyzed, and a complete motion vector is determined, then the field delay, required, can be calculated. Preferably, the calculated field delay is averaged with previous delays to filter "noisy" values and also to prevent the field delay from changing too quickly. As noted above, the faster the movement the smaller the delay between the images to each eye. Therefore, smaller field delay values are used in scenes with large motion vectors, whereas larger delays are used in scenes with little lateral movement. That is, there is an inverted relationship in the preferred embodiment between the delay and the momentum. When a scene change is determined, the history of the field delays must be readjusted to zero, as if the movement had not previously occurred. In the first motion detection when a non-zero field delay is calculated, while the history of the field delays is still zero, the full history of the field delay is adjusted to the calculated field delay. This allows the system to immediately display the correct field delay when movement is detected. b) Implementation of the Field Delay Motion parallax can be generated in the physical equipment and computer program by storing scanned images in the memory. Preferably, the digitized images can be stored in a buffer and a single input indicator used with two output indicators, one for the left eye image and one for the right eye image. The guide eye image memory indicator is held at or near the memory indicator of the current input image while the image memory indicator of the delayed eyes is further adjusted downward from the buffer to produce an output delayed Many images can be stored, up to 8-10 video fields is typical in video applications. The delay is dependent on the speed of the movement analyzed in the image. The maximum field delay is when there is a minimum movement 2) Forced Parallax (Lateral Displacement) A forced parallax can be created by entering a lateral displacement between: i) An exact copy of an image and of itself. ii) The two fields of a video frame. iii) Two frames of a film sequence iv) A transformed copy of an image and its original There is a negative side shift when moving the left image to the right and the right image to the left for the same amount (it establishes a depth of the field that begins from the plane of screen and that it persecutes in front of this one) and a positive lateral displacement when moving the left image to the left and the right image to the right by the same amount (establishes a depth of field that starts from the plane of the screen and backs up behind it). You can reduce the forced parallax to improve the stereoscopic effect for a stationary object, in front of a panoramic effect where the object is "placed" closer to the plane of the screen and the second plane is "pushed back" from the plane of the object definite. 3) Paralleling Zones Because the majority of the scenes are seen with the second plane in the upper part and the foreground in the background, it is possible to accentuate the depths of a scene by "triangulating" the forced parallax. This is done by laterally displacing the upper part of the image rather than the background of an image, thereby accentuating the depth of the backward front observed in a scene. Another technique is to use a combination of parallax of motion and forced parallax in different parts of the image. For example, by dividing the image vertically in half and applying different parallax shifts on each side, a scene as viewed from a moving train below a track on the track has the correct stereoscopic effect. Otherwise, one side will always appear in inverted 3D 4) Image rotation around the Y-axis When an object is moving towards the viewer in a real-world scene, the object is rotated slightly in view for each eye. The spin effect is more pronounced as the effect moves closer. Translating this rotation in the pairs of stereoscopic images defines the effect as follows: i) Movement towards the viewer. - The left image is rotated vertically around its central axis in a counter-clockwise direction and the right image in a clockwise direction. ii) Movement away from the viewer. - The left image is rotated vertically around its central axis in the clockwise direction and the right image in the counterclockwise direction. Therefore, by the rotation of the image, the perspective of the objects in the image is changed selectively so that depth is perceived. When this technique is combined with forced parallax for certain scenes, the combined effect provides stereoscopic, very powerful depth reference points.
) Processing of objects The processing of objects is performed to further improve the stereoscopic effect, particularly fixed images, by separating the objects and the background, so that these items can be processed independently. It is most effective when the objects are large, of small number and occupy different levels, of depth through the depth of the field. A database can be used for object mapping and object tracking to set trends so that an object can be "cut" digitally from its background and appropriate measurements are taken to improve the stereoscopic effect. Once the processing has taken place, the object "sticks" back to the same position in the second plane, again. This can be called the "cut and paste" technique and is useful in the conversion process.
By integrating the processes of object mapping, tracking, cutting and gluing, a powerful tool is available to handle and process objects and process the second. flat . Another object processing technique is the stratification of objects that defines an independent depth module for each object in motion. The object can then be placed anywhere in an image because the fill detail in the background has been defined when the object was not in that position. This is not generally possible with a fixed object unless the filling of the second plane is interposed. A more important issue in stereoscopic conversion is the correction of accommodation / convergence and inverted 3D imbalances that cause discomfort to the viewer. The processing of objects in the preferred mode allows corrections to this problem as well. a) Distortion and Mesh Formation. - This object processing technique allows an object to be cut and pasted into a distorted mesh to improve depth perception. By distorting an object in the image of the left eye to the right and by distorting the same object in the image of the right eye to the left, thus creating the object parallax, the object can be made to appear much closer to a viewer when using a stereoscopic display device. b) Object barrel. - This technique is a specific form of mesh distortion and refers to a technique of cutting an object from the image and winding it in a half barrel to see it placed only. This makes the object appear to have depth by making the central portion of the object appear closer than the edges of the object. c) Improvement of the Edge of the object. - When emphasizing the edges of an object there is greater differentiation between the second plane or other objects in the image. The stereoscopic effect is improved in many applications by this technique. ^ _ d) Improvement of the brightness of the object. - In any image, the eye is always drawn to the largest and brightest objects. By modifying the luminance of the object, the object can be emphasized more on the second plane, improving the effect is ereoscopic. e) Rotation of the object Around the Y-axis.
- The rotation of the object around the Y-axis refers to a process similar to that of the rotation of the image around the Y-axis, except that this time the rotation does not rotate in the object only. If the object in the pair of stereoscopic images is "cut" from its second plane and rotated slightly, the change in the perspective generated by the rotation, such as depth, is perceived. 3D OPTIMIZATION 1) Benchmarks or Edges When a normal television or video monitor is used to display stereoscopic images the eye continuously observes the edge of the monitor or screen and these are perceived as a reference point or fixation point for all the depth perception. That is, all objects are perceived at a depth behind or in front of this reference point. If the edge of the monitor is not easily seen due to poor ambient lighting or due to its dark color, then this reference point can be lost and the eye can continuously search for a fixation point in the 3D domain. Under prolonged stereoscopic vision, this can cause fatigue to the eye and a diminished perception of depth. A display system on the front or rear projection screen can also suffer from the same problems. Therefore, the present invention also preferably defines a common edge or reference point within a viewed image. Ideally, the reference plane is adjusted to the screen level and the entire depth behind this level is perceived. This has the advantage of improving the stereoscopic effect in many images. This reference point can be a simple video edge or reference graphic and, for example, it can be of the following types: i) A simple, colored video edge around the perimeter ^ of the image. ii) A colored, complex video edge consisting of two or more concentric edges that may have opaque or transparent sections therebetween. For example, a mesh edge 2-3 centimeters wide or a wide outer edge with two thin inner edges. iii) A partial edge that can occupy any edge, or either of two horizontal or vertical edges. iv) A logo or other graphic located at some point within the image. v) An image within an image. vi) A combination of any of the above. What is essential in this modality is that in the eyes of the viewer a point of reference is disproportioned by which the depth of the objects in the image can be perceived. If an edge or graph is added at the 3D generation level, then it can be specified to provide a reference point at a particular depth when creating left and right edges that move laterally to each other. This allows the fixation reference point ^ to move in space to a point anywhere behind or in front of the screen level. Edges or graphics defined without parallax for the left and right eyes will be perceived at the level of the screen. This is the preferred mode of the present invention. An image border or reference chart can be inserted at the 3D generation point or it can be defined externally and set in synchronization at the output of the stereoscopic image for display. This image border or reference graph can be black, white or colored, simple or in pattern, opaque, translucent or transparent to the second plane of the image, or it can be static or dynamic. While a static edge is appropriate in most cases, in some circumstances a dynamic movement border can be used for the improvement of movement. 2) Parallax adjustment Control of Depth Sensitivity Stereoscopic images' viewed through a stereoscopic display device automatically define a depth range (called depth acuity) that can be increased or decreased by modifying the time and amount of parallax applied to the image or object. It has been found that different viewers have comfort levels of stereoscopic vision, variables based on the depth interval or amount of stereopsis defined by the pairs of stereoscopic images. That is, while some viewers prefer a pronounced stereoscopic effect with a greater depth interval, others prefer an image with minimal depth. To adjust the level of depth sensitivity and comfort of vision, many techniques can be used, specifically: i) Vary the amount of motion parallax by varying the field delay. ii) Vary the amount of forced parallax to an image. iii) Vary the amount of parallax applied to the objects. By reducing the maximum level of parallax, you can reduce the depth range, improving the comfort of vision for those with faculties of perception that have greater sensitivity, or stereoscopy. 3) Parallax Smoothing Parallax smoothing is the process to maintain the total amount of parallax (motion parallax plus forced parallax) as a continuous function. Changes in the field delay for specific motion types, this, simple panoramic effect and foreground object movement, cause discontinuities in the amount of motion parallax produced, which are seen as "jump" in the stereoscopic images by the viewer. Discontinuities only occur in the image produced for the trailing eye, since the guiding eye is presented with a non-delayed image. These discontinuities can be compensated by adjusting the forced parallax or object parallax in an equal or opposite direction for the trailing eye, thereby maintaining a continuous total parallax. The forced parallax or object parallax then smoothly adjusts back to its normal value, ready for the next change in field delay. The adjustments made to parallax forced by parallax smoothing are a function of changing the field delay, movement type and motion vector. To implement parallax smoothing, forced parallax for the left and right eye images must be adjusted independently. 4) Parallax Modulation The forced parallax technique to create a stereoscopic effect can also be used to moderate the amount of stereopsis detected by the viewer. This is done by varying the setting of the forced parallax between a minimum and maximum limit over a short time such that the perceived depth of an object or image varies over time. Ideally, forced parallax is modulated between its minimum and maximum settings every 0.5 to 1 second, this allows a viewer to adjust their level of stereoscopic sensitivity.
) Synthesis of Movement When creating the pseudo-movement, by randomly moving the second plane in small, undetectable increments, the perceived depth of the objects in the foreground is emphasized. The objects in the foreground are "cut" from the second plane, the second plane is altered in a pseudo-random way by one of the later techniques and then the object in the foreground is "pasted" back to the second plane ready for display. Any of the following techniques can be used: i) Luminance values are varied on a pseudo-random basis. ii) Chrominance values are varied on a pseudo-random basis. iii) add pseudo-random noise to the second plane to create a movement. 6) Analysis and correction of inverted 3D The inverted 3D occurs when the order of depth of the objects created by it parallax is perceived to be different from that which corresponds to the order of depth in the real world. This leads in general to the discomfort of the viewer and must be corrected. When monoscopic images are converted into pairs of stereoscopic images, inverted 3D can be produced by: i) counter-movements, objects that move to the left and right in the same image. ii) objects and background that moves in different directions. iii) Many objects that move at variable speeds. Inverted 3D is corrected by analyzing the nature of the movement of the objects in an image and then manipulating each object individually using techniques and mesh distortion so that the object parallax corresponds to the expected visual perception norms. 7) Miscellaneous techniques By modifying the perspective of an object within an image and by improving many of the lower depth reference points, the stereoscopic effect can be emphasized. The later technique operates all using the technique of "cut and paste". That is, a close-up object is "cut", improved, and then "pasted" back into the background. a) Shadows - The shadowing gives an object perspective. b) Close-up / background. - By blur the second plane, through blur or fogging, an object in the foreground can be emphasized, while the blur of the foreground object can emphasize the depth of the second plane. c) Improvement of Edge. - The edges help to differentiate an object from its second plane. d) Texture Correlation. - Help differentiate the object from the second plane.
MODULE 4. - 3D MEDIUM (TRANSMISSION AND STORAGE) Regarding Module 1, Modules 4 and 5 are not essential to the present invention. Module 4 provides the transmission and / or storage of stereoscopic images. The transmission medium can be adapted for a particular application. For example, the following can be used: 1) Local Transmission. - can be via coaxial cable 2) Network television transmission. It can be via. i) Cable. ii) Satellite. iii) Terrestrial 3) Digital Network. - INTERNET, etc. 4) Storage of Stereoscopic Images (3D) An image storage medium can be used for the storage of image data for subsequent transmission or display and may include: i) Analog Storage. - Movie tape, etc. ii) digital storage. - Laser disk, Hard disk, CD-ROM, Magneto-optical disc, DAT, Digital Video Cartridge (DVC), DVD.
MODULE 5. - EXHIBITION OF 3D As for the transmission medium, the display medium may be dependent on the requirements of the application and may include: 1) Decoding device A decoding device by definition is a small device of electronic components which receives, decodes, provides auxiliary interfaces and finally makes transmissions to suit the application. It can incorporate the following: a) Video or RF receiver. b) Stereoscopic decoder (3D) to provide separate transfers of left and right images and to devices mounted on the head or other stereoscopic displays where separate video channels are required. c) Improvement of the Resolution. - Line Dubbing / Pixel Interpolation. d) Synchronization of the Sealing or Sequential Lenses. e) Stereoscopic depth sensitivity control circuit set. f) Auxiliary interface. - remote control with features such as a 2D / 3D switch and depth control. g) Audio interface. - Audio transfer, headset connection. h) Decoding of the access channel, cable and pay television applications, i) RF video transfers. 2) Stereoscopic displays The use of special lenses or gears to provide separate images to the left and right eyes that includes: a) Polarization lenses. - Linear and Circular Polarizers. b) Anaglyphic lenses. - Colored lenses. - red / green, etc. c) LCD shutter lens. d) Sequential Color Lenses. e Head Mounted Devices (HMD). - Top gear equipped with two miniature video monitors (one for each eye), VR headsets. 3) Autostereoscopic Exhibits a) Display systems based on a retroreflective video projector / screen. b) Volumetric display systems. c) Display systems based on lenticular lenses. d) Display systems based on a holographic Optical Element (HÓE).
PREFERRED MODE In summary, the present invention provides in a preferred embodiment a system that is capable of introducing sequences of monoscopic images in a digital format, or in an analog format in which case an analog-to-digital conversion process is comprised. These image data are then subjected to an image analysis method, by which the monoscopic images are compressed, if this is required for the particular application. By comparing the blocks of pixels in an image, with corresponding blocks in an adjacent image, and by omitting the minimum mean square error for each block, one can determine, the movement within the image. After motion detection, regions of an image are identified for similar characteristics, such as image brightness, color, or motion, pattern continuity, and edge. Then, the data is subjected to motion analysis in order to determine the nature of the movement in the image. This movement analysis takes the form of the determination of the direction, speed, type, depth and position of any movement in the image. This movement is then categorized into a number of categories that include whether the movement is a complete scene change, a simple panoramic effect, a complex panoramic effect, an object moving in a stationary second plane, a stationary object in front of a second plane in motion, or if there is no movement completely. The additional actions are then determined based on these categories to convert the monoscopic images into pairs of appropriate stereoscopic images for addition in a stereoscopic display device. In the preferred mode, once the monoscopic images are analyzed, if a scene change or a complex panoramic effect is detected, no further analysis of that particular scene is required, instead the Field Delay and the history of the Field Delay both are readjusted to zero. Then an object detection process is applied to the new scene in order to try and identify objects within that scene. Once these objects are identified, then the processing of the object takes place. No objects are identified, then the image is passed to the additional processing using forced parallax and 3D optimization. If the catagorized movement during the image analysis is not a scene change, then additional analysis of that scene is required. If the additional analysis of that scene results in the movement being categorized as a simple panoramic effect, then it is necessary to apply a field delay in accordance with the principles of motion parallax. Then, it goes to additional processing. If the movement is not categorized as a simple panoramic effect, but rather as a moving object in a stationary second plane, then again a Field Delay will have to be applied according to the principles of motion parallax. In this regard, once the parallax of the movement has been applied, it is necessary to consider whether the objects have a uniform direction. If the objects move in a uniform direction, then additional processing is carried out at a later stage. If the objects do not have a uniform direction, then it is necessary to perform additional processing of the object on the selected objects within that scene to correct the inverted 3D effect. This can be achieved through the use of distortion and mesh formation techniques. If the movement is categorized as being a stationary object in a second plane in motion, then it is necessary to consider whether the second plane has a large variation in depth. If not, then a Field Delay is applied within the object that has priority using the principles of motion parallax. However, the second plane does not have much variation in depth, so a field delay is applied with the second plane having priority as opposed to the object, again using the parallax principles of motion. In this case, then it is also necessary to perform additional processing of the object in the foreground object to correct the inverted 3D effect before it is passed to further processing. If a movement is not detected, then it is considered below if an object in the scene was found from some previous movement. If this is the case, then the object is processed in that selected object. If not, then an object detection process is applied to that particular scene in order to try to identify some object in it. If an object is identified, then object processing is performed on that particular object, otherwise forced parallax and 3D Optimization are performed. Where object processing is required, objects are identified, marked and tracked and then processed by using mesh information distortion techniques, barrel formation of the object, improvement of the edge, modification of the brightness and rotation of the object. In all cases once the movement has been categorized and the primary techniques for converting to stereoscopic images have been applied, then an additional amount of parallax or lateral displacement called forced parallax is applied to the image. It is noted that in the preferred embodiment, the forced parallax is applied to each image, not only for purposes of depth stabilization, but to provide an underlying stereoscopic effect that all images are seen as having depth from behind or from the front of the image. reference plane of the stereoscopic display device, generally the front of the monitor screen. The advantages of applying forced parallax are that the system is better able to cope with changes in the detected movement category without causing sudden changes in the depth perception of the spectators.Once the forced parallax has been applied to the image, the image is passed for 3D Optimization. Again, this is not necessary in order to see a stereoscopic image, however, Optimization improves the perception of depth of the image by the viewer. 3D Optimization can take a number of forms including the addition of reference points or edges, parallax modulation, parallax smoothing and parallax adjustment to alter the depth sensitivity of any particular viewer. The image can also be optimized by modifying the values of illuminance or chrominance, in a pseudo-random way so that the movement of the second behind the objects in the foreground can be observed so that the perception of depth is improved. It is also possible to analyze the inverted 3D so that spectator fatigue is minimized. Additional techniques such as shading, fogging or blurring of the first and second planes and improvement of the edge of the image can also be carried out at this stage. Once the image has been optimized then it is transmitted to the appropriate display device. This transmission can take a number of forms including cable, coaxial cable, satellite or any other form of signal transmission from one point to another. It is also possible that the image can be stored before it is sent to a display device. The display device can take a number of forms, and only needs to be appropriate for the application in mind. For example, it is possible to use existing video monitors with a decoder device in order to separate the left and right images, increase the scanning speed and synchronize the viewing lenses. Alternatively, dedicated stereoscopic display apparatus incorporating the use of lenses or head gears can be used to provide the stereoscopic images or alternatively, a self-stereoscopic display device can be used. It is contemplated that the present invention will have application in theaters, cinemas, video galleries, cable or network TV, in the educational area, particularly in the multimedia industry and in many other areas such as theme parks and other entertainment applications.
It is noted that in relation to this date, the best method known by the applicant to carry out the present invention is that which is clear from the present description of the invention. Having described the invention as above, the content of the following is claimed as property:

Claims (68)

  1. CLAIMS 1. A method to convert monoscopic images to see in three dimensions, characterized because it includes the steps of: receiving the monoscopic images; analyze the monoscopic images to determine the characteristics of the images; process the monoscopic images based on the determined characteristics of the image; transfer the processed images to suitable stereoscopic and / or storage display systems; wherein the analysis of the monoscopic images to determine the characteristics includes determining the movement of the monoscopic images by: dividing each image into a plurality of blocks, wherein corresponding blocks in an adjacent image are misaligned horizontally and / or vertically; and ___ compare each block with the corresponding blocks to find the minimum mean square error and thus the movement of the block.
  2. 2. A method according to claim 1, characterized in that the processing of monoscopic images uses at least one of the following methods: motion parallax, forced parallax, parallax zones, image rotation and / or object processing.
  3. 3. A method according to claim 1 or claim 2, characterized in that the monoscopic image is digitized before any analysis or processing is performed.
  4. 4. A method according to any of claims 1 to 3, characterized in that the monoscopic image is compressed before any analysis.
  5. 5. A method according to any of claims 1 to 4, characterized in that the monoscopic image can be scaled before any analysis.
  6. 6. A method according to claim 5, characterized in that the scale factor by which the monoscopic image is scaled is dependent on the digital video resolution of each image
  7. 7. A method according to any of claims 1 to 6, characterized in that it generalises first and second successive images for continuity before determining the characteristics of the image.
  8. 8. A method according to claim 7, characterized in that continuity is determined by comparing the average values of luminances between the first and second successive images.
  9. 9. A method according to claim 8, characterized in that continuity is not assumed when the difference in average luminance values exceeds 30.
  10. 10. A method according to any of the rei indications 7 to 9, characterized in that the few upper lines of the successive images are compared to aid in the determination of the continuity.
  11. 11. A method according to any of claims 1 to 10, characterized in that image processing where continuity is not determined includes producing a field delay to an eye such that the image lacking continuity is seen by an eye before it it is seen by the other eye of a spectator.
  12. 12. A method according to any of claims 1 to 11, characterized in that during the analysis of the monoscopic images, the objects within the images are defined to assist during the processing.
  13. 13. A method according to claim 12, characterized in that the objects are identified by comparing the luminance value, chrominance value __, movement vector and / or image detail of adjacent pixels or groups of pixels.
  14. 14. A method according to any of claims 1 to 12, characterized in that during the analysis of the monoscopic image, the movement of the objects within the images is determined to aid processing.
  15. 15. A method according to claim 14, characterized in that the movement of the image and / or objects is categorized into one of a predetermined range of motion categories.
  16. 16. A method according to claim 14, characterized in that a motion vector is defined for each image based on a comparison of the motion nature detected with the predefined categories of motion ranging from no movement to a complete change of scene.
  17. 17. A method according to claim 15 or claim 16, characterized in that the movement categories include: scene change, simple panoramic effect, complex panoramic object, moving object, moving second plane, and no movement-.
  18. 18. A method according to any preceding claim, characterized in that any block without details is not compared with the corresponding blocks.
  19. 19. A method according to claim 7, characterized in that continuity is not assumed when the comparison of most of the blocks with the corresponding blocks has resulted in large error values.
  20. 20. A method according to any of claims 1 to 19, characterized in that the processing of the image includes the use of motion parallax by introducing a field delay such that an eye of a viewer sees the image before the other eye of the viewer.
  21. 21. A method according to claim 20, characterized in that the amount of movement is inversely proportional to the field delay.
  22. 22. A method according to claim 20 or claim 21, characterized in that the field delays are stored, and the field delay for each new image is averaged against the previous image delays
  23. 23. A method according to claim 22, characterized in that the stored field delays are erased when continuity is not detected.
  24. 24. A method according to any of claims 1 to 23, characterized in that the processing of the image includes the use of forced parallax by introducing a lateral change through the displacement of the left and right eye images. ^ _
  25. 25. A method according to any of claims 1 to 24 characterized in that the processing of the image includes the use of parallax zones when introducing a lateral shift greater than a portion of the image.
  26. 26. A method according to claim 25, characterized in that an upper portion of the image is laterally displaced by an amount greater than a portion of the background of the image.
  27. 27. A method according to claim 25, characterized in that a different parallax shift is applied to the left side of the image as opposed to the right side of the image.
  28. 28. A method according to any of claims 1 to 27, characterized in that the processing of the image includes a combination of the forced parallax and the moving parallax in various parts of the image.
  29. 29. A method according to any of claims 1 to 28, characterized in that the processing of the image includes rotation of the left and right eye images around the axis of the and in an equal amount in the opposite direction.
  30. 30. A method according to any of claims 1 to 29, characterized in that the processing of the image includes the use of at least one of the following object processing techniques: distortion and mesh formation; barrel formation of the object - improvement of the edge of the object; improvement of the brightness of the object; and / or rotation of the object.
  31. 31. A method according to any of claims 1 to 30, characterized in that the processed image is further processed by applying a final forced parallax to the processed image.
  32. 32. A method according to claim 31, characterized in that the degree of forced parallax is determined by the amount of parallax added during image processing, such that the total parallax added during processing and the forced parallax, is substantially equal to total parallax of the adjacent images.
  33. 33. A method according to claim 31 or claim 32, characterized in that the final forced parallax degree is modulated between predetermined minimum and maximum settings over a predetermined time frame.
  34. 34. A method according to any of claims 1 to 33, characterized in that the processed image is optimized to further improve the processed images before transferring the images to the stereoscopic display and / or storage system.
  35. 35. A method according to any of claims 1 to 34, characterized in that a reference point is added to the processed image.
  36. 36. A method according to claim 35, characterized in that the reference point is at least one of: an edge around the perimeter of the image; a plurality of concentric edges; a partial edge; a logo; and / or an image.
  37. 37. A method according to any of claims 1 to 36, characterized in that the amount of depth added to the monoscopic images during the processing of the images can be adjusted in response to the preference of the viewers
  38. 38. A method according to any of claims 1 to 37, characterized in that the second plane of the image moves randomly in small increments that are not consciously perceived by the viewer.
  39. 39. A method according to any of claims 1 to 38, characterized in that the image is tested for inverted 3D and the objects are manipulated individually to compensate for any inverted 3D.
  40. 40. A method according to any of claims 1 to 39, characterized in that the techniques of cutting and gluing are used to further emphasize the stereoscopic effect.
  41. 41. An image conversion system for converting monoscopic images for viewing in three dimensions, characterized in that it includes: an input means adapted to receive monoscopic images; a means of preliminary analysis to determine if there is any continuity between a first image and a second image in the sequence of monoscopic images; a secondary analysis means to receive monoscopic images that have a continuity, and analyze the images to determine at least one of the speed of direction in movement, or the depth, size and position of the objects, where the analysis of the monoscopic images to determine a movement includes the steps of: dividing each image into a plurality of blocks, wherein the corresponding blocks in an adjacent image are misaligned horizontally and / or vertically, and comparing each block with the corresponding blocks to find the minimum square root mean error and in this way the movement of the block; and a first processing means for processing the monso-opic images based on the data received from the preliminary analysis means and / or the secondary analysis means.
  42. 42. An image conversion system according to claim 41, characterized in that it also includes a transmission means capable of transferring the processed images to a stereoscopic display system or a storage system.
  43. 43. An image conversion system according to claim 41 and rei indication 42, characterized in that the first processing means processes the images by using at least one of: Motion parallax, forced parallax, parallax zones, rotation of the image or processing of objects.
  44. 44. An image conversion system according to any of claims 41 to 43, characterized in that the second processing means is provided to further process the images received from the first processing means.
  45. 45. An image conversion system according to claim 44, characterized in that the second processing means uses forced parallax, to further process the image.
  46. 46. An image conversion system according to any of claims 41 to 45, characterized in that a third processing means is provided for optionally enhancing the images before transmitting the converted images to stereoscopic display device.
  47. 47. An image conversion system according to claim 46, characterized in that the third processing means improves the images by using at least one of: reference points, parallax adjustment, parallax smoothing, parallax modulation, motion synthesis, 3D inverted correction or cutting and pasting techniques.
  48. 48. A system according to any of claims 41 to 47, characterized in that the input means is further adapted to digitize the monoscopic images.
  49. 49. A system according to any of claims 41 to 48, characterized in that it also includes a compression means adapted to compress the monoscopic images before analysis by the first analysis means.
  50. 50. A system according to any of claims 41 to 49, characterized in that it also includes a means of scaling adapted to scale the image, monoscopic before analysis po? the first means of analysis.
  51. A system according to claim 50, characterized in that the scale factor by which the monoscopic image is scaled is dependent on the digital video resolution of each image.
  52. 52. A system according to any of claims 41 to 51, characterized in that the preliminary analysis means is able to determine objects within the images.
  53. 53. A system according to any of claims 41 to 52, characterized in that the preliminary analysis means is able to determine the movement of the images and / or the movement of objects within the images.
  54. 54. A system according to claim 52 or claim 53, characterized in that the preliminary analysis means is able to categorize the movement into one of a predetermined range of movement categories.
  55. 55. A system according to claim 58, characterized in that the movement categories include at least one of scene change, simple panoramic effect, complex panoramic effect, moving object, second plane in movement, and no movement.
  56. 56. A system according to any of claims 41 to 55, characterized in that it also includes: means for controlling the level of depth added to the monoscopic images.
  57. 57. A system according to any of claims 41 to 56, characterized in that it also includes a means for adding a reference point to the processed image.
  58. 58. A system according to any of claims 41 to 57, characterized in that it also includes a means to optimize the stereoscopic image to further improve the stereoscopic effect.
  59. 59. A system for converting monoscopic images to see in three dimensions, characterized in that it includes: a first module adapted to receive a monoscopic image; a second module adapted to receive the monoscopic image and analyze the monoscopic image to create the date of the image, wherein the analysis of the monoscopic image includes determining the movement of the monoscopic images by: dividing each image into a plurality of blocks where the corresponding blocks in an adjacent image are misaligned horizontally and / or vertically, and compare each block with the corresponding blocks to find the minimum mean square error and thus the movement of the block; a third module adapted to create pairs of stereoscopic images from the monoscopic image using at least one predetermined technique selected as a fusion of the image data; _ a fourth module adapted to transfer the pairs of stereoscopic images to a stereoscopic display medium; a fifth module consisting of a stereoscopic display medium.
  60. 60. A system according to claim 59, characterized in that the first module is further adapted to convert any analog image into a digital image.
  61. 61. A system according to claim 59 and 60, characterized in that the second module is adapted to detect any object in a scene and expects a determination as to the speed, direction and movement of any of these objects.
  62. 62. A system according to any of claims 59 to 61, characterized in that the image is compressed before any analysis.
  63. 63. A system according to any of claims 59 to 62, characterized in that the third module further includes an optimization step for further improving the stereoscopic image pairs before transmitting the stereoscopic image pairs to the stereoscopic display means.
  64. 64. A system according to any of claims 59 to 63, characterized in that the system can be suspended for further processing between the modules.
  65. 65. A system according to any of the rei indications 59 to 64, characterized in that the fourth module also includes a storage means for storing pairs of stereoscopic images to be displayed in the stereoscopic display medium at a later time.
  66. 66. A stereoscopic display system that includes the provision of a point of reference for the viewer.
  67. ^ 67. A method substantially as described hereinabove, with reference to the accompanying drawings.
  68. 68. A system substantially as described hereinabove, with reference to the accompanying drawings. / *% r APPARATUS FOR THE PROCESSING OF IMAGES METHOD AND SUMMARY OF THE INVENTION An image conversion system for converting monoscopic images to see in three dimensions including: an input means (1) 5 adapted to receive the monoscopic images; a means of preliminary analysis to determine if there is any continuity between a first image and a second image of a sequence of monoscopic images; a means of secondary analysis (2) 10 to receive monoscopic images that have a continuity, and analyze the images to determine the speed and direction of the movement, and the depth, size and position of the objects; a first processing means (3) for processing 15 the monoscopic images based on the data received from the preliminary analysis means or the secondary analysis means, a second processing means capable of further processing images received from the first medium of 20 processing; a transmission means (4) capable of transferring the processed images to a stereoscopic display system (5). 25
MXPA/A/2000/002201A 1997-09-02 2000-03-02 Image processing method and apparatus MXPA00002201A (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
POPO8944 1997-09-02

Publications (1)

Publication Number Publication Date
MXPA00002201A true MXPA00002201A (en) 2001-03-05

Family

ID=

Similar Documents

Publication Publication Date Title
US6496598B1 (en) Image processing method and apparatus
USRE39342E1 (en) Method for producing a synthesized stereoscopic image
US9485487B2 (en) Method and apparatus for generating a signal for a display
EP2188672B1 (en) Generation of three-dimensional movies with improved depth control
JP4427032B2 (en) Image conversion and encoding technology
US8514275B2 (en) Three-dimensional (3D) display method and system
Devernay et al. Stereoscopic cinema
US20130057644A1 (en) Synthesizing views based on image domain warping
US20120176481A1 (en) Processing image data from multiple cameras for motion pictures
NZ589170A (en) Stereoscopic editing for video production, post-production and display adaptation
CN103563363A (en) Automatic conversion of a stereoscopic image in order to allow a simultaneous stereoscopic and monoscopic display of said image
JP2009500963A (en) System and method for capturing visual and non-visual data for multi-dimensional video display
JP6207640B2 (en) 2D image stereoscopic display device
KR100503276B1 (en) Apparatus for converting 2D image signal into 3D image signal
MXPA00002201A (en) Image processing method and apparatus
KR101433082B1 (en) Video conversing and reproducing method to provide medium feeling of two-dimensional video and three-dimensional video
AU8964598A (en) Image processing method and apparatus
CN113676731A (en) Method for compressing VR video data
Didier et al. Use of a Dense Disparity Map to Enhance Quality of Experience Watching Stereoscopic 3D Content at Home on a Large TV Screen
Millin et al. Three dimensions via the Internet
Hallpike The role of ocula in stereo post production
PIRTLE THE FAR DARK SHORE: SCULPTURAL DEPTH AND RELIEF IN DIGITAL STEREOSCOPIC VIDEO AND COMPOSITED STEREOSCOPIC COMPUTER GRAPHICS