EP2156410A1 - Verfahren zum darstellen von bildobjekten in einem virtuellen dreidimensionalen bildraum - Google Patents
Verfahren zum darstellen von bildobjekten in einem virtuellen dreidimensionalen bildraumInfo
- Publication number
- EP2156410A1 EP2156410A1 EP08734459A EP08734459A EP2156410A1 EP 2156410 A1 EP2156410 A1 EP 2156410A1 EP 08734459 A EP08734459 A EP 08734459A EP 08734459 A EP08734459 A EP 08734459A EP 2156410 A1 EP2156410 A1 EP 2156410A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- image object
- views
- image
- virtual
- viewer
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/20—Perspective computation
Definitions
- the invention relates to a method for displaying image objects in a virtual three-dimensional image space, wherein the position of a viewer of the image object is detected for the method.
- VR virtual reality
- the term "virtual reality” refers to the representation and simultaneous perception of reality and its physical properties in a mostly in real-time computer-generated interactive virtual environment.
- the technical possibilities for example in terms of a comprehensive perception of physical properties, are still limited
- their use in aircraft simulators in the training of pilots, in the creation of virtual prototypes in industry, in the performance of ergonomic tests, for the visualization of buildings, in medical diagnostics, in the simulation of operations, hard-to-reach virtual visits, edutainment, or the like.
- HMDs Head-Mounted Displays
- CAVEs Camera Automatic Virtual Environments
- a presentation with an experienced spatial depth is helpful.
- two views of an object are generated and displayed from slightly different positions (stereo projection).
- the distance between the two positions often corresponds to the distance of the eyes of the beholder.
- the two views must be fed to the correct eye become.
- active and passive procedures include, for example, shutter glasses, which can be switched transparent and dark at high speed. These are used in conjunction with a monitor that alternately displays an image for the left eye and an image for the right eye.
- the glasses are synchronized with the monitor, the correct image is transmitted to each eye.
- Passive techniques include anaglyph and polarization techniques in which two views of a small spaced image are superposed in an image. Using color or polarized filter glasses, these image views can be separated again.
- An autostereoscopic monitor which allow the user to perceive spatial depth of the objects presented without the use of special aids such as spectacles or the like.
- An autostereoscopic monitor has a very fine image matrix, in front of which an optical means, usually in the form of lenticular or parallax barrier systems, is directly attached. Due to the special geometry of the optical means is achieved that certain pixels of the image matrix are emitted in a defined spatial direction. By selectively controlling the pixels, images for the left and the right eye can be displayed simultaneously and independently of each other. The quality of the three-dimensional impression is the higher, the better the two views can be perceived separately. This can be achieved by limiting the solid angle in which a three-dimensional perception is possible.
- tracking systems which continuously detect the position of the viewer.
- the pixels on the image matrix or the position of the optical means are readjusted by slight shifting, so that the spatially narrow viewing angle tracks the movement of the observer.
- the known methods have the disadvantage that their use is usually perceived as unnatural and uncomfortable.
- the systems provide the most natural impressions that can completely encapsulate the real world. This happens, for example, in HMDs where the user only sees the virtual reality. Here, the actions of the user can be fully or at least largely discussed.
- HMDs are usually difficult, complicated to attach and expensive.
- complete encapsulation of the real world requires extensive sensor technology that records the actions of the user and, in particular, his line of sight.
- the present invention is therefore based on the object, a method of the type mentioned in such a way and further, that as realistic as possible representation of an image object in a virtual image space, especially when interacting with the image object, can be reached.
- the above object is achieved by the features of claim 1.
- the method in question is characterized in that the position of the viewer is included in a calculation and / or selection of views of the image object such that the real position of the viewer and / or its change in the virtual image space and mapped to control at least a virtual camera is used, wherein a view of the viewer of the image object is recorded by a virtual camera.
- a particularly realistic representation is possible if the movement of the observer can essentially be transferred to the representation of the virtual reality. If, for example, a viewer moves to the left in front of a display device, he expects, analogously to the real world, that the three-dimensional image object appears rotated by a certain solid angle. As the viewer moves closer to the screen, he expects different details of the image object to become more apparent or individual views to change. If a viewer uses a tool for interacting with the image object, for example in the form of a selection process in the virtual image space, then it is expected that the user will stop the tool at the selected three-dimensional point of the image object when the viewer moves.
- This position or change transmitted into the virtual space is used according to the invention to control at least one virtual camera, wherein a virtual camera corresponds to a view represented to the viewer.
- a virtual camera corresponds to a view represented to the viewer.
- two virtual cameras would be provided which generate the views for one eye each of the observer.
- the position of the virtual cameras in the virtual space corresponds to the real position of the Beholder.
- Using the virtual cameras makes it particularly easy to "virtualize" a movement of a viewer, in particular to transfer the most varied processes into virtual reality, whereby all aspects outlined above can be implemented or only individual ones can be realized that are particularly desired for the respective application.
- a detection of the position of the observer or its change will take place in such a way that the position of the eyes of the observer is detected.
- the detection of the eyes in a camera image is relatively simple and corresponding methods are well known in practice.
- the views of the observer from the displayed image object can be determined very simply and directly.
- a wide range of possible applications can be covered.
- the system need not be aware of any more precise physiognomic characteristics of the viewer. Simplifying, however, the position of the head could be used. This can be the Determine change in position quite accurately. Also, the position detection can be done with often sufficiently good accuracy.
- the presentation should then be perceived as particularly realistic if the views of the viewer are calculated in real time.
- a soft real time should be sufficient, because here, for example, individual missing intermediate images are not perceived too clearly.
- the views of the image object could be recalculated when the position of the observer changes.
- the position changes are preferably detected three-dimensionally, transformed into the virtual image space and used to control one or more virtual cameras. In this way, the views of the image object can be presented to the viewer in a realistic manner.
- the recalculation of the views could be in affine transformations of the image object.
- Affine mappings are images between two vector spaces, where colinearities and spacing are preserved.
- these could consist of displacements, rotations, compressions, extensions or combinations thereof.
- Corresponding algorithms for calculating the transformations are well known in practice.
- an understeered tracking can take place, in which a change in the position of the observer by a certain solid angle to a lesser rotation of the image object is transmitted.
- Changing the viewing angle or distance to a three-dimensional image object will change various details of the views. If a viewer moves parallel to the screen layer, different areas of the image object may only become visible after a certain offset between the viewing positions. As a simple example, consider a cube whose surface is parallel to the screen plane. If a viewer moves to the left or to the right, he will be able to see the left or the right surface of the cube only from a certain position. The same applies to the offset upwards and downwards. The deck or floor area will only become visible from a certain position.
- complex entities will have areas that obscure each other depending on the viewing position. In another position, however, they could be visible. In order to leave as realistic a impression as possible on the viewer, these areas, which are visible depending on the position, could respectively be generated or adapted accordingly. This can be realized according to an embodiment of the invention by recalculating these areas.
- a three-dimensional model of the image object could be present.
- This three-dimensional model could be realized in many different ways. For example, if the image object is generated as a virtual object, then the three-dimensional information will most likely already be in a simple manner. Become real If objects are transformed into the virtual image space, it may be necessary to generate three-dimensional models, for example from existing images and views, or by means of 3D scans of the object. Again, a variety of methods are known from practice. Thus, for example, image objects are approximated by polygons and small image sections are mapped into the subpolygons.
- the method according to the invention is preferably used in connection with the representation on an autostereoscopic display device. It will be advantageous if, in addition to the calculation of the views in dependence on the position or the movement of the viewer in addition an accurate control of the viewing angle is made. This is done - as described above - by suitably driving the luminous dots behind the optical means of the autostereoscopic display device.
- the adaptation can be carried out as a control loop in parallel or sequentially to the recalculation of the views. It is important to distinguish that in the readjustment only in a small range pixels are moved. A complete recreation of views of the image object is not done here.
- the method need not necessarily be used in conjunction with three-dimensional display devices. So it is quite possible to use a standard monitor and to display the views of the image object only monoscopically. Here a virtual camera would suffice, which only produces a view of the image object.
- the method may also be used in conjunction with a selector that allows interaction with the image object or parts thereof.
- This selection device is preferably freely movable in the image space. With this selection device, the image object or parts thereof can be selected, marked, moved, edited, rotated or otherwise influenced. Such a selection device is described in the applicant's European patent EP 1 025 520 B1.
- the selection device could be formed by a relatively arbitrary object whose three-dimensional position and optionally orientation is determined by means of a suitable system.
- a stereoscopically operating camera system could be used, with which the object is detected.
- the object to be tracked could be realized by a stylus, any tool with which the viewer interacts with the image object, or the like.
- the viewer could also use a finger as a selector. This can be interacted naturally with individual areas of the image object.
- the illustrated image object appears to float in front of the display device. If a viewer selects a point of the image object and then moves away from the previous observation point, the selected point seems to move away from the finger.
- This can be taken into account in the method according to the invention in that the view of the image object is calculated not only as a function of the position of the observer but also as a function of the position of the selection device. When selecting a point of the image object, therefore, it could be determined which image areas the observer sees lying behind the selection device. These image areas would then have to remain behind the selection device even when the observer moves.
- the selector is used as a reference point in a transformation of the image object or the control of the virtual camera.
- the selection device could influence the image object.
- "touching" the image object by the selector could cause the image object to be deformed or virtually immersed in the image object
- Such deformations, shifts, or other transformations of the image object could be done in addition to the recalculations of the views.
- the calculations of the views on mobile hardware components could be performed.
- only individual components of the entire system could be made mobile and connected via a preferably wireless connection with a more or less stationary component.
- the views could be calculated on a standard computer and transmitted via a radio link to a mobile display device.
- both the calculations and the presentation could take place on a mobile device.
- a mobile device can be implemented in different ways. It can include a laptop as well as a powerful PDA (Personal Digital Assistant) or devices designed specifically for this application.
- standard software components can be used on the mobile device.
- Fig. 2 shows a change of view on movement of the viewer on the display device
- FIG 3 shows the maintenance of a selected point upon movement of the observer in a method according to the invention.
- FIGS. 1 to 3 show exemplary arrangements which are suitable for the application of a method according to the invention.
- like reference numerals designate corresponding components.
- the display device 1 comprises an autostereoscopic display device, in which the image object 2 seems to float in front of the display device.
- a viewer whose eyes 3 are shown in the figures views the image object 2 displayed on the display device 1.
- a position detection in the form of a stereoscopically operating camera system continuously determines the position of the eyes 3 of the observer. In this case, the coordinates of one or both eyes 3 with respect to the display device 1 are determined both in the horizontal and in the vertical direction. In addition, the distance of the eyes 3 to the display device 1 is determined.
- two views of the image object 2 are suitably displayed with a corresponding offset, so that in front of the display device 1, a virtual three-dimensional image space is spanned.
- the image object 2 is displayed.
- the position of the eyes 3 of the observer determined by the position detection is transmitted into the virtual image space. Since the most realistic possible representation of the image object 2 on the display device 1 is to be achieved, the position of the eyes 3 corresponds to the position of two virtual cameras in the virtual image space. When the eyes 3 are moved by a certain amount, the virtual cameras are also shifted by this amount in the virtual image space. If necessary, the scaling of the virtual image space must be taken into account.
- Fig. 1 shows the movement of the viewer to the left, parallel to the screen plane of the display device 1.
- the virtual cameras are shifted to the left, thereby changing the views generated by the virtual cameras.
- the virtual cameras then generate images from a perspective shifted to the left, whereby a view of the image object 2 rotated to the right with a rotation axis perpendicular to the plane of the drawing is produced.
- These views generated by the two virtual cameras are in turn converted into images suitable for the display device and displayed on the display device 1.
- the image object 2 rotates on the display device 1. This gives the observer the impression that he can actually view the image object 2 rotated by a certain solid angle.
- a quasi-holographic impression of the image object is generated.
- FIG. 2 shows the procedure when a viewer moves towards the display device 1. This can cause several different effects.
- the viewing angles will change to individual locations of the image object 2.
- the magnification of the image object 2 will increase.
- by approaching the Observers may only visible individual components of the image object, which were concealed by other parts of the image object at the farther distant position.
- further details could become visible as the observer approaches.
- the disparities, ie the distances of corresponding pixels in the views for the left and the right eye change. All these effects can be taken into account by using the virtual cameras in the display.
- FIG. 3 shows the case that a part of the image object 2 is marked by means of a selector 4.
- the selector 4 is formed here by a finger of the hand of the observer.
- the viewer has marked in the virtual image space in the figure marked with a circle area 5. If the observer moves to the left, then, without corresponding tracking, the marked area 5 apparently also moves to the left.
- a detection unit for detecting the position of the selection device 4 first determines the position of the selection device with respect to the display device 1. Again, using virtual cameras, it can be determined which area 5 in the virtual image space is marked by the selection device 4. This marked area 5 remains when moving the viewer to the left as a fixed point. In the transformation of the image object 2, therefore, not only the position of the eyes 3 of the observer is considered, but also the selected region 5 is used as a fixed point of the transformation.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computing Systems (AREA)
- Geometry (AREA)
- Computer Graphics (AREA)
- General Physics & Mathematics (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE200710023506 DE102007023506A1 (de) | 2007-05-18 | 2007-05-18 | Verfahren zum Darstellen von Bildobjekten in einem virtuellen dreidimensionalen Bildraum |
PCT/DE2008/000572 WO2008141596A1 (de) | 2007-05-18 | 2008-04-04 | Verfahren zum darstellen von bildobjekten in einem virtuellen dreidimensionalen bildraum |
Publications (1)
Publication Number | Publication Date |
---|---|
EP2156410A1 true EP2156410A1 (de) | 2010-02-24 |
Family
ID=39661378
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP08734459A Withdrawn EP2156410A1 (de) | 2007-05-18 | 2008-04-04 | Verfahren zum darstellen von bildobjekten in einem virtuellen dreidimensionalen bildraum |
Country Status (3)
Country | Link |
---|---|
EP (1) | EP2156410A1 (de) |
DE (1) | DE102007023506A1 (de) |
WO (1) | WO2008141596A1 (de) |
Families Citing this family (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102010010002A1 (de) | 2010-03-02 | 2011-09-08 | Geuder Ag | Verfahren zur Durchführung einer virtuellen Operation zu Trainingszwecken |
DE102010010001A1 (de) | 2010-03-02 | 2011-09-08 | Geuder Ag | Verfahren zur Entwicklung und virtuellen Erprobung eines chirurgischen Instruments |
US10979672B1 (en) | 2020-10-20 | 2021-04-13 | Katmai Tech Holdings LLC | Web-based videoconference virtual environment with navigable avatars, and applications thereof |
US11070768B1 (en) | 2020-10-20 | 2021-07-20 | Katmai Tech Holdings LLC | Volume areas in a three-dimensional virtual conference space, and applications thereof |
US10952006B1 (en) | 2020-10-20 | 2021-03-16 | Katmai Tech Holdings LLC | Adjusting relative left-right sound to provide sense of an avatar's position in a virtual space, and applications thereof |
US11095857B1 (en) | 2020-10-20 | 2021-08-17 | Katmai Tech Holdings LLC | Presenter mode in a three-dimensional virtual conference space, and applications thereof |
US11076128B1 (en) | 2020-10-20 | 2021-07-27 | Katmai Tech Holdings LLC | Determining video stream quality based on relative position in a virtual space, and applications thereof |
US11457178B2 (en) | 2020-10-20 | 2022-09-27 | Katmai Tech Inc. | Three-dimensional modeling inside a virtual video conferencing environment with a navigable avatar, and applications thereof |
US11743430B2 (en) | 2021-05-06 | 2023-08-29 | Katmai Tech Inc. | Providing awareness of who can hear audio in a virtual conference, and applications thereof |
US11184362B1 (en) | 2021-05-06 | 2021-11-23 | Katmai Tech Holdings LLC | Securing private audio in a virtual conference, and applications thereof |
US11651108B1 (en) | 2022-07-20 | 2023-05-16 | Katmai Tech Inc. | Time access control in virtual environment application |
US12009938B2 (en) | 2022-07-20 | 2024-06-11 | Katmai Tech Inc. | Access control in zones |
US11876630B1 (en) | 2022-07-20 | 2024-01-16 | Katmai Tech Inc. | Architecture to control zones |
US12022235B2 (en) | 2022-07-20 | 2024-06-25 | Katmai Tech Inc. | Using zones in a three-dimensional virtual environment for limiting audio and video |
US11928774B2 (en) | 2022-07-20 | 2024-03-12 | Katmai Tech Inc. | Multi-screen presentation in a virtual videoconferencing environment |
US11741664B1 (en) | 2022-07-21 | 2023-08-29 | Katmai Tech Inc. | Resituating virtual cameras and avatars in a virtual environment |
US11700354B1 (en) | 2022-07-21 | 2023-07-11 | Katmai Tech Inc. | Resituating avatars in a virtual environment |
US11711494B1 (en) | 2022-07-28 | 2023-07-25 | Katmai Tech Inc. | Automatic instancing for efficient rendering of three-dimensional virtual environment |
US11704864B1 (en) | 2022-07-28 | 2023-07-18 | Katmai Tech Inc. | Static rendering for a combination of background and foreground objects |
US11562531B1 (en) | 2022-07-28 | 2023-01-24 | Katmai Tech Inc. | Cascading shadow maps in areas of a three-dimensional environment |
US11682164B1 (en) | 2022-07-28 | 2023-06-20 | Katmai Tech Inc. | Sampling shadow maps at an offset |
US11956571B2 (en) | 2022-07-28 | 2024-04-09 | Katmai Tech Inc. | Scene freezing and unfreezing |
US11593989B1 (en) | 2022-07-28 | 2023-02-28 | Katmai Tech Inc. | Efficient shadows for alpha-mapped models |
US11776203B1 (en) | 2022-07-28 | 2023-10-03 | Katmai Tech Inc. | Volumetric scattering effect in a three-dimensional virtual environment with navigable video avatars |
US11748939B1 (en) | 2022-09-13 | 2023-09-05 | Katmai Tech Inc. | Selecting a point to navigate video avatars in a three-dimensional environment |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6154723A (en) * | 1996-12-06 | 2000-11-28 | The Board Of Trustees Of The University Of Illinois | Virtual reality 3D interface system for data creation, viewing and editing |
EP1025520B1 (de) | 1997-10-30 | 2002-08-28 | Dr. Baldeweg Aktiengesellschaft | Verfahren und vorrichtung zur bearbeitung von bildobjekten |
US7274380B2 (en) * | 2001-10-04 | 2007-09-25 | Siemens Corporate Research, Inc. | Augmented reality system |
-
2007
- 2007-05-18 DE DE200710023506 patent/DE102007023506A1/de not_active Withdrawn
-
2008
- 2008-04-04 WO PCT/DE2008/000572 patent/WO2008141596A1/de active Application Filing
- 2008-04-04 EP EP08734459A patent/EP2156410A1/de not_active Withdrawn
Non-Patent Citations (1)
Title |
---|
See references of WO2008141596A1 * |
Also Published As
Publication number | Publication date |
---|---|
WO2008141596A1 (de) | 2008-11-27 |
DE102007023506A1 (de) | 2008-11-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2156410A1 (de) | Verfahren zum darstellen von bildobjekten in einem virtuellen dreidimensionalen bildraum | |
DE69524332T2 (de) | Vorrichtung zur dreidimensionalen Bildwiedergabe | |
WO2009062492A2 (de) | Verfahren zum darstellen von bildobjekten in einem virtuellen dreidimensionalen bildraum | |
EP2027728B1 (de) | Verfahren und vorrichtung zur pseudoholographischen bilderzeugung | |
EP1763845B1 (de) | Verfahren und vorrichtung zur bestimmung von optischen überdeckungen mit ar-objekten | |
DE19953595B4 (de) | Verfahren und Vorrichtung zur Verarbeitung dreidimensionaler Bilder | |
EP2977961B1 (de) | Verfahren und Kommunikationseinrichtung zum Erstellen und/oder Ändern virtueller Objekte | |
DE112016005343T5 (de) | Elektronische Anzeigestabilisierung unter Verwendung von Pixelgeschwindigkeiten | |
CN101803395A (zh) | 3d显示器的呈现改善 | |
EP3427474B1 (de) | Bildverarbeitungsverfahren, bildverarbeitungsmittel und bildverarbeitungsvorrichtung zur erzeugung von abbildungen eines teils eines dreidimensionalen raums | |
EP1964413A1 (de) | Verfahren und anordnung zum monoskopischen darstellen wenigstens eines bereiches eines bildes auf einer autostereoskopischen anzeigevorrichtung sowie informationswiedergabegerät mit einer solchen anordnung | |
WO2018185201A2 (de) | Mikroskopanordnung zur aufnahme und darstellung dreidimensionaler bilder einer probe | |
DE102021117453B3 (de) | Verfahren zum Betreiben einer Datenbrille in einem Kraftfahrzeug während einer Fahrt, entsprechend betreibbare Datenbrille, Prozessorschaltung sowie Kraftfahrzeug | |
DE69837165T2 (de) | Verfahren und gerät für automatische animation von dreidimensionalen grafischen szenen für verbesserte 3-d visualisierung | |
DE69715816T2 (de) | Videoanzeigesystem zum Darstellen einer virtuellen dreidimensionalen Bildanzeige | |
WO2012140397A2 (en) | Three-dimensional display system | |
DE102011008886A1 (de) | Verfahren und Vorrichtung zur Stereobasis-Erweiterung von stereoskopischen Bildern und Bildfolgen | |
DE102014115363A1 (de) | Virtuelles Zeichen in realer Umgebung | |
DE102018209377A1 (de) | Verfahren zur Darstellung von AR-/VR-Inhalten auf einem mobilen Endgerät und mobiles Endgerät, auf dem AR-/VR-Inhalte dargestellt werden | |
WO2013164208A1 (de) | Verfahren zur ausführung beim betreiben eines mikroskops und mikroskop | |
DE102017112190A1 (de) | Anzeigesystem und Verfahren zum Erzeugen eines scheinbar dreidimensionalen Bilds eines Objekts | |
DE10056978A1 (de) | Verfahren zur Erzeugung eines stereographischen Bildes | |
WO2021180932A2 (de) | Verfahren und vorrichtung zum präzisen auswählen einer raumkoordinate mittels eines digitalen bildes | |
DE102013213492A1 (de) | Bildanpassung für kontaktanaloge Darstellungen auf Datenbrillen | |
DE102011112617A1 (de) | Kooperativer 3D-Arbeitsplatz |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20091217 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MT NL NO PL PT RO SE SI SK TR |
|
AX | Request for extension of the european patent |
Extension state: AL BA MK RS |
|
RIN1 | Information on inventor provided before grant (corrected) |
Inventor name: SCHMITT, THOMAS Inventor name: OPEL, WOLFGANG Inventor name: BOETTCHER, STEFFEN |
|
DAX | Request for extension of the european patent (deleted) | ||
17Q | First examination report despatched |
Effective date: 20101213 |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: SPATIAL VIEW INC |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20120821 |