WO2003019474A2 - Darstellung von anwenderinformationen - Google Patents
Darstellung von anwenderinformationen Download PDFInfo
- Publication number
- WO2003019474A2 WO2003019474A2 PCT/DE2002/002956 DE0202956W WO03019474A2 WO 2003019474 A2 WO2003019474 A2 WO 2003019474A2 DE 0202956 W DE0202956 W DE 0202956W WO 03019474 A2 WO03019474 A2 WO 03019474A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- camera
- user information
- information
- image
- image information
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/183—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source
Definitions
- the invention relates to a system and a method for
- Visualization systems are usually used to inform a user about the current status of a process. You are able to display the current process status of a system via recorded process values and status data of a process control program with changing text or graphic elements (e.g. dynamic bars) as user information.
- the process values are recorded using appropriate sensors.
- the user information is limited to that which can be recorded via the sensors or which is reflected in the status of the control program - but not everything can be recorded with the help of sensors. For this reason, video technology is increasingly being used. With the help of a recorded video image, it enables the user to visualize the visible status of the process and the process environment on the visualization system.
- the invention has for its object to improve the simultaneous display of user information and image information of an environment.
- This object is achieved by a system for displaying user information with a camera for recording image information of a section of an environment, a zoom device for changing the size of the recorded section according to a zoom factor and / or a device for three-dimensional alignment of the camera according to a space vector - with a computer unit for calculating location coordinates of the image information based on spatial coordinates of the camera and / or the control variables zoom factor and space vector, for assigning user information to the location coordinates and for calculating positions of images of the image information on a display surface of a visualization device and
- This object is achieved by a method for displaying user information, in which
- Image information of a section of an environment is recorded with a camera, the size of the recorded section being changeable with a zoom device in accordance with a zoom factor and / or the camera being aligned three-dimensionally with a device in accordance with a space vector, a computer unit - Location coordinates of the image information based on
- an image processing unit calculates the image information and the user information for playback with the visualization device and for the correct display of the user information on the display surface at the positions of the images of the image information with location coordinates to which the respective user information is assigned.
- the system or method according to the invention enables the dynamic insertion of user information - e.g. B. of process values, status information of a control program - in the user-visualized image of a section of an environment.
- This image is recorded by a camera which is movable and / or by means of a zoom device offers the possibility of changing the size of the image section.
- the camera therefore does not have to have a fixed image section, but a free definition of the image section (orientation and / or zoom factor) is possible.
- the user information to be faded in does not have to relate to a static image in the sense of camera orientation and zoom factor, but rather receives a reference to the real spatial coordinates of the image information in the area currently captured by the camera.
- the user information on the currently visible location section is automatically displayed at the correct position.
- the dynamic overlays do not change their position in relation to the images of the image information (for example, objects) visible on the display surface of the visualization device a changed viewing angle of the camera, ie when the camera moves (rotations or inclinations, zoom factor).
- the computer unit contains a control unit for controlling the camera, the zoom device and / or the device for three-dimensional alignment of the camera in accordance with the control variables zoom factor or space vector.
- the control variables are thus already known to the computer unit and can be used directly by it to calculate the location coordinates of the
- Image information of the section of the environment can be used.
- a particular user friendliness can be achieved in that the image processing unit is used to select and display the user information depending on the zoom factor. It is conceivable. B. that in a wide-angle shot only for individual objects on the display surface user information, for. B. Object names can be shown. If the camera zooms in on these objects, detailed information could be displayed, e.g. B. level, temperature or the like. The detailed information would currently be read from an operating and monitoring system.
- the user information in this embodiment is thus designed as a combination of static and dynamic information.
- any other data sources can be connected, for example the connection to databases with static information or
- the camera is advantageously designed as a video camera and the visualization device as a screen.
- the image data supplied by the video camera are processed by the image processing unit for display on the screen.
- control unit for controlling the camera, the zoom device and the device for three-dimensional alignment of the camera have means for operation by a user. So that the camera, regardless of the computer unit z. B. can also be moved with a remote control.
- the user information is superimposed on the display surface in accordance with a mapping rule.
- a mapping rule contains specific rules, formats and links according to which the respective user information is displayed.
- FIG. 1 shows a schematic overview of a system for displaying user information
- FIG 3 shows views of a display surface of a 5 visualization device with different control variables, space vector and zoom factor.
- FIG. 1 shows an embodiment of a system for
- a camera 1 captures image information 2 of a Section of the surroundings of the camera 1.
- the image information 2 is the view of a tank 21 with a valve 22.
- the viewing angle 23 of the camera 1, which captures the section of the surroundings, is shown in a stylized manner.
- the camera 1 is mounted on a device 4 for three-dimensional alignment of the camera and has a zoom device 3.
- the camera 1 and the device 4 are connected to a computer unit 5.
- the computer unit 5 contains a control unit 10 and a display area 7.
- the computer unit 5 contains user information 6, which in the example is provided via a process connection 20 of measuring points 17, 18.
- the user information 6 is linked in an image processing unit 9 with location coordinates 12 and is shown as a display 16 on the display surface 7 together with an image 13 of the image information 2.
- the computer unit also has various input means for a user, a computer mouse 14, a keyboard 15 and other means 11 for operation by a user.
- the camera 1 captures the objects 21, 22 lying in its viewing angle 23 as image information 2.
- the viewing angle 23 can be adjusted with a zoom device 3, e.g. B. a focal length adjustment, adjustable in its opening angle and by rotating or tilting the camera 1 in its orientation.
- the different size of the opening angle of the camera 1 is called the zoom factor and is an important control variable of the system.
- the camera 1 captures a larger or smaller section of its surroundings.
- the camera 1 is fastened on a device 4 for three-dimensional alignment and is therefore rotatable about two of its axes of movement.
- the device 4 for three-dimensional alignment is driven, for. B.
- the movement of the device 4, the adjustment of the zoom device 3 and the functions of the camera 1 are controlled by a control unit 10 of the computer unit 5.
- the orientation of the camera 1 in the room is described by the control variable room vector.
- the camera 1 and the device 4 for three-dimensional alignment return actual values of the space vector and zoom factor to the computer unit.
- the placement of the camera 1 in the room is defined in the form of spatial coordinates of the camera 1.
- the computer unit has further information about the surroundings of camera 1 available, e.g. B. in the form of a model which describes the essential points of the objects 21, 22 of the environment in the form of spatial coordinates or as vectors.
- the computer unit 5 thus has enough information available to determine the location coordinates 12 of the image information 2 captured by the camera 1.
- the location coordinates 12 are calculated from the control variables zoom factor and space vector and - in the case of linear movements - the space coordinates of the camera 1.
- the result of this calculation determines the viewing angle 23 of the camera 1 in terms of its size and its position in space.
- the image information 2 is processed by an image processing unit 9 of the computer unit 5 such that it can be represented as a two-dimensional image 13 of the objects 21, 22 on the display surface 7 of the visualization device. Based on the calculation of the location coordinates 12, the information is also available at which position on the
- Playback surface 7 the image 13 of the image information 2 or the objects 21, 22 is reproduced.
- the user information 6 is assigned to specific location coordinates 12.
- the image processing unit 9 of the computer unit 5 recognizes that image information 2 of objects 21, 22 are recorded by the camera 1 with these specific location coordinates 12, then it displays the corresponding user information 6 together with the image 13 on the display surface 7. Since the position of the image 13 of the objects 21, 22 is known, the user information 6 assigned to them via the location coordinates 12 can also be in the correct position, e.g. B. be displayed in the immediate vicinity of the image 13.
- the actual values of the control variables room vector and zoom factor change continuously, and the detected section of the surroundings also changes accordingly.
- the position of the image 13 on the display surface 7 changed as a result can also be calculated and the user information 6 can continue to be displayed in the same position relative to the image 13, even if its position on the display surface is thereby shifted.
- the location coordinates 12 are thus assigned to the user information 6 and the current orientation (space vector) of the camera 1, the current zoom factor and - in the case of linear movement of the camera 1 in space - the space coordinates of the camera 1 (ie their placement in space) are known, so the display and the placement of the user information 6 for the overlay technology can currently be calculated, so that the user information 6 for the currently visible area is shown in the correct place.
- the user information 6 can be dynamic or static information or a combination thereof. Dynamic information is, for example, process values.
- Dynamic information is, for example, process values.
- a system consisting of a tank 21 and a valve 22. On the tank
- a temperature sensor 17 is attached, and on the valve
- the recorded process values of temperature or valve opening are transmitted via the process connection 20 to the computer unit 5, where they are then used as user information 6 Are available and displayed in the correct position in the image of objects 21, 22.
- the image of objects 21, 22 displayed to the user is thus enriched with user information 6 with the additionally displayed process variables.
- the user can operate the computer unit 5 with input means 14, 15 and also has the option of directly specifying the orientation and the zoom factor of the camera 1 by means 11 for operation.
- the camera 1 is designed as a video camera 27, the computer unit 5 as a personal computer 28 and the visualization device as a screen 29.
- the device 4 for three-dimensional alignment, on which the video camera 27 is attached, is in this
- Embodiment designed as a turning and tilting device 30 The degrees of freedom of the video camera 27 are indicated by arrows 31.
- the personal computer 28 Via a camera control, the personal computer 28 has the possibility of aligning the controllable video camera 27 with regard to zoom and position.
- the image information recorded by the video camera is sent as a video signal 26 to the personal computer 28 or to a so-called frame grabber card contained therein. With the help of the frame grabber card and the corresponding software, it is possible to display the video image of the video camera 27 on the
- the turning and tilting device 30 pan, tilt
- the zoom device 3 of the video camera 27 are connected via an RS232 connection 24 to a serial interface 25 of the personal computer 28.
- VISCA a corresponding protocol
- the video camera 27 can be moved by software and the resulting viewing angles can be read out.
- the video camera 27 can also be moved independently of the personal computer using a remote control, which is not shown in FIG. 2. Since the associated data for rotation, tilt and zoom factor are read out from the video camera 27 for each video frame to be displayed on the screen 29, it is possible to do so Fade in user information dynamically in the correct position, regardless of whether the video camera 27 was moved by software or by remote control.
- z. B Show supporting texts.
- the particular advantage of the proposed system and method thus lies in the dynamic insertion of information into the video image, taking into account the area currently captured by the video camera 27.
- the dynamic overlays do not change their position from those on the
- Video image of visible objects during movements (rotations or inclinations, zoom factor) of the video camera 27 Only in the context of the lens distortion of the video camera 27 and the perspective distortion do the dynamic insertions move slightly compared to the visible objects.
- Visualization device 8 with a display surface 7 at different viewing angles of a camera 1 according to FIG 1 in a system according to the invention. The one on the
- Display area 7 projected image of a camera 1 show an arrangement of control cabinets.
- a supplementary text 16 is shown in the displayed image on a control cabinet on the opening lever 19.
- the angle of view has changed slightly by rotating the camera 1.
- Image be united. This is possible because the current position and zoom settings of camera 1 are also read out for each video image. In addition, depending on the zoom, more or less data can be shown in the image. It is conceivable. B. that in a
- Tankl Wide angle shot only individual objects are named (e.g. Tankl, control cabinet2). If you zoom in on these elements, you could Detailed information about this is displayed (e.g. Tankl: fill level 3 m). These data would currently be read out from an operating and monitoring system.
- the invention thus relates to a system and a method for displaying user information, in which the simultaneous display of user information and image information of an environment is improved.
- the system contains a camera 1 for capturing image information 2 of a section of an environment, a zoom device 3 for changing the size of the captured section in accordance with a zoom factor and / or a device 4 for three-dimensionally aligning the camera 1 in accordance with a space vector, a computer unit 5 for calculating location coordinates 12 of the image information 2 on the basis of spatial coordinates of the camera 1 and / or the control variables zoom factor and space vector, for assigning user information 6 to the location coordinates 12 and for calculating positions of images 13 of the image information 2 on a display surface 7 of a visualization device 8 and an image processing unit 9 for processing the image information 2 and the user information 6 for reproduction with the visualization device 8 and for correctly displaying the user information 6 on the display surface 7 at the positions of the images 13 of the image information 2 with location coordinates 12, to which the respective user information 6 is assigned.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Closed-Circuit Television Systems (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
Description
Claims
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| EP02754527A EP1419483A2 (de) | 2001-08-24 | 2002-08-12 | Darstellung von anwenderinformationen |
| US10/784,836 US20040227818A1 (en) | 2001-08-24 | 2004-02-24 | System and an associated method for displaying user information |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| DE10141521A DE10141521C1 (de) | 2001-08-24 | 2001-08-24 | Darstellung von Anwenderinformationen |
| DE10141521.4 | 2001-08-24 |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US10/784,836 Continuation US20040227818A1 (en) | 2001-08-24 | 2004-02-24 | System and an associated method for displaying user information |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| WO2003019474A2 true WO2003019474A2 (de) | 2003-03-06 |
| WO2003019474A3 WO2003019474A3 (de) | 2003-08-28 |
Family
ID=7696487
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/DE2002/002956 Ceased WO2003019474A2 (de) | 2001-08-24 | 2002-08-12 | Darstellung von anwenderinformationen |
Country Status (4)
| Country | Link |
|---|---|
| US (1) | US20040227818A1 (de) |
| EP (1) | EP1419483A2 (de) |
| DE (1) | DE10141521C1 (de) |
| WO (1) | WO2003019474A2 (de) |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10089534B2 (en) * | 2016-12-16 | 2018-10-02 | Adobe Systems Incorporated | Extracting high quality images from a video |
Family Cites Families (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| GB9119964D0 (en) * | 1991-09-18 | 1991-10-30 | Sarnoff David Res Center | Pattern-key video insertion |
| JP3179623B2 (ja) * | 1992-04-29 | 2001-06-25 | キヤノン インフォメーション システムズ リサーチ オーストラリア プロプライエタリー リミテツド | ビデオムービー |
| US5488675A (en) * | 1994-03-31 | 1996-01-30 | David Sarnoff Research Center, Inc. | Stabilizing estimate of location of target region inferred from tracked multiple landmark regions of a video image |
| DE19710727A1 (de) * | 1997-03-14 | 1998-09-17 | Sick Ag | Überwachungseinrichtung |
| DE19900884A1 (de) * | 1999-01-12 | 2000-07-20 | Siemens Ag | System und Verfahren zum Bedienen und Beobachten eines Automatisierungssystems mit Prozeßvisualisierung und Prozeßsteuerung durch virtuelle Anlagenmodelle als Abbild einer realen Anlage |
| DE10005213A1 (de) * | 2000-02-05 | 2001-08-16 | Messer Griesheim Gmbh | Überwachungssystem und Verfahren zum Fernüberwachen von Messgrößen |
-
2001
- 2001-08-24 DE DE10141521A patent/DE10141521C1/de not_active Expired - Fee Related
-
2002
- 2002-08-12 EP EP02754527A patent/EP1419483A2/de not_active Ceased
- 2002-08-12 WO PCT/DE2002/002956 patent/WO2003019474A2/de not_active Ceased
-
2004
- 2004-02-24 US US10/784,836 patent/US20040227818A1/en not_active Abandoned
Also Published As
| Publication number | Publication date |
|---|---|
| US20040227818A1 (en) | 2004-11-18 |
| EP1419483A2 (de) | 2004-05-19 |
| DE10141521C1 (de) | 2003-01-09 |
| WO2003019474A3 (de) | 2003-08-28 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| DE60102907T2 (de) | Überwachungsvorrichtung für ein kamera-überwachungssystem | |
| DE69332282T2 (de) | Verfahren und Gerät für Benutzersteuerung durch Ermittlung der nächsten Prozesszustände aus dem aktuellen Zustand und durch visuelle Darstellung der abgeleiteten nächsten Zustände | |
| DE69835185T2 (de) | Kamerasteuersystem | |
| DE69520329T2 (de) | Kamerasteuervorrichtung und -Verfahren | |
| DE69429028T2 (de) | Rundumsicht Überwachungssystem mit bewegungsloser Kamera | |
| DE69233439T2 (de) | Überwachungsvorrichtung mit Steuerung der Kamera und der Linsenmontage | |
| EP2196892B1 (de) | Verfahren und Vorrichtung zum Anzeigen von Informationen | |
| EP1141799B1 (de) | System und verfahren zum bedienen und beobachten eines automatisierungssystems | |
| DE69629144T2 (de) | Verfahren und Vorrichtung zur Kamerakontrolle | |
| DE60225933T2 (de) | Tragbare virtuelle realität | |
| DE60031059T2 (de) | Videokamerabeobachtungssystem und Vorrichtung zur Anzeige von Bildern von dieser Videobeobachtungskamera | |
| DE19836681A1 (de) | Stereoskopisches Aufnahme- und Wiedergabesystem | |
| DE102021123234A1 (de) | Teleoperator-arbeitsplatz | |
| DE19932217A1 (de) | Steuervorrichtung | |
| EP2822813B1 (de) | Kraftfahrzeug mit einem elektronischen rückspiegel | |
| EP2822814A1 (de) | Kraftfahrzeug mit einem elektronischen rückspiegel | |
| DE69421654T2 (de) | Interaktives Bildanzeigesystem mit Cursorsteuerung | |
| DE10246652B4 (de) | Verfahren zum Betrieb eines Darstellungssystems in einem Fahrzeug | |
| EP3418839B1 (de) | Verfahren zur überwachung einer automatisierungsanlage | |
| DE10141521C1 (de) | Darstellung von Anwenderinformationen | |
| DE102006049981A1 (de) | Bedienungshilfe für eine Vorrichtung zum Behandeln von Behältnissen | |
| DE19811286C2 (de) | Kamerabewegungssteuerung | |
| DE10117030C2 (de) | Verfahren zur Darstellung von bildhaft strukturierten Informationen auf einem Bildschirm | |
| DE4138453A1 (de) | Prozessbeobachtungssystem und fensteranzeigeverfahren hierfuer | |
| DE60108918T2 (de) | Interaktives Verfahren und Vorrichtung für Bildrundfunk einer beweglichen Videokamera |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AK | Designated states |
Kind code of ref document: A2 Designated state(s): US |
|
| AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR IE IT LU MC NL PT SE SK TR Kind code of ref document: A2 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FR GB GR IE IT LU MC NL PT SE SK TR |
|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
| DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
| REEP | Request for entry into the european phase |
Ref document number: 2002754527 Country of ref document: EP |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2002754527 Country of ref document: EP |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 10784836 Country of ref document: US |
|
| WWP | Wipo information: published in national office |
Ref document number: 2002754527 Country of ref document: EP |