CN110543236A - Machine room monitoring system and method based on virtual reality technology - Google Patents

Machine room monitoring system and method based on virtual reality technology Download PDF

Info

Publication number
CN110543236A
CN110543236A CN201910701438.5A CN201910701438A CN110543236A CN 110543236 A CN110543236 A CN 110543236A CN 201910701438 A CN201910701438 A CN 201910701438A CN 110543236 A CN110543236 A CN 110543236A
Authority
CN
China
Prior art keywords
scene
information
machine room
virtual reality
interaction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN201910701438.5A
Other languages
Chinese (zh)
Inventor
马骏峰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Suzhou Wave Intelligent Technology Co Ltd
Original Assignee
Suzhou Wave Intelligent Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Suzhou Wave Intelligent Technology Co Ltd filed Critical Suzhou Wave Intelligent Technology Co Ltd
Priority to CN201910701438.5A priority Critical patent/CN110543236A/en
Publication of CN110543236A publication Critical patent/CN110543236A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/10Constructive solid geometry [CSG] using solid primitives, e.g. cylinders, cubes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/20Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/04Indexing scheme for image data processing or generation, in general involving 3D image data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/61Scene description
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2219/00Indexing scheme for manipulating 3D models or images for computer graphics
    • G06T2219/20Indexing scheme for editing of 3D models
    • G06T2219/2016Rotation, translation, scaling

Abstract

The invention discloses a machine room monitoring system based on a virtual reality technology, which comprises a scene construction module, a view conversion module, an interaction control module and an information transmission module, wherein the scene construction module is used for constructing a scene; the scene construction module is used for constructing a virtual reality scene of the machine room to be monitored to obtain scene data; the view conversion module is used for displaying the scene data according to a preset rule; the interaction control module is used for converting the interaction intention of the user into corresponding interaction information and acquiring the monitoring information of the displayed scene according to the interaction information; the information transmission module is used for transmitting the monitoring information of the machine room to be monitored. By the scheme of the invention, a user can obtain more convenient interactive experience based on a virtual reality technology when monitoring the machine room.

Description

Machine room monitoring system and method based on virtual reality technology
Technical Field
the invention relates to the technical field of virtual reality, in particular to a machine room monitoring system and method based on a virtual reality technology.
background
At present, in the 5G era, network resources are no longer the bottleneck of data transmission, which provides technical feasibility for Virtual Reality (Virtual Reality) technology implementation. How to provide a computer lab monitored control system based on virtual reality technique makes the user can obtain mutual experience more conveniently when monitoring the computer lab, makes the user can enjoy the mutual experience that is close to the real world.
Disclosure of Invention
In order to solve the technical problem, the invention provides a machine room monitoring system and method based on a virtual reality technology, which can enable a user to obtain more convenient interactive experience through an interactive method based on a virtual reality scene when monitoring a machine room.
in order to achieve the aim, the invention provides a machine room monitoring system based on a virtual reality technology, which comprises a scene construction module, a view conversion module, an interaction control module and an information transmission module, wherein the scene construction module is used for constructing a scene;
The scene construction module is used for constructing a virtual reality scene of the machine room to be monitored to obtain scene data; the virtual reality scene comprises one or more of the following devices in the machine room to be monitored: the system comprises a cabinet, a case, a rack, a server, a storage and a router;
the view conversion module is used for displaying the scene data according to a preset rule;
The interaction control module is used for converting the interaction intention of the user into corresponding interaction information and acquiring the monitoring information of the displayed scene according to the interaction information;
And the information transmission module is used for transmitting the monitoring information of the machine room to be monitored.
In an exemplary embodiment, the scene building module is specifically configured to:
exporting relevant information of the model established by the modeling software;
converting the derived related information into scene data of a predetermined format; wherein the relevant information of the model comprises one or more of the following: vertices of the model, normal vectors, texture maps.
In an exemplary embodiment, the converting the interaction intention of the user into the corresponding interaction information includes:
Acquiring direction data or rotation angle data of the movement of a user;
and mapping the acquired direction data or angle data into corresponding interactive information through a corresponding algorithm.
in an exemplary embodiment, the obtaining the monitoring information of the displayed scene according to the interaction information includes:
When the interaction information indicates that a user views first equipment in the virtual reality scene, controlling the transmission module to acquire monitoring information of second equipment acquired in the machine room to be monitored; wherein the second device corresponds to the first device.
In an exemplary embodiment, the displaying the scene data according to a preset rule includes:
And displaying the scene data of the corresponding visual angle according to the requirement of the user for observing the scene.
In order to solve the above problem, the present invention further provides a method for monitoring a machine room based on a virtual reality technology, where the method includes:
constructing a virtual reality scene of a machine room to be monitored to obtain scene data; the virtual reality scene comprises one or more of the following devices in the machine room to be monitored: the system comprises a cabinet, a case, a rack, a server, a storage and a router;
Displaying the scene data according to a preset rule;
acquiring the interaction intention of a user, converting the interaction intention into corresponding interaction information, and acquiring the monitoring information of a displayed scene according to the interaction information;
and transmitting the monitoring information of the machine room to be monitored.
in an exemplary embodiment, the constructing a virtual reality scene of a machine room to be monitored to obtain scene data includes:
exporting relevant information of the model established by the modeling software;
Converting the derived related information into scene data of a predetermined format; wherein the relevant information of the model comprises one or more of the following: vertices of the model, normal vectors, texture maps.
In an exemplary embodiment, the obtaining the interaction intention of the user and converting into the corresponding interaction information includes:
Acquiring direction data or rotation angle data of the movement of a user;
And mapping the acquired direction data or angle data into corresponding interactive information through a corresponding algorithm.
in an exemplary embodiment, the obtaining the monitoring information of the displayed scene according to the interaction information includes:
when the interaction information indicates that a user views first equipment in the virtual reality scene, acquiring monitoring information of second equipment collected in the machine room to be monitored according to the interaction information; wherein the second device corresponds to the first device.
In an exemplary embodiment, the displaying the scene data according to a preset rule includes:
and displaying the scene data of the corresponding visual angle according to the requirement of the user for observing the scene.
compared with the prior art, the machine room monitoring system based on the virtual reality technology disclosed by the embodiment of the invention comprises a scene construction module, a view conversion module, an interaction control module and an information transmission module; the scene construction module is used for constructing a virtual reality scene of the machine room to be monitored to obtain scene data; the view conversion module is used for displaying the scene data according to a preset rule; the interaction control module is used for converting the interaction intention of the user into corresponding interaction information and acquiring the monitoring information of the displayed scene according to the interaction information; the information transmission module is used for transmitting the monitoring information of the machine room to be monitored. By the scheme of the invention, when a user monitors the machine room, more convenient interactive experience and interactive experience close to the real world can be obtained through the interactive method based on the virtual reality scene.
Additional features and advantages of the invention will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention. The objectives and other advantages of the invention will be realized and attained by the structure particularly pointed out in the written description and claims hereof as well as the appended drawings.
Drawings
the accompanying drawings are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the example serve to explain the principles of the invention and not to limit the invention.
Fig. 1 is a schematic diagram of a machine room monitoring system based on a virtual reality technology in an embodiment of the invention;
FIG. 2 is a flow chart of a machine room monitoring method based on a virtual reality technology in the embodiment of the invention;
fig. 3 is a flowchart illustrating a machine room monitoring method based on a virtual reality technology according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, embodiments of the present invention will be described in detail below with reference to the accompanying drawings. It should be noted that the embodiments and features of the embodiments in the present application may be arbitrarily combined with each other without conflict.
The steps illustrated in the flow charts of the figures may be performed in a computer system such as a set of computer-executable instructions. Also, while a logical order is shown in the flow diagrams, in some cases, the steps shown or described may be performed in an order different than here.
Fig. 1 is a schematic diagram of a machine room monitoring system based on a virtual reality technology according to the present invention, and according to the schematic diagram, the machine room monitoring system based on a virtual reality technology of the present embodiment includes: the system comprises a scene construction module 101, a view conversion module 102, an interaction control module 103 and an information transmission module 104;
the scene construction module 101 is configured to construct a virtual reality scene of a machine room to be monitored, and obtain scene data.
In this embodiment, a virtual reality scene of a machine room to be monitored includes one or more of the following devices in the machine room to be monitored: cabinet, machine case, frame, server, storage, router. Any equipment to be monitored in the machine room may be included, and the type of the equipment is not limited thereto.
In an exemplary embodiment, information about a model created by modeling software is derived; converting the derived related information into scene data of a predetermined format; wherein the relevant information of the model comprises one or more of the following: vertices of the model, normal vectors, texture maps. The modeling software can be third-party modeling software, relevant information of a model created by the third-party modeling software is exported and converted into data information which can be identified by the operation environment of the monitoring system, the relevant information is analyzed, and a scene is restored and displayed. The running environment of the monitoring system can be a browser end running on a browser, and can also be a client running on a windows desktop, an android device and an ios device. The model established by the modeling software cannot be normally identified in the operating environment, information such as a vertex, a normal vector, a texture map and the like of the model needs to be extracted, the information is converted according to a data format which can be identified by a related software package and provided by the operating environment of the system or a self-defined data format, and scene data is obtained and used for displaying the model established by the third-party modeling software in the operating environment of the monitoring system.
and the view conversion module 102 is configured to display the scene data according to a preset rule.
in this embodiment, the view conversion module 102 is used to display the scene data obtained by the scene construction module 101 according to a preset rule. The preset rule can be set according to the terminal equipment used by the user, the observation mode, the interaction habit of the user and the like.
in an exemplary embodiment, the displaying the scene data according to a preset rule includes: and displaying the scene data of the corresponding visual angle according to the requirement of the user for observing the scene. The virtual reality scene for constructing the machine room to be monitored is a three-dimensional world, and the three-dimensional world is observed by simulating human eyes through a camera. If the user desires to view the virtual reality scene at the first-person perspective, the view of the virtual reality scene may be switched to the camera at the first-person perspective according to the tools provided by the associated graphics software development kit. If the user needs to observe the virtual reality scene at the third person perspective, the view of the virtual reality scene can be switched to the camera at the third person perspective according to the tools provided by the relevant graphic software development kit. If the user needs to observe the virtual reality scene at the aerial view angle, the view of the virtual reality scene can be switched to the aerial view angle according to the tools provided by the related graphic software development kit.
And the interaction control module 103 is configured to convert the interaction intention of the user into corresponding interaction information, and acquire monitoring information of the displayed scene according to the interaction information.
in this embodiment, the interaction intention of the user may obtain the direction, the acceleration, and the rotation angular velocity of the limb movement of the user through the sensor, convert the obtained data into corresponding interaction information through corresponding operations, and obtain the monitoring information of the displayed scene according to the interaction information.
in an exemplary embodiment, converting the interaction intention of the user into corresponding interaction information includes: acquiring direction data or rotation angle data of the movement of a user; and mapping the acquired direction data or angle data into corresponding interactive information through a corresponding algorithm. The specific implementation mode can be realized by the following steps:
When acquiring the direction data of the user movement, the implementation mode is as follows:
the method for acquiring the mobile interaction intention of a user in a virtual reality scene comprises the following steps: acquiring interactive intentions of front, back, left and right movement of a user in a scene through a direction input key on the mobile terminal; and mapping the interaction intention data to obtain corresponding interaction data. The user can input the key information according to the preset setting and the user requirement through the direction input key on the mobile terminal. Once the key is input, the specific setting corresponding to the fixed distance moved by the user can be set as required. For example: pressing the key once, which is equivalent to the user moving 0.1 meter; when the user moves 0.2 meters forward in the virtual reality scene, the up input key may be pressed 2 times in succession.
when obtaining the angle data of the user rotation, the implementation mode is as follows:
The method includes the steps of obtaining a rotation interaction intention of a user in a virtual reality scene, and comprises the following steps: acquiring the rotation angle of a user in a scene through a gyroscope sensor to obtain angle data; and mapping the angle data through integral processing to obtain corresponding interactive data.
The gyroscope sensor is located in a wearing device of a user, wherein the wearing device can be a mobile terminal.
the specific implementation manner of acquiring the rotation angle of the user in the scene through the gyroscope sensor can be as follows: the motion state of the gyro sensor is monitored, and the angle of the user rotating in the scene can be determined according to the reading of the gyro sensor.
The motion state of the gyro sensor can be divided into two types: a "moving" state and an "unmoving" state. When the equipment where the gyroscope sensor is positioned is in a large-amplitude motion state, the equipment is judged to be in a motion state; when the equipment where the gyroscope sensor is located is in a static state or in uniform motion, the equipment is judged to be in an 'unmoved' state.
First, a preset threshold value is acquired.
next, the motion state of the gyro sensor is judged. When the reading of the gyroscope sensor is larger than a set threshold value, judging that the gyroscope sensor is in a motion state; and when the reading of the gyroscope sensor is smaller than a set threshold value, judging that the gyroscope sensor is in a non-motion state.
And finally, determining the rotating angle in the user scene according to the motion state of the gyroscope sensor.
And when the reading of the gyroscope sensor is larger than a set threshold value, judging that the gyroscope sensor is in a motion state, and using the output of the reading of the gyroscope sensor as the output data of the motion state of the current equipment.
and when the reading of the gyroscope sensor is smaller than a set threshold value, judging that the gyroscope sensor is in a non-motion state, and using the angle value calculated by the accelerometer and the magnetometer as the motion state output data of the current equipment. In this embodiment, the specific implementation of the angle values calculated using the accelerometer and the magnetometer may be as follows:
Firstly, reading of three axes of accelerometers ax, ay and az is obtained;
Secondly, after reading the readings of the three axes ax, ay and az by the accelerometer, calculating the instantaneous angle around the x axis and the instantaneous angle around the y axis by the following formulas:
The instantaneous angle around the x-axis is: θ ═ arcsin (-ax/g), where θ denotes the instantaneous angle around the x-axis and g denotes the acceleration of gravity;
the instantaneous angle around the y-axis is: phi is arctan (ay/az), where phi denotes the instantaneous angle around the y-axis;
thirdly, after calculating the instantaneous angle around the x axis and the instantaneous angle around the y axis, acquiring the reading of the magnetometer;
fourthly, calculating the angle of rotation around the z axis according to the reading of the magnetometer:
Where ψ represents an angle of rotation about the z-axis, and Mx, My, and Mz represent outputs of the magnetometer x, y, and z axes, respectively.
in an exemplary embodiment, the obtaining the monitoring information of the displayed scene according to the interaction information includes: when the interaction information indicates that a user views first equipment in the virtual reality scene, controlling the transmission module to acquire monitoring information of second equipment acquired in the machine room to be monitored; wherein the second device corresponds to the first device. The equipment in the virtual scene corresponds to the equipment in the real machine room one by one, the first equipment is the virtual equipment in the virtual scene, and the second equipment is the real equipment in the machine room to be monitored. A sensor, a camera and the like for monitoring can be installed in the real machine room to be monitored so as to obtain monitoring information.
And the information transmission module 104 is configured to transmit monitoring information for the machine room to be monitored.
in this embodiment, the interaction control module acquires monitoring information of a displayed scene according to the interaction information; and the information transmission module transmits the acquired related monitoring information and displays the related monitoring information in a virtual reality scene.
in addition, the application provides an embodiment of a method for monitoring a machine room based on a virtual reality technology, and the embodiment of the method corresponds to the embodiment of the device shown in fig. 1.
In order to solve the above problem, as shown in fig. 2, the present invention further provides a method for monitoring a machine room based on a virtual reality technology.
step 200: and constructing a virtual reality scene of the machine room to be monitored to obtain scene data.
In this embodiment, the virtual reality scene includes one or more of the following devices in the machine room to be monitored: cabinet, machine case, frame, server, storage, router.
In an exemplary embodiment, information about a model created by modeling software is derived; converting the derived related information into scene data of a predetermined format; wherein the relevant information of the model comprises one or more of the following: vertices of the model, normal vectors, texture maps. The modeling software can be third-party modeling software, relevant information of a model created by the third-party modeling software is exported and converted into data information which can be identified by the operation environment of the monitoring system, the relevant information is analyzed, and a scene is restored and displayed. The running environment of the monitoring system can be a browser end running on a browser, and can also be a client running on a windows desktop, an android device and an ios device. The model established by the modeling software cannot be normally identified in the operating environment, information such as a vertex, a normal vector, a texture map and the like of the model needs to be extracted, the information is converted according to a data format which can be identified by a related software package and provided by the operating environment of the system or a self-defined data format, and scene data is obtained and used for displaying the model established by the third-party modeling software in the operating environment of the monitoring system.
Step 201: and displaying the scene data according to a preset rule.
In this embodiment, the preset rule may be set according to a terminal device used by the user, an observation mode, an interaction habit of the user, and the like.
In an exemplary embodiment, the displaying the scene data according to a preset rule includes: and displaying the scene data of the corresponding visual angle according to the requirement of the user for observing the scene. The virtual reality scene for constructing the machine room to be monitored is a three-dimensional world, and the three-dimensional world is observed by simulating human eyes through a camera. If the user desires to view the virtual reality scene at the first-person perspective, the view of the virtual reality scene may be switched to the camera at the first-person perspective according to the tools provided by the associated graphics software development kit. If the user needs to observe the virtual reality scene at the third person perspective, the view of the virtual reality scene can be switched to the camera at the third person perspective according to the tools provided by the relevant graphic software development kit. If the user needs to observe the virtual reality scene at the aerial view angle, the view of the virtual reality scene can be switched to the aerial view angle according to the tools provided by the related graphic software development kit.
Step 202: and acquiring the interaction intention of the user, converting the interaction intention into corresponding interaction information, and acquiring the monitoring information of the displayed scene according to the interaction information.
In this embodiment, the interaction intention of the user may obtain the direction, the acceleration, and the rotation angular velocity of the limb movement of the user through the sensor, convert the obtained data into corresponding interaction information through corresponding operations, and obtain the monitoring information of the displayed scene according to the interaction information.
In an exemplary embodiment, converting the interaction intention of the user into corresponding interaction information includes: acquiring direction data or rotation angle data of the movement of a user; and mapping the acquired direction data or angle data into corresponding interactive information through a corresponding algorithm. The specific implementation mode can be realized by the following steps:
When acquiring the direction data of the user movement, the implementation mode is as follows:
the method for acquiring the mobile interaction intention of a user in a virtual reality scene comprises the following steps: acquiring interactive intentions of front, back, left and right movement of a user in a scene through a direction input key on the mobile terminal; and mapping the interaction intention data to obtain corresponding interaction data. The user can input the key information according to the preset setting and the user requirement through the direction input key on the mobile terminal. Once the key is input, the specific setting corresponding to the fixed distance moved by the user can be set as required. For example: pressing the key once, which is equivalent to the user moving 0.1 meter; when the user moves 0.2 meters forward in the virtual reality scene, the up input key may be pressed 2 times in succession.
when obtaining the angle data of the user rotation, the implementation mode is as follows:
the method includes the steps of obtaining a rotation interaction intention of a user in a virtual reality scene, and comprises the following steps: acquiring the rotation angle of a user in a scene through a gyroscope sensor to obtain angle data; and mapping the angle data through integral processing to obtain corresponding interactive data.
The gyroscope sensor is located in a wearing device of a user, wherein the wearing device can be a mobile terminal.
The specific implementation manner of acquiring the rotation angle of the user in the scene through the gyroscope sensor can be as follows: the motion state of the gyro sensor is monitored, and the angle of the user rotating in the scene can be determined according to the reading of the gyro sensor.
the motion state of the gyro sensor can be divided into two types: a "moving" state and an "unmoving" state. When the equipment where the gyroscope sensor is positioned is in a large-amplitude motion state, the equipment is judged to be in a motion state; when the equipment where the gyroscope sensor is located is in a static state or in uniform motion, the equipment is judged to be in an 'unmoved' state.
First, a preset threshold value is acquired.
Next, the motion state of the gyro sensor is judged. When the reading of the gyroscope sensor is larger than a set threshold value, judging that the gyroscope sensor is in a motion state; and when the reading of the gyroscope sensor is smaller than a set threshold value, judging that the gyroscope sensor is in a non-motion state.
and finally, determining the rotating angle in the user scene according to the motion state of the gyroscope sensor.
And when the reading of the gyroscope sensor is larger than a set threshold value, judging that the gyroscope sensor is in a motion state, and using the output of the reading of the gyroscope sensor as the output data of the motion state of the current equipment.
and when the reading of the gyroscope sensor is smaller than a set threshold value, judging that the gyroscope sensor is in a non-motion state, and using the angle value calculated by the accelerometer and the magnetometer as the motion state output data of the current equipment. In this embodiment, the specific implementation of the angle values calculated using the accelerometer and the magnetometer may be as follows:
firstly, reading of three axes of accelerometers ax, ay and az is obtained;
secondly, after reading the readings of the three axes ax, ay and az by the accelerometer, calculating the instantaneous angle around the x axis and the instantaneous angle around the y axis by the following formulas:
the instantaneous angle around the x-axis is: θ ═ arcsin (-ax/g), where θ denotes the instantaneous angle around the x-axis and g denotes the acceleration of gravity; (ii) a
The instantaneous angle around the y-axis is: phi is arctan (ay/az), where phi denotes the instantaneous angle around the y-axis;
thirdly, after calculating the instantaneous angle around the x axis and the instantaneous angle around the y axis, acquiring the reading of the magnetometer;
fourthly, calculating the angle of rotation around the z axis according to the reading of the magnetometer:
Where ψ represents an angle of rotation about the z-axis, and Mx, My, and Mz represent outputs of the magnetometer x, y, and z axes, respectively.
In an exemplary embodiment, the obtaining the monitoring information of the displayed scene according to the interaction information includes: when the interaction information indicates that a user views first equipment in the virtual reality scene, controlling the transmission module to acquire monitoring information of second equipment acquired in the machine room to be monitored; wherein the second device corresponds to the first device. The equipment in the virtual scene corresponds to the equipment in the real machine room one by one, the first equipment is the virtual equipment in the virtual scene, and the second equipment is the real equipment in the machine room to be monitored. A sensor, a camera and the like for monitoring can be installed in the real machine room to be monitored so as to obtain monitoring information.
step 203: and transmitting the monitoring information of the machine room to be monitored.
in this embodiment, the interaction control module acquires monitoring information of a displayed scene according to the interaction information; and the information transmission module transmits the acquired related monitoring information and displays the related monitoring information in a virtual reality scene.
As shown in fig. 3, an exemplary embodiment of the present invention is implemented as follows:
step 300: and (4) deriving relevant information of the model established by the modeling software.
In this embodiment, the relevant information of the model includes one or more of the following: vertices of the model, normal vectors, texture maps.
step 301: the derived related information is converted into scene data of a predetermined format.
In the embodiment, the derived related information is converted into scene data in a corresponding format of the operating environment in the virtual reality system.
step 302: and displaying the scene data according to a preset rule.
in this embodiment, after a virtual reality scene of a machine room to be monitored is constructed, scene data of a corresponding view angle is displayed according to a requirement of a user for observing the scene.
Step 303: acquiring direction data or rotation angle data of the movement of a user; and mapping the acquired direction data or angle data into corresponding interactive information through a corresponding algorithm.
in this embodiment, acquiring a mobile interaction intention of a user in a virtual reality scene includes: acquiring interactive intentions of front, back, left and right movement of a user in a scene through a direction input key on the mobile terminal; and mapping the interaction intention data to obtain corresponding interaction data.
The method includes the steps of obtaining a rotation interaction intention of a user in a virtual reality scene, and comprises the following steps: acquiring the rotation angle of a user in a scene through a gyroscope sensor to obtain angle data; and mapping the angle data through integral processing to obtain corresponding interactive data.
Step 304: and acquiring the monitoring information of the displayed scene according to the interaction information.
In this embodiment, when the interaction information indicates that a user views a first device in the virtual reality scene, acquiring monitoring information of a second device acquired in the machine room to be monitored according to the interaction information; wherein the second device corresponds to the first device.
Step 305: and transmitting the monitoring information of the machine room to be monitored.
based on a specific example of this embodiment, the method is applied to a machine room monitoring system of a virtual reality technology, and the machine room scene is displayed as a virtual reality scene for a user to view, so that the user can enjoy better interactive experience.
It will be understood by those of ordinary skill in the art that all or some of the steps of the methods, systems, functional modules/units in the devices disclosed above may be implemented as software, firmware, hardware, and suitable combinations thereof. In a hardware implementation, the division between functional modules/units mentioned in the above description does not necessarily correspond to the division of physical components; for example, one physical component may have multiple functions, or one function or step may be performed by several physical components in cooperation. Some or all of the components may be implemented as software executed by a processor, such as a digital signal processor or microprocessor, or as hardware, or as an integrated circuit, such as an application specific integrated circuit. Such software may be distributed on computer readable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media). The term computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data, as is well known to those of ordinary skill in the art. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, Digital Versatile Disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by a computer. In addition, communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media as known to those skilled in the art.

Claims (10)

1. A machine room monitoring system based on a virtual reality technology is characterized by comprising a scene construction module, a view conversion module, an interaction control module and an information transmission module;
the scene construction module is used for constructing a virtual reality scene of the machine room to be monitored to obtain scene data; the virtual reality scene comprises one or more of the following devices in the machine room to be monitored: the system comprises a cabinet, a case, a rack, a server, a storage and a router;
the view conversion module is used for displaying the scene data according to a preset rule;
the interaction control module is used for converting the interaction intention of the user into corresponding interaction information and acquiring the monitoring information of the displayed scene according to the interaction information;
And the information transmission module is used for transmitting the monitoring information of the machine room to be monitored.
2. The machine room monitoring system based on the virtual reality technology as claimed in claim 1, wherein the scene construction module is specifically configured to:
exporting relevant information of the model established by the modeling software;
Converting the derived related information into scene data of a predetermined format; wherein the relevant information of the model comprises one or more of the following: vertices of the model, normal vectors, texture maps.
3. The machine room monitoring system based on virtual reality technology according to claim 1, wherein the converting of the interaction intention of the user into corresponding interaction information comprises:
Acquiring direction data or rotation angle data of the movement of a user;
And mapping the acquired direction data or angle data into corresponding interactive information through a corresponding algorithm.
4. The machine room monitoring system based on virtual reality technology according to claim 3, wherein the obtaining of the monitoring information of the displayed scene according to the interaction information includes:
When the interaction information indicates that a user views first equipment in the virtual reality scene, controlling the transmission module to acquire monitoring information of second equipment acquired in the machine room to be monitored; wherein the second device corresponds to the first device.
5. the machine room monitoring system based on the virtual reality technology as claimed in claim 1, wherein the displaying the scene data according to a preset rule comprises:
And displaying the scene data of the corresponding visual angle according to the requirement of the user for observing the scene.
6. a method for monitoring a machine room based on a virtual reality technology is characterized by comprising the following steps:
Constructing a virtual reality scene of a machine room to be monitored to obtain scene data; the virtual reality scene comprises one or more of the following devices in the machine room to be monitored: the system comprises a cabinet, a case, a rack, a server, a storage and a router;
Displaying the scene data according to a preset rule;
Acquiring the interaction intention of a user, converting the interaction intention into corresponding interaction information, and acquiring the monitoring information of a displayed scene according to the interaction information;
And transmitting the monitoring information of the machine room to be monitored.
7. The method for monitoring the machine room based on the virtual reality technology according to claim 6, wherein the constructing a virtual reality scene of the machine room to be monitored to obtain scene data comprises:
Exporting relevant information of the model established by the modeling software;
Converting the derived related information into scene data of a predetermined format; wherein the relevant information of the model comprises one or more of the following: vertices of the model, normal vectors, texture maps.
8. the method for monitoring the machine room based on the virtual reality technology according to claim 6, wherein the obtaining of the interaction intention of the user and the conversion into the corresponding interaction information comprises:
Acquiring direction data or rotation angle data of the movement of a user;
And mapping the acquired direction data or angle data into corresponding interactive information through a corresponding algorithm.
9. The method for monitoring the machine room based on the virtual reality technology according to claim 8, wherein the obtaining of the monitoring information of the displayed scene according to the interaction information includes:
When the interaction information indicates that a user views first equipment in the virtual reality scene, acquiring monitoring information of second equipment collected in the machine room to be monitored according to the interaction information; wherein the second device corresponds to the first device.
10. The method for monitoring the machine room based on the virtual reality technology according to claim 6, wherein the displaying the scene data according to a preset rule comprises:
And displaying the scene data of the corresponding visual angle according to the requirement of the user for observing the scene.
CN201910701438.5A 2019-07-31 2019-07-31 Machine room monitoring system and method based on virtual reality technology Withdrawn CN110543236A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910701438.5A CN110543236A (en) 2019-07-31 2019-07-31 Machine room monitoring system and method based on virtual reality technology

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910701438.5A CN110543236A (en) 2019-07-31 2019-07-31 Machine room monitoring system and method based on virtual reality technology

Publications (1)

Publication Number Publication Date
CN110543236A true CN110543236A (en) 2019-12-06

Family

ID=68709936

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910701438.5A Withdrawn CN110543236A (en) 2019-07-31 2019-07-31 Machine room monitoring system and method based on virtual reality technology

Country Status (1)

Country Link
CN (1) CN110543236A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111659106A (en) * 2020-06-10 2020-09-15 网易(杭州)网络有限公司 Game interaction method, device, equipment and storage medium
CN113242410A (en) * 2021-04-30 2021-08-10 北京京东乾石科技有限公司 Monitoring alarm method and device, monitoring display method and device and storage medium
CN113868102A (en) * 2020-06-30 2021-12-31 伊姆西Ip控股有限责任公司 Method, electronic device and computer program product for information display

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106445777A (en) * 2016-08-31 2017-02-22 广州和运科技发展有限公司 Machine room smart 3D inspection system
CN106657906A (en) * 2016-12-13 2017-05-10 国家电网公司 Information equipment monitoring system with function of self-adaptive scenario virtual reality
CN107346174A (en) * 2017-06-15 2017-11-14 武汉秀宝软件有限公司 A kind of exchange method and system of actual environment and virtual environment
CN109660508A (en) * 2018-10-23 2019-04-19 平安壹钱包电子商务有限公司 Data visualization method, electronic device, computer equipment and storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106445777A (en) * 2016-08-31 2017-02-22 广州和运科技发展有限公司 Machine room smart 3D inspection system
CN106657906A (en) * 2016-12-13 2017-05-10 国家电网公司 Information equipment monitoring system with function of self-adaptive scenario virtual reality
CN107346174A (en) * 2017-06-15 2017-11-14 武汉秀宝软件有限公司 A kind of exchange method and system of actual environment and virtual environment
CN109660508A (en) * 2018-10-23 2019-04-19 平安壹钱包电子商务有限公司 Data visualization method, electronic device, computer equipment and storage medium

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111659106A (en) * 2020-06-10 2020-09-15 网易(杭州)网络有限公司 Game interaction method, device, equipment and storage medium
CN111659106B (en) * 2020-06-10 2023-03-31 网易(杭州)网络有限公司 Game interaction method, device, equipment and storage medium
CN113868102A (en) * 2020-06-30 2021-12-31 伊姆西Ip控股有限责任公司 Method, electronic device and computer program product for information display
CN113242410A (en) * 2021-04-30 2021-08-10 北京京东乾石科技有限公司 Monitoring alarm method and device, monitoring display method and device and storage medium

Similar Documents

Publication Publication Date Title
KR101777820B1 (en) Perception based predictive tracking for head mounted displays
US9355452B2 (en) Camera and sensor augmented reality techniques
CN106846497B (en) Method and device for presenting three-dimensional map applied to terminal
CN110543236A (en) Machine room monitoring system and method based on virtual reality technology
CN107329671B (en) Model display method and device
CN108318027B (en) Method and device for determining attitude data of carrier
CN108960947A (en) Show house methods of exhibiting and system based on virtual reality
TWI783472B (en) Ar scene content generation method, display method, electronic equipment and computer readable storage medium
CN108389264A (en) Coordinate system determines method, apparatus, storage medium and electronic equipment
JP2021185484A (en) Display of object based on multiple models
CN109992111B (en) Augmented reality extension method and electronic device
US11080943B2 (en) Method and apparatus for displaying with 3D parallax effect
KR102199772B1 (en) Method for providing 3D modeling data
CN113407084A (en) Display content updating method, head-mounted display device and computer readable medium
US10295403B2 (en) Display a virtual object within an augmented reality influenced by a real-world environmental parameter
EP3961571A1 (en) Virtual paintbrush implementing method and apparatus, and computer readable storage medium
CN111179438A (en) AR model dynamic fixing method and device, electronic equipment and storage medium
KR102314782B1 (en) apparatus and method of displaying three dimensional augmented reality
CN106802716B (en) Data processing method of virtual reality terminal and virtual reality terminal
CN112987924A (en) Method, apparatus, device and storage medium for device interaction
CN110544316B (en) Virtual reality playback method, system, equipment and storage medium
KR20210050997A (en) Method and apparatus for estimating pose, computer-readable storage medium and computer program for controlling the holder device
CN112825198B (en) Mobile tag display method, device, terminal equipment and readable storage medium
CN115641397A (en) Method and system for synthesizing and displaying virtual image
CN114862997A (en) Image rendering method and apparatus, medium, and computer device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication

Application publication date: 20191206

WW01 Invention patent application withdrawn after publication