CN106789198B - Computing node management method and system - Google Patents
Computing node management method and system Download PDFInfo
- Publication number
- CN106789198B CN106789198B CN201611116226.3A CN201611116226A CN106789198B CN 106789198 B CN106789198 B CN 106789198B CN 201611116226 A CN201611116226 A CN 201611116226A CN 106789198 B CN106789198 B CN 106789198B
- Authority
- CN
- China
- Prior art keywords
- computing node
- node
- information
- central server
- computing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/02—Standardisation; Integration
- H04L41/0246—Exchanging or transporting network management information using the Internet; Embedding network management web servers in network elements; Web-services-based protocols
- H04L41/0253—Exchanging or transporting network management information using the Internet; Embedding network management web servers in network elements; Web-services-based protocols using browsers or web-pages for accessing management information
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/02—Standardisation; Integration
- H04L41/0246—Exchanging or transporting network management information using the Internet; Embedding network management web servers in network elements; Web-services-based protocols
- H04L41/0273—Exchanging or transporting network management information using the Internet; Embedding network management web servers in network elements; Web-services-based protocols using web services for network management, e.g. simple object access protocol [SOAP]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/08—Configuration management of networks or network elements
- H04L41/0803—Configuration setting
- H04L41/0823—Configuration setting characterised by the purposes of a change of settings, e.g. optimising configuration for enhancing reliability
- H04L41/0826—Configuration setting characterised by the purposes of a change of settings, e.g. optimising configuration for enhancing reliability for reduction of network costs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/08—Configuration management of networks or network elements
- H04L41/0889—Techniques to speed-up the configuration process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/10—Protocols in which an application is distributed across nodes in the network
Landscapes
- Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Computer And Data Communications (AREA)
Abstract
The invention relates to the field of cloud computing, in particular to a computing node management method and system. The method comprises the following steps: s1, a central server receives computing node information sent by a web management interface; s2, the central server side inquires whether the computing node really exists and is available according to the received computing node information, if yes, the central server side inquires whether record information related to a host where the computing node is located exists in a central database, if yes, physical information of the computing node is obtained and updated through an interface of livirt, and if not, the step S3 is executed; and S3, the central server mounts the stored directory on a new computing node, initializes the configuration of the computing node, and calls an interface of livirt to acquire and update the physical information of the computing node. The method and the system shield the difference of the virtualization layer hypervisor, can simultaneously carry out unified management on the computing nodes, are simple and quick, greatly reduce the management cost and improve the working efficiency.
Description
Technical Field
The invention relates to the field of cloud computing, in particular to a computing node management method and system.
Background
At present, the bottom layer of virtualization has various implementation modes, and a lot of virtualization systems are born in order to realize the unified management of virtual machines. Currently, there are two main types of virtualization systems for computing node management:
one is a single-machine version virtualization system represented by virt-manager. The single-machine version virtualization system cannot realize the public use of a plurality of computer points and cannot form the concept of a resource pool due to the limitation of a single machine, so that the virtualization scale is greatly limited, and most of the virtualization systems only aim at one hypervisor and are weak in general use.
And secondly, a virtualization system with multiple computing nodes represented by openstack and the like. The OpenStack project is an open-source cloud computing platform, and cloud computing developers and technicians all over the world jointly create the OpenStack project. OpenStack provides an infrastructure as a service (IaaS) solution through a set of related services. Each service provides an Application Programming Interface (API), facilitating such integration. For a virtualization system represented by openstack, the defects of complex deployment and high expansion difficulty of computing nodes exist, and convenient and rapid deployment cannot be realized.
Disclosure of Invention
In order to overcome at least one of the above-mentioned drawbacks (shortcomings) of the prior art, the present invention provides a method for managing a compute node that can quickly add a compute node.
The invention also provides a computing node management system for rapidly increasing computing nodes.
In order to solve the technical problems, the technical scheme of the invention is as follows:
a computing node management method comprises the following steps:
s1, a central server receives computing node information sent by a web management interface;
s2, the central server side inquires whether the computing node really exists and is available according to the received computing node information, if yes, the central server side inquires whether record information related to a host corresponding to the computing node exists in a central database, if yes, physical information of the computing node is obtained and updated through an interface of livirt, the computing node information is marked as available, and if not, the step S3 is executed;
and S3, the central server mounts the stored directory on a new computing node, initializes the configuration of the computing node, and calls an interface of livirt to acquire and update the physical information of the computing node.
The invention manages the computing nodes by using the central server, the central server can inquire the record information of the computing nodes through the database, updates different computing nodes by using the livirt interface, and can acquire and store the information of the new computing nodes by using the livirt interface while increasing the new computing nodes.
In the above solution, the computing node information sent by the web management interface in step S1 is input to the web management interface by means of external input. The method comprises the steps of obtaining external requirements through a web management interface, and then informing a center server side to execute through the web management interface, so that cloud operation is achieved.
In the above scheme, the computing node information includes an IP address and a login password of the computing node.
In the above scheme, in step S2, the central server queries whether the computing node is really present and available by using ping according to the IP address of the computing node.
In the above scheme, in step S3, the central server mounts the stored directory to the new computing node in a manner of remotely calling a shell command and in a manner of glusterfs or nfs.
In the above solution, in step S3, the configuration initialization of the computing node is implemented by executing a script remotely.
In the above scheme, the method further comprises:
and S4, when the computing node needs to be deleted, the central server side marks the computing node information recorded in the central database as unavailable.
The deletion of the nodes only needs to mark the information of the computing nodes in the central database, and the computing nodes do not need to be deleted directly, thereby providing convenience for the subsequent increase of the computing nodes.
A computing node management system comprises a central server, wherein the central server comprises a receiving module, an inquiry updating module and a node adding module:
the receiving module is used for receiving the computing node information sent by the web management interface;
the query updating module is used for querying whether the computing node really exists and is available according to the received computing node information, if so, querying whether record information related to a host where the computing node is located exists in a central database, if so, acquiring and updating physical information of the computing node through an interface of livirt, and if not, notifying the node adding module;
and the node adding module is used for mounting the stored directory on a new computing node, initializing the configuration of the computing node, and calling an interface of livirt to acquire and update the physical information of the computing node.
The system of the invention utilizes the livirt interface to carry out unified management on different computing nodes by arranging the query updating module, the difference of the virtualization layer hypervisor is not required to be considered in the increasing process of the computing nodes, the operation is simple and quick, the management cost is greatly reduced, and the working efficiency is improved.
In the above solution, the central server further includes:
and the node deleting module is used for marking the computing node information recorded by the central database as unavailable when the computing node is required to be deleted.
In the above scheme, when the node adding module mounts the storage directory to the new computing node, the storage directory is specifically mounted to the new computing node in a mode of remotely calling a shell command and in a mode of glusterfs or nfs.
Compared with the prior art, the technical scheme of the invention has the beneficial effects that:
the invention utilizes libvirt interface for managing the computing nodes, shields the hypervisor difference of the virtualization layer, can realize flexible management on the computing nodes, is convenient and quick, greatly reduces the management cost and improves the working efficiency.
Drawings
Fig. 1 is a flowchart of a method for managing a compute node according to an embodiment of the present invention.
Fig. 2 is an architecture diagram of a computing node management system according to an embodiment of the present invention.
Fig. 3 is an architecture diagram of a cloud computing resource virtualization system according to embodiment 3 of the present invention.
Detailed Description
The drawings are for illustrative purposes only and are not to be construed as limiting the patent;
for the purpose of better illustrating the embodiments, certain features of the drawings may be omitted, enlarged or reduced, and do not represent the size of an actual product;
it will be understood by those skilled in the art that certain well-known structures in the drawings and descriptions thereof may be omitted.
In the description of the present invention, it is to be understood that, furthermore, the terms "first" and "second" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implying any number of technical features indicated. Thus, a defined feature of "first", "second", may explicitly or implicitly include one or more of that feature. In the description of the present invention, "a plurality" means two or more unless otherwise specified.
In the description of the present invention, it should be noted that, unless otherwise explicitly specified or limited, the terms "mounted" and "connected" are to be interpreted broadly, e.g., as being either fixedly connected, detachably connected, or integrally connected; can be mechanically or electrically connected; they may be connected directly or indirectly through intervening media, so to speak, as communicating between the two elements. The specific meaning of the above terms in the present invention can be understood in specific cases to those skilled in the art.
The technical solution of the present invention is further described below with reference to the accompanying drawings and examples.
Example 1
Fig. 1 is a flowchart of a method for managing a compute node according to an embodiment of the present invention. Referring to fig. 1, a method for managing a compute node according to this embodiment includes:
s101, a central server receives computing node information sent by a web management interface; the computing node information sent by the web management interface is input into the web management interface in an external input mode. After the information required by the computing node is input through the web management interface, the web management interface is transmitted to the central server through an http protocol. The information of the computing node comprises but is not limited to an IP address and a login password of the computing node, and the setting of the login password is mainly used for secret-free initialization of the computing node, so that the central shared storage area can be initialized conveniently through a remote command and the initialization of a node virtualization environment can be facilitated.
S102, the central server side inquires whether a computing node really exists and is available according to received computing node information, if yes, the central server side inquires whether record information related to a host corresponding to the computing node exists in a central database, if yes, physical information of the computing node is obtained and updated through a livirt interface, the computing node information is marked as available, and if not, the step S103 is executed; whether the computing node really exists and is available can be achieved through a ping mode, and after receiving the computing node information, the central server side can determine whether the computing node really exists and is available through the IP address of the central server side in a ping mode. In this step, the central server determines whether the computing node exists really and is available, in which the purpose is that the deleted computing node information may exist in the central database, if the computing node needs to be added, the central server only needs to update the physical information of the computing node in the central database and mark the physical information as available, so that the addition of the computing node can be completed, and if the computing node needing to be added does not have the record information in the central database, the computing node needs to be initialized by executing step S103. The physical information of the computing node mentioned in this step includes, but is not limited to, the disk size, cpu, memory, network card information, and the like of the computing node, and the central server automatically acquires and updates the physical information.
In this step, the livirt interface supports the dynamic migration in the backend mirroring mode by modifying the implementation manner of the livirt dynamic migration interface.
S103, the central server mounts the stored directory on a new computing node, initializes the configuration of the computing node, and calls an interface of livirt to acquire and update the physical information of the computing node. The stored catalog is positioned in the central shared storage, and the content comprises the following components: the method comprises the steps of initializing script information needed by a computing node, creating a mirror image template and a related configuration file template needed by a virtual machine, managing files of a virtual machine cluster network, and mirror images and snapshot files of other virtual machines. The central storage is realized through the central shared storage, so that the management of data is convenient on one hand, and the virtual machine can be quickly migrated from one to another on the other hand. The mounting process is that the central server mounts the stored directory to a new computing node in a remote shell command calling mode and in a glusterfs or nfs mode, and the configuration initialization of the computing node is realized in a remote script executing mode. S104, when the computing node needs to be deleted, the central server side marks the computing node information recorded in the central database as unavailable. The deleting step does not need to delete the computing nodes directly, but marks the computing node information in the central database, and the mode aims to provide convenience for the later re-addition of the computing nodes. The deletion requirement of the computing node can be acquired through the web management interface, and then the central server side is informed to carry out the computing node deletion step. When the deleted computing node is subsequently restarted, the method can be realized only by executing the computing node adding step. The central server side can judge whether to add or delete the computing node according to the information mark bit carried by the computing node information sent by the web.
In the specific implementation process, after the central shared storage network is unexpectedly disconnected for a period of time and the network is restored again, the virtual machine cannot be identified, and therefore a mounting mode of gluterfs + nfs is adopted in step S103. The replication tape roll of the gluterfs is unstable, the problem is easy to occur in the production process, and the problem that the virtual machine cannot be identified after the central shared network is disconnected and recovered can be well solved by using the replication tape roll + raid0 instead.
In the method of the present invention, the information stored in the central database includes, but is not limited to, host information, virtual machine information, template information, and cluster network related information.
The method for managing the computing nodes can realize the rapid addition and deletion of the computing nodes, and the addition and deletion process shields the difference of the virtualization layer hypervisor, thereby greatly reducing the management cost and improving the working efficiency.
Example 2
The invention further provides a computing node management system on the basis of the embodiment 1. Referring to fig. 2, a computing node management system in this embodiment specifically includes a central server 200, where the central server 200 includes a receiving module 201, an inquiry updating module 202, a node adding module 203, and a node deleting module 204.
A receiving module 201, configured to receive computing node information sent from a web management interface; wherein the computing node information includes, but is not limited to, an IP address of the computing node, a login password.
The query updating module 202 is configured to query whether a computing node exists and is available according to received computing node information, if so, query whether record information related to a host corresponding to the computing node exists in a central database, and if so, acquire and update physical information of the computing node through an interface of livirt and mark the computing node information as available, where the physical information includes but is not limited to information of a disk size, a cpu, a memory, and a network card of the computing node; if not, notifying the node to increase the module; in order to determine whether a computing node exists and is available, the query and update module 202 may be implemented in a ping manner, and after receiving the computing node information, the query and update module 202 may determine whether the computing node exists and is available in a ping manner through its IP address.
The node adding module 203 is configured to mount the stored directory on a new computing node, initialize configuration of the computing node, and call an interface of livirt to obtain and update physical information of the computing node. When the node adding module 203 mounts the storage directory to the new computing node, the storage directory is specifically mounted to the new computing node in a mode of remotely calling a shell command and in a mode of glusterfs or nfs. The stored catalog is positioned in the central shared storage, and the content comprises the following components: the method comprises the steps of initializing script information needed by a computing node, creating a mirror image template and a related configuration file template needed by a virtual machine, managing files of a virtual machine cluster network, and mirror images and snapshot files of other virtual machines. The central storage is realized through the central shared storage, so that the management of data is convenient on one hand, and the virtual machine can be quickly migrated from one to another on the other hand.
And the node deleting module 204 is configured to mark the computing node information recorded in the central database as unavailable when the computing node needs to be deleted. The system of the present invention does not need to delete the computing node directly, but marks the computing node information in the node deletion module 204, which aims to provide convenience for the later re-addition of the computing node. The deletion requirement of the computing node can be acquired through the web management interface, and then the node deletion module 204 of the center server is informed to perform the computing node deletion step. When the deleted computing node needs to be restarted later, the method can be realized only by the node adding module 203. The central server 200 may determine whether to add or delete a compute node according to an information flag carried in the compute node information sent by the web.
The system of the invention utilizes the livirt interface to carry out unified management on different computing nodes by arranging the query updating module, the difference of the virtualization layer hypervisor is not required to be considered in the increasing process of the computing nodes, the operation is simple and quick, the management cost is greatly reduced, and the working efficiency is improved.
Example 3
On the basis of embodiment 2, this embodiment further describes embodiment 2 in combination with a cloud computing resource virtualization system where a central server is located.
The cloud computing resource virtualization system comprises a management end, a central service layer and a computing node end, wherein the management end is provided with a web management interface 100, the central server layer is provided with a central service end 200, a central database 300 and a central shared storage area 400, and the computing node end is provided with at least one computing node 500;
the Web management interface 100 is used for a manager to input information required by the computing node, including an IP address, a login password and the like of the computing node, and the Web management interface 100 transmits the input information to the central server 200 through an http protocol;
the central server 200 is responsible for processing incoming and outgoing information from the web management interface 100. The central server 200 firstly analyzes the received information, extracts and stores the required information, and the information storage is completed through the central database 300; the central server 200 then determines whether the computing node is actually present and available by pinging the computing node's IP address. If the physical information is available, the central database 300 is searched first, whether the relevant records of the host corresponding to the computing node exist is checked, if the relevant records of the host corresponding to the computing node exist, the relevant interfaces of livirt acquire and update the physical information of the computing node, wherein the physical information comprises but is not limited to the information of the disk size, the cpu, the memory and the network card of the computing node; if no relevant record is found in the central database 300, the central server 200 mounts the directory of the central shared storage area 400 to the new computing node 500 in a manner of remotely calling a shell command and in a manner of glusterfs or nfs, initializes relevant configuration of the computing node 500 by remotely executing a script, and then calls a relevant interface to acquire and update relevant information of the computing node.
When the computing node 500 is deleted, the central server 200 sets the flag bit of the record related to the computing node 500 to be deleted in the central database 300 as unavailable, so that convenience is provided for the subsequent re-addition of the computing node 500.
The same or similar reference numerals correspond to the same or similar parts;
the positional relationships depicted in the drawings are for illustrative purposes only and are not to be construed as limiting the present patent;
it should be understood that the above-described embodiments of the present invention are merely examples for clearly illustrating the present invention, and are not intended to limit the embodiments of the present invention. Other variations and modifications will be apparent to persons skilled in the art in light of the above description. And are neither required nor exhaustive of all embodiments. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present invention should be included in the protection scope of the claims of the present invention.
Claims (5)
1. A method for managing a compute node, comprising the steps of:
s1, a central server receives computing node information sent by a web management interface;
s2, the central server side inquires whether the computing node really exists and is available according to the received computing node information, if yes, the central server side inquires whether record information related to a host corresponding to the computing node exists in a central database, if yes, physical information of the computing node is obtained and updated through an interface of libvirt, the computing node information is marked as available, and if not, the step S3 is executed;
s3, the central server side mounts the stored directory on a new computing node, initializes the configuration of the computing node, and calls an interface of libvirt to acquire and update physical information of the computing node;
s4, when the computing node needs to be deleted, the central server side marks the computing node information recorded by the central database as unavailable;
the computing node information sent by the web management interface in the step S1 is input to the web management interface by means of external input; the computing node information comprises an IP address and a login password of the computing node; in the step S2, the central server side inquires whether the computing node really exists and is available or not in a ping mode according to the IP address of the computing node;
in step S2, the central server determines whether the compute node exists and is available, where the purpose of the determination is that the deleted compute node may exist in the central database, and if the compute node needs to be added, the central server only needs to update the physical information of the compute node in the central database and mark the update as available, so as to complete the addition of the compute node, and if the compute node that needs to be added does not have record information in the central database, the central server needs to execute step S3 to initialize the compute node.
2. The method as claimed in claim 1, wherein in step S3, the central server mounts the stored directory onto the new compute node by means of a remote call shell command and by means of glusterfs or nfs.
3. The method for managing computing nodes according to claim 1, wherein in step S3, the initialization of configuration of the computing nodes is realized by executing scripts remotely.
4. A computing node management system is characterized by comprising a central server, wherein the central server comprises a receiving module, an inquiry updating module and a node adding module:
the receiving module is used for receiving the computing node information sent by the web management interface;
the query updating module is used for querying whether the computing node really exists and is available according to the received computing node information, if so, querying whether record information related to a host where the computing node is located exists in a central database, if so, acquiring and updating physical information of the computing node through an interface of libvirt, and if not, informing the node adding module;
the node adding module is used for mounting the stored directory on a new computing node, initializing the configuration of the computing node, and calling an interface of livirt to acquire and update physical information of the computing node;
the center server also comprises:
the node deleting module is used for marking the computing node information recorded by the central database as unavailable when the computing node is required to be deleted;
the computing node information sent by the web management interface is input into the web management interface in an external input mode; the computing node information comprises an IP address and a login password of the computing node; the central server side inquires whether the computing node really exists and is available or not in a ping mode according to the IP address of the computing node;
the central server side judges whether the computing node really exists and is available, and aims to solve the problem that deleted computing node information possibly exists in a central database, if the computing node needs to be added, the central server side only needs to update physical information of the computing node in the central database and mark the updated physical information as available to complete the addition of the computing node, if the computing node needing to be added does not have record information in the central database, the central server side needs to mount a stored directory on a new computing node, initialize the configuration of the computing node, and call an interface of libvirt to acquire and update the physical information of the computing node to initialize the computing node.
5. The system according to claim 4, wherein the node adding module mounts the stored directory to the new computing node by calling a shell command remotely and by using glusterfs or nfs.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201611116226.3A CN106789198B (en) | 2016-12-07 | 2016-12-07 | Computing node management method and system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201611116226.3A CN106789198B (en) | 2016-12-07 | 2016-12-07 | Computing node management method and system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106789198A CN106789198A (en) | 2017-05-31 |
CN106789198B true CN106789198B (en) | 2020-10-02 |
Family
ID=58882177
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201611116226.3A Active CN106789198B (en) | 2016-12-07 | 2016-12-07 | Computing node management method and system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106789198B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107391640B (en) * | 2017-07-11 | 2020-12-08 | 浪潮云信息技术股份公司 | Method for realizing automatic deployment of SQL Server database mirror mode |
CN111240895A (en) * | 2019-12-31 | 2020-06-05 | 深圳证券通信有限公司 | OpenStack-oriented node batch backup system method |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103699430A (en) * | 2014-01-06 | 2014-04-02 | 山东大学 | Working method of remote KVM (Kernel-based Virtual Machine) management system based on J2EE (Java 2 Platform Enterprise Edition) framework |
CN104657215A (en) * | 2013-11-19 | 2015-05-27 | 南京鼎盟科技有限公司 | Virtualization energy-saving system in Cloud computing |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10467036B2 (en) * | 2014-09-30 | 2019-11-05 | International Business Machines Corporation | Dynamic metering adjustment for service management of computing platform |
-
2016
- 2016-12-07 CN CN201611116226.3A patent/CN106789198B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104657215A (en) * | 2013-11-19 | 2015-05-27 | 南京鼎盟科技有限公司 | Virtualization energy-saving system in Cloud computing |
CN103699430A (en) * | 2014-01-06 | 2014-04-02 | 山东大学 | Working method of remote KVM (Kernel-based Virtual Machine) management system based on J2EE (Java 2 Platform Enterprise Edition) framework |
Non-Patent Citations (2)
Title |
---|
《libvirt详解》;chenyulancn;《https://blog.csdn.net/chenyulancn/article/details/8916452》;20130512;全文 * |
《高清互动电视云服务平台及其关键技术研究》;孙亮等;《中国有线电视》;20130620;全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN106789198A (en) | 2017-05-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11226847B2 (en) | Implementing an application manifest in a node-specific manner using an intent-based orchestrator | |
US11829263B2 (en) | In-place cloud instance restore | |
US11113158B2 (en) | Rolling back kubernetes applications | |
US11016935B2 (en) | Centralized multi-cloud workload protection with platform agnostic centralized file browse and file retrieval time machine | |
US11243792B2 (en) | Image file conversion method and apparatus | |
US11347684B2 (en) | Rolling back KUBERNETES applications including custom resources | |
EP3767471A1 (en) | Provisioning and managing replicated data instances | |
US9864791B2 (en) | Flow for multi-master replication in distributed storage | |
US20080195827A1 (en) | Storage control device for storage virtualization system | |
US8745342B2 (en) | Computer system for controlling backups using wide area network | |
CN110019514A (en) | Method of data synchronization, device and electronic equipment | |
CN110532123A (en) | The failover method and device of HBase system | |
CN105446831A (en) | Server-Free backup method in conjunction with SAN | |
US10620871B1 (en) | Storage scheme for a distributed storage system | |
US9461884B2 (en) | Information management device and computer-readable medium recorded therein information management program | |
CN111582824B (en) | Cloud resource synchronization method, device, equipment and storage medium | |
CN109407975B (en) | Data writing method, computing node and distributed storage system | |
US20160308871A1 (en) | Network element data access method and apparatus, and network management system | |
CN106789198B (en) | Computing node management method and system | |
CN100571172C (en) | A kind of multi-vendor alarm and control system and alarm method thereof in the telecommunication management network | |
JP2009251756A (en) | Client device, distributed file system, shared resource multiplexing method, and program | |
US10338849B2 (en) | Method and device for processing I/O request in network file system | |
US20150074116A1 (en) | Indexing attachable applications for computing systems | |
US10976952B2 (en) | System and method for orchestrated application protection | |
US10848405B2 (en) | Reporting progress of operation executing on unreachable host |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |