CN110533181B - Rapid training method and system for deep learning model - Google Patents

Rapid training method and system for deep learning model Download PDF

Info

Publication number
CN110533181B
CN110533181B CN201910676874.1A CN201910676874A CN110533181B CN 110533181 B CN110533181 B CN 110533181B CN 201910676874 A CN201910676874 A CN 201910676874A CN 110533181 B CN110533181 B CN 110533181B
Authority
CN
China
Prior art keywords
training
server
file server
control interface
environment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910676874.1A
Other languages
Chinese (zh)
Other versions
CN110533181A (en
Inventor
赵铭
林镇锋
易文峰
杨育
杨正刚
李小芬
徐文娟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Southern Power Grid Digital Platform Technology Guangdong Co ltd
Original Assignee
China Southern Power Grid Digital Platform Technology Guangdong Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Southern Power Grid Digital Platform Technology Guangdong Co ltd filed Critical China Southern Power Grid Digital Platform Technology Guangdong Co ltd
Priority to CN201910676874.1A priority Critical patent/CN110533181B/en
Publication of CN110533181A publication Critical patent/CN110533181A/en
Application granted granted Critical
Publication of CN110533181B publication Critical patent/CN110533181B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/54Interprogram communication
    • G06F9/546Message passing systems or structures, e.g. queues
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/54Indexing scheme relating to G06F9/54
    • G06F2209/548Queue

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Electrically Operated Instructional Devices (AREA)

Abstract

A method and a system for quickly training a deep learning model. The rapid training method comprises the following steps: pre-installing a training environment in a training server; the training server invokes training codes which are pre-led into the file server from the file server to train the deep learning model; and connecting the file server with a display terminal displaying a control interface, wherein the control interface comprises a training code updating tab. Thus, since the training environment is preinstalled in the training server, the installation and preparation of the environment are not required when training the deep learning model, and the working efficiency can be greatly improved compared with the original mode (a minimum of 7 days is required for one ordinary skilled person to install and deploy the environment). In addition, when the training server needs to be updated, the latest training code coverage operation can be automatically downloaded in the file server, so that the training code is updated rapidly.

Description

Rapid training method and system for deep learning model
Technical Field
The invention relates to the technical field of information, in particular to a method and a system for quickly training a deep learning model.
Background
Deep learning is a new field in machine learning research, and the motivation is to build and simulate a neural network for analysis learning of the human brain, which mimics the mechanism of the human brain to interpret data such as image, sound and text. Deep learning models, such as convolutional neural networks (Convolutional Neural Network, CNN), require extensive data training to be practical. According to the model training of the self-defined deep learning in the past, the following steps are needed: writing codes, installing an environment, performing local test operation, performing server operation, adjusting parameters of the server codes, and tracking training results. It takes a lot of time and effort to install the environment, adjust the parameters, and track the training results each time. How to accelerate the training process of deep learning models has become an urgent need in the industry.
Disclosure of Invention
Aiming at the problems in the prior art, the invention provides a rapid training method of a deep learning model in order to improve the training efficiency and reduce the training time of the deep learning model, and the environment construction, the code updating and the parameter adjustment can be processed in one interface by the rapid training method, so that the adjustment can be completed in a few minutes each time, and the working efficiency is greatly improved.
The technical scheme adopted for solving the technical problems is as follows: the quick training method for the deep learning model comprises the following steps:
pre-installing a training environment in a training server;
the training server invokes training codes which are pre-led into the file server from the file server to train the deep learning model;
and connecting the file server with a display terminal displaying a control interface in a signal way, wherein the control interface comprises a training code updating tab.
In the rapid training method of the deep learning model provided by the invention, training data generated in the training process of the deep learning model is stored in the file server, and the file server displays the training data through the control interface.
In the rapid training method of the deep learning model provided by the invention, the log generated in the training process of the deep learning model is stored in the file server, and the file server displays the log through the control interface.
In the rapid training method of the deep learning model provided by the invention, the control interface further comprises a training parameter adjustment tab, and the updated training parameters are transmitted to the training server through the message received by the message queue.
In the rapid training method of the deep learning model provided by the invention, the training parameter adjustment tab comprises a learning rate adjustment tab and an iteration number tab.
In the rapid training method of the deep learning model provided by the invention, relevant dependent components or software are provided in the pre-installed training environment.
In the rapid training method of the deep learning model provided by the invention, the rapid training method comprises the step of directly carrying out self-definition through the control interface when relevant dependent components or software are absent in the preloaded training environment.
In the method for quickly training a deep learning model provided by the invention, the step of directly customizing the deep learning model through the control interface when relevant dependent components or software are absent in the preloaded training environment comprises the following steps:
the training server automatically generates a dockerfile according to the existing dependent component and automatically downloads the related dependent component through the dockerfile;
adding the base image to the new dependent components generates a custom running environment for the user.
In order to solve the technical problem, the invention also provides a rapid training system of the deep learning model, which comprises a training server, a file server in signal connection with the training server and a display terminal connected with the file server, wherein the training server is preloaded with a training environment, training codes are stored in the file server, and a control interface is displayed on the display terminal and comprises a training code updating tab.
In the rapid training system of the deep learning model provided by the invention, the control interface also comprises a training parameter adjustment tab, and the updated training parameters are transmitted to the training server through the message received by the message queue; the training parameter adjustment tab includes a learning rate adjustment tab and an iteration number tab.
By implementing the rapid training method of the deep learning model, the following beneficial effects can be achieved:
1. the rapid training method comprises the following steps: pre-installing a training environment in a training server; the training server invokes training codes which are pre-led into the file server from the file server to train the deep learning model; and connecting the file server with a display terminal displaying a control interface in a signal way, wherein the control interface comprises a training code updating tab. Thus, since the training environment is preinstalled in the training server, the installation and preparation of the environment are not required when training the deep learning model, and the working efficiency can be greatly improved compared with the original mode (a minimum of 7 days is required for one ordinary skilled person to install and deploy the environment). In addition, the training codes are uploaded to the training server through the file server, if the training codes need to be updated, the new training codes are directly uploaded to the file server, and the training server can automatically download the latest training codes in the file server for covering and running, so that the updating of the training codes is completed rapidly.
2. Training data generated in the deep learning model training process are stored in the file server, and the file server displays the training data through the control interface. In this way, the training process can be tracked through the control interface without viewing in the training server.
3. And the log generated in the deep learning model training process is stored in the file server, and the file server displays the log through the control interface. Therefore, the log generated in the training process can be directly displayed in the control interface, and is convenient to read.
4. The control interface also comprises a training parameter adjustment tab, and the updated training parameters are transmitted to the training server through the information received by the information queue. Preferably, the training parameter adjustment tab includes a learning rate adjustment tab and an iteration number tab. In this way, the parameter adjustment in the training process can be completed by operating the training code updating tab card in the control interface, and each time the parameter is adjusted, the parameter adjustment can be completed in 5 minutes, and compared with the original mode (the minimum time for each adjustment in the prior art is 1 hour), the working efficiency can be greatly improved.
5. And when relevant dependent components or software are absent in the preloaded training environment, directly performing the self-defining step through the control interface. The step of directly customizing the control interface when the relevant dependent components or software are absent in the preloaded training environment comprises the following steps: the training server automatically generates a dockerfile according to the existing dependent component and automatically downloads the related dependent component through the dockerfile; adding the base image to the new dependent components generates a custom running environment for the user. Compared with the original environment modification technology which needs to take 2 days at least, the environment is customized by the process only by a few minutes, so that the working efficiency is greatly improved.
Correspondingly, the rapid training system of the deep learning model comprises a training server, a file server in signal connection with the training server and a display terminal connected with the file server, wherein a training environment is preloaded in the training server, training codes are stored in the file server, a control interface is displayed on the display terminal, and the control interface comprises a training code updating tab. It can be seen that the rapid training system corresponds to the rapid training method, and thus, the same technical effects can be achieved by implementing the rapid training system.
Drawings
FIG. 1 is a flow chart showing steps of a fast training method according to a first embodiment of the present invention;
FIG. 2 is a diagram of an updated training parameter interface of a control interface according to a first embodiment of the present invention;
FIG. 3 is a customized interface of the control interface according to a first embodiment of the present invention;
fig. 4 is a block diagram of a rapid training system according to a second embodiment of the present invention.
Detailed Description
For a clearer understanding of technical features, objects and effects of the present invention, a detailed description of embodiments of the present invention will be made with reference to the accompanying drawings. It will be apparent that the described embodiments are only some, but not all, embodiments of the invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Example 1
The embodiment provides a rapid training method of a deep learning model, referring to fig. 1, the rapid training method includes the following steps:
pre-installing a training environment in a training server;
the training server invokes training codes which are pre-led into the file server from the file server to train the deep learning model;
and (3) connecting the file server with a display terminal displaying a control interface, wherein the control interface comprises a training code updating tab, as shown in fig. 2.
The training server may employ a server or a cluster of servers.
The file server may employ a server or a cluster of servers.
Because the training environment is preinstalled in the training server, the environment is not required to be installed and prepared when the deep learning model is trained, and the working efficiency can be greatly improved compared with the original mode (the time for installing and deploying the environment by one ordinary technician is minimum by 7 days). In addition, the training codes are uploaded to the training server through the file server, if the training codes need to be updated, the new training codes are directly uploaded to the file server, the training code updating option card is selected, and the training server can automatically download the latest training code coverage operation in the file server, so that the training codes are updated rapidly. Referring to fig. 2, fig. 2 is an updated training parameter interface of the control interface, and the training code update tab is the "update FTP training code" option in fig. 2.
It should be noted that in order to ensure that the training process is efficiently and smoothly completed, it is required that dependent components or software related to the training environment are already provided in the pre-installed training environment.
Of course, if the pre-installed training environment lacks relevant dependent components or software, the training environment can be customized based on the pre-installed training environment. Specifically, the rapid training method comprises the step of directly customizing through the control interface when relevant dependent components or software are absent in the preloaded training environment. The step of directly customizing the control interface when the relevant dependent components or software are absent in the preloaded training environment comprises the following steps: the training server automatically generates a dockerfile according to the existing dependent component and automatically downloads the related dependent component through the dockerfile; adding the base image to the new dependent components generates a custom running environment for the user. The control interface includes an environment custom interface as shown in fig. 3, where the environment custom interface includes a "select base environment" field, an "existing dependency" field, an "add dependency" field, an "environment name" field, an "environment version" field, and an "environment description field. We can perform the above steps in the environment custom interface.
Further, the control interface further includes a training parameter adjustment tab (see fig. 2, where the training parameter adjustment tab is displayed as a "start parameter"), and the updated training parameter is transmitted to the training server through a message received by the message queue. Preferably, referring to fig. 2, the training parameter adjustment tab includes at least a learning rate adjustment tab and an iteration number tab. In this way, the parameter adjustment in the training process can be completed by operating the training code update tab in the control interface, and each time the parameter is adjusted, the parameter adjustment can be completed in 5 minutes, and compared with the original processing mode to the training server (the minimum time required for each adjustment in the prior art is 1 hour), the working efficiency can be greatly improved.
Further, training data generated in the deep learning model training process are stored in the file server, and the file server displays the training data through the control interface. In this way, the training process can be tracked through the control interface without viewing in the training server.
Further, the log generated in the deep learning model training process is stored in the file server, and the file server displays the log through the control interface. Therefore, the log generated in the training process can be directly displayed in the control interface, and is convenient to read.
In summary, the implementation of the rapid training method for the deep learning model provided by the invention has at least the following beneficial effects:
1. the file server is used for uniformly managing training codes and training data, the file server is connected with the display terminal and the training server, the training server can acquire the training codes from the file server, and the training data can be displayed through the display terminal. In this way, the training codes are conveniently updated and the training data is conveniently tracked.
2. The control interface on the display terminal comprises a training parameter adjustment tab, so that adjustment of the training parameters is not required to be processed by the training server, and the time consumption for parameter adjustment is greatly shortened.
3. The training server presets a training environment and supports a fast custom training environment when the preset training environment lacks relevant dependent components or software.
Example two
The embodiment provides a rapid training system for a deep learning model. The rapid training method of embodiment one may be implemented by the rapid training system.
In this embodiment, as shown in fig. 4, the fast training system includes a training server, a file server connected with the training server by signals, and a display terminal connected with the file server, where the training server is preloaded with a training environment, the file server stores training codes, and the display terminal displays a control interface, and the control interface includes a training code update tab. The control interface also comprises a training parameter adjustment tab, and the updated training parameters are transmitted to the training server through the information received by the information queue; the training parameter adjustment tab includes a learning rate adjustment tab and an iteration number tab. The training server may employ a server or a cluster of servers. The file server may employ a server or a cluster of servers. Because the training environment is preinstalled in the training server, the environment is not required to be installed and prepared when the deep learning model is trained, and the working efficiency can be greatly improved compared with the original mode (the time for installing and deploying the environment by one ordinary technician is minimum by 7 days). In addition, the training codes are uploaded to the training server through the file server, if the training codes need to be updated, the new training codes are directly uploaded to the file server, and the training server can automatically download the latest training codes in the file server for covering and running, so that the updating of the training codes is completed rapidly. It should be noted that in order to ensure that the training process is efficiently and smoothly completed, it is required that dependent components or software related to the training environment are already provided in the pre-installed training environment.
Of course, if the pre-installed training environment lacks relevant dependent components or software, the training environment can be customized based on the pre-installed training environment. Specifically, the rapid training method comprises the step of directly customizing through the control interface when relevant dependent components or software are absent in the preloaded training environment. The step of directly customizing the control interface when the relevant dependent components or software are absent in the preloaded training environment comprises the following steps: the training server automatically generates a dockerfile according to the existing dependent component and automatically downloads the related dependent component through the dockerfile; adding the base image to the new dependent components generates a custom running environment for the user.
In this embodiment, the control interface further includes a training parameter adjustment tab, and the updated training parameter is transmitted to the training server through a message received by the message queue. Preferably, referring to fig. 2, the training parameter adjustment tab includes at least a learning rate adjustment tab and an iteration number tab. In this way, the parameter adjustment in the training process can be completed by operating the training code update tab in the control interface, and each time the parameter is adjusted, the parameter adjustment can be completed in 5 minutes, and compared with the original processing mode to the training server (the minimum time required for each adjustment in the prior art is 1 hour), the working efficiency can be greatly improved.
In this embodiment, training data generated in the training process of the deep learning model is stored in the file server, and the file server displays the training data through the control interface. In this way, the training process can be tracked through the control interface without viewing in the training server.
In this embodiment, the log generated in the training process of the deep learning model is stored in the file server, and the file server displays the log through the control interface. Therefore, the log generated in the training process can be directly displayed in the control interface, and is convenient to read.
The embodiments of the present invention have been described above with reference to the accompanying drawings, but the present invention is not limited to the above-described embodiments, which are merely illustrative and not restrictive, and many forms may be made by those having ordinary skill in the art without departing from the spirit of the present invention and the scope of the claims, which are to be protected by the present invention.

Claims (2)

1. A training method of a deep learning model, the training method comprising the steps of:
pre-installing a training environment in a training server;
the training server invokes training codes which are pre-led into the file server from the file server to train the deep learning model;
the file server is in signal connection with a display terminal displaying a control interface, wherein the control interface comprises a training code updating option card;
training data generated in the deep learning model training process are stored in the file server, and the file server displays the training data through the control interface;
the logs generated in the training process of the deep learning model are stored in the file server, and the file server displays the logs through the control interface;
the control interface also comprises a training parameter adjustment tab, and the updated training parameters are transmitted to the training server through the information received by the information queue;
the training parameter adjustment tab comprises a learning rate adjustment tab and an iteration number tab;
pre-installed dependent components or software have been provided in the training environment;
the training method comprises the step of directly customizing through the control interface when relevant dependent components or software are absent in the preloaded training environment;
the step of directly customizing the control interface when the relevant dependent components or software are absent in the preloaded training environment comprises the following steps:
the training server automatically generates a dockerfile according to the existing dependent component and automatically downloads the related dependent component through the dockerfile;
adding a new dependent component to the basic mirror image to generate a user-defined running environment, and processing a training parameter adjustment tab and a training code update tab in an interface;
and uploading the training codes to the training server through the file server, if the training codes need to be updated, directly uploading the new training codes to the file server, and automatically downloading the latest training codes in the file server by the training server for covering and running, thereby completing the updating of the training codes.
2. A training system of a deep learning model, which is used for the training method of the deep learning model according to claim 1, and is characterized in that the training system comprises a training server, a file server connected with the training server in a signal manner and a display terminal connected with the file server, wherein the training server is preloaded with a training environment, training codes are stored in the file server, and a control interface is displayed on the display terminal and comprises a training code updating tab;
the control interface also comprises a training parameter adjustment tab, and the updated training parameters are transmitted to the training server through the information received by the information queue; the training parameter adjustment tab comprises a learning rate adjustment tab and an iteration number tab;
training data generated in the deep learning model training process are stored in the file server, and the file server displays the training data through the control interface;
the logs generated in the training process of the deep learning model are stored in the file server, and the file server displays the logs through the control interface;
if the pre-installed training environment lacks relevant dependent components or software, the training environment is customized on the basis of the pre-installed training environment.
CN201910676874.1A 2019-07-25 2019-07-25 Rapid training method and system for deep learning model Active CN110533181B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910676874.1A CN110533181B (en) 2019-07-25 2019-07-25 Rapid training method and system for deep learning model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910676874.1A CN110533181B (en) 2019-07-25 2019-07-25 Rapid training method and system for deep learning model

Publications (2)

Publication Number Publication Date
CN110533181A CN110533181A (en) 2019-12-03
CN110533181B true CN110533181B (en) 2023-07-18

Family

ID=68661939

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910676874.1A Active CN110533181B (en) 2019-07-25 2019-07-25 Rapid training method and system for deep learning model

Country Status (1)

Country Link
CN (1) CN110533181B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111310934B (en) * 2020-02-14 2023-10-17 北京百度网讯科技有限公司 Model generation method and device, electronic equipment and storage medium
CN112035419A (en) * 2020-08-12 2020-12-04 湖北世纪创新科技有限公司 Novel data center visualization algorithm
CN112615937A (en) * 2020-12-28 2021-04-06 南方电网深圳数字电网研究院有限公司 Data transmission method, data training method, server, electronic device and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105808500A (en) * 2016-02-26 2016-07-27 山西牡丹深度智能科技有限公司 Realization method and device of deep learning
CN106529673A (en) * 2016-11-17 2017-03-22 北京百度网讯科技有限公司 Deep learning network training method and device based on artificial intelligence
CN106663037A (en) * 2014-06-30 2017-05-10 亚马逊科技公司 Feature processing tradeoff management
CN106951926A (en) * 2017-03-29 2017-07-14 山东英特力数据技术有限公司 The deep learning systems approach and device of a kind of mixed architecture
CN109558940A (en) * 2018-11-09 2019-04-02 深圳市康拓普信息技术有限公司 A kind of management method and system of deep learning model training

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6535713B1 (en) * 1996-05-09 2003-03-18 Verizon Services Corp. Interactive training application

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106663037A (en) * 2014-06-30 2017-05-10 亚马逊科技公司 Feature processing tradeoff management
CN105808500A (en) * 2016-02-26 2016-07-27 山西牡丹深度智能科技有限公司 Realization method and device of deep learning
CN106529673A (en) * 2016-11-17 2017-03-22 北京百度网讯科技有限公司 Deep learning network training method and device based on artificial intelligence
CN106951926A (en) * 2017-03-29 2017-07-14 山东英特力数据技术有限公司 The deep learning systems approach and device of a kind of mixed architecture
CN109558940A (en) * 2018-11-09 2019-04-02 深圳市康拓普信息技术有限公司 A kind of management method and system of deep learning model training

Also Published As

Publication number Publication date
CN110533181A (en) 2019-12-03

Similar Documents

Publication Publication Date Title
CN110533181B (en) Rapid training method and system for deep learning model
CN101453469B (en) System and method for dynamically generating user interfaces for network client devices
CN101505320B (en) Graphic user interface sharing method, system and tool
CN104093077B (en) Method, Apparatus and system that multiple terminals is interconnected
US9064001B2 (en) Method and apparatus for a frequently-asked questions portal workflow
US20110154305A1 (en) System and method for remotely compiling multi-platform native applications for mobile devices
CN103150513B (en) The method of the implantation information in interception application program and device
JP6788057B2 (en) Methods and devices for transmitting information
US20100180270A1 (en) Extensibility for business accelerator authoring tools
CN109542468B (en) Application program release method and device and electronic equipment
CN107463405B (en) Application program updating method and device
CN111459539A (en) Continuous integration pipeline operation method and device based on mirror layering
CN105138387A (en) Method, device and system for compiling software
CN103677956A (en) Method and device for simulating application operation of mobile equipment on calculating equipment
CN112015654A (en) Method and apparatus for testing
CN111679811A (en) Web service construction method and device
CN106681759A (en) Software updating method and mobile terminal
CN111459506B (en) Deep learning platform cluster deployment method and device, medium and electronic equipment
CN113157270A (en) Page adaptation method and device
CN112559158A (en) Micro-service timing task scheduling method and device
CN117311784A (en) AI algorithm container visual manufacturing method, system, medium and equipment
CN107193597A (en) A kind of application program update method and terminal
CN114564223A (en) Application program hot updating method and device
CN109976773A (en) The dispositions method and device of game test environment
CN114374456A (en) Outdoor LED screen adaptation method, device and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information

Address after: 518034 501, 502, 601 and 602, building D, wisdom Plaza, Qiaoxiang Road, Gaofa community, Shahe street, Nanshan District, Shenzhen, Guangdong

Applicant after: China Southern Power Grid Digital Platform Technology (Guangdong) Co.,Ltd.

Address before: 518034 501, 502, 601 and 602, building D, wisdom Plaza, Qiaoxiang Road, Gaofa community, Shahe street, Nanshan District, Shenzhen, Guangdong

Applicant before: China Southern Power Grid Shenzhen Digital Power Grid Research Institute Co.,Ltd.

Address after: 518034 501, 502, 601 and 602, building D, wisdom Plaza, Qiaoxiang Road, Gaofa community, Shahe street, Nanshan District, Shenzhen, Guangdong

Applicant after: China Southern Power Grid Shenzhen Digital Power Grid Research Institute Co.,Ltd.

Address before: 518034 501, 502, 601 and 602, building D, wisdom Plaza, Qiaoxiang Road, Gaofa community, Shahe street, Nanshan District, Shenzhen, Guangdong

Applicant before: SHENZHEN COMTOP INFORMATION TECHNOLOGY Co.,Ltd.

CB02 Change of applicant information
GR01 Patent grant
GR01 Patent grant