CN109740765B - Machine learning system building method based on Amazon network server - Google Patents
Machine learning system building method based on Amazon network server Download PDFInfo
- Publication number
- CN109740765B CN109740765B CN201910106145.2A CN201910106145A CN109740765B CN 109740765 B CN109740765 B CN 109740765B CN 201910106145 A CN201910106145 A CN 201910106145A CN 109740765 B CN109740765 B CN 109740765B
- Authority
- CN
- China
- Prior art keywords
- amazon
- machine learning
- zeppelin
- aws
- storage platform
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Stored Programmes (AREA)
Abstract
The invention discloses a machine learning system building method based on an Amazon network server, which comprises the following steps: creating an Amazon EMR cluster by adopting an AWS bot 3 interface, wherein the Amazon EMR cluster is configured with a Zeppelin storage platform; copying precompiled Spark task code from Amazon S3 onto a Master machine of the Amazon EMR cluster; registering a memory path of the Spark task code on the Master machine into a Spark interpreter of the Zeppelin storage platform through a service interface of the Zeppelin storage platform, and registering a code warehouse of Zeppelin Notebook of the Zeppelin storage platform into the Amazon S3; the required machine learning instance is created by the AWS. The technical scheme provided by the invention can rapidly complete the construction of the required system and improve the development efficiency.
Description
Technical Field
The invention relates to the technical field of computer network resource management, in particular to a machine learning system building method based on an Amazon network server.
Background
Amazon Web servers (Amazon Web Services, AWS) are cloud computing resource management platforms operated by Amazon corporation to provide various types of AWS resources, such as Amazon elastic computing network cloud (AWS EC2, amazon Elastic Compute Cloud) service resources, amazon simple storage service (Amazon S3, amazon Simple Storage Service) resources, and the like, to enterprises in a remote Web service manner. The AWS EC2 service resource can enable a user to remotely use a computer system consisting of different types of virtual computers in a mode of renting the virtual computers (namely examples), any application software required by the user can be operated in the computer system, and meanwhile, the user can create, operate and terminate the AWS EC2 service at any time; amazon S3 services may be used for network data storage. The hosted Hadoop framework provided by Amazon EMR allows users to handle large amounts of data in multiple dynamically extensible Amazon EC2 instances. Clusters are a collection of Amazon EC2 instances, which are also the core components of Amazon EMR.
Based on various basic services provided by the AWS, developers can quickly build various computing environments for realizing cloud computing, big data, machine learning and the like. How to respond quickly to product demand under controllable cost resources to accomplish machine learning goals becomes a factor that needs to be prioritized by developers. The machine learning involves a lot of content, and mainly comprises tasks such as data acquisition, data management, data characteristic processing, model selection, super-parameter searching, model training and the like. The data acquisition, management and the like in the early stage can be completed by using software tools such as Flume, spark, kinesis, kafka, elastic Search and the like; subsequent processing of data features, model-related training tasks, etc. require the use of a computing engine and computing platform such as Spark, tensorflow. After the machine learning task target is determined, the developer also needs to perform a plurality of links such as data verification, model training, AB test and the like, and the links need to quickly call the AWS resource for calculation and verification. Specifically, the creation of EMR and EC2 needs to be completed in the amazon background, then relevant codes are deployed, a Spark computing engine and the like are used for submitting computing tasks or starting a Python programming language script to debug data features and models, and relevant data results and indexes are observed. The entire development cycle itself is time consuming, as different models may require different types of data, and model tuning may take a period of time. If the AWS resources are recovered due to insufficient investment of the early cost, the developer needs to re-create the EMR and the EC2, so that the development period is further prolonged, and meanwhile, the workload of the developer is greatly increased due to the process of re-creating the EMR and the EC 2. All of the above problems result in low working efficiency for developers.
Disclosure of Invention
The invention aims to provide a machine learning system building method based on an Amazon network server, which can quickly complete the building of a required system and improve the development efficiency.
In order to achieve the above purpose, the invention adopts the following technical scheme:
a machine learning system building method based on an Amazon network server comprises the following steps: creating an Amazon EMR cluster by adopting an AWS bot 3 interface, wherein the Amazon EMR cluster is configured with a Zeppelin storage platform; copying precompiled Spark task code from Amazon S3 onto a Master machine of the Amazon EMR cluster; registering a memory path of the Spark task code on the Master machine into a Spark interpreter of the Zeppelin storage platform through a service interface of the Zeppelin storage platform, and registering a code warehouse of Zeppelin Notebook of the Zeppelin storage platform into the Amazon S3; the required machine learning instance is created by the AWS.
Preferably, the creating the required machine learning instance through AWS includes: selecting an AWS EC2 instance, and creating a mirror image for the AWS EC2 instance; the mirror image comprises a predetermined software tool; and creating a required machine learning instance by adopting the AWS bot 3 interface according to the mirror image.
Further, the method further comprises the following steps: the AWS bot 3 interface is adopted to specify the IP address of the required machine learning instance, and/or the AWS bot 3 interface is adopted to specify the identity of the required machine learning instance.
Further, the method further comprises the following steps: and adding a Jupyter Notebook function in the required machine learning example, and setting the Jupyter Notebook function to be in a starting-up self-starting state.
Preferably, the predetermined software tool is Tensorflow.
Further, after registering the code repository of Zeppelin Notebook of the Zeppelin storage platform in the Amazon S3, the method further comprises: and restarting the Zeppelin storage platform.
Preferably, the precompiled Spark task code is copied from Amazon S3 onto the Master machine of the Amazon EMR cluster using aws cli Shell commands.
According to the machine learning system building method based on the Amazon network server, spark task codes are synchronized to a Master machine of an Amazon EMR cluster, and then the Spark task codes can be called in subsequent Zeppelin Notebook; registering the code repository of Zeppelin Notebook into the Amazon S3 can quickly synchronize to Zeppelin Notebook again when the Amazon EMR clusters are reestablished after they are released. In addition, when the machine learning example is created, the mirror technology is adopted, so that the re-created EC2 example is quickly configured through mirror; the invention adopts the AWS bot 3 interface to carry out various operations, and is also greatly convenient for developers. Therefore, the technical scheme provided by the invention can rapidly complete the construction of the required system and improve the development efficiency.
Drawings
FIG. 1 is a flow chart of a method according to a first embodiment of the invention;
fig. 2 is a flowchart of a method according to a second embodiment of the invention.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings, in order to make the objects, technical solutions and advantages of the present invention more apparent.
Fig. 1 is a flowchart of a method according to a first embodiment of the present invention, including:
boto is a python-based SDK of AWS (of course also supporting SDKs of other languages, such as Ruby, java, etc.), and allows developers to use amazon, etc. services like S3 and EC2, etc. when writing software, boto provides a simple, object-oriented API, as well as low-level service access.
The Zeppelin is a distributed Key-Value storage platform, and aims at high performance and large clusters, and the platform is not an end point but a starting point, so that on the basis of the Zeppelin, not only can KV database access be provided, but also more complex protocol requirements can be met through simple one-layer conversion.
in this step, the precompiled Spark task code is copied from Amazon S3 to the Master machine of the Amazon EMR cluster using aws cli Shell command. The Spark task code mainly refers to a jar file.
after the step is completed, the method further comprises the step of restarting the Zeppelin storage platform.
For Zeppelin Notebook, if an operation needs to be performed, a trigger call is performed through a service interface of the Zeppelin storage platform, and the trigger call mainly completes a Spark task related to Zeppelin Notebook to process data features, for example, normalize or vectorize data, etc., and the finally processed data can be saved in an Amazon S3 or a RedShift data warehouse so as to be read by a subsequent machine learning task.
The machine learning example is used for completing tasks such as training a machine learning model, and data required for training the machine learning model is the data subjected to the data feature processing and is read from the Amazon S3 or the RedShift data warehouse.
Fig. 2 is a flowchart of a method according to a second embodiment of the present invention, which is different from the first embodiment in that in step 104, a mirroring method is used to create a required machine learning instance, and specifically, the creating, by AWS, the required machine learning instance includes:
step 1041, selecting an AWS EC2 instance, and creating a mirror image for the AWS EC2 instance; the mirror image comprises a predetermined software tool;
the image needs to include a currently mainstream machine learning platform or toolkit, the predetermined software tools including, but not limited to Tensorflow, MXnet, theano, pytorch, CNTK, caffe. In this embodiment, we use Tensorflow as the machine learning platform. Related toolkits may be installed in a designated Conda environment based on Conda or Pip tools as needed before creating the image.
Step 1042, creating a required machine learning instance using the AWS bot 3 interface according to the mirror image.
And starting a corresponding working platform environment (such as a TensorFlow Python3.6 version environment in the example) in the required machine learning example, and executing a Jupyter Notebook or a customized Python script which needs to be run so as to finish the machine requirement target.
After the machine learning instance is created, further comprising: the AWS bot 3 interface is adopted to specify the IP address of the required machine learning instance, and/or the AWS bot 3 interface is adopted to specify the identity of the required machine learning instance. Further, the method further comprises the following steps: and adding a Jupyter Notebook function in the required machine learning example, and setting the Jupyter Notebook function to be in a starting-up self-starting state.
The method of the present invention is described in further detail below in connection with the specific requirements of the development effort:
in this embodiment, the EMR cluster needs to be started to complete the data feature processing in a certain business model. The feature processing here is to complete the primary feature processing with sufficient computing resources in the EMR cluster, in hopes of reducing the time expenditure and resource utilization that would otherwise be required to process the primary feature if the model processing stage resources were relatively small.
Because AWS provides a very large number of EMR instance types, the actual service can complete setting the designated resources according to its own needs, for example, set EMR cluster names, use r4. Xlage type instances, number 50, use bid instance types to obtain AWS idle resources with less cost, set security groups of clusters, zeppelin version, zeppelin for synchronizing S3 addresses of notbook, custom Jar files synchronized from S3 after EMR cluster creation is completed, etc.
The data processing flow in Zeppelin Notebook needs to rely on the actual processing logic code, and the code of this block can be subjected to custom development, and the compiling and packaging work of the code can be completed by publishing the code to a Jenkins platform and the like. The Jar file is then synchronized to the Master machine of the EMR cluster. Since Zeppelin cannot automatically identify a custom Jar file, it is necessary to register the custom Jar file in the Spark interpreter of the Zeppelin storage platform. After the registration is successful, the Zeppelin storage platform is restarted, and then the call to the relevant logic code in the custom Jar can be completed in the subsequent Zeppelin Notebook.
After registering the user-defined Jar file, restarting the Zeppeilin service, and then accessing the WEB page of the Zeppeilin to perform data characteristic processing operation. The flow of the process will vary with the different models and the different demands of the business. The current flow may be generally described as reading data in Amazon S3 or a data warehouse, organizing related data into Spark dataframes, completing the transformation of the feature data using Spark ML toolkit, and storing the transformed data in a storage medium such as S3 or a data warehouse.
In daily development, the EMR cluster resources are released continuously in consideration of factors such as cost, and Zeppelin Notebook in the EMR cluster needs to be stored in time for subsequent editing and calling. For this reason, in the present invention, a code repository is configured for the Zeppelin storage platform, and Zeppelin Notebook in this embodiment is saved to the bucket and path specified by Amazon S3. Such that subsequently re-established EMR clusters will again be synchronized Zeppelin Notebook from the AWS-specified buckets and paths for re-editing and recall by the developer.
The saved Zeppelin Notebook is called for as part of the daily process flow, and is executed in a fixed period (e.g., daily or weekly, etc.) to output the data characteristic process results. The modification of the variable in the specified Paragraph in Zeppelin Notebook (e.g., the import of a specified date, etc. parameter) can be done and the call completed through the service interface of the Zeppelin storage platform. After the Notebook completes the call, the number of Paragraph successes, failures is returned for caller awareness.
In this embodiment, an AWS EC2 needs to be started to complete the debugging of the service model. The characteristic data stored in the data warehouse or S3 after the characteristic processing is carried out on the data can be called out for use. The specific usage mode may be that the data in the data warehouse or the S3 is read in the processing logic of the service model and loaded into the memory. The machine learning model uses the TensorFlow toolkit to complete reprocessing of feature data. Reprocessing here refers to converting the feature data into data that can be directly read and recognized by the model.
The AWS provides a plurality of EC2 instance types, including general type, computation optimization type, memory optimization type, acceleration computation type, storage optimization type, and the like. The model in this case assumes that deep learning is required to select the EC2 instance of the acceleration calculation type. While considering economic benefits, bidding examples may be selected for use based on the parameters of the present invention. The benefit of the bid instance is that in most cases free resources can be selected from the market of the AWS and the computing resources obtained at a lower cost. In addition to instance type and whether or not to use bidding, the present invention also provides for setting the size of the EC2 local storage space, the setting of the AWS original mirror.
By default, the present invention will complete a fast EC2 boot process based on custom mirroring without reinstalling the Python's dependent toolkit. Thereby saving time spent installing the tool pack. The Python-dependent toolkit may involve functions like database access, numerical calculation, graph drawing, parameter searching, etc. The instance types provided by the AWS for machine learning may not themselves contain these toolkits, so configuration in conjunction with the business is required, and the designated environment of Conda is started for installation after EC2 machine creation is complete, so that the business model is invoked.
Of course, after the EC2 is created, the lP address or the ID of the EC2 may be specified, and the toolkit list may be updated again, so as to complete the installation again. This function mainly takes into account the need to extend existing functions in the course of model discovery.
The juyter Notebook function provides an exploration function for completing a machine learning business model in a WEB page using the Python language. In this embodiment, after the juyter Notebook function is automatically started, the WEB page of the juyter can be directly accessed, the Notebook is created, relevant development work is performed in the Notebook, and the tool kit is used to complete index drawing of the service model.
The principle of the method of the invention is as follows:
in the existing machine learning task, the data feature processing needs relatively more machine resources, such as a CPU and a memory, and the types of resources needed by different data processing modes also differ, one is computationally intensive, one is memory intensive, or both. Traditional data centers cannot provide computing resources or different types of computing resources anytime anywhere due to limited hardware resources. Cloud computing offers a very large number of possibilities. The Amazon-based cloud computing service solves the problem of resource limitation of the traditional data center, and simultaneously provides a plurality of services so as to solve the requirements in different scenes.
When the machine learning task is carried out by means of the AWS service, the task of calculating resources to solve the data characteristic processing is firstly applied through the EMR service of the AWS, and the data characteristic processing taking Spark as a calculation frame needs to use a plurality of EMR examples to carry out distributed calculation. In order to solve the different feature processing and more conveniently perform data verification and debugging, a Spark calculation task is generally required to be submitted on EMR, and data is written into services such as S3 or a data warehouse. This commit process includes encoding, compiling, publishing, performing file synchronization, committing jobs, etc., and modifying a computing DAG takes a significant amount of time to wait, delaying development efficiency. By using the Zeppein service, the arrangement, debugging and verification of the computing DAG can be completed in the Zepplin Notebook without compiling, publishing, synchronizing, submitting Job and the like again for the DAG, thereby greatly improving the efficiency.
As a machine learning task, tensorflow is used as a platform tool kit in the embodiment, and processed characteristic data is required to be used as model input to complete tuning, parameter debugging, model training, reasoning and the like of a model. The current learning targets such as deep learning and reinforcement learning depend on the CPU and the GPU. For this purpose, computing resources need to be acquired based on AWS EC2 cluster services, and Python procedures are performed in order to accomplish specific business objectives.
Both parts require creating computing resources, configuring a toolkit environment, configuring code synchronization policies, executing computing DAGs or machine learning targets, which require a complete set of automation mechanisms to complete.
The beneficial effects of the invention are as follows:
1. in contrast to the traditional model, the created EC2 does not require reinstallation and configuration environment due to the mirror image used in the present invention. While it takes 11 minutes to complete the EC2 resource for machine learning in the conventional mode, the time spent in the present invention is reduced to 6.5 minutes, which is greatly reduced.
2. Since machine learning requires a related exploration at different stages, for example, a third party Python toolkit is required. The invention provides a function based on configuration, and the installation of a new tool kit can be completed by only adding related parameters in configuration parameters and executing updating operation without reapplying EC2 machine resources.
3. Due to the requirements of different models on data processing, the calculation DAG also needs to be adjusted in time in the feature processing process. The invention supports updating operation of the Jar file relied on in Zeppelin Notebook, and does not need to reapply EMR resources.
4. The invention provides an interface for executing Zeppelin Notebook, which needs to be operated daily after the business model is determined, and the call to the appointed Zeppelin Notebook can be completed through the corresponding interface. At the same time, through a further interface, the running of a specific Python program in EC2 can be initiated. Is more convenient and efficient.
The foregoing is merely illustrative of the present invention, and the present invention is not limited thereto, and any person skilled in the art will readily recognize that variations or substitutions are within the scope of the present invention.
Claims (6)
1. The machine learning system building method based on the Amazon network server is characterized by comprising the following steps of:
creating an Amazon EMR cluster by adopting an AWS bot 3 interface, wherein the Amazon EMR cluster is configured with a Zeppelin storage platform;
copying precompiled Spark task code from Amazon S3 onto a Master machine of the Amazon EMR cluster;
registering a memory path of the Spark task code on the Master machine into a Spark interpreter of the Zeppelin storage platform through a service interface of the Zeppelin storage platform, and registering a code warehouse of Zeppelin Notebook of the Zeppelin storage platform into the Amazon S3;
creating a required machine learning instance through AWS; creating a required machine learning instance by adopting a mirror image method; the method specifically comprises the following steps of;
selecting an AWS EC2 instance, and creating a mirror image for the AWS EC2 instance; the mirror image comprises a predetermined software tool;
and creating a required machine learning instance by adopting the AWS bot 3 interface according to the mirror image.
2. The amazon-network-server-based machine learning system building method of claim 1, further comprising:
the AWS bot 3 interface is adopted to specify the IP address of the required machine learning instance, and/or the AWS bot 3 interface is adopted to specify the identity of the required machine learning instance.
3. The amazon-network-server-based machine learning system building method of claim 1, further comprising:
and adding a Jupyter Notebook function in the required machine learning example, and setting the Jupyter Notebook function to be in a starting-up self-starting state.
4. The amazon web server-based machine learning system building method of claim 1, wherein the predetermined software tool is a Tensorflow.
5. The Amazon web server-based machine learning system building method of claim 1, further comprising, after said registering the code repository of Zeppelin Notebook of the Zeppelin storage platform into the Amazon S3:
and restarting the Zeppelin storage platform.
6. The Amazon web server-based machine learning system building method of claim 1, wherein precompiled Spark task code is copied from Amazon S3 to a Master machine of the Amazon EMR cluster using aws cli Shell commands.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910106145.2A CN109740765B (en) | 2019-01-31 | 2019-01-31 | Machine learning system building method based on Amazon network server |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910106145.2A CN109740765B (en) | 2019-01-31 | 2019-01-31 | Machine learning system building method based on Amazon network server |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109740765A CN109740765A (en) | 2019-05-10 |
CN109740765B true CN109740765B (en) | 2023-05-02 |
Family
ID=66367221
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910106145.2A Active CN109740765B (en) | 2019-01-31 | 2019-01-31 | Machine learning system building method based on Amazon network server |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109740765B (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117215635A (en) * | 2019-06-28 | 2023-12-12 | 杭州海康威视数字技术股份有限公司 | Task processing method, device and storage medium |
CN111159270A (en) * | 2019-12-31 | 2020-05-15 | 杭州依图医疗技术有限公司 | Method, system, computing device and storage medium for scheduling Zeppelin tasks |
CN111158672B (en) * | 2019-12-31 | 2023-04-21 | 浪潮云信息技术股份公司 | Integrated interactive Elastic MapReduce job management method |
CN112181644B (en) * | 2020-09-21 | 2021-08-06 | 上海微亿智造科技有限公司 | Method, system and device for cross-domain machine learning component Jupitter |
CN112631527A (en) * | 2021-01-07 | 2021-04-09 | 上海明略人工智能(集团)有限公司 | Juypter notewood code remote storage method and device based on k8s multi-tenant |
CN112906907B (en) * | 2021-03-24 | 2024-02-23 | 成都工业学院 | Method and system for layering management and distribution of machine learning pipeline model |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6772141B1 (en) * | 1999-12-14 | 2004-08-03 | Novell, Inc. | Method and apparatus for organizing and using indexes utilizing a search decision table |
CN104243563A (en) * | 2014-09-03 | 2014-12-24 | 河海大学 | Quick parallel system allocation method for AWS platform |
CN106547910A (en) * | 2016-11-25 | 2017-03-29 | 山东浪潮商用系统有限公司 | Moosefs realizes the multistage storage method of file based on Tachyon |
CN108037984A (en) * | 2017-11-28 | 2018-05-15 | 深圳前海微众银行股份有限公司 | Method for managing resource, system and the readable storage medium storing program for executing of data analysis |
CN108595473A (en) * | 2018-03-09 | 2018-09-28 | 广州市优普计算机有限公司 | A kind of big data application platform based on cloud computing |
CN109255440A (en) * | 2017-07-11 | 2019-01-22 | 上海有孚网络股份有限公司 | The method that predictive maintenance is carried out to Electric Power Generating Equipment based on recurrent neural network (RNN) |
CN109284184A (en) * | 2018-03-07 | 2019-01-29 | 中山大学 | A kind of building method of the distributed machines learning platform based on containerization technique |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180341956A1 (en) * | 2017-05-26 | 2018-11-29 | Digital River, Inc. | Real-Time Web Analytics System and Method |
-
2019
- 2019-01-31 CN CN201910106145.2A patent/CN109740765B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6772141B1 (en) * | 1999-12-14 | 2004-08-03 | Novell, Inc. | Method and apparatus for organizing and using indexes utilizing a search decision table |
CN104243563A (en) * | 2014-09-03 | 2014-12-24 | 河海大学 | Quick parallel system allocation method for AWS platform |
CN106547910A (en) * | 2016-11-25 | 2017-03-29 | 山东浪潮商用系统有限公司 | Moosefs realizes the multistage storage method of file based on Tachyon |
CN109255440A (en) * | 2017-07-11 | 2019-01-22 | 上海有孚网络股份有限公司 | The method that predictive maintenance is carried out to Electric Power Generating Equipment based on recurrent neural network (RNN) |
CN108037984A (en) * | 2017-11-28 | 2018-05-15 | 深圳前海微众银行股份有限公司 | Method for managing resource, system and the readable storage medium storing program for executing of data analysis |
CN109284184A (en) * | 2018-03-07 | 2019-01-29 | 中山大学 | A kind of building method of the distributed machines learning platform based on containerization technique |
CN108595473A (en) * | 2018-03-09 | 2018-09-28 | 广州市优普计算机有限公司 | A kind of big data application platform based on cloud computing |
Non-Patent Citations (6)
Title |
---|
"启动集群时,在EMR上配置Zeppelin的Spark解释器";Thinbug;《https://www.thinbug.com/q/45328671》;20170726;1-4 * |
"如何使用EMRSparkML与Zeppelin搭建机器学习服务技术-AWS.PDF";zhaoxiaoj;《https://max.book118.com/html/2018/1127/8040000130001134.shtm》;20181203;1-28 * |
Cerebral Blood Flow Monitoring Using IoT Enabled Cloud Computing for mHealth Applications;Beulah Preethi Vallur等;《FICC 2018: Advances in Information and Communication Networks》;20181227;第887卷;578–590 * |
Consolidating billions of Taxi rides with AWS EMR and Spark in the Cloud : Tuning, Analytics and Best Practices;Alex Kaplunovich等;《2018 IEEE International Conference on Big Data (Big Data)》;20190124;4501-4507 * |
Hadoop平台性能优化的研究与实现;杨浩;《中国优秀硕士学位论文全文数据库信息科技辑》;20160115(第(2016)01期);I138-341 * |
基于Hadoop的遗传算法在TSP中的研究;曹立禄;《中国优秀硕士学位论文全文数据库信息科技辑》;20180715(第(2018)07期);I140-97 * |
Also Published As
Publication number | Publication date |
---|---|
CN109740765A (en) | 2019-05-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109740765B (en) | Machine learning system building method based on Amazon network server | |
Li et al. | Serverless computing: state-of-the-art, challenges and opportunities | |
Sahni et al. | A hybrid approach to live migration of virtual machines | |
US8887158B2 (en) | Dynamic cluster expansion through virtualization-based live cloning | |
Rodríguez et al. | CPPC: a compiler‐assisted tool for portable checkpointing of message‐passing applications | |
CN107329799A (en) | A kind of fusion Docker containers and the system of KVM virtualization technology | |
KR20230106719A (en) | Containerized deployment of microservices based on monolithic legacy applications | |
RU2658190C2 (en) | Controlling runtime access to application programming interfaces | |
CN105630488A (en) | Docker container technology-based continuous integration realizing method | |
US8156211B2 (en) | Transitioning from dynamic cluster management to virtualized cluster management | |
US20100162245A1 (en) | Runtime task with inherited dependencies for batch processing | |
US8407713B2 (en) | Infrastructure of data summarization including light programs and helper steps | |
US20210365457A1 (en) | Graph database and methods with improved functionality | |
CN112948025B (en) | Data loading method and device, storage medium, computing equipment and computing system | |
US20170322792A1 (en) | Updating of operating system images | |
CN110928659B (en) | Numerical value pool system remote multi-platform access method with self-adaptive function | |
CN112286633A (en) | Virtual machine creating method, device, equipment and storage medium based on CloudStack platform | |
Cai et al. | Deployment and verification of machine learning tool-chain based on kubernetes distributed clusters: This paper is submitted for possible publication in the special issue on high performance distributed computing | |
CN115629860A (en) | Software parameter tuning method, container management platform, storage medium and system | |
US11163603B1 (en) | Managing asynchronous operations in cloud computing environments | |
Wu et al. | An automatic artificial intelligence training platform based on kubernetes | |
CN112181403A (en) | Development, operation and maintenance integration realization method, device and equipment and readable storage medium | |
US20230297346A1 (en) | Intelligent data processing system with metadata generation from iterative data analysis | |
CN115878711B (en) | Redis connection pool realization method, device, equipment and medium | |
Rong | Design and Implementation of Operating System in Distributed Computer System Based on Virtual Machine |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |