CN108170770A - A kind of analyzing and training platform based on big data - Google Patents
A kind of analyzing and training platform based on big data Download PDFInfo
- Publication number
- CN108170770A CN108170770A CN201711428840.8A CN201711428840A CN108170770A CN 108170770 A CN108170770 A CN 108170770A CN 201711428840 A CN201711428840 A CN 201711428840A CN 108170770 A CN108170770 A CN 108170770A
- Authority
- CN
- China
- Prior art keywords
- data
- analysis
- row
- matrix
- value
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000012549 training Methods 0.000 title claims abstract description 13
- 238000004458 analytical method Methods 0.000 claims abstract description 47
- 238000007726 management method Methods 0.000 claims abstract description 34
- 238000011068 loading method Methods 0.000 claims abstract description 16
- 238000012545 processing Methods 0.000 claims abstract description 14
- 238000013499 data model Methods 0.000 claims abstract description 9
- 230000002688 persistence Effects 0.000 claims abstract description 8
- 238000013461 design Methods 0.000 claims abstract description 7
- 238000005314 correlation function Methods 0.000 claims abstract description 4
- 230000010354 integration Effects 0.000 claims abstract description 4
- 239000011159 matrix material Substances 0.000 claims description 54
- 238000000034 method Methods 0.000 claims description 46
- 230000006870 function Effects 0.000 claims description 43
- 239000013598 vector Substances 0.000 claims description 27
- 238000006243 chemical reaction Methods 0.000 claims description 15
- 238000004422 calculation algorithm Methods 0.000 claims description 12
- 238000000513 principal component analysis Methods 0.000 claims description 12
- 230000017105 transposition Effects 0.000 claims description 12
- 230000009471 action Effects 0.000 claims description 10
- 230000008859 change Effects 0.000 claims description 9
- YHXISWVBGDMDLQ-UHFFFAOYSA-N moclobemide Chemical compound C1=CC(Cl)=CC=C1C(=O)NCCN1CCOCC1 YHXISWVBGDMDLQ-UHFFFAOYSA-N 0.000 claims description 9
- 238000006116 polymerization reaction Methods 0.000 claims description 9
- 230000004044 response Effects 0.000 claims description 9
- 238000003780 insertion Methods 0.000 claims description 8
- 230000037431 insertion Effects 0.000 claims description 8
- 238000000354 decomposition reaction Methods 0.000 claims description 6
- 230000008676 import Effects 0.000 claims description 6
- 238000004321 preservation Methods 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 6
- 230000000717 retained effect Effects 0.000 claims description 6
- 230000001932 seasonal effect Effects 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 4
- 238000012098 association analyses Methods 0.000 claims description 3
- 230000001419 dependent effect Effects 0.000 claims description 3
- 238000003064 k means clustering Methods 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 238000001543 one-way ANOVA Methods 0.000 claims description 3
- 239000002356 single layer Substances 0.000 claims description 3
- 238000007619 statistical method Methods 0.000 claims description 3
- 238000012795 verification Methods 0.000 claims description 3
- 238000007405 data analysis Methods 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 5
- 238000010586 diagram Methods 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 238000003860 storage Methods 0.000 description 3
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 239000004615 ingredient Substances 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 239000012141 concentrate Substances 0.000 description 1
- 238000007418 data mining Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/22—Indexing; Data structures therefor; Storage structures
- G06F16/2282—Tablespace storage structures; Management thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/24—Querying
- G06F16/245—Query processing
- G06F16/2458—Special types of queries, e.g. statistical queries, fuzzy queries or distributed queries
- G06F16/2465—Query processing support for facilitating data mining operations in structured databases
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/28—Databases characterised by their database models, e.g. relational or object models
- G06F16/283—Multi-dimensional databases or data warehouses, e.g. MOLAP or ROLAP
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/28—Databases characterised by their database models, e.g. relational or object models
- G06F16/284—Relational databases
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2135—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
- G06F18/232—Non-hierarchical techniques
- G06F18/2321—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
- G06F18/23213—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Software Systems (AREA)
- Fuzzy Systems (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The present invention relates to a kind of analyzing and training platform based on big data, including with lower module:Web Analysis Service modules, for publication and analysis result, the Pages Design based on browser;Automation services assembly module, the calling of data task is with automatically updating;Management and integrated services module, the centralized deployments correlation functions such as data integration, user management, Study document management, data source information management;Statistical fractals module provides advanced statistics computing engines and Deployment Manager;Application data service module, is directly integrated with enterprise data source:SAP R/3, SAP B/W, UF ERP, Oracle EBS, Kingdee ERP;Further include persistence architecture load-on module;The persistence architecture load-on module includes flat file data processing unit, database file data loading unit and creates data model unit.
Description
Technical field
The invention belongs to big data technical fields, and in particular to a kind of analyzing and training platform based on big data.
Background technology
Big data is that a kind of big arrive of scale is well beyond traditional database software in terms of acquisition, storage, management, analysis
The data acquisition system of means capability range.Big data does not use traditional Stochastic Analysis Method, but is divided using all data
Analysis is handled.Data scale with magnanimity, various data type, quick stream compression and the low four big feature of value density.
Data scale:Data volume is huge to concentrate storage/centralized calculation that can not handle huge data volume;
Various data type:Type and source diversification daily record/picture/video/document/geographical location;
Quick stream compression:Analyze and process the timely and effective analysis of the fast mass data of speed;
Value density is low:Value density is low, the carry out complexity depth analysis of the high a large amount of irrelevant information of commercial value, deep
Dig value.
The strategic importance of big data technology does not simultaneously lie in people and has grasped how many huge data, but by these by
The data information of grasp carries out some specialized process with big data technology.In fact, if big data ratio is done a kind of industry, that
, key that this big data industry is generated profit, in that improve some working abilities to big data, by this plus
The ability of work and then the value for realizing big data.In fact, people study big data, seek to using big data research and reality
Existing its certain value, especially some departments of commercial business, the meaning for more having its tangible to excavating big data research.
However in the prior art but without providing special big data experience system for the demand.This is the prior art
Shortcoming.
Therefore, for drawbacks described above of the prior art, provide design a kind of analyzing and training platform based on big data;With
Above-mentioned technical problem is solved, is necessary.
Invention content
It is an object of the present invention in view of the above-mentioned drawbacks of the prior art, to provide design a kind of based on big data
Analyzing and training platform, to solve above-mentioned technical problem.
To achieve the above object, the present invention provides following technical scheme:
A kind of analyzing and training platform based on big data, which is characterized in that including with lower module:
Web Analysis Service modules, for publication and analysis result, the Pages Design based on browser;
Automation services assembly module, the calling of data task is with automatically updating;
Management and integrated services module, the collection such as data integration, user management, Study document management, data source information management
Chinese style disposes correlation function;
Statistical fractals module provides advanced statistics computing engines and Deployment Manager;
Application data service module, is directly integrated with enterprise data source:SAP R/3, SAP B/W, UF ERP,
Oracle EBS, Kingdee ERP;
Further include persistence architecture load-on module;
The persistence architecture load-on module includes flat file data processing unit, database file data load
Unit and establishment data model unit;
The flat file data processing unit includes the following steps:
S1.1:Plane txt file application
(1) select " file "->" interpolation data table "->" addition " carries out data loading;
(2) addition analysis text file;
(3) Study document is opened;
(4) " determining " is clicked, imports text;
S1.2:Plane excel file applications
(1) select " file "->" interpolation data table "->" addition " carries out data loading;
(2) Excel Study documents are added;
(3) Excel Study documents are opened;
(4) " determining " is clicked, imports Excel Study documents;
S1.3:Change data step:
(1) " file "-" replacement data table " or selection " insertion "-" row " or selection " insertion "-" row " are selected;
(2) conversion regime is selected;
(3) a variety of normalization methods can be written as expression formula or as switch process during interpolation data table;
(4) conversion can apply in the dialog box of interpolation data, " interpolation data table " dialog box or from external data work
In the insertion column or row of tool;" conversion "-" standardization "-" addition " is clicked to show control as described below;
S1.4:Data are pre-processed, calculate row, standardization processing operation;
S1.5:Transposition data
(1) selection " conversion "-" transposition "-" addition " carries out data transposition;
(2) row identifier:Each unique value in selected identity column or level can form a line in generation table;
(3) column heading:Each unique value in selected classification row or level can be to each polymerization in generated tables of data
Method forms a newline;
(4) value (%V) and polymerization (%M):The row of data value are calculated from it.Value in generated tables of data is according to row
The method selected under " polymerization " in selector menu is calculated;
S1.6:Data are put in reverse
(1) selection " conversion "-" reverse is put "-" addition " carries out data inverse transposition;
(2) will by row:Selected row, wherein the information for reversing and putting data set should be transferred to comprising unconverted;
(3) row to be converted:Selected row, wherein including the value that be merged into single row.The column name of these row will
Class label in new category row as generation;
(4) classification column name:The row for having selected the information provided in the row converted can be summarised in by keying in one
Title;
(5) value column name:Key in a column name that can show the information type included in new value row;
The database file data loading unit includes the following steps:
S1.7:The step of data being opened using OLE DB
(1) select " file "->" interpolation data table "->" addition "->" other "->" database ";
(2) it in " opening database " dialog box, clicks to select " OleDb Data Provider ";
(3) it clicks " configuration ";
S1.8:The step of data being opened using ODBC
(1) select " file "->" interpolation data table "->" addition "->" other "->" database ";
(2) it in " opening database " dialog box, clicks to select " Odbc Data Provider ";
(3) it clicks " configuration ".
The establishment data model unit includes the following steps:
S1.9:Interpolation data connects
S1.91:In library the step of interpolation data source:
(1) " tool " selection " management data connection " option, pop-up " management data connection " dialog box are clicked in menu bar;
(2) " data source " option of selection " addition is new ", selects data source types from list.
(3) according to selected data source types, corresponding information is filled in, is connected to data source, selects database simultaneously " determining ";
(4) " explanation " is optional fills out for addition in " data source dialog box ";
(5) " saving as library item mesh " dialog box will be shown by clicking " preservation ";
(6) newdata source is stored in designated position in library.
S1.92:The step of interpolation data connects in library:
(1) " tool " selection " management data connection " option is clicked in menu bar;
(2) " data connection " option of selection " addition new projects ", selection " the source connections of data in the library " choosing from list
;
(3) according to selected type of data connection, corresponding information is filled in, is connected to data source, selects database simultaneously " true
It is fixed ", pop up " view in connection " dialog box;
(4) in " the available table in database " list, the table used in the analyzing and training platform of big data is double-clicked;
(5) after the completion of, click OK;Pop up " data connection setting " dialog box;The tables of data added can be shown in
In " spreadsheet views " list;
(6) the input connection description in " connection description " frame facilitates other users to understand and uses;
(7) it clicks " preservation " and data connection is stored in designated position in library;
S1.10:Data connection is used in analysis
(1) selection " interpolation data table " option in " file " is clicked, pops up " interpolation data table " dialog box;
(2) it clicks " addition " and is connected to " data connection in library " option, pop-up " selection data connection " dialog box;
(3) your data connection to be used, and " determining " are selected in library;
(4) in " interpolation data table " dialog box, by choosing check box, selection is by which of data connection view
It is added to new data table;
(5) " loading method " and " importing tables of data " is selected still " to be retained in tables of data outer ", also may specify whether
On-demand loading data;
(6) " determining " is clicked;
S1.10:The editor of data connection
S1.101 is connected in library edit data:
(1) selection " tool " is clicked " management data connection ";
(2) your data connection to be edited is selected, is then clicked " editor ";It will show " data connection setting " dialog box;
(3) it makes change and preserves data connection;
S1.102 is in library edit data source:
(1) selection " tool " is clicked " management data connection ";
(2) your data source to be edited is selected, " data source setting " dialog box will be shown by then clicking " editor ";
(3) it makes change and preserves data source;
S1.11:Self-defined inquiry
(1) using tool>Manage data connection or file>Interpolation data table ... creates one to the new of relational database
Then data connection selects necessary content, until showing " view in connection " dialog box;
(2) it in " view in connection " dialog box, selects " self-defined inquiry ">" newly-built self-defined inquiry ";
(3) Query Name is keyed in " self-defined inquiry ";
(4) it is keyed in and inquired using the language of selected database;
(5) verification is clicked;
(6) browsing result row, it is ensured that list required all results row, and ensure that they have correct data type;
(7) click OK;
S1.12:Create tables of data relationship
For establishing the relationship between tables of data in data connection;When there are relationships between tables of data, established by relationship
Data model;
The establishment of S1.121 tables of data " relationship ":
(1) in " view in connection ", the tables of data of " relationship " is established in selection;
(2) select " relationship "->" newly-built relationship " opening relationships;
(3) click OK;
S1.122 tables " relationship " is checked:
Plus sige represents that the table establishes one or more structural relations with other tables in database, to the relationship of checking
Structure clicks plus sige so that view is unfolded;
It further includes big data and excavates module, the big data is excavated module and included:
Principal component analysis function prcomp (x ...)
Prcomp.default (x, retx=TRUE, center=TRUE, scale.=FALSE, tol=
NULL,...)
Parameter declaration:
The specified numeric type or complex matrix for being used for analyzing of x acquiescences;
Retx logical variables designate whether to return to rotary variable;
Center logical variables are designated whether variable centralization;
Scale. logical variable is designated whether variable standardization;
Tol numeric type variables, are used to specify precision, and the value less than the numerical value will be ignored;
Principal component analysis refers to turn to multi objective a kind of statistical analysis technique of a few overall target, the master of generation
Ingredient can reflect the linear combination of the most information, usually original variable of original variable;
Principal component information inquiry function summary.aov (object, intercept=FALSE, split,
Expand.split=TRUE, keep.zero.df=TRUE ...);
Summary functions can extract principal component information, provide minimum value, maximum value, quartile and numeric type variable
The Frequency statistics of mean value and factor vector sum logical type vector;
Parameter declaration:
object:It is inherited from the model object of class avo;
Intercept:This option is only applicable under single-layer model default situations.Intercept=FALSE
expand.split:Whether it is continuing in level;
Expand.split=TRUE
keep.zero.df:Whether original data length is retained;
Linear least square fitting lsfit (x, y, wt=NULL, intercept=T, tolerance=1.e-07,
Yname=NULL)
It is suitble to weighted least-squares method multiple regression;Return explanatory variable matrix estimation coefficient and residual error and QR points
The list of solution;
Parameter declaration:
X vectors or explanatory variable matrix;
Y response variables can be a matrix;
Wt optional parameters, the execution weight vectors of weighted least-squares method;
Whether intercept should use intercept item;
Tolerance tolerances will be used in matrix decomposition;
Yname is used for the title of response variable;
Factorial analysis factanal (x, factors, data=NULL, covmat=NULL, n.obs=NA, subset,
Na.action, start=NULL, scores=c (" none ", " regression ", " Bartlett "), rotation="
Varimax ", control=NULL ...)
Factanal functions can set out from sample, sample variance matrix or sample correlation matrix and do Factor minute to data
Analysis;
Parameter declaration:
X is the matrix or data frame being made of data.
Factors is the number of the factor;
Data data frames or matrix, this is only used when x is formula;
Covmat is the covariance matrix of sample or the correlation matrix of sample;
The method that x.scores represents factor score.Rotation represents rotation, the title of the rotation function used;
Principal component analysis anticipation function predict.prcomp (object, newdata ...)
By the prediction model of principal component analysis, prediction processing is carried out to data;
Parameter declaration:
object:The object of class prcomp;
newdata:For the data matrix or numerical value analyzed;
Variance analysis calculating aov (formula, data=NULL, projections=FALSE, qr=TRUE,
Contrasts=NULL ...)
The class " aov " of the variance analysis of designated model is fitted differential pattern analysis;
Parameter declaration:
Formula represents the formula of variance analysis, is x~A in one-way analysis of variance;
Data represents to do the data frame of variance analysis;
Projections is logical value, indicates whether to return to prediction result;
Qr is logic flag, indicates whether to return to Orthogonal Decomposition;
Regression forecasting function lm (formula, data, weights, subset, na.action, method=" qr ",
Model=FALSE, x=FALSE, y=FALSE, contrasts=NULL ...)
Lm () function can return fitting as a result, it can be used for being returned, single stratigraphic analysis, variance and covariance point
Analysis;
Parameter introduction:
A kind of model of fit of Formula;
Mono- optional data frame of Data, list;
The subset for the observation that Subset chooses;
Model, x, y, qr logical value are fitted in the model component of object and return to model framework, matrix, response and matrix
QR decompose;
Hierarchical cluster function hclust (d, method=" complete ", members=NULL)
Hclust can perform hierarchical clustering in distance or Similarity Structure;
Parameter declaration:
D represents distance structure or distance matrix;
Method provides the character string of clustering method;
Complete longest distance methods;
Members is the vector of NULL or d length, and default value is that the value of all elements is all 1;
K mean clusters kmeans (x, centers, iter.max=10, nstart=1, algorithm=c ())
K-means clustering algorithms are divided into K class using by the matrix X of N*P so that in class between object away from
From maximum, and the distance between class minimum;
Parameter declaration:
X is the matrix or data frame being made of data;
Centers is the center of the number either initial classes of cluster;
It is 10 when iter.max maximum iterations are default;
The number of nstart random collections.
Algorithm specifies the algorithm for cluster calculation;
Association analysis function confint (object, parm, level=0.95 ...)
Confint functions can obtain the confidence interval of model parameter;
Parameter declaration:
The model that object is suitble to;
Parm character string vectors refer to the parameter of required interval estimation;
Level represents that the confidence level of confidence interval is had to be between 0 and 1;
ARIMA time series modeling functions arima (x, order=c (0,0,0), seasonal=list (order=c
(0,0,0), period=NA), xreg=NULL, include.mean=TRUE, transform.pars=TRUE, fixed
=NULL, init=NULL, method=c (" CSS-ML ", " ML ", " CSS "), n.cond, optim.method="
BFGS ", optim.control=list (), kappa=1e+06)
ARIMA models refer to nonstationary time series being converted into stationary time series, then by dependent variable only to it
Lagged value and the present worth and lagged value of stochastic error returned established model;
Parameter declaration:
X digital vectors or univariate time series;
Order integer vectors are the exponent number (p, d, q) of model;
Seasonal is designated whether as seaconal model;Whether fixed designated models are fixed model, if preset parameter is
0;
Xreg time serieses, vector or regression matrix;
Function of time time (x ...)
cycle(x,...)
The time point in time series a cycle or position are provided, returns to single time series or other objects;
Mono- time series object of x;
Function of time ar (x, aic=TRUE, order.max=NULL, method=c (" yule- walker ",
“burg”,“ols”,
" mle ", " yw "), na.action=na.fail, series=deparse (substitute (x)) ...)
It is suitble to the autoregression model of time series, autoregression model can be fitted to time series;
Parameter declaration:
x:From the single argument or Multivariate Time Series of " ts " Similar integral;
order.max:Autoregressive maximum sequence is suitble to time series;
na.action:Handle the function of missing values.
The beneficial effects of the present invention are Internet era, big data develops like a raging fire, and data and information content are to refer to
Number mode rises, and people carry out data higher level analysis, and then are preferably used.The distributed storage of cloud computing
Technical support is provided with computing capability;And the core of big data is data processing, data mining technology, which becomes, efficiently utilizes number
According to, find value core technology.It is skilled to grasp big data analysis tool, it is good at big data analysis result combination enterprise
Sale and operation management practice are new requirements.The analyzing and training platform of big data is the big data analysis software of latest generation,
Multiple types of data can be carried out quickly analyzing and handling, can met in management of different nature and processes of research & development to a large amount of
The analysis and decision requirement of data.The characteristics of it is maximum is by a variety of dynamic figures and screening conditions, quickly to a large amount of
Data are analyzed and are handled, can generate including block diagram, curve graph, pie chart, scatter plot, constitutional diagram, map, arborescence,
Thermal map, box-shaped figure, summary sheet and crosstab etc. are a variety of to show form, and all figures can provide numerous data analysis dimensions
Degree, supports the access and display of a variety of client end interfaces and web interface.In addition, the platform also has the characteristics that:
There is data warehouse to foundation, existing data warehouse can be made full use of;Avoid existing IT inputs that from can continuing to produce
Raw bigger value;Support big data quantity analysis;Support access strategy on demand;Data warehouse data is loaded directly into, avoids repeating to build
The workload of mould;Various table components are provided:Table, crosstab, figure table, summary sheet;Abundant figure is provided:Bar chart, broken line
Figure, constitutional diagram, pie chart, scatter plot, three-dimensional scatter plot, map, arborescence, thermal map, parallel coordinates figure, box must scheme;Text is provided
Component;Drag and drop formula visualized operation, instant analysis result are shown.In addition, design principle of the present invention is reliable, simple in structure, tool
There is very extensive application prospect.
It can be seen that compared with prior art, the present invention improve with prominent substantive distinguishing features and significantly, implement
Advantageous effect be also obvious.
Specific embodiment
Below by specific embodiment, the present invention will be described in detail, and following embodiment is explanation of the invention, and
The invention is not limited in implementation below.
A kind of analyzing and training platform based on big data provided by the invention, which is characterized in that including with lower module:
Web Analysis Service modules, for publication and analysis result, the Pages Design based on browser;
Automation services assembly module, the calling of data task is with automatically updating;
Management and integrated services module, the collection such as data integration, user management, Study document management, data source information management
Chinese style disposes correlation function;
Statistical fractals module provides advanced statistics computing engines and Deployment Manager;
Application data service module, is directly integrated with enterprise data source:SAP R/3, SAP B/W, UF ERP,
Oracle EBS, Kingdee ERP;
Further include persistence architecture load-on module;
The persistence architecture load-on module includes flat file data processing unit, database file data load
Unit and establishment data model unit;
The flat file data processing unit includes the following steps:
S1.1:Plane txt file application
(1) select " file "->" interpolation data table "->" addition " carries out data loading;
(2) addition analysis text file;
(3) Study document is opened;
(4) " determining " is clicked, imports text;
S1.2:Plane excel file applications
(1) select " file "->" interpolation data table "->" addition " carries out data loading;
(2) Excel Study documents are added;
(3) Excel Study documents are opened;
(4) " determining " is clicked, imports Excel Study documents;
S1.3:Change data step:
(1) " file "-" replacement data table " or selection " insertion "-" row " or selection " insertion "-" row " are selected;
(2) conversion regime is selected;
(3) a variety of normalization methods can be written as expression formula or as switch process during interpolation data table;
(4) conversion can apply in the dialog box of interpolation data, " interpolation data table " dialog box or from external data work
In the insertion column or row of tool;" conversion "-" standardization "-" addition " is clicked to show control as described below;
S1.4:Data are pre-processed, calculate row, standardization processing operation;
S1.5:Transposition data
(1) selection " conversion "-" transposition "-" addition " carries out data transposition;
(2) row identifier:Each unique value in selected identity column or level can form a line in generation table;
(3) column heading:Each unique value in selected classification row or level can be to each polymerization in generated tables of data
Method forms a newline;
(4) value (%V) and polymerization (%M):The row of data value are calculated from it.Value in generated tables of data is according to row
The method selected under " polymerization " in selector menu is calculated;
S1.6:Data are put in reverse
(1) selection " conversion "-" reverse is put "-" addition " carries out data inverse transposition;
(2) will by row:Selected row, wherein the information for reversing and putting data set should be transferred to comprising unconverted;
(3) row to be converted:Selected row, wherein including the value that be merged into single row.The column name of these row will
Class label in new category row as generation;
(4) classification column name:The row for having selected the information provided in the row converted can be summarised in by keying in one
Title;
(5) value column name:Key in a column name that can show the information type included in new value row;
The database file data loading unit includes the following steps:
S1.7:The step of data being opened using OLE DB
(1) select " file "->" interpolation data table "->" addition "->" other "->" database ";
(2) it in " opening database " dialog box, clicks to select " OleDb Data Provider ";
(3) it clicks " configuration ";
S1.8:The step of data being opened using ODBC
(1) select " file "->" interpolation data table "->" addition "->" other "->" database ";
(2) it in " opening database " dialog box, clicks to select " Odbc Data Provider ";
(3) it clicks " configuration ".
The establishment data model unit includes the following steps:
S1.9:Interpolation data connects
S1.91:In library the step of interpolation data source:
(1) " tool " selection " management data connection " option, pop-up " management data connection " dialog box are clicked in menu bar;
(2) " data source " option of selection " addition is new ", selects data source types from list.
(3) according to selected data source types, corresponding information is filled in, is connected to data source, selects database simultaneously " determining ";
(4) " explanation " is optional fills out for addition in " data source dialog box ";
(5) " saving as library item mesh " dialog box will be shown by clicking " preservation ";
(6) newdata source is stored in designated position in library.
S1.92:The step of interpolation data connects in library:
(1) " tool " selection " management data connection " option is clicked in menu bar;
(2) " data connection " option of selection " addition new projects ", selection " the source connections of data in the library " choosing from list
;
(3) according to selected type of data connection, corresponding information is filled in, is connected to data source, selects database simultaneously " true
It is fixed ", pop up " view in connection " dialog box;
(4) in " the available table in database " list, the table used in the analyzing and training platform of big data is double-clicked;
(5) after the completion of, click OK;Pop up " data connection setting " dialog box;The tables of data added can be shown in
In " spreadsheet views " list;
(6) the input connection description in " connection description " frame facilitates other users to understand and uses;
(7) it clicks " preservation " and data connection is stored in designated position in library;
S1.10:Data connection is used in analysis
(1) selection " interpolation data table " option in " file " is clicked, pops up " interpolation data table " dialog box;
(2) it clicks " addition " and is connected to " data connection in library " option, pop-up " selection data connection " dialog box;
(3) your data connection to be used, and " determining " are selected in library;
(4) in " interpolation data table " dialog box, by choosing check box, selection is by which of data connection view
It is added to new data table;
(5) " loading method " and " importing tables of data " is selected still " to be retained in tables of data outer ", also may specify whether
On-demand loading data;
(6) " determining " is clicked;
S1.10:The editor of data connection
S1.101 is connected in library edit data:
(1) selection " tool " is clicked " management data connection ";
(2) your data connection to be edited is selected, is then clicked " editor ";It will show " data connection setting " dialog box;
(3) it makes change and preserves data connection;
S1.102 is in library edit data source:
(1) selection " tool " is clicked " management data connection ";
(2) your data source to be edited is selected, " data source setting " dialog box will be shown by then clicking " editor ";
(3) it makes change and preserves data source;
S1.11:Self-defined inquiry
(1) using tool>Manage data connection or file>Interpolation data table ... creates one to the new of relational database
Then data connection selects necessary content, until showing " view in connection " dialog box;
(2) it in " view in connection " dialog box, selects " self-defined inquiry ">" newly-built self-defined inquiry ";
(3) Query Name is keyed in " self-defined inquiry ";
(4) it is keyed in and inquired using the language of selected database;
(5) verification is clicked;
(6) browsing result row, it is ensured that list required all results row, and ensure that they have correct data type;
(7) click OK;
S1.12:Create tables of data relationship
For establishing the relationship between tables of data in data connection;When there are relationships between tables of data, established by relationship
Data model;
The establishment of S1.121 tables of data " relationship ":
(1) in " view in connection ", the tables of data of " relationship " is established in selection;
(2) select " relationship "->" newly-built relationship " opening relationships;
(3) click OK;
S1.122 tables " relationship " is checked:
Plus sige represents that the table establishes one or more structural relations with other tables in database, to the relationship of checking
Structure clicks plus sige so that view is unfolded;
It further includes big data and excavates module, the big data is excavated module and included:
Principal component analysis function prcomp (x ...)
Prcomp.default (x, retx=TRUE, center=TRUE, scale.=FALSE, tol=
NULL,...)
Parameter declaration:
The specified numeric type or complex matrix for being used for analyzing of x acquiescences;
Retx logical variables designate whether to return to rotary variable;
Center logical variables are designated whether variable centralization;
Scale. logical variable is designated whether variable standardization;
Tol numeric type variables, are used to specify precision, and the value less than the numerical value will be ignored;
Principal component analysis refers to turn to multi objective a kind of statistical analysis technique of a few overall target, the master of generation
Ingredient can reflect the linear combination of the most information, usually original variable of original variable;
Principal component information inquiry function summary.aov (object, intercept=FALSE, split,
Expand.split=TRUE, keep.zero.df=TRUE ...);
Summary functions can extract principal component information, provide minimum value, maximum value, quartile and numeric type variable
The Frequency statistics of mean value and factor vector sum logical type vector;
Parameter declaration:
object:It is inherited from the model object of class avo;
Intercept:This option is only applicable under single-layer model default situations.Intercept=FALSE
expand.split:Whether it is continuing in level;
Expand.split=TRUE
keep.zero.df:Whether original data length is retained;
Linear least square fitting l sfit (x, y, wt=NULL, intercept=T, tolerance=1.e-07,
Yname=NULL)
It is suitble to weighted least-squares method multiple regression;Return explanatory variable matrix estimation coefficient and residual error and QR points
The list of solution;
Parameter declaration:
X vectors or explanatory variable matrix;
Y response variables can be a matrix;
Wt optional parameters, the execution weight vectors of weighted least-squares method;
Whether intercept should use intercept item;
Tolerance tolerances will be used in matrix decomposition;
Yname is used for the title of response variable;
Factorial analysis factanal (x, factors, data=NULL, covmat=NULL, n.obs=NA, subset,
Na.action, start=NULL, scores=c (" none ", " regression ", " Bartlett "), rotation="
Varimax ", control=NULL ...)
Factanal functions can set out from sample, sample variance matrix or sample correlation matrix and do Factor minute to data
Analysis;
Parameter declaration:
X is the matrix or data frame being made of data.
Factors is the number of the factor;
Data data frames or matrix, this is only used when x is formula;
Covmat is the covariance matrix of sample or the correlation matrix of sample;
The method that x.scores represents factor score.Rotation represents rotation, the title of the rotation function used;
Principal component analysis anticipation function predict.prcomp (object, newdata ...)
By the prediction model of principal component analysis, prediction processing is carried out to data;
Parameter declaration:
object:The object of class prcomp;
newdata:For the data matrix or numerical value analyzed;
Variance analysis calculating aov (formula, data=NULL, projections=FALSE, qr=TRUE,
Contrasts=NULL ...)
The class " aov " of the variance analysis of designated model is fitted differential pattern analysis;
Parameter declaration:
Formula represents the formula of variance analysis, is x~A in one-way analysis of variance;
Data represents to do the data frame of variance analysis;
Projections is logical value, indicates whether to return to prediction result;
Qr is logic flag, indicates whether to return to Orthogonal Decomposition;
Regression forecasting function lm (formula, data, weights, subset, na.action, method=" qr ",
Model=FALSE, x=FALSE, y=FALSE, contrasts=NULL ...)
Lm () function can return fitting as a result, it can be used for being returned, single stratigraphic analysis, variance and covariance point
Analysis;
Parameter introduction:
A kind of model of fit of Formula;
Mono- optional data frame of Data, list;
The subset for the observation that Subset chooses;
Model, x, y, qr logical value are fitted in the model component of object and return to model framework, matrix, response and matrix
QR decompose;
Hierarchical cluster function hclust (d, method=" complete ", members=NULL)
Hclust can perform hierarchical clustering in distance or Similarity Structure;
Parameter declaration:
D represents distance structure or distance matrix;
Method provides the character string of clustering method;
Complete longest distance methods;
Members is the vector of NULL or d length, and default value is that the value of all elements is all 1;
K mean clusters kmeans (x, centers, iter.max=10, nstart=1, algorithm=c ())
K-means clustering algorithms are divided into K class using by the matrix X of N*P so that the distance between object in class
Maximum, and the distance between class minimum;
Parameter declaration:
X is the matrix or data frame being made of data;
Centers is the center of the number either initial classes of cluster;
It is 10 when iter.max maximum iterations are default;
The number of nstart random collections.
Algorithm specifies the algorithm for cluster calculation;
Association analysis function confint (object, parm, level=0.95 ...)
Confint functions can obtain the confidence interval of model parameter;
Parameter declaration:
The model that object is suitble to;
Parm character string vectors refer to the parameter of required interval estimation;
Level represents that the confidence level of confidence interval is had to be between 0 and 1;
ARIMA time series modeling functions arima (x, order=c (0,0,0), seasonal=list (order=c
(0,0,0), period=NA), xreg=NULL, include.mean=TRUE, transform.pars=TRUE, fixed
=NULL, init=NULL, method=c (" CSS-ML ", " ML ", " CSS "), n.cond, optim.method="
BFGS ", optim.control=list (), kappa=1e+06)
ARIMA models refer to nonstationary time series being converted into stationary time series, then by dependent variable only to it
The present worth and lagged value of lagged value and stochastic error are returned established model;
Parameter declaration:
X digital vectors or univariate time series;
Order integer vectors are the exponent number (p, d, q) of model;
Seasonal is designated whether as seaconal model;Whether fixed designated models are fixed model, if preset parameter is
0;
Xreg time serieses, vector or regression matrix;
Function of time time (x ...)
cycle(x,...)
Provide the time point in time series a cycle or position, return single time series or other.
Claims (1)
1. a kind of analyzing and training platform based on big data, which is characterized in that including with lower module:
Web Analysis Service modules, for publication and analysis result, the Pages Design based on browser;
Automation services assembly module, the calling of data task is with automatically updating;
Management and integrated services module, the centralizations such as data integration, user management, Study document management, data source information management
Dispose correlation function;
Statistical fractals module provides advanced statistics computing engines and Deployment Manager;
Application data service module, is directly integrated with enterprise data source:SAP R/3, SAP B/W, UF ERP, Oracle
EBS, Kingdee ERP;
Further include persistence architecture load-on module;
The persistence architecture load-on module includes flat file data processing unit, database file data loading unit
And create data model unit;
The flat file data processing unit includes the following steps:
S1.1:Plane txt file application
(1) select " file "->" interpolation data table "->" addition " carries out data loading;
(2) addition analysis text file;
(3) Study document is opened;
(4) " determining " is clicked, imports text;
S1.2:Plane excel file applications
(1) select " file "->" interpolation data table "->" addition " carries out data loading;
(2) Excel Study documents are added;
(3) Excel Study documents are opened;
(4) " determining " is clicked, imports Excel Study documents;
S1.3:Change data step:
(1) " file "-" replacement data table " or selection " insertion "-" row " or selection " insertion "-" row " are selected;
(2) conversion regime is selected;
(3) a variety of normalization methods can be written as expression formula or as switch process during interpolation data table;
(4) conversion can apply in the dialog box of interpolation data, " interpolation data table " dialog box or from external data tool
It is inserted into column or row;" conversion "-" standardization "-" addition " is clicked to show control as described below;
S1.4:Data are pre-processed, calculate row, standardization processing operation;
S1.5:Transposition data
(1) selection " conversion "-" transposition "-" addition " carries out data transposition;
(2) row identifier:Each unique value in selected identity column or level can form a line in generation table;
(3) column heading:Each unique value in selected classification row or level can be to each polymerization in generated tables of data
Form a newline;
(4) value and polymerization:The row of data value are calculated from it.Value in generated tables of data is according in column selector menu
The method selected under " polymerization " is calculated;
S1.6:Data are put in reverse
(1) selection " conversion "-" reverse is put "-" addition " carries out data inverse transposition;
(2) will by row:Selected row, wherein the information for reversing and putting data set should be transferred to comprising unconverted;
(3) row to be converted:Selected row, wherein including the value that be merged into single row.The column name of these row will act as
Class label in the new category row of generation;
(4) classification column name:The column name for having selected the information provided in the row converted can be summarised in by keying in one;
(5) value column name:Key in a column name that can show the information type included in new value row;
The database file data loading unit includes the following steps:
S1.7:The step of data being opened using OLE DB
(1) select " file "->" interpolation data table "->" addition "->" other "->" database ";
(2) it in " opening database " dialog box, clicks to select " OleDb Data Provider ";
(3) it clicks " configuration ";
S1.8:The step of data being opened using ODBC
(1) select " file "->" interpolation data table "->" addition "->" other "->" database ";
(2) it in " opening database " dialog box, clicks to select " Odbc Data Provider ";
(3) it clicks " configuration ".
The establishment data model unit includes the following steps:
S1.9:Interpolation data connects
S1.91:In library the step of interpolation data source:
(1) " tool " selection " management data connection " option, pop-up " management data connection " dialog box are clicked in menu bar;
(2) " data source " option of selection " addition is new ", selects data source types from list.
(3) according to selected data source types, corresponding information is filled in, is connected to data source, selects database simultaneously " determining ";
(4) " explanation " is optional fills out for addition in " data source dialog box ";
(5) " saving as library item mesh " dialog box will be shown by clicking " preservation ";
(6) newdata source is stored in designated position in library.
S1.92:The step of interpolation data connects in library:
(1) " tool " selection " management data connection " option is clicked in menu bar;
(2) " data connection " option of selection " addition new projects ", selection " the source connections of data in the library " option from list;
(3) according to selected type of data connection, corresponding information is filled in, is connected to data source, selects database simultaneously " determining ", bullet
Go out " view in connection " dialog box;
(4) in " the available table in database " list, the table used in the analyzing and training platform of big data is double-clicked;
(5) after the completion of, click OK;Pop up " data connection setting " dialog box;The tables of data added can be shown in " data
In table view " list;
(6) the input connection description in " connection description " frame facilitates other users to understand and uses;
(7) it clicks " preservation " and data connection is stored in designated position in library;
S1.10:Data connection is used in analysis
(1) selection " interpolation data table " option in " file " is clicked, pops up " interpolation data table " dialog box;
(2) it clicks " addition " and is connected to " data connection in library " option, pop-up " selection data connection " dialog box;
(3) your data connection to be used, and " determining " are selected in library;
(4) in " interpolation data table " dialog box, by choosing check box, which of data connection view is added in selection
For new data table;
(5) " loading method " and " importing tables of data " is selected still " to be retained in tables of data outer ", also may specify whether on demand
Load data;
(6) " determining " is clicked;
S1.10:The editor of data connection
S1.101 is connected in library edit data:
(1) selection " tool " is clicked " management data connection ";
(2) your data connection to be edited is selected, is then clicked " editor ";It will show " data connection setting " dialog box;
(3) it makes change and preserves data connection;
S1.102 is in library edit data source:
(1) selection " tool " is clicked " management data connection ";
(2) your data source to be edited is selected, " data source setting " dialog box will be shown by then clicking " editor ";
(3) it makes change and preserves data source;
S1.11:Self-defined inquiry
(1) using tool>Manage data connection or file>Interpolation data table ... creates a new data for arriving relational database
Connection, then selects necessary content, until showing " view in connection " dialog box;
(2) it in " view in connection " dialog box, selects " self-defined inquiry ">" newly-built self-defined inquiry ";
(3) Query Name is keyed in " self-defined inquiry ";
(4) it is keyed in and inquired using the language of selected database;
(5) verification is clicked;
(6) browsing result row, it is ensured that list required all results row, and ensure that they have correct data type;
(7) click OK;
S1.12:Create tables of data relationship
For establishing the relationship between tables of data in data connection;When there are relationships between tables of data, data are established by relationship
Model;
The establishment of S1.121 tables of data " relationship ":
(1) in " view in connection ", the tables of data of " relationship " is established in selection;
(2) select " relationship "->" newly-built relationship " opening relationships;
(3) click OK;
S1.122 tables " relationship " is checked:
Plus sige represents that the table establishes one or more structural relations with other tables in database, to check relationship knot
Structure clicks plus sige so that view is unfolded;
It further includes big data and excavates module, the big data is excavated module and included:
Principal component analysis function prcomp (x ...)
Prcomp.default (x, retx=TRUE, center=TRUE, scale.=FALSE, tol=NULL ...)
Parameter declaration:
The specified numeric type or complex matrix for being used for analyzing of x acquiescences;
Retx logical variables designate whether to return to rotary variable;
Center logical variables are designated whether variable centralization;
Scale. logical variable is designated whether variable standardization;
Tol numeric type variables, are used to specify precision, and the value less than the numerical value will be ignored;
Principal component analysis refers to turn to multi objective a kind of statistical analysis technique of a few overall target, the principal component of generation
It can reflect the linear combination of the most information, usually original variable of original variable;
Principal component information inquiry function summary.aov (object, intercept=FALSE, split, expand.split
=TRUE, keep.zero.df=TRUE ...);
Summary functions can extract principal component information, provide the mean value of minimum value, maximum value, quartile and numeric type variable,
And the Frequency statistics of factor vector sum logical type vector;
Parameter declaration:
object:It is inherited from the model object of class avo;
Intercept:This option is only applicable under single-layer model default situations.Intercept=FALSE
expand.split:Whether it is continuing in level;
Expand.split=TRUE
keep.zero.df:Whether original data length is retained;
Linear least square fitting lsfit (x, y, wt=NULL, intercept=T, tolerance=1.e-07, yname=
NULL)
It is suitble to weighted least-squares method multiple regression;It returns the estimation coefficient of explanatory variable matrix and residual error and QR decomposes
List;
Parameter declaration:
X vectors or explanatory variable matrix;
Y response variables can be a matrix;
Wt optional parameters, the execution weight vectors of weighted least-squares method;
Whether intercept should use intercept item;
Tolerance tolerances will be used in matrix decomposition;
Yname is used for the title of response variable;
Factorial analysis factanal (x, factors, data=NULL, covmat=NULL, n.obs=NA, subset,
Na.action, start=NULL, scores=c (" none ", " regression ", " Bartlett "), rotation="
Varimax ", control=NULL ...)
Factanal functions can set out from sample, sample variance matrix or sample correlation matrix and do factorial analysis to data;
Parameter declaration:
X is the matrix or data frame being made of data.
Factors is the number of the factor;
Data data frames or matrix, this is only used when x is formula;
Covmat is the covariance matrix of sample or the correlation matrix of sample;
The method that x.scores represents factor score.Rotation represents rotation, the title of the rotation function used;
Principal component analysis anticipation function predict.prcomp (object, newdata ...)
By the prediction model of principal component analysis, prediction processing is carried out to data;
Parameter declaration:
object:The object of class prcomp;
newdata:For the data matrix or numerical value analyzed;
Variance analysis calculates aov (formula, data=NULL, projections=FALSE, qr=TRUE, contrasts
=NULL ...)
The class " aov " of the variance analysis of designated model is fitted differential pattern analysis;
Parameter declaration:
Formula represents the formula of variance analysis, is x~A in one-way analysis of variance;
Data represents to do the data frame of variance analysis;
Projections is logical value, indicates whether to return to prediction result;
Qr is logic flag, indicates whether to return to Orthogonal Decomposition;
Regression forecasting function lm (formula, data, weights, subset, na.action, method=" qr ", model=
FALSE, x=FALSE, y=FALSE, contrasts=NULL ...)
Lm () function can return fitting as a result, it can be used for returned, single stratigraphic analysis, variance and covariance analysis;
Parameter introduction:
A kind of model of fit of Formula;
Mono- optional data frame of Data, list;
The subset for the observation that Subset chooses;
Model, x, y, qr logical value are fitted in the model component of object and return to model framework, matrix, QR points of response and matrix
Solution;
Hierarchical cluster function hclust (d, method=" complete ", members=NULL) hclust can be in distance or phase
Like performing hierarchical clustering in property structure;
Parameter declaration:
D represents distance structure or distance matrix;
Method provides the character string of clustering method;
Complete longest distance methods;
Members is the vector of NULL or d length, and default value is that the value of all elements is all 1;
K mean cluster kmeans (x, centers, iter.max=10, nstart=1, algorithm=c ())
K-means clustering algorithms are divided into K class using by the matrix X of N*P so that the distance between object is most in class
Greatly, the distance between class minimum;
Parameter declaration:
X is the matrix or data frame being made of data;
Centers is the center of the number either initial classes of cluster;
It is 10 when iter.max maximum iterations are default;
The number of nstart random collections.
Algorithm specifies the algorithm for cluster calculation;
Association analysis function confint (object, parm, level=0.95 ...)
Confint functions can obtain the confidence interval of model parameter;
Parameter declaration:
The model that object is suitble to;
Parm character string vectors refer to the parameter of required interval estimation;
Level represents that the confidence level of confidence interval is had to be between 0 and 1;
ARIMA time series modeling functions arima (x, order=c (0,0,0), seasonal=list (order=c (0,0,
0), period=NA), xreg=NULL, include.mean=TRUE, transform.pars=TRUE, fixed=
NULL, init=NULL, method=c (" CSS-ML ", " ML ", " CSS "), n.cond, optim.method=" BFGS ",
Optim.control=list (), kappa=1e+06)
ARIMA models refer to nonstationary time series being converted into stationary time series, then by dependent variable only to its lag
The present worth and lagged value of value and stochastic error are returned established model;
Parameter declaration:
X digital vectors or univariate time series;
Order integer vectors are the exponent number (p, d, q) of model;
Seasonal is designated whether as seaconal model;Whether fixed designated models are fixed model, if preset parameter is 0;
Xreg time serieses, vector or regression matrix;
Function of time time (x ...)
cycle(x,...)
The time point in time series a cycle or position are provided, returns to single time series or other objects;
Mono- time series object of x;
Function of time ar (x, aic=TRUE, order.max=NULL, method=c (" yule-walker ", " burg ",
“ols”,
" mle ", " yw "), na.action=na.fail, series=deparse (substitute (x)) ...)
It is suitble to the autoregression model of time series, autoregression model can be fitted to time series;
Parameter declaration:
x:From the single argument or Multivariate Time Series of " ts " Similar integral;
order.max:Autoregressive maximum sequence is suitble to time series;
na.action:Handle the function of missing values.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711428840.8A CN108170770A (en) | 2017-12-26 | 2017-12-26 | A kind of analyzing and training platform based on big data |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711428840.8A CN108170770A (en) | 2017-12-26 | 2017-12-26 | A kind of analyzing and training platform based on big data |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108170770A true CN108170770A (en) | 2018-06-15 |
Family
ID=62520729
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711428840.8A Pending CN108170770A (en) | 2017-12-26 | 2017-12-26 | A kind of analyzing and training platform based on big data |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108170770A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109508317A (en) * | 2018-10-31 | 2019-03-22 | 武汉光谷联众大数据技术有限责任公司 | A kind of Large Volume Data and service management system |
CN110660461A (en) * | 2019-09-23 | 2020-01-07 | 广州市番禺区中心医院(广州市番禺区人民医院、广州市番禺区心血管疾病研究所) | Cross-platform medical data information uploading system based on artificial intelligence |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103544299A (en) * | 2013-10-30 | 2014-01-29 | 刘峰 | Construction method for commercial intelligent cloud computing system |
CN103605732A (en) * | 2013-11-19 | 2014-02-26 | 北京京东尚科信息技术有限公司 | Data warehouse, data warehouse system and data warehouse construction method based on Infobright |
CN105549982A (en) * | 2016-01-14 | 2016-05-04 | 国网山东省电力公司物资公司 | Automated development platform based on model configuration |
CN105843842A (en) * | 2016-03-08 | 2016-08-10 | 东北大学 | Multi-dimensional gathering querying and displaying system and method in big data environment |
US20160267503A1 (en) * | 2013-07-01 | 2016-09-15 | Salespredict Sw Ltd. | System and method for predicting sales |
CN106022477A (en) * | 2016-05-18 | 2016-10-12 | 国网信通亿力科技有限责任公司 | Intelligent analysis decision system and method |
CN106340161A (en) * | 2016-08-25 | 2017-01-18 | 山东联科云计算科技有限公司 | Public security early warning system based on big data |
CN107103050A (en) * | 2017-03-31 | 2017-08-29 | 海通安恒(大连)大数据科技有限公司 | A kind of big data Modeling Platform and method |
CN107491553A (en) * | 2017-08-31 | 2017-12-19 | 武汉光谷信息技术股份有限公司 | A kind of data digging method and system |
-
2017
- 2017-12-26 CN CN201711428840.8A patent/CN108170770A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160267503A1 (en) * | 2013-07-01 | 2016-09-15 | Salespredict Sw Ltd. | System and method for predicting sales |
CN103544299A (en) * | 2013-10-30 | 2014-01-29 | 刘峰 | Construction method for commercial intelligent cloud computing system |
CN103605732A (en) * | 2013-11-19 | 2014-02-26 | 北京京东尚科信息技术有限公司 | Data warehouse, data warehouse system and data warehouse construction method based on Infobright |
CN105549982A (en) * | 2016-01-14 | 2016-05-04 | 国网山东省电力公司物资公司 | Automated development platform based on model configuration |
CN105843842A (en) * | 2016-03-08 | 2016-08-10 | 东北大学 | Multi-dimensional gathering querying and displaying system and method in big data environment |
CN106022477A (en) * | 2016-05-18 | 2016-10-12 | 国网信通亿力科技有限责任公司 | Intelligent analysis decision system and method |
CN106340161A (en) * | 2016-08-25 | 2017-01-18 | 山东联科云计算科技有限公司 | Public security early warning system based on big data |
CN107103050A (en) * | 2017-03-31 | 2017-08-29 | 海通安恒(大连)大数据科技有限公司 | A kind of big data Modeling Platform and method |
CN107491553A (en) * | 2017-08-31 | 2017-12-19 | 武汉光谷信息技术股份有限公司 | A kind of data digging method and system |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109508317A (en) * | 2018-10-31 | 2019-03-22 | 武汉光谷联众大数据技术有限责任公司 | A kind of Large Volume Data and service management system |
CN109508317B (en) * | 2018-10-31 | 2023-06-09 | 陕西合友网络科技有限公司 | High-capacity data and service management system |
CN110660461A (en) * | 2019-09-23 | 2020-01-07 | 广州市番禺区中心医院(广州市番禺区人民医院、广州市番禺区心血管疾病研究所) | Cross-platform medical data information uploading system based on artificial intelligence |
CN110660461B (en) * | 2019-09-23 | 2023-03-24 | 广州市番禺区中心医院(广州市番禺区人民医院、广州市番禺区心血管疾病研究所) | Cross-platform medical data information uploading system based on artificial intelligence |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107704608A (en) | A kind of OLAP multidimensional analyses and data digging system | |
US9639814B2 (en) | Automated default dimension selection within a multidimensional enterprise software system | |
CN106202536B (en) | Global metadata standardized platform system and its construction method based on XBRL | |
US8533152B2 (en) | System and method for data provenance management | |
US7593955B2 (en) | Generation of aggregatable dimension information within a multidimensional enterprise software system | |
CN110134724A (en) | A kind of the data intelligence extraction and display system and method for Building Information Model | |
CN109684330A (en) | User's portrait base construction method, device, computer equipment and storage medium | |
US20140181154A1 (en) | Generating information models in an in-memory database system | |
CN103631882A (en) | Semantization service generation system and method based on graph mining technique | |
US9043337B1 (en) | Data mining and model generation using an in-database analytic flow generator | |
CN103226743A (en) | TRL-based technical maturity assessment information processing method for aeronautic equipment | |
US20090055429A1 (en) | Method and system for data collection | |
Kalampokis et al. | Linked open cube analytics systems: Potential and challenges | |
CN108170770A (en) | A kind of analyzing and training platform based on big data | |
CN115640300A (en) | Big data management method, system, electronic equipment and storage medium | |
CN110377668A (en) | Data analysing method and system | |
Delchev et al. | Big Data Analysis Architecture | |
Wang et al. | Leveraging relational database for fast construction of OPC UA information model | |
Alhaj Ali et al. | Distributed data mining systems: techniques, approaches and algorithms | |
CN107562909A (en) | A kind of big data analysis system and its analysis method for merging search and calculating | |
CN111813555B (en) | Super-fusion infrastructure layered resource management system based on internet technology | |
Fisun et al. | Implementation of the information system of the association rules generation from OLAP-cubes in the post-relational DBMS caché | |
Jiang | Research and practice of big data analysis process based on hadoop framework | |
Zhang et al. | Application of SQL server in data mining | |
Dijkman et al. | Advanced queueing models for quantitative business process analysis |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180615 |
|
RJ01 | Rejection of invention patent application after publication |