WO2016001709A1 - Big data centralized intelligence system - Google Patents

Big data centralized intelligence system Download PDF

Info

Publication number
WO2016001709A1
WO2016001709A1 PCT/IB2014/002769 IB2014002769W WO2016001709A1 WO 2016001709 A1 WO2016001709 A1 WO 2016001709A1 IB 2014002769 W IB2014002769 W IB 2014002769W WO 2016001709 A1 WO2016001709 A1 WO 2016001709A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
system
step
risk management
user
Prior art date
Application number
PCT/IB2014/002769
Other languages
French (fr)
Inventor
Sensharma PRITHVIRAJ
Original Assignee
Jpmorgan Chase Bank, N.A.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jpmorgan Chase Bank, N.A. filed Critical Jpmorgan Chase Bank, N.A.
Priority to PCT/IB2014/002769 priority Critical patent/WO2016001709A1/en
Publication of WO2016001709A1 publication Critical patent/WO2016001709A1/en

Links

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09CCODING OR CIPHERING APPARATUS FOR CRYPTOGRAPHIC OR OTHER PURPOSES INVOLVING THE NEED FOR SECRECY
    • G09C1/00Apparatus or methods whereby a given sequence of signs, e.g. an intelligible text, is transformed into an unintelligible sequence of signs by transposing the signs or groups of signs or by replacing them by others according to a predetermined system
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/283Multi-dimensional databases or data warehouses, e.g. MOLAP or ROLAP

Abstract

Embodiments of the present invention are directed to a system, and method for a central intelligence system for managing, analyzing, and maintaining large scale, connected information systems. The centralized information system may receive data from servers, databases, mainframes, processes, and other technological assets. A user is able to use the centralized information system to run analyses on the data associated with the connected systems, including: historical analysis, real-time analysis, and predictive modeling. The system can monitor the data and automatically correct identified errors without the need of human intervention. The centralized information system can also generate risk management profiles and automatically modify data to conform to the risk management profiles.

Description

BIG DATA CENTRALIZED INTELLIGENCE SYSTEM

TECHNICAL HELD

(00011 The present, invention relates to systems and methods for managing and maintaining large scale, connected information systems including servers, databases, mainframes, and other ec nologicai assets.

BACKGROUND OF THE IN VENTION

[0002] System Administrators ("Administrators") at large corporations must support, a vast array of technological assets, virtual and physical. Administrators are usually part of a simple operational hierarchy, .sometimes based on the asset type they support. For example, server administrators maintain servers, while storage administrators support large memory devices.

[0003] System administration requires a significant number of human resources. Administrators focus much of their time resolving infrastructure incidents, implementing changes and requests. Resources are often wasted to locate causes of failure and to react to sudden and unexpected maintenance or execution problems. Unfortunately, many times these human assets arc restricte by these reactive tasks. Administrators arc not able to detect patterns between events or provide comprehensive solutions that would provide better ways of maintaining the connected infrastructure, either due to the reactive nature of work, or because they are limited by the human ability to make complex inferences from large data sets.

[0004] Furthermore, many big organizations maintain multiple and redundant data systems, which increases the complexity of data management. Further, over the last few years, there has been a rapid, increase in the interconnectedness of devices and in the volume, velocity, and variety of data ihey produce. Administrators, therefore, find it increasingly difficult to effectively maintain their assets without understanding the complete picture of the technological asset landscape.

[OO0S] Currently, .many organizations take a silo approach to managing and maintaining their technological assets, Administrators are divided into homogenous groups with similar objectives, usually by similar technologies or functions. The interconnections between, these groups are usually limited, and only increase on an ad hoc basis. Any co-relations made between events or concepts that, apply to multiple groups are made by Administrators, not software or management systems. Such connections are usually limited to a localized level, and also limited by human ability. j'0O06J While the architectural hierarchies of Administrator groups can remain the same, how the Administrators engage with the same data landscape has to evolve. A need exists for a suite of related analytical, tools and techniques to create a single holistic picture of an increasingly connected environment

BRIEF SUMMARY OF THE IN VENTION 0007| In one aspect of the invention, a coraputer^raplemenied centralized intelligence system is provided for managing, analyzing, and maintaining large scale, connected information systems. The system may include a computer memory storing a Big Data Engine for (a) generating historical analysis, real-time analysis, and forecast modeling regarding connected information systems; (b) automatically correcting errors and self-learning; and (c) creating a risk management profile, The system is linked to at least one connected, information system, and

— 7— transmits and receives data from the at least one connected information system. The system includes a user interface for accepting and interpreting user input and providing information to the user regarding the at least one connected information system. The system, also includes computer processing components for accessing the stored historical analysis, real-time analysis, forecast modeling, correction, and risk management programs and executing the programs to allow the user to perform managing, analyzing, and maintaining tasks through the user interface. f O S) in an additional aspect of the invention, a computer-implemented centralized intelligence system method is provided for managing, analyzing, and maintaining large scale, connected information systems. The method may include storing a Big Data Engine in a computer memory, where the Big Data Engine (a) generates historical analysis, real-time analysis, and forecast modeling regarding connected information systems; (b) automatically corrects data; and (c) creates a risk management profile. The method also implements computer processing components for accessing the stored historical analysis, real-time analysts, forecast modeling, correction, and risk management programs and executing the programs to allow the user to perform managing, analyzing, and maintaining tasks through a user .interface.

BRIEF DESCRIPTION OF DRAWINGS j.0009] The present invention will now he described, by way of example only, with, reference to the accompanying Figures, in which:

[0010] FIG. .1 is a block diagram illustrating an operating environment for a centralized intelligence system in accordance with an embodiment of the invention.

I'OOil] FIG. 2 is an exemplary embodiment illustrating various components of a centralized intelligence system and an exemplary process flow between the components. (0012) FIG. 3 is a block diagram illustrating various components of a Big Data Engine in accordance with an embodiment of the invention.

1.0013.1 FIG. 4A is an exemplary embodiment illustrating a user initiated reporting method associated with the Historical Analysis Engine,

(001.4) FIG. 4B is an exemplary embodiment illustrating a user initiated reporting method associated with, the Real-Time Analysis Engine.

[0015] FIG. 4C is an exemplary embodiment illustrating a user initiated reporting method associated with the Forecast Modeling Engine.

(00161 FIG. 5 is an exemplary embodiment illustrating an automatic correcting method associated with the Correction Engine.

[0017] FIG, 6 is an exemplary embodiment illustrating a risk management method associated with the Risk Management Engine.

[0018] The following describes in detail various embodiments of the present invention. One of ordinary skill in the art would understand that standard programming and engineering techniques may be used to produce such embodiments including software, firmware, hardware, or any combination thereof to implement the disclosed, subject matter. The attached Figures depict exemplary embodiments and are meant to be understood in view of the detai ls disclosed herein.

DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS

[0019] Embodiments of the present invention are directed to a system and method for a centralized intelligence system ("CIS") that provides an integrated and holistic view of large scale, connected information systems including servers, databases, mainframes, and other technological assets.

1.00201 FIG, I is a block diagram illustrating an operating environment for the CIS 10 in accordance with an embodiment of the invention. The CIS is connected through a oommunica tions medium of a network 80 to one or more assets including databases 20, servers 30, inainfram.es 40, as well, as other physical or virtual assets 50. The CIS can also be connected to other processes 60 that link the same or different assets. The CIS can also provide an interface with human assets 70. The CIS may be connected to the one or more assets to receive data feeds.

{'0021] The network 80 is preferably a wired, or wireless local area network (LAN), but may be or include other types of networks. For example, the network 80 may include the .Internet, a wide area network (WAN), a wireless personal area network (PAN). When used in a LAN' networking environment, computers and other technical assets may be connected to the LAN through a network interface or adapter. When used in a WAN networking environment, computers and other technical assets typically include a modem, or other communication mechanism. Modems may be internal or external, and may be connected to the system bus via the user-input interface, or other appropriate mechanism. Computers and technical assets may be connected over the internet, an Intranet, Extranet, Ethernet, or any other system that provides communication . Some suitable communications protocols may include TCP/IP, UDP, or OSI for example. For wireles communications, communications protocols may include Bluetooth, Zigbee, I 'Da or other suitable protocol. Furthermore, components of the system may communicate through a combination of wired or wireless paths. (0022) The databases 20 can include any structure for storing and organizing data. A database supports the definition, creation, querying, update, and administration, of data. Well- known database management systems include Oracle, dBase, SQL, MySQL, Microsoft SQLServer, Microsoft Access, and IBM. DB2. The servers 30 can include any system that responds to requests across a computer network or helps provide a network service. Typical servers include database servers, communications servers, mail servers, print servers, file servers, web servers, DNS servers, prox servers, financial services servers, and gaming servers. The mainframes 40 can include any computer system for bulk data processing. Examples of mainframes include the IBM ^Enterprise System models.

(0023) The other assets 50 can include any technical asset not covered by databases 20, servers 30, or m inframes 40. Such assets may include personal computers, laptops, ultrabooks, mobile devices, tablets, or any other computational devices. Other assets 50 may also include sensors or input/output devices like keyboards, mice, cameras, video cameras, microphones, telephones, iris scanners, fingerprint readers, key card readers, or any other device capable of creating data and/or transmitting a data stream,

(0024) The other processes 60 can include IT service management software that links technical assets like database incident management processes, data integration processes, knowledge management applications, infrastructure management tools, and other processes, software, and tools well known in the art. The other processes 60 may link to the same assets, some of the assets, or none of the assets that link to the CIS. The other processes 60 may connect to a plurality of different networks and more than one OS.

(002-5) The human assets 70 are preferably Administrators, information regarding the Administrators activities including types of tasks done, time spent doing tasks, and for which assets can be valuable. The human assets 70 could also include any other users that interact with the technical assets linked, to the CIS, such as subscribers, employees, patients, clients, or agents. Their usage information can be equally valuable.

[0026] in a preferred embodiment, the above described assets are linked to the CIS 10 such that, they transmit, data to the CIS 10 and receive data from the CIS .10. The CIS 10 may he facilitated through a computer implemented application. The application may run locally or be available over a network like the Intemet or a local area network (LAN). The CIS 10 provides system users with an interactive experience in real time. The details of the CIS 10 are further described in connections with FIGs. 2-6. 0027] FIG. 2 is an exemplary embodiment illustrating various components of a CIS system and an exemplary process flow between the components, The CIS 10 may include, for example, a Big Data Analytics Engine 100, World View Interlace 200, and Data Stream Linkages 300. The Big Data Analytics Engine 100 drives the analysis of the data provided via the Data Stream Linkages 300. The World View Interface 200 provides a graphical, user interface for a. user to view and interact with, the analyses provided by the Big Data Analytics Engine 100.

10028] The Data Stream. Linkages 300 may accept input and send output to the linked assets. As set forth in respect to FIG. 1 , the linked assets may include databases 20, servers 30, mainframes 40, other physical or virtual assets 50, other processes 60, and human assets 70.

[0029] The World View interface 200 operates as an. interface between the CIS 10 and the user. It processes user input in order to provide the high-level management of the linked assets, in embodiments, the World View Interface 200 allows a user to interact with the linked assets and their data at all resolution levels. The input may take various forms and. preferably includes tactile or touch input and voice input. The input may also be keyboard input, motion input, and for the purpose of security monitoring, may include retinal input or thumbprint input. The World View Interface 200 is operational to sense any input that may be provided, from user interface tools on any devices connected to the CIS .10. As used herein, the World View Interface 200 may include any hardware, software, or combination of hardware and software used by a processing machine that allows a user to interact with the computer processor or processing machine.

[0030] In embodiments, the World View Interface 200 can also act as a virtual collaborator that assists Administrators with, their daily tasks. The World View Interface 200 utilizes advances in speech recognition, natural language programming, and machine Seaming to interact with human actors. I t can analyze data in real-time and assist in decision -making by providing a seamless and universal interactive experience for the Administrator. Thus, Administrators without analytical backgrounds can make sense of the data they administer,

[0031] FIG. 3 is a block diagram illustrating various components of the Big Data Analytics Engine 100 in accordance with, an embodiment of the invention. The Big Data Analytics Engine 100 may include, for example, a Historical Analysis Engine 1 10, Real-Time Analysis Engine 120. Forecast Modeling Engine 130. Correction Engine 140, and Risk Management Engine 150. It should be understood that although these components are shown as separate components communicating with one another, software components such as Historical Analysis Engine .1.10 may be stored by a computer memory and accessed and executed by a computer processor in operation. Parallel processing can be utilized to run the components at the same time. While the computer processing components and. computer memory are not shown separately, their functionality is integrated with the components shown and they are fully described herein. (0032) The Historical Analysis Engine 1 10 may provide information to a user through the World View interface 200 regarding the historical performance of linked assets. For example, a user may input a query regarding the historical performance of a functional area that utilizes one or more of the assets connected to the CIS 1.0. The Historical Analysis Engine 110 can utilize machine learning algorithms, pattern detection algorithms, image recognition algorithms, and other algorithms to investigate large volumes of data and provide meaningful analysis. Using the Historical Analysis Engine 1.10, tasks such as root cause analysis could be fully automated. In. a preferred embodiment, the Historical Analysis Engine 1 1.0 provides a historical performance report to the user via the World View interface 200.

[0033] The Real-Time Analysis Engine 120 may provide information to a user through the World View Interface 200 regarding the real-time performance of linked assets. For example, a user may input a query regarding the real-time functionality of a certain line of business that, utilizes one or more assets connected to the CIS 10. The Real-Time Analysis Engine 120 can utilize machine learning algorithms, pattern, detection algorithms, image recognition algorithms, and other algorithms to provide meaningful information regarding a real-time infrastructure event. It may also call on the Historical Analysis Engine 110 to perform historical analysis regarding data related to a specific real-time event. In a preferred embodiment, the Real-Time Analysis Engine 120 can continuously provide up-to-date reports to the user via the World View interface 200 for a given real-time event or events.

1.0034] The Forecast Modeling Engine 130 may provide information to a user through the World. View Interface 200 regarding a predicted performance of linked assets. For example, a user may input a query regarding the future functionality of certain assets connected to the CIS 10. The Forecast Modeling Engine 130 can utilize machine learning algorithms, pattern detection algorithms, image recognition algorithms, and other algorithms to provide a prediction regarding the performance of the specific assets. It may also call on the Historical Analysis Engine 1 1.0 to perform historical analysis regarding data related to certain assets or functions. In a preferred embodiment, the Forecast Modeling Engine .130 provides a predictive report to the user via the World View interface 200.

{0 35J An Administrator could review the one or more reports generated by the Historical Analysis Engine 110» the Real-Time Analysis Engine 120, and/or the Forecast Modeling Engine 130, to more effectively (1) fix a known problem; (2) react to a real-time infrastructure event; or (3) allocate resources to test future infrastructure changes. For example, Administrators may be upgrading one or more systems linked to the CIS 10. To understand where possible breaks may occur, the Admi istrators run the Historical Analysis Engine 110 to produce a report pinpointing possible sonrceis) of failure. Then they run the Forecast Modeling Engine 130 to predict, for example, how many human resources are needed to implement the change, how much time it will take, how other connected systems will be affected, and what changes are needed in the system infrastructure to effectively implement the upgrade. Once the upgrade commences, the Real-Time Analysis Engine 120 can be used to monitor the status of the upgrade and its effect on the other connected systems in real time.

[0036] Referring back to FIG. 3, the Correction Engine 140 may provide information to a user through the World View Interface 200 regarding proactive or reactive steps with respect to unwanted errors associated with, certain assets. For example, the Correction Engine 140 may continuously monitor the data input to the CIS 10 via the Data Stream Linkages 300. The Correction Engine 1 0 can utilize, among other algorithms, machine learning algorithms, pattern detection algorithms, image ecognition algorithms, and other algorithms to analyze the data and recognize correctable errors or unwanted deviations. It may also call on the Historical Analysis Engine 1 10, Real-Time Analysis Engine 120, and/or Forecast Modeling Engine 130 to perform specialized analysis of the data, in a preferred embodiment the Correction Engine 140 employs an intelligent automated mechanism tor pro-active correction of an error as it originates and/or for minimum reactive action after the error is identified. In another embodiment, the Correction Engine 140 curates an infrastructure with no human involvement, reducing both human cost of capital and possibility of human error.

[0037] The Risk Management Engine 150 may provide information to a user through the World View Interface 200 regarding models for risk management based on planned and unplanned future events with real-life constraints. For example, the Risk Management Engine .150 may interface with a database storing regulatory provisions via the Data Stream, linkages 300. The Risk Management Engine 150 can utilize, among other algorithms, machine learning algorithms, partem detection algorithms, image recognition algorithms, and other algorithms to analyze the regulatory provisions. It may also call on the Historical Analysis Engine I. Hh Real- Time Analysis Engine 120, and/or Forecast Modeling Engine 130 to perform specialized analysis of the linked assets' data and the regulatory provisions. In a preferred embodiment, the Risk Management Engine 1.50 produces a risk management profile report to the user via the World View interface 200. In another preferred embodiment, the Risk Management Engine 1 0 can interact with the Correction Engine 140 to flag or even pro-aetiveiy correct deviations deemed to be high risk.

(0038) FlGs. 4A-4C are flow charts illustrating user initiated reporting methods associated with the Historical Analysis Engine 1 10, Real-Time Analysis Engine 120, and/or Forecast Modeling Engine 130. All three methods begin in step 400 and the World View Interface 200 provides the user with an option to run a historical analysis report real-time analysis report, or forecast modeling report in step 401.

[0039] As illustrated in Fig, 4 A, if the user selects a historical analysis report, the Historical Analysis Engine 1 10, via the World View interface 200 in step 10, provides the user with options regarding the historical, analysis to be conducted. The options may include which linked assets are involved, assets and/or data associated with a functional area, assets and/or data associated with a line of business, assets and/or data associated with a particular event, etc. The Historical Analysis Engine 1 10 connects with the selected assets in step 411. In step 4 12, pattern detection algorithms are implemented on the selected data, in step 414, machine learning algorithms are implemented on the selected data. As can be seen in the illustration, the pattern detection algorithms and machine learning algorithms can be run at the same time and share information. The processes can run in serial or parallel on a single processor, multiple processors, or a distributed computational system, The output of the algorithms is synthesized in step 4.16. A report is generated to be output to the user via the World View Interface 200 in step 418. The reports can provide interactive visual displays of related, normalized data in a meaningful and accessible format. The method ends in. step 420,

[0040] As illustrated in Fig. 4B, if the user selects a real-time analysis report in step 401 , the Real-Time Analysis Engine 120, via the World View Interface 200 in step 430, provides the user with options regarding the real-time analysis to be conducted. The options may include which linked assets arc involved, assets and/or data associated with a functional area, assets and/or data associated with a line of business, assets and/or data associated with a particular real-time event, etc. The Real-Time Analysis Engine 120 connects with the selected assets in step 43 1 , In step 432, pattern detection algorithms are implemented on the selected data. In step 434, machine learning algorithms are implemented on the selected data. As can be seen in the illustration, the pattern detection algorithms and machine learning algorithms can be run at the same time and share information. The processes can run in. serial or parallel on a single processor, multiple processors, or a distributed computational system. The output of the algorithms is synthesized in step 436. A report is generated to he output to the user via the World. View interface 200 in step 438. The user is prompted if the real-time analysis should he terminated in step 440. ff the user selects no, the Real-Time Analysis Engine .120 loops back to step 431. hi embodiments, the user is not notified every time the method reaches step 440, rather he is able to input his preference that the real-time analysis should terminate and this input is recognized anytime step 440 is reached. In this way, the real-time analysis report is continuously updated until the user terminates the process. The reports can provide real-time interactive visual, displays of related, normalized data in a meaningful and accessible format. The method ends at step 442.

[0041 j As illustrated in Fig. 4C, if the user selects a forecast modeling report i n step 401 , the Forecast Modeling Engine 130, via the World View interface 200 in. step 450, provides the user with options regarding the predictive analysis to be conducted. The options may include which linked assets are involved, assets and/or data associated with a functional area, assets and/or data associated with a fine of business, assets and/or data associated with a particular event, etc. Other options may include hypothetical criteria, where the user may input a number of variables at different resolution levels. The Forecast Modeling Engine 130 connects with the selected assets in step 45 1. m step 452, the Historical Analysis Engine 1 0 is invoked and relevant information, is automatically supplied to the method. In step 454 pattern detection algorithms are implemented on the selected data and/or variables supplied by the user, in step 456, machine learning algorithms are implemented on the selected data and/or variables. As can be seen in the illustration, the pattern detection algorithms and machine learning algorithms can be run at the same time and share information. The processes can. run in serial or parallel on a single processor, multiple processors, or a distributed computational system, in step 458, the output, of the algorithms is synthesized with the information generated by Historical Analysis Engine. A report is generated to be output to the user via the World View Interface 200 in step 460. The reports can provide predictions in interactive visual displays of related, normalized, data in a meaningful, and accessible format, The method ends at step 462.

[0042] FIG, 5 is a flow chart illustrating an automatic correcting method associated with the Correction Engine 140. The method begins at step 500 and automatically the Correction Engine 1.40 continuously monitors connected assets in step 502. At step 504. the Correction. Engine .140 detects a correctable error/deviation. In step 506, the Correction Engine 140 determines whether the error requires the attentio of an Administrator, In step 508, it notifies an Administrator via the World View Interface 200 that an error has been detected. It then can ask whether the error should be automatically corrected. If yes, the Correction Engine 140 make the correction n. step 510 and logs the change in step 532, If the error does not require human intervention, the Correction Engine 10 can proceed from step 506 to step 510 to automatically correct the error. A report can be generated i step 5.14 that compiles change logs for output via the World View Interface 200. The method ends in. step .16,

[0043 j FIG. 6 is a. flow chart illustrating a risk management method associated with the Risk Management Engine 150, The method begins in step 600 and automatically the Risk Management Engine 150 connects to a regulations database/server asset in step 602. In step 604 it utilizes a machine learning algorithm to create a regulatory framework applicable to the relevant data owned by assets linked to CIS 10 via Data Stream Linkages 300. In. step 606 the Risk Management Engine 150 runs an analysis on the linked assets' data in relation to the regulatory framework, hi step 608, it creates a risk management report and outputs it to a user via the World View Interface 200 in step 6.10. in optional step 612, the Risk Management Engine 150 calls the Correction Engine 140 to automatically correct identified data that does not conform with the risk management report. The method ends in step 614.

1.0044] Generally, it should be noted that the components depicted and described herein above may be, or include, a computer or multiple computers. Although the components are shown as discrete units, ail components may be interconnected or combined. The components may be described in the general, context, of computer-executable instructions, such, as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, etc, that perform particular tasks or implement particular abstract data types.

(0045) As described above, embodiments of the system and processes of the invention are described. The system of the invention or portions of the system of the invention may be in the form of a "processing machine," i.e. a tangibly embodied machine, such as a general purpose computer or a special purpose computer, for example. As used herein, the term "processing machine'5 is to be understood to include at least one processor that uses at least one memory. The at least one memory stores a set of instructions. The instructions may be either permanently or temporarily stored in the memory or memories of the processing machine. The processor executes the instructions that arc stored in the memory or memories in order to process data. The set of instructions may include various instructions that perform a particular task or tasks, such as any of the processing as described herein. Such a set of instructions for performing a particular task may be characterized as a program, software program, or simply software. (0046) As noted, above, the processing machine, which may be constituted, for example, by the particular system and/or systems described above, executes the instructions that are stored in the memory or .memories to process data. This processing of data may be in response to commands by a user or users of the processing machine, in. response to previous processing, in response to a request by another processing machine and/or any other input, for example.

|O047] As noted above, the processing machine used to implement the invention may be a general purpose computer. Bowever, the processing machine described above may also utilize (or be in the form of) any of a wide variety of other technologies including a special purpose computer, a computer system including a microcomputer, mini-computer or mainframe for example, a programmed microprocessor, a micro-controller, a peripheral integrated circuit element, a CSIC (Consumer Specific integrated Circuit) or ASIC (Application Specific- integrated Circuit) or other integrated circuit, a logic circuit, a digital signal processor, a programmable logic device such as a FPGA, PLD, PLA or PAL, or any other device or arrangement of devices that is capable of implementing the steps of the processes of the invention.

(0048) The processing machine used to implement the invention may u li e a suitable operating system. Thus, embodiments of the invention may include a processing machine running the Microsoft Windows Vista operating system, the Microsoft Windows XP operating system, the Microsoft Windows NT operating system, the Microsoft Windows 2000 operating system, the Microsoft Windows 7 operating system, the Microsoft Windows 8 operating system, the Unix operating system, the Linux operating system, the Xenix operating system, the IBM AIX operating system, the Hewlett-Packard. UX operating system, the Novell Netware operating system, the Su Microsystems Solaris operating system, the OS/2 operating system, the BeOS operating system, the Apple Macintosh OS X operating system, the Apache operating system, a GpenSiep operating system, the Apple iOS operating system, the Google Android operating system, or another operating system, or platform.

[0049] it is appreciated that, in order to practice the method of the invention as described above, it is not necessary that the processors and/or the memories of the processing machine he physically located in the same geographical, place. That is, each of the processors and the memories used by the processing machine may be located in geographically distinct locations and connected so as to communicate in any suitable manner. Additionally, it is appreciated that each of ihe processor and/or the memory may be composed of different physical pieces of equipment. Accordingly, it is not necessary that the processor be one single piece of equipment in one location and that the memory be another single piece of equipment in another location. That is, it is contemplated that, the processor may be two pieces of equipment in two different, physical locations. The two distinct pieces of equipment may be connected in any suitable manner. Additionally, the memory may include two or more portions of memory in two or more physical locations.

| OS0| To explain further, processing as described above is performed by various components and various memories. However, it is appreciated that the processing performed by two distinct components as described above may, in accordance with a further embodiment of the invention, be performed by a single component. Further, the processing performed by one distinct component as described above may be performed by two distinct components. In a similar manner, the memory storage performed by two distinct memory portions as described above may, in accordance with a further embodiment of the invention, be performed by a single memory portion. Further, the memory storage performed by one distinct memory portion as described above may be performed by two memory portions.

1.0051.1 Further, various technologies may be used to provide communication between the various processors and/or memories, as well as to allow the processors and/or the memories of the invention to communicate with any other entity; i.e., so as to obtain further instructions or to access and use remote memory stores, for example. Such, technologies used to provide such communication might include a network, the Internet, Intranet, Extranet, LAN, an Ethernet, or any client server system that provides communication, for example. Such communications technologies may use any suitable protocol, such as TCP/IP, UDP, or GSI, for example.

{'0052] As described above, a set of instructions is used in the processing of the invention. The set of instructions may be in the form of a program or software. The software may be in the form of system software or application software, tor example. The software might also be in the form, of a collection of separate programs, a program module within a larger program, or a portion of a program module, for example. The software used might also include modular programming in the form of object oriented programming. The software tells the processing machine what to do with the data being processed.

10053] Further, it is appreciated that the instructions or set of instructions used in the implementation and operation of the invention may be in a suitable form such that the processing machine may read the instructions. For example, the instructions that form, a program may be in. the form of a suitable programming language, which, is converted to machine language or object code to allow the processor or processors to read the instructions. That is, written lines of programming code or source code, in a particular programming language, are converted to machine language using a compiler, assembler or interpreter. The machine language is binary coded machine instructions that axe specific to a particular type of processing machine, i.e., to a particular type of computer, for example. The computer understands the machine language.

[0054] Any suitable programming language may he used in accordance with the various embodiments of the invention, illustratively, the programming language used may include assembly language, Ada, API,, Basic, C, C-H-, COBOL, dBase, Forth., Fortran, Java, Modula-2, Pascal Prolog, REXX, Visual. Basic, and/or JavaScript, for example. Further, it is not necessary that a single type of instructions or single programming language be utilized in conjunction with the operation of the system and method of the invention. Rather, any number of different programming languages may be utilized as is necessary or desirable. 0 55] Also, the instructions and/or data used in the practice of the invention may utilize any compression, hashing, or encryption technique or algorithm, as may be desired. An encryption module might be used to encrypt data. Further, files or other data may be decrypted using a suitable decryption module, for example.

|0056] As described above, the invention may illustratively be embodied in. the form of a processing machine, including a computer or computer system, for example, that includes at least one memory. It is to be appreciated that the set of instructions, i.e., the software tor example, that enables the computer operating system to perform the operations described above may be contained on any of a wide variety of media or medium, as desired, f rther, the data that is processed by the set of instructions might also be contained on any of a wide variety of media or medium.. That is, the particular medium., i.e., the memory in the processing machine, utilized to hold the set of instructions and/or the data used in the invention may take on any of a variety of physical forms or transmissions, for example. Illustratively, the medium may be in the form of paper, paper transparencies, a compact disk, a DVD, an integrated circuit, a hard disk, a floppy disk, an optical disk, a magnetic tape, a RAM, a ROM, a PROM, a EPROM, a wire, a cable, a fiber, communications channel, a satellite transmissions or other remote transmission, as well as any other medium or source of data that may be read by the processors of the invention.

[0057] Further, the memory or memories used in the processing machine that implements the invention may be in any of a wide variety of forms to allow the memory to hold instructions, data, or other information, as is desired. Thus, the memory might be in the form of a database to hold data. The database might use any desired arrangement of files such as a flat file arrangemen or a relational database arrangement, for example. f 00.581 In the system and method of the in entions a variety of "user interfaces" may be utilized to allow a user to interface with the processing machine or machines that are used to implement the invention. As used herein, a user interface includes any hardware, software, or combination of hardware and software used by the processing machine that allows a user to interact with the processing machine. A user interface may be in the form of a dialogue screen for example. A user interface may also include any of a mouse, touch screen, keyboard, voice reader, voice recognizer, dialogue screen, menu box, list, checkbox, toggle switch, a pushbutton, or any other device that allows a user to receive information regarding the operation of the processing machine as it processes a set of instructions and/or provide the processing machine with information. Accordingly, the user interface is any device that provides communication between a user and a processing machine. The information provided by the user to the processing machine through the user interface may be in the form of a command, a selection of data, or some other input, for example.

|0059j As discussed above, a user interface is utilized by the processing machine that performs a set of instructions such that the processing machine processes data for a user. The user interface is typically used by the processing machine for interacting with a user either to convey information or receive information from the user. However, it should be appreciated that in accordance with some embodiments of the system and method of the invention, it is not necessary that a human user actually interact with a user interface used, by the processing machine of the invention. Rather, it is also contemplated that the user interface of the invention might interact, i.e., convey and receive information, with another processing machine, rather than a human, user. Accordingly, the other processing machine might he characterized as a user. Further, it is contemplated that a user interface utilized in the system and method of the invention may interact partially with another processing machine or processing machines, while also interacting partially with a human user.

[00601 It will be readily understood by those persons skilled in the art that the present invention is susceptible to broad utility and application. Many embodiments and adaptations of the present invention other than those herein described, as well as many variations, modifications and equivalent arrangements, will be apparent from or reasonably suggested by the present invention and foregoing description thereof, without departing from the substance or scope of the invention.

[0061] Accordingly, while the present invention has been described here in detail in relation to its exemplary embodiments, it is to be understood that this disclosure is only illustrative and exemplary of the present invention and is made to provide an enabling disclosure of the invention. Accordingly, the foregoing disclosure is not intended to be construed or to limit the present invention or otherwise to exclude any other such embodiments, adaptations, variations, modifications and equivalent arrangements.

Claims

1 , A computing system for the centralized administration of at least one connected information system, the system comprising: a computer memory storing instructions including a Big Data Engine; at least one data link, wherein the data link transmits and receives data from the at least one connected information system; a world view interface for allowing a user to interact with the computing system; at least one computer processor executing the stored instructions to perform steps including: analyzing data associated with the at least one connected information system; generating a report based on the analyzed data; outputtio.g the report to a user via the world view interface; and automatically correcting data associated with the at least one connected information system.
2, The system of claim 1 , wherein the step of analyzing includes performing at least one of historical analysis, real-time analysts, and forecast modeling.
3. The system of claim 2, wherein the steps of generating a report and ©inputting the report are performed continuously
4. The system of claim 2, wherein the step of analyzing includes performing at least a plurality of the historical analysis, real-time analysis,, and forecast modeling in parallel.
5. The system of claim 2, wherein the step of analysing implements machine learning and pattern detection algorithms,
6. The system of claim 1 , wherein the step of automatically correcting data includes notifying a user that an error has been detected.
7. The system of claim I , wherein the step of automatically correcting data utilizes machine learning and. pattern detection algorithms.
8. The system of claim 2, wherein the at least one processor performs additional steps including: creating a risk management profile.
9. The syste of claim 8.. wherein the step of creating a risk management profile includes utilizing machine learning algorithms to produce a risk management profile,
10. The system of claim 9, wherein the step of creating a risk management profile includes automatically correcting data identified by the risk management profile,
11. A comp er-implemented method for the centralized administration of at least one connected information system, the method comprising: storing instructions in at least one computer memory, the instructions including a Big Data Engine; and implementing at least one computer processor for accessing the stored instructions in the computer and executing the instructions to perform steps including: analyzing data associated with, the at least one connected information system; generating a report based on the analyzed data; outputting the report to a user; and automatically correcting data associated with the at least one connected i nfo rotation sys tern .
12. The method of claim 1 1 , wherein the step of analyzing includes performing at least one of historical analysis, real-time analysis, and forecast modeling.
13. The method of claim 12, wherein the steps of generating a report and outputting the report are performed continuously,
14. The method of claim 12, wherein the step of analyzing includes performing at least a plurality of the historical analysis, real-time analysis, and forecast modeling in parallel.
15. The method of claim 12, wherein the step of analyzing implements machine learning and pattern detection algorithms.
16. The method of claim 1 1. wherein the step of automatically correcting dat includes notifying a user that an error has been detected.
17. The method of claim 1 1 , wherein the step of automatically correcting data utilizes machine learning and pattern detection algorithms.
18. The method of claim 12, wherein the at least one processor performs additional steps including: creating a risk management profile.
19. The method of claim 18, wherein the step of creating a risk management profile includes utilizing .machine learning algorithms to produce a. risk management, profile.
20. The method of claim 19, wherein, the step of creating a risk management profile includes autom tically correcting data identified by the risk management profile.
2 i . A computing system for the centralized administration of at least one connected information system, the system comprising: a computer memory storing instructions including a Big Data Engine; at least one data link, wherein the data link transmits and receives data from the at least one connected, information system; a world view interface for allowing a user to interact with the computing system; at least one computer processor executing the stored instructions to perform steps including: analyzing data associated with the at least one connected information system by performing at. least one of historical analysis, real-time analysis, and forecast modeling using machine learning and pattern detection algorithms; generating a report based, on the analyzed data; outputting the report to a user via t e world view i terface; automatically correcting data associated with the at least one connected information system by utilizing machine learning and pattern detection, algorithms; and creating a risk management profile and automatically correcting data identified by the risk management profile,
22. A computer-implemented method for the centralized administration of at least one connected information system, the method comprising: storing instructions in at least one computer memory, the instructions including a Big Data Engine; and implementing at least one computer processor for accessing the stored instructions in the computer and executing the instructions to perform steps including: analyzing data associated with the at least one connected information system by performing at (east one of historical analysis, real-time analysis, and forecast modeling using machine learning and pattern detection algorithms; generating a report based on the analyzed data; outputting the report to a user; automatically correcting data associated with the at least one connected information system by utilizing machine learning and pattern, detection algorithms; and creating a risk management profile and automatically correcting data identified by the risk management prof le.
_ _
PCT/IB2014/002769 2014-07-01 2014-07-01 Big data centralized intelligence system WO2016001709A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/IB2014/002769 WO2016001709A1 (en) 2014-07-01 2014-07-01 Big data centralized intelligence system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/IB2014/002769 WO2016001709A1 (en) 2014-07-01 2014-07-01 Big data centralized intelligence system

Publications (1)

Publication Number Publication Date
WO2016001709A1 true WO2016001709A1 (en) 2016-01-07

Family

ID=55018504

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2014/002769 WO2016001709A1 (en) 2014-07-01 2014-07-01 Big data centralized intelligence system

Country Status (1)

Country Link
WO (1) WO2016001709A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090150213A1 (en) * 2007-12-11 2009-06-11 Documental Solutions, Llc. Method and system for providing customizable market analysis
US20120124197A1 (en) * 2010-11-17 2012-05-17 At&T Intellectual Property I, L.P. Method, system, and computer program product for providing network data flow analysis and routing
US20140067408A1 (en) * 2012-09-04 2014-03-06 Douglas Brian Aach Systems and methods for use in training medical treatments to healthcare professionals

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090150213A1 (en) * 2007-12-11 2009-06-11 Documental Solutions, Llc. Method and system for providing customizable market analysis
US20120124197A1 (en) * 2010-11-17 2012-05-17 At&T Intellectual Property I, L.P. Method, system, and computer program product for providing network data flow analysis and routing
US20140067408A1 (en) * 2012-09-04 2014-03-06 Douglas Brian Aach Systems and methods for use in training medical treatments to healthcare professionals

Similar Documents

Publication Publication Date Title
JP6549351B2 (en) Collection of data within process control system and distribution to big data machines
US20060059253A1 (en) Architectures for netcentric computing systems
US5367473A (en) Expert system for computer system resource management
US7996814B1 (en) Application model for automated management of software application deployments
US6411922B1 (en) Problem modeling in resource optimization
US20080098454A1 (en) Network Management Appliance
EP3029596A1 (en) Cyber threat monitor and control apparatuses, methods and systems
CN1940951B (en) Safety loophole information aggregation
US20110289440A1 (en) Techniques for evaluating and managing cloud networks
JP6595980B2 (en) Computer system and computer-implemented method
US7917536B2 (en) Systems, methods and computer program products for managing a plurality of remotely located data storage systems
CN105917690B (en) (NFV) system of intermodule communication, method and computer program in a network is virtualized based on network function
DE102015101370A1 (en) Managing big data in process control systems
US20170060574A1 (en) Edge Intelligence Platform, and Internet of Things Sensor Streams System
Wen et al. Fog orchestration for internet of things services
US9558220B2 (en) Big data in process control systems
US20130132551A1 (en) Reduction of alerts in information technology systems
CN107690623B (en) Automatic abnormality detection and solution system
WO2013048986A1 (en) Method, system and program product for allocation and/or prioritization of electronic resources
JP2006500654A (en) Adaptive problem determination and recovery in computer systems
US20160098037A1 (en) Data pipeline for process control system anaytics
CN1737790A (en) Dynamic schema for unified plant model
US10379842B2 (en) Edge computing platform
CN102859510B (en) Automated recovery and escalation in complex distributed applications
US20120166623A1 (en) Performing dependency analysis on nodes of a business application service group

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14896546

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase in:

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14896546

Country of ref document: EP

Kind code of ref document: A1