WO2018112023A1 - Systems and methods for aggregating, filtering, and presenting streaming data - Google Patents

Systems and methods for aggregating, filtering, and presenting streaming data Download PDF

Info

Publication number
WO2018112023A1
WO2018112023A1 PCT/US2017/066068 US2017066068W WO2018112023A1 WO 2018112023 A1 WO2018112023 A1 WO 2018112023A1 US 2017066068 W US2017066068 W US 2017066068W WO 2018112023 A1 WO2018112023 A1 WO 2018112023A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
client
streaming data
query
snapshot
Prior art date
Application number
PCT/US2017/066068
Other languages
French (fr)
Inventor
Ilya Slavin
Matthew Alistair LEGGE
Reed Alpert
Jonathan V. TOM
Original Assignee
Jpmorgan Chase Bank, N.A.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US15/378,501 external-priority patent/US10657137B2/en
Application filed by Jpmorgan Chase Bank, N.A. filed Critical Jpmorgan Chase Bank, N.A.
Priority to AU2017378245A priority Critical patent/AU2017378245B2/en
Priority to EP17881926.4A priority patent/EP3555741A4/en
Priority to CN201780085598.0A priority patent/CN110249322B/en
Priority to JP2019531996A priority patent/JP7048614B2/en
Publication of WO2018112023A1 publication Critical patent/WO2018112023A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2455Query execution
    • G06F16/24568Data stream processing; Continuous queries

Definitions

  • the present disclosure generally relates to systems and methods for aggregating, filtering, and presenting streaming data.
  • a method for presenting streaming data may include (1 ) receiving, at web services layer for a server comprising at least one computer processor, a query from a client, wherein the query comprises a plurality of parameters; (2) a data caching layer for the server receiving streaming data from at least one predefined streaming data source; (3) the data caching layer conflating the streaming data for each of the plurality of parameters; (4) the data caching layer aggregating the conflated data; (5) the data caching layer generating a snapshot of the conflated data by simultaneously running the query against the conflated data; and (6) outputting the snapshot to the client.
  • the parameters may include a specific descriptor for at least one of a security and an investment.
  • the query may further include an identification of a streaming data source.
  • the streaming data may include market data.
  • the web services layer may output the snapshot that is delayed by a predetermined amount of time, wherein the period of time is based on one or more rale associated with the streaming data.
  • the method may further include an entitlement services layer for the server verifying that the client is authorized to access information responsive to the query.
  • the method may further include
  • the snapshot may be accurate for the at least one of the security and the investment for a certain time.
  • the snapshot may include a proper state for the at least one of the security and the investment for a certain time.
  • a system for presenting streaming data may include a plurality of streaming data sources; a data loader for each streaming data source, the data loader receiving streaming data from the streaming data source; a data caching layer the receives the streaming data from the data loaders; and a web services layer comprising at least one computer processor in communication with the data caching layer.
  • the web services layer may receive a query from a client, wherein the query comprises a plurality of parameters; the data caching layer may conflate the streaming data for each of the plurality of parameters; the data caching layer may aggregate the conflated data; the data caching layer may generate a snapshot of the conflated data by simultaneously running the query against the conflated data; and the snapshot may be output to the client,
  • the parameters may include a specific descriptor for at least one of a security and an investment.
  • the query may include an identification of a streaming data source.
  • the streaming data may include market data.
  • the web services layer may output the snapshot that is delayed by a predetermined amount of time.
  • the period of time may be based on one or more rules associated with the streaming data.
  • the query may be received from at least one of a cloud application and a local application.
  • the system may further include an entitlement services layer for the server verifying that the client is authorized to access information responsive to the query.
  • the entitlement services layer may further authenticate the client based on at least one client credential that is received from the client.
  • the snapshot may be accurate for the at least one of the security and the investment for a certain time
  • the snapshot may represent a proper state for the at least one of the security and the investment for a certain time.
  • Figure 1 depicts a system for aggregating, filtering, and presenting streaming data according to one embodiment
  • Figure 2 depicts a method for aggregating, filtering, and presenting streaming data according to one embodiment
  • Figure 3 depicts a strategic flow is provided according to one embodiment.
  • Figure 4 depicts an exemplary process flow according to one embodiment.
  • Embodiments are directed to systems and methods for aggregating, filtering, and presenting streaming data.
  • data from internal and external sources may be pushed into a high-speed data aggregation engine.
  • Clients may make a simple web services request, specifying their credentials, source of information, and list of requested information. For example, in a financial institution, clients may submit a list of securities and fields. The client then receives a single response with the information that the client needs. This reduces or eliminates high-speed stream management that all applications have to support as well as the required infrastructure and development.
  • any suitable aggregation engine may be used as is necessary and/or desired.
  • the aggregation engine may permit Structured Query Language ("SQL”)-style content filtering on individual fields within data payloads, allowing for near realtime analytics on a wide array of products.
  • SQL Structured Query Language
  • Embodiments may provide some or all of the following: (1) increased price advantage of Cloud-based applications through lower data transit costs; (2) large scale reduction of technology costs through decreased infrastructure requirements and shorter development cycles for data-driven applications; (3) more fungible developer workforce due to simplification of API and integration with spreadsheets; (4) advanced content filtering across disparate sources permits development of real-time application features that were cost prohibitive in the past; (5) facilitated distributed analytics through merging of external and internal data; and (6) the ability to age, or delay, the delivery of real-time data for individuals that do not need the data in realtime to reduce the cost of that data. Other benefits may also be provided.
  • Embodiments address these and other challenges while giving end users and developers an interface with filtering.
  • FIG. 1 a system for aggregating, filtering, and presenting streaming data is disclosed according to one embodiment.
  • System 100 may include a plurality of client access points 110, web services layer 120, data caching layer 130, data loaders 140 l5 140 2 , ... 140 n , and data sources 150i, 150 2 , ... 150 n .
  • Data sources 150 may receive data from one or more streaming data source, such as internal and external market data, industry news, etc.
  • Data loaders 140 may receive data from data sources 150.
  • each data loader 140 may "feed" streaming data from one or more data source 150 into data caching layer 130.
  • Each data loader 140 may further be in communication with web services layer 120.
  • each data loader 140 may communicate with control plane sendee 126, which may instruct one or more data loader 140 to create a new subscription, retry a failure, etc.
  • Data caching layer 130 may service requests from, web services layer 120.
  • data caching layer 130 may function as both a database and a message bus.
  • data caching layer 130 may also filter the data and may provide a "snapshot" of requested data.
  • An example of a suitable data caching layer 130 is the
  • AMPS Advanced Message Processing System
  • Web services layer 120 may interface with clients using one or more client access points 1 10.
  • web services layer may provide sendees, such as configuration management sendee 122, monitoring service 124, control plane layer 126, and entitlement services 128. Other services may be provided as is necessary and/or desired.
  • configuration management service 122 may provide configuration data for one or more data loader 140, such as what data sources 150 to connect to, how to connect to data sources 150, etc.
  • configuration management service 122 may provide run-time configuration information, for example, connection information (e.g., hosnport) for data loaders 140, web services layer 120, entitlement sendees 128, what data feeds to load, what web connections to support (http or https), SSL certificate locations, connection and thread pool sizing, logging intervals, environment (dev/test/prod), etc.
  • monitoring service 124 may monitor a status of one or more data source 150, and may re-route requests if a data source becomes unavailable.
  • Control plane sendee 150 may interface with one or more data loaders 140.
  • control plane service 150 may pass directives between services, allowing both automated instructions (e.g., monitoring sendee 124 notices a failure of a data loader and instructs web service layer 120 to switch servers within data aggregation layer 130), or manual directives (e.g., an operations team wants to take a service offline).
  • entitlement services 128 may verify that the client is allowed to access the data that is requested.
  • client access points 1 10 may include, for example, cloud applications, local application, user interfaces, APIs, etc. In one embodiment, client access point 1 10 may provide, for example, a spreadsheet as its output to a client.
  • system 100 may include the ability to provide a "subscription" for a client to receive updates.
  • updates to a complex query may be smaller in size than re-submitting the query, and receiving a full result set several times a second.
  • a client may request information of interest, via an interface, such as a cloud application, a local application, an upload, etc.
  • the client may also provide credentials and identify the source of information and a list of requested information.
  • machine learning may be used to identify the sources based on past, queries,
  • the client may use a web service protocol to request the information.
  • the web service protocol may use, for example, SQL-like syntax for filtering information.
  • the query may include an identification of a stock, industry, an field/area of interest, etc.
  • the query may be used in filters.
  • the request may be a subscription (e.g., to a stock, industry, etc.).
  • the request may specify a period or condition for each snapshot to be provided.
  • the client may be authorized to access the requested data. In one embodiment, this may involve an entitlement check to determine if the client is permitted access to the requested information,
  • the client may be further authenticated based on credentials that may be provided as part of the request.
  • the request may be provided to the data caching layer, which, in step 240, may filter the data received from streaming data sources.
  • the data caching layer may also "snapshot" the data, that is, provide a data state at a particular instance in time.
  • business logic may reduce the number of fields of data that are presented to the client.
  • the filtered data may be conflated.
  • data from more than one source that has been filtered may be aggregated.
  • a stock may change in price from 70.00 to 70.05, back to 69.90 and up to 70.01.
  • client would only receive 70.01 since that was the proper state of the price at the end of the minute. This value, however, will not be the last value as it is not always the proper data point, in particular for a dataset with multiple fields, such as Bid and Ask,
  • a data snapshot of the data may be generated.
  • a snapshot may be for a collection of securities, and may be accurate across that collection at that point of time.
  • the query may be r n simultaneously across the data thai has been aggregated (e.g., the securities of interest) rather than sequentially.
  • the snapshot may be output to the client via a web application, local application, API, spreadsheet, etc.
  • the delivery of the snapshot may be delayed by a predetermined amount of time in order to "age" the data.
  • the timing for data deliver ⁇ ' may determine whether the data can be delivered for free, or if there is a fee for the data. Not all clients require realtime data; for example, an investment manager may not need real-time data when meeting prospective clients.
  • the system may determine how long to delay the output of the snapshot, which may be based on rules associated with the data and/or the data source, and may delay the snapshot by at least this amount of time.
  • an auditable trail of the age of the data in the snapshot may be maintained.
  • Embodiments disclosed herein may separate the mechanics of the various processes performed by multiple lines of business from content that is relevant to solving the problem. That content, such as various lots in clients' portfolios, changes resulting from trades, money transfers, additions, withdrawals, etc. may be treated as generic internally published data, anonymized, and provided to one or more system/method for aggregating, filtering, and presenting streaming data. Examples of such a system/method are disclosed in U.S. Patent Application Ser. No. 15/378,501, the disclosure of which is hereby incorporated, by reference, in its entirety.
  • market data from various venues may be conflated inside of one or more system/method for aggregating, filtering, and presenting streaming data, and may be merged with portfolio information. As data goes through several merging stages, each lot becomes priced against real-time market data, then aggregated into a real-time portfolio views.
  • customers may use a custom rales engine to subscribe to "alerts" relevant to the business.
  • the system may send a notice to the client when a particular stock reaches a price threshold, when a concentration risk of a particular holding breaches a predefined level, when a valuation of the lot or portfolio rises/falls by a certain percentage, etc.
  • Strategic flow includes customers, a financial institution interface, a gateway service, one or more graphical user interface, a perimeter gateway, a Complex Event Processor (CEP) engine, an internal data platform (e.g., an internal data platform that may provide secure reliable messaging between publishes and consumers. Access may be provided using Open MAM A API), one or more market, data feeds (e.g., NASDAQ Basic, which carries NASDAQ and N YSE market data.
  • CEP Complex Event Processor
  • a commercial feed handler may be used; OPRA, Pink Sheets, Trade Web, etc.), an Anonymizer a data injector, a security master, a portfolio management system, intraday activity monitors (e.g., intraday activity, order management systems, and cash flow systems), an alerting engine, and a notification platform.
  • the CEP engine may process data for applications that require market data and one or more of the following: (1) are in configurations that cannot support high-rate streaming data; (2) only require prices a few times a day but need those prices to have a very small hysteresis; (3) need to submit a large number of symbols at a single time (as opposed to streaming subscriptions that each take a single symbol); (4) aggregation of market and analytical data from a number of sources (e.g., Reuters RMDS, Direct Feeds, Internal Data, LOB market analytics); or (5) need the ability to filter the data on dynamic queries taking into account the values of one or more fields, or can use market data that has been delayed a determined time to lower its cost.
  • sources e.g., Reuters RMDS, Direct Feeds, Internal Data, LOB market analytics
  • the anonymizer component may strip identity of clients from portfolio contents prior to injection into the CEP engine.
  • a reverse process allows processes to look up the ID associated with a specific client's portfolio they are interested in.
  • the data injector may provide a single interface to one or more technical assets. It may map reference data (e.g., symbology) between systems and communicate with both and an Internal Data Platform (e.g., to send portfolio information) and the CEP engine (e.g. ,to ensure instruments of interest are on the watch list).
  • an Internal Data Platform e.g., to send portfolio information
  • the CEP engine e.g. ,to ensure instruments of interest are on the watch list.
  • the portfolio management system may be the source of portfolio content data for clients' holdings. It may publish daily refreshes of portfolios of interest, publish portfolios on demand, as well as maintain the "official" start of day prices for all lots/portfolios.
  • intraday activity monitors may monitor activity such as that originating from Order Management Systems, Cash Flow engines, etc. It may publish updates into the CEP engine as portfolio changes since start of day to reflect changes in valuation in real-time.
  • the alerting engine may translate client directions from the financial institution website into subscriptions against the CEP engine and other back-end systems. Persists alerts across days, listens for signals from the CEP engine that conditions were satisfied, performs desired actions, disables aciioried alerts, etc.
  • Figure 4 depicts an exemplary process flow according to one embodiment.
  • the data source layer may receive market data (e.g., market data streams) and may provide the data to the data loader layer, where it may be loaded into the data caching layer using one or more data loader.
  • the CEP Engine may access the data from the data caching layer, and may provide configuration management, data filter processing, monitoring services, control plane services, entitlement services, etc. In one embodiment, the CEP Engine may provide services to customers via the customer access layer.
  • Data Set 1 Equity instrument from direct feed: [0074] ⁇ snapSymbol>JPM.N ⁇ /snapSymbol> [0075] ⁇ snapS tatu s>ok ⁇ /s n apS tatu s>
  • the instrument price is an average of wBidPrice and wAskPrice.
  • Data Set 3 Bond instrument from vendor feed:
  • Data sets do not need to exist in the same CEP instance.
  • Data sets 1 and 3 may exist on the same CEP instance where the portfolio view is created.
  • Data set 2 may be on another instance due to volume demands of ingestion.
  • Portfolios [00122] An example portfolio design is provided.
  • the portfolio may be stored in its own table, called Portfolios:
  • portfolio pricing may se the following pseud o code:
  • the first layer may receive data from market data publishers. For example, for both PortfolioPricing and CommodityFutures, there may be one table per data feed/source. On PortfolioPricing, these may be named “Equity” and “Bond” and on Commodity Futures, it may be named “Comm”. These tables are not transaction logged nor replicated and the "shape" of the data in these tables is whatever that data source defines.
  • AUPrices which may act as the normalized union of price data from all sources. Data may be normalized and published to AUPrices from the Comm, Equity and Bond tables. In embodiments, this information may be conflated in order to reduce the maximum data rate.
  • the first step in the chain of actions may be to parse the underlying data and extract values of interest into variables. Extraction of values may be specified as a projection expression. For example when processing a message from Equity, the line
  • ⁇ Value>price (/price/wBidPrice+/price/wAskPrice) / 2 ⁇ /Value> computes the average of the /wBidPrice and /wAskPrice and places the result in the variable ⁇ ⁇ price ⁇ ⁇ .
  • the next step of the action is to publish a new message containing the ⁇ ⁇ symbol] ⁇ and the normalized ⁇ ⁇ price ⁇ ), along with a "source" field that indicates the source of the pricing data. This may be published to the AUPrices table and may be formatted as JSON.
  • PortfolioPricing instance has to deal with, while giving PortfolioPricing a complete view of all of the security prices needed to price portfolios.
  • the portfolio may be modeled as a table "Portfolio", with one row per portfolio per security.
  • Each message in Portfolio may have a portfolio ID, a security, and the weight of that security in that portfolio.
  • This table is essentially the input data for the portfolios of interest, and the system may dynamically re-compute portfolio prices as the contents of the portfolio are adjusted, added, and/or removed.
  • This computation is done in the next layer, the aggregations. This may be modeled as two Views. The first view is a join between
  • PricedPortfolio Portfolio and AlTPrices, called “PricedPortfolio.” There may be one message here for ever message in Portfolio, but because of the Join, current price for each security may be included.
  • a top PricedPortfolio may be the 2nd view that computes the portfolio price, "PricedPortfolioAgg.” This is in essence a translation of the pseudocode (above) over to a view, containing one message per portfolio with the portfolio ID, the count of securities in the portfolio, and the SUM of the /price's/weight of the securities in the portfolio.
  • the system may receive portfolio data (e.g., by push), and then be receive queries/subscriptions to PricedPortfio and PricedPortfolioAgg to see the dynamically updated portfolio prices as the underlying market data changes.
  • each portfolio's official start of day valuation may be submitted as part of a daily load into a PortfolioRef table. This permits the PricedPortfolioAgg view to include a calculation of an unrealized start of day gain/loss by comparing current valuation and the reference data submitted into the PortfolioRef table. In one embodiment, a percentage change from start of day portfolio valuation may be calculated.
  • an alerting engine may set up subscriptions into the CEP engine that maps human-identifiable concepts to an underlying data structure. Arithmetic expressions and formulas in SQL language may he used to define when the alerting engine should receive a notification of an event. Upon data being received on a specific subscription, the alerting engine will map it back to a desired action (e.g., sending a message, such as an email or SMS, popping up a window on Advisor's user interface, etc.) and may execute that action. It may be further be configured to remain as an active alert, or disable itself by dropping a subscription.
  • a desired action e.g., sending a message, such as an email or SMS, popping up a window on Advisor's user interface, etc.
  • Non-limiting examples of alerts include portfolio gain/loss above/below dollar amount or percentage, specific holding's value exceeds a specific threshold, etc.
  • Advisors representing multiple clients, may maintain alerts covering multiple portfolios of interest for a holistic view of their clients' holdings.
  • cash may be represented in a portfolio as a number of units of underlying currency, such as the U.S. Dollar.
  • Market data for the currency may be static (such as for U.S. Dollar for Dollar- denominated accounts), or represent current foreign exchange rates.
  • a similar mechanism may be used for hypothetical modeling of portfolio changes.
  • a separate CEP engine or a side table may be used to store the final representation of portfolio's hypothetical contents, while the same aggregation mechanisms can be used to price the portfolio.
  • a separate table in the CEP engine may store stock to sector relationships that may be used to build dynamic risk models to help customers manage risk. Multiple risk models may be defined to facilitate different risk appetites.
  • a separate subscription may be set up on the table containing aggregated views that may then be extracted and preserved in a database.
  • This historical record may be used as an audit trail, or to train machine learning systems to generate automated investment advice to customers and advisors.
  • the CEP engine may be used for real-time margin management, since it is aware of the current valuation status of portfolio lots as well as availability of monetary instruments.
  • the system of the invention or portions of the system of the invention may be in the form of a "processing machine,” such as a general purpose computer, for example.
  • processing machine such as a general purpose computer, for example.
  • the term "processing machine” is to be understood to include at least one processor that uses at least one memory.
  • the at least one memory stores a set of instructions.
  • the instructions may be either permanently or temporarily stored in the memory or memories of the processing machine.
  • the processor executes the instructions that are stored in the memory or memories in order to process data.
  • the set of instructions may include various instructions that perform a particular task or tasks, such as those tasks described above. Such a set of instructions for performing a particular task may be characterized as a program, software program, or simply software,
  • the processing machine may be a specialized processor
  • the processing machine executes the instructions that are stored in the memory or memories to process data.
  • This processing of data may be in response to commands by a user or users of the processing machine, in response to previous processing, in response to a request by another processing machine and/or any other input, for example.
  • the processing machine used to implement the invention may be a general purpose computer.
  • the processing- machine described above may also utilize any of a wide variety of other technologies including a special purpose computer, a computer system including, for example, a microcomputer, mini-computer or mainframe, a programmed microprocessor, a micro-controller, a peripheral integrated circuit element, a CSIC (Customer Specific Integrated Circuit) or ASIC (Application Specific Integrated Circuit) or other integrated circuit, a logic circuit, a digital signal processor, a programmable logic device such as a FPGA, PLD, PL A or PAL, or any other device or arrangement of devices that is capable of implementing the steps of the processes of the invention,
  • the processing machine used to implement the invention may utilize a suitable operating system.
  • embodiments of the invention may include a processing machine running the iOS operating system, the OS X operating system, the Android operating system, the Microsoft WindowsTM operating systems, the Unix operating system, the Linux operating system, the Xenix operating system, the IBM ATXTM operating system, the Hewlett- Packard UXTM operating system, the Novell NetwareTM operating system, the Sun Microsystems SolarisTM operating system, the OS/2TM operating- system, the BeOSTM 1 operating system, the Macintosh operating system, the Apache operating system, an OpenStepTM operating system or another operating system or platform.
  • processors and/or the memories of the processing machine it is not necessary that the processors and/or the memories of the processing machine be physically located in the same geographical place. That is, each of the processors and the memories used by the processing machine may be located in geographically distinct locations and connected so as to communicate in any suitable manner.
  • each of the processor and/or the memory may be composed of different physical pieces of equipment. Accordingly, it is not necessary that the processor be one single piece of equipment in one location and that the memory be another single piece of equipment in another location. That is, it is contemplated that the processor may be two pieces of equipment in two different physical locations. The two distinct pieces of equipment may be connected in any suitable manner. Additionally, the memory may include two or more portions of memory in two or more physical locations.
  • processing is performed by various components and various memories.
  • the processing performed by two distinct components as described above may, in accordance with a further embodiment of the invention, be performed by a single component.
  • the processing performed by one distinct component as described above may be performed by two distinct components, hi a similar manner, the memory storage performed by two distinct memory portions as described above may, in accordance with a further embodiment of the invention, be performed by a single memory portion.
  • the memory storage performed by one distinct memory portion as described above may be performed by two memory portions,
  • Such technologies used to provide such communication mieht include a network, the Internet. Intranet.
  • Extranet LAN
  • Ethernet wireless communication via cell tower or satellite
  • client server system that provides communication
  • Such communications technologies may use any suitable protocol such as TCP/IP, UDP, or OSI, for example.
  • a set of instructions may be used in the processing of the invention.
  • the set of instructions may be in the form of a program or software.
  • the software may be in the form of system software or application software, for example.
  • the software might also be in the form of a collection of separate programs, a program module within a larger program, or a portion of a program module, for example.
  • the software used might also include modular programming in the form of object oriented programming. The software tells the processing machine what to do with the data being processed.
  • the instructions or set of instructions used in the implementation and operation of the invention may be in a suitable form such that the processing machine may read the instructions.
  • the instructions that form a program may be in the form of a suitable programming language, which is converted to machine language or object code to allow the processor or processors to read the instructions. That is, written lines of programming code or source code, in a particular programming language, are converted to machine language using a compiler, assembler or interpreter.
  • the machine language is binary coded machine instructions that are specific to a particular type of processing machine, i.e., to a particular type of computer, for example. The computer understands the machine language.
  • programming language used may include assembly language, Ada, APL, Basic, C, C++, COBOL, dBase, Forth, Fortran, Java, Modula-2, Pascal, Prolog, REXX, Visual Basic, and/or JavaScript, for example. Further, it is not necessary that a single type of instruction or single programming language be utilized in conjunction with the operation of the system and method of the invention. Rather, any number of different programming languages may be utilized as is necessary and/or desirable.
  • the instructions and/or data used in the practice of the invention may utilize any compression or encryption technique or algorithm, as may be desired.
  • An enciyption module might be used to encrypt data.
  • files or other data may be decrypted using a suitable decryption module, for example.
  • the invention may illustratively be embodied in the form of a processing machine, including a computer or computer system, for example, that includes at least one memory.
  • the set of instructions i.e., the software for example, that enables the computer operating system to perform the operations described above may be contained on any of a wide variety of media or medium, as desired.
  • the data that is processed by the set of instructions might also be contained on any of a wide variety of media or medium. That is, the particular medium, i.e., the memory in the processing machine, utilized to hold the set of instructions and/or the data used in the invention may take on any of a variety of physical forms or transmissions, for example.
  • the medium may be in the form of paper, paper transparencies, a compact disk, a DVD, an integrated circuit, a hard disk, a floppy disk, an optical disk, a magnetic tape, a RAM, a ROM, a PROM, an EPROM, a wire, a cable, a fiber, a communications channel, a satellite transmission, a memory card, a SIM card, or other remote transmission, as well as any other medium or source of data that may be read by the processors of the invention.
  • the memory or memories used in the processing machine that implements the invention may be in any of a wide variety of forms to allow the memory to hold instructions, data, or other information, as is desired.
  • the memory might be in the form of a database to hold data.
  • the database might use any desired arrangement of files such as a flat file arrangement or a relational database arrangement, for example.
  • a user interface includes any hardware, software, or combination of hardware and software used by the processing machine that allows a user to interact with the processing machine.
  • a user interface may be in the form of a dialogue screen for example.
  • a user interface may also include any of a mouse, touch screen, keyboard, keypad, voice reader, voice recognizer, dialogue screen, menu box, list, checkbox, toggle switch, a pushbutton or any other device that allows a user to receive information regarding the operation of the processing machine as it processes a set of instructions and/or provides the processing machine with information.
  • the user interface is any device that provides communication between a user and a processing machine.
  • the information provided by the user to the processing machine through the user interface may be in the form of a command, a selection of data, or some other input, for example.
  • a user interface is utilized by the processing machine that performs a set of instructions such that the processing machine processes data for a user.
  • the user interface is typically used by the processing machine for interacting with a user either to convey information or receive information from the user.
  • the user interface of the invention might interact, i.e., convey and receive information, with another processing machine, rather than a human user. Accordingly, the other processing machine might be characterized as a user.
  • a user interface utilized in the system and method of the invention may interact partially with another processing machine or processing machines, while also interacting partially with a human user.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Information Transfer Between Computers (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Financial Or Insurance-Related Operations Such As Payment And Settlement (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Processing (AREA)

Abstract

Systems and methods for aggregating, filtering, and presenting streaming data are disclosed. In one embodiment, a method for presenting streaming data may include (1 ) receiving, at web services layer for a server comprising at least one computer processor, a query from a client, wherein the query comprises a plurality of parameters; (2) a data caching layer for the server receiving streaming data from at least one predefined streaming data source; (3) the data caching layer conflating the streaming data for each of the plurality of parameters; (4) the data caching layer aggregating the conflated data; (5) the data caching layer generating a snapshot of the conflated data by simultaneously running the query against the conflated data; and (6) outputting the snapshot to the client.

Description

BACKGROUND OF THE INVENTION
1 . Field of the Invention
[0001] The present disclosure generally relates to systems and methods for aggregating, filtering, and presenting streaming data.
2. Description of The Related Art
[0002] Both public and private clouds have become increasingly popular computing environments. These environments, however, do not support streaming data due to variable latency between the host system and the remote operating system. In addition, the cost of pushing terabytes of streaming market data into various clouds incurs very high "transit" fees from providers, decreasing the value proposition of using clouds for data- driven applications.
SUMMARY OF THE INVENTION
[0003] Systems and methods for aggregating, filtering, and presenting streaming data are disclosed. In one embodiment, a method for presenting streaming data may include (1 ) receiving, at web services layer for a server comprising at least one computer processor, a query from a client, wherein the query comprises a plurality of parameters; (2) a data caching layer for the server receiving streaming data from at least one predefined streaming data source; (3) the data caching layer conflating the streaming data for each of the plurality of parameters; (4) the data caching layer aggregating the conflated data; (5) the data caching layer generating a snapshot of the conflated data by simultaneously running the query against the conflated data; and (6) outputting the snapshot to the client.
[0004] In one embodiment, the parameters may include a specific descriptor for at least one of a security and an investment.
[0005] In one embodiment, the query may further include an identification of a streaming data source.
[0006] In one embodiment, the streaming data may include market data.
[0007] In one embodiment, the web services layer may output the snapshot that is delayed by a predetermined amount of time, wherein the period of time is based on one or more rale associated with the streaming data.
[0008] In one embodiment, the method may further include an entitlement services layer for the server verifying that the client is authorized to access information responsive to the query.
[0009] In one embodiment, the method may further include
authenticating the client based on at least one client credential that is received from the client.
[0010] In one embodiment, the snapshot may be accurate for the at least one of the security and the investment for a certain time.
[0011] In one embodiment, the snapshot may include a proper state for the at least one of the security and the investment for a certain time.
[0012] According to another embodiment, a system for presenting streaming data may include a plurality of streaming data sources; a data loader for each streaming data source, the data loader receiving streaming data from the streaming data source; a data caching layer the receives the streaming data from the data loaders; and a web services layer comprising at least one computer processor in communication with the data caching layer. The web services layer may receive a query from a client, wherein the query comprises a plurality of parameters; the data caching layer may conflate the streaming data for each of the plurality of parameters; the data caching layer may aggregate the conflated data; the data caching layer may generate a snapshot of the conflated data by simultaneously running the query against the conflated data; and the snapshot may be output to the client,
[0013] In one embodiment, the parameters may include a specific descriptor for at least one of a security and an investment.
[0014] In one embodiment, the query may include an identification of a streaming data source.
[0015] In one embodiment, the streaming data may include market data.
[0016] In one embodiment, the web services layer may output the snapshot that is delayed by a predetermined amount of time.
[0017] In one embodiment, the period of time may be based on one or more rules associated with the streaming data.
[0018] In one embodiment, the query may be received from at least one of a cloud application and a local application.
[0019] In one embodiment, the system may further include an entitlement services layer for the server verifying that the client is authorized to access information responsive to the query. The entitlement services layer may further authenticate the client based on at least one client credential that is received from the client.
[0020] In one embodiment, the snapshot may be accurate for the at least one of the security and the investment for a certain time,
[0021] In one embodiment, the snapshot may represent a proper state for the at least one of the security and the investment for a certain time.
BRIEF DESCRIPTION OF THE DRAWINGS
[0022] For a more complete understanding of the present invention, the objects and advantages thereof, reference is now made to the following descriptions taken in connection with the accompanying drawings in which:
[0023] Figure 1 depicts a system for aggregating, filtering, and presenting streaming data according to one embodiment;
[0024] Figure 2 depicts a method for aggregating, filtering, and presenting streaming data according to one embodiment;
[0025] Figure 3 depicts a strategic flow is provided according to one embodiment; and
[0026] Figure 4 depicts an exemplary process flow according to one embodiment.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
[0027] Several embodiments of the present invention and their advantages may be understood by referring to Figures 1-4.
[0028] Embodiments are directed to systems and methods for aggregating, filtering, and presenting streaming data. [0029] In one embodiment, data from internal and external sources may be pushed into a high-speed data aggregation engine. Clients may make a simple web services request, specifying their credentials, source of information, and list of requested information. For example, in a financial institution, clients may submit a list of securities and fields. The client then receives a single response with the information that the client needs. This reduces or eliminates high-speed stream management that all applications have to support as well as the required infrastructure and development.
[0030] In one embodiment, any suitable aggregation engine may be used as is necessary and/or desired. In one embodiment, the aggregation engine may permit Structured Query Language ("SQL")-style content filtering on individual fields within data payloads, allowing for near realtime analytics on a wide array of products.
[0031] Embodiments may provide some or all of the following: (1) increased price advantage of Cloud-based applications through lower data transit costs; (2) large scale reduction of technology costs through decreased infrastructure requirements and shorter development cycles for data-driven applications; (3) more fungible developer workforce due to simplification of API and integration with spreadsheets; (4) advanced content filtering across disparate sources permits development of real-time application features that were cost prohibitive in the past; (5) facilitated distributed analytics through merging of external and internal data; and (6) the ability to age, or delay, the delivery of real-time data for individuals that do not need the data in realtime to reduce the cost of that data. Other benefits may also be provided.
[0032] In the financial industry, because market data and messaging involves a large amount of data, clients are expected to do significant work associated with dealing with payload coming out of the ticker plants. They often have to deal with hardware and software challenges related to large volumes of information, conflate it in order for users to derive benefits, aggregate across many threads and systems to support portfolio
management, etc. Embodiments address these and other challenges while giving end users and developers an interface with filtering.
[0033] Referring to Figure 1, a system for aggregating, filtering, and presenting streaming data is disclosed according to one embodiment.
System 100 may include a plurality of client access points 110, web services layer 120, data caching layer 130, data loaders 140l5 1402, ... 140n, and data sources 150i, 1502, ... 150n. Data sources 150 may receive data from one or more streaming data source, such as internal and external market data, industry news, etc.
[0034] Although three data loaders 140 and data sources 150 are illustrated in Figure 1 , it should be recognized that a greater or fewer number of data loaders 140 and data sources 150 may be provided as is necessary and/or desired.
[0035] Data loaders 140 may receive data from data sources 150. In one embodiment, each data loader 140 may "feed" streaming data from one or more data source 150 into data caching layer 130. Each data loader 140 may further be in communication with web services layer 120. For example, each data loader 140 may communicate with control plane sendee 126, which may instruct one or more data loader 140 to create a new subscription, retry a failure, etc.
[0036] Data caching layer 130 may service requests from, web services layer 120. In one embodiment, data caching layer 130 may function as both a database and a message bus. [0037] In one embodiment, data caching layer 130 may also filter the data and may provide a "snapshot" of requested data.
[0038] An example of a suitable data caching layer 130 is the
Advanced Message Processing System (AMPS) from 60East Technologies.
[0039] Web services layer 120 may interface with clients using one or more client access points 1 10. In one embodiment, web services layer may provide sendees, such as configuration management sendee 122, monitoring service 124, control plane layer 126, and entitlement services 128. Other services may be provided as is necessary and/or desired.
[0040] In one embodiment, configuration management service 122 may provide configuration data for one or more data loader 140, such as what data sources 150 to connect to, how to connect to data sources 150, etc. In one embodiment, configuration management service 122 may provide run-time configuration information, for example, connection information (e.g., hosnport) for data loaders 140, web services layer 120, entitlement sendees 128, what data feeds to load, what web connections to support (http or https), SSL certificate locations, connection and thread pool sizing, logging intervals, environment (dev/test/prod), etc.
[0041] In one embodiment, monitoring service 124 may monitor a status of one or more data source 150, and may re-route requests if a data source becomes unavailable.
[0042] Control plane sendee 150 may interface with one or more data loaders 140. In one embodiment, control plane service 150 may pass directives between services, allowing both automated instructions (e.g., monitoring sendee 124 notices a failure of a data loader and instructs web service layer 120 to switch servers within data aggregation layer 130), or manual directives (e.g., an operations team wants to take a service offline).
[0043] In one embodiment, entitlement services 128 may verify that the client is allowed to access the data that is requested.
[0044] In one embodiment, client access points 1 10 may include, for example, cloud applications, local application, user interfaces, APIs, etc. In one embodiment, client access point 1 10 may provide, for example, a spreadsheet as its output to a client.
[0045] In one embodiment, system 100 may include the ability to provide a "subscription" for a client to receive updates. In one embodiment, updates to a complex query may be smaller in size than re-submitting the query, and receiving a full result set several times a second.
[0046] Referring to Figure 2, a method for aggregating, filtering, and presenting streaming data is disclosed according to one embodiment.
[0047] In step 210, a client may request information of interest, via an interface, such as a cloud application, a local application, an upload, etc. In one embodiment, the client may also provide credentials and identify the source of information and a list of requested information. In another embodiment, machine learning may be used to identify the sources based on past, queries,
[0048] In one embodiment, the client may use a web service protocol to request the information. The web service protocol may use, for example, SQL-like syntax for filtering information. An example query that may be submitted is "(/BID / /ASK >= (0.05 * /TRDPRC....1) )", which means "where value of BID divided by value of ASK is greater or equal to 5% of the value of "TRDPRC_1."
[0049] In a financial system, the query may include an identification of a stock, industry, an field/area of interest, etc. The query may be used in filters.
[0050] In one embodiment, the request may be a subscription (e.g., to a stock, industry, etc.). In one embodiment, the request may specify a period or condition for each snapshot to be provided.
[0051] In step 220, the client may be authorized to access the requested data. In one embodiment, this may involve an entitlement check to determine if the client is permitted access to the requested information,
[0052] In one embodiment, the client may be further authenticated based on credentials that may be provided as part of the request.
[0053] In step 230, the request may be provided to the data caching layer, which, in step 240, may filter the data received from streaming data sources. The data caching layer may also "snapshot" the data, that is, provide a data state at a particular instance in time.
[0054] In one embodiment, business logic may reduce the number of fields of data that are presented to the client.
[0055] In step 250, the filtered data may be conflated. For example, data from more than one source that has been filtered may be aggregated. For example, in a minute, a stock may change in price from 70.00 to 70.05, back to 69.90 and up to 70.01. With a conflated source, client would only receive 70.01 since that was the proper state of the price at the end of the minute. This value, however, will not be the last value as it is not always the proper data point, in particular for a dataset with multiple fields, such as Bid and Ask,
[0056] In step 260, a data snapshot of the data may be generated. In one embodiment, a snapshot may be for a collection of securities, and may be accurate across that collection at that point of time. For example, the query may be r n simultaneously across the data thai has been aggregated (e.g., the securities of interest) rather than sequentially.
[0057] In step 270, the snapshot may be output to the client via a web application, local application, API, spreadsheet, etc.
[0058] In one embodiment, the delivery of the snapshot may be delayed by a predetermined amount of time in order to "age" the data. For example, the timing for data deliver}' may determine whether the data can be delivered for free, or if there is a fee for the data. Not all clients require realtime data; for example, an investment manager may not need real-time data when meeting prospective clients.
[0059] Thus, in one embodiment, the system may determine how long to delay the output of the snapshot, which may be based on rules associated with the data and/or the data source, and may delay the snapshot by at least this amount of time. In one embodiment, an auditable trail of the age of the data in the snapshot may be maintained.
[0060] Embodiments disclosed herein may separate the mechanics of the various processes performed by multiple lines of business from content that is relevant to solving the problem. That content, such as various lots in clients' portfolios, changes resulting from trades, money transfers, additions, withdrawals, etc. may be treated as generic internally published data, anonymized, and provided to one or more system/method for aggregating, filtering, and presenting streaming data. Examples of such a system/method are disclosed in U.S. Patent Application Ser. No. 15/378,501, the disclosure of which is hereby incorporated, by reference, in its entirety.
[0061] In embodiments, market data from various venues may be conflated inside of one or more system/method for aggregating, filtering, and presenting streaming data, and may be merged with portfolio information. As data goes through several merging stages, each lot becomes priced against real-time market data, then aggregated into a real-time portfolio views.
[0062] In embodiments, customers may use a custom rales engine to subscribe to "alerts" relevant to the business. For example, the system may send a notice to the client when a particular stock reaches a price threshold, when a concentration risk of a particular holding breaches a predefined level, when a valuation of the lot or portfolio rises/falls by a certain percentage, etc.
[0063] Referring to Figure 3, a strategic flow is provided according to one embodiment. Strategic flow mav include customers, a financial institution interface, a gateway service, one or more graphical user interface, a perimeter gateway, a Complex Event Processor (CEP) engine, an internal data platform (e.g., an internal data platform that may provide secure reliable messaging between publishes and consumers. Access may be provided using Open MAM A API), one or more market, data feeds (e.g., NASDAQ Basic, which carries NASDAQ and N YSE market data. A commercial feed handler may be used; OPRA, Pink Sheets, Trade Web, etc.), an Anonymizer a data injector, a security master, a portfolio management system, intraday activity monitors (e.g., intraday activity, order management systems, and cash flow systems), an alerting engine, and a notification platform. [0064] In one embodiment, the CEP engine may process data for applications that require market data and one or more of the following: (1) are in configurations that cannot support high-rate streaming data; (2) only require prices a few times a day but need those prices to have a very small hysteresis; (3) need to submit a large number of symbols at a single time (as opposed to streaming subscriptions that each take a single symbol); (4) aggregation of market and analytical data from a number of sources (e.g., Reuters RMDS, Direct Feeds, Internal Data, LOB market analytics); or (5) need the ability to filter the data on dynamic queries taking into account the values of one or more fields, or can use market data that has been delayed a determined time to lower its cost.
[0065] In one embodiment, the anonymizer component may strip identity of clients from portfolio contents prior to injection into the CEP engine. A reverse process allows processes to look up the ID associated with a specific client's portfolio they are interested in.
[0066] The data injector may provide a single interface to one or more technical assets. It may map reference data (e.g., symbology) between systems and communicate with both and an Internal Data Platform (e.g., to send portfolio information) and the CEP engine (e.g. ,to ensure instruments of interest are on the watch list).
[0067] In one embodiment, the portfolio management system may be the source of portfolio content data for clients' holdings. It may publish daily refreshes of portfolios of interest, publish portfolios on demand, as well as maintain the "official" start of day prices for all lots/portfolios.
[0068] In one embodiment, intraday activity monitors may monitor activity such as that originating from Order Management Systems, Cash Flow engines, etc. It may publish updates into the CEP engine as portfolio changes since start of day to reflect changes in valuation in real-time.
[0069] The alerting engine may translate client directions from the financial institution website into subscriptions against the CEP engine and other back-end systems. Persists alerts across days, listens for signals from the CEP engine that conditions were satisfied, performs desired actions, disables aciioried alerts, etc.
[0070] Figure 4 depicts an exemplary process flow according to one embodiment. In one embodiment, the data source layer may receive market data (e.g., market data streams) and may provide the data to the data loader layer, where it may be loaded into the data caching layer using one or more data loader. The CEP Engine may access the data from the data caching layer, and may provide configuration management, data filter processing, monitoring services, control plane services, entitlement services, etc. In one embodiment, the CEP Engine may provide services to customers via the customer access layer.
[0071] EXAMPLE
[0072] A non-limiting example is provided below.
[0073] Data Set 1: Equity instrument from direct feed: [0074] <snapSymbol>JPM.N</snapSymbol> [0075] < snapS tatu s>ok</s n apS tatu s>
[0076] <wBidPrice>87 ,790Q00</wBidPrice>
[0077] <wBidSize> 18</wBidSize> [0078] <wAskPrice>87.80Q00Q</wAskPrice>
[0079] <wAskSize>4</wAskSize>
[0080] <wInstrumentType>Stock</wIn strumentType>
[0081] < wOpenPrice>90.360000</wOpenPrice>
[0082] <wAdjPrevClose>90.070000</wAdjPrevClose>
[0083] <wPrevClosePrice>90.070000</wPrevClosePrice>
[0084] <wBidOpen>90.340000</wBidOpen>
[0085] <wAskOpen>90.400000</wAskOpen>
[0086] The instrument price is an average of wBidPrice and wAskPrice.
[0087] Data Set 2: Commodity future from direct feed:
[0088] <snapSymbol>CLJ7.NYM</snapSymbol>
[0089] < snapS tatu s>ok</ s napS tatu s>
[0090] <wTssueSymbol>CLJ7.NYM</wIssueSymbol>
[0091] <wBidHigh>48.620000</wBidHigh>
[0092] <wAskLow>47.400000</wAskLow>
[0093] < wAskPrice>47.4100G0</w AskPrice>
[0094] < wA skSize>5 </w AskSizo
< wB idPrice>47.390000</wB idPrice> [0096] <wBidSize>52</wBidSize>
[0097] <wB idLo w>47.340000</wB idLo w>
[0098] <wAskHigh>48.750000</wAskHigh>
[0099] < wOpenPrice>47.930000</wOpenPrice>
[00100] <wOpenTime>2017-03-21
17:45:27,229603</wOpenTime>
[00101] < CfiCode>FCMXSX</wCfiCode>
[00102] <wCurrency>USD</wCurrency>
[00103] <wEntryStatus>0</wEntryStatus>
[00104] Same rule for price applies
[00105] Data Set 3: Bond instrument from vendor feed:
[00106] <snapSymbol>ISIN_USN82008AK46„MODEL.FUSE</ snapSymbol>
[00107] <snapStatus>ok</snapStatus>
[00108] <CLJENT_OFFER>0.000000</CLIENT_OFFER>
[00109] <CLIENT_BID>0.000000</CLJENT_BID>
[00110] <POSITION>0,000000</POSITION>
[00111] <COUPON_RATE>2.000000</COUPON„RATE>
[00112] <STRING2>HG -
Industrials/Defense/Chemicals</STRING2> <MKT_PRICE_2>1000000.000000</MKT_PRICE_2>
<MKT .PRICE,J>1000000.000000</MKT .PRICE...1>
[00115] <MATURD ATE> 15 Sep 2023</MATURDATE>
[00116] <ISIN...,BENCHMARK>US912828 W556</1SIN.. BENC
HMARK>
7] <CURVE_TYPE>EMEA< CURVE_TYPE> [00118] <UPDATE_TM> 12:31 : 15.000</UPDATE_TM>
[00119] <lTEM...ID/>
[00120] Price is in field MKT_PRICE_1
Data sets do not need to exist in the same CEP instance. For example, Data sets 1 and 3 may exist on the same CEP instance where the portfolio view is created. Data set 2 may be on another instance due to volume demands of ingestion.
[00122] An example portfolio design is provided. The portfolio may be stored in its own table, called Portfolios:
[00123] :
[00124] {
[00125] "portfolio" : "12345", "version" : 3.2.6,
[00127] "holdings" : [ [00128] { "source" : "Data Set 1 ", "security" : "JPM.N", "weight"
: 150 },
[00129] { "source" : "Data Set 2", "security" : "CLJ7.NYM", " weight" : 200 ) ,
[00130] { "source" : "Data Set 3", "security" :
"ISIN_USN82008AK46_MODEL,FUSE", " weight" : 2 } j
[00133] For example, portfolio pricing may se the following pseud o code:
[00134] for specific portfolio ID
[00135] for each asset identifier
[00136] total price ~ total price + (asset weight *
as setjpirce [as set identifier J )
[00137] return total price
[00138] In other words, a single value representing the real-time price of the portfolio, which will change as data changes, may be returned.
[00139] In one embodiment, there may be four layers of tables defined for both instances,
[00140] The first layer may receive data from market data publishers. For example, for both PortfolioPricing and CommodityFutures, there may be one table per data feed/source. On PortfolioPricing, these may be named "Equity" and "Bond" and on Commodity Futures, it may be named "Comm". These tables are not transaction logged nor replicated and the "shape" of the data in these tables is whatever that data source defines.
[00141] At the next layer is a replicated table that may be called
"AUPrices," which may act as the normalized union of price data from all sources. Data may be normalized and published to AUPrices from the Comm, Equity and Bond tables. In embodiments, this information may be conflated in order to reduce the maximum data rate.
[00142] The first step in the chain of actions may be to parse the underlying data and extract values of interest into variables. Extraction of values may be specified as a projection expression. For example when processing a message from Equity, the line
<Value>price=(/price/wBidPrice+/price/wAskPrice) / 2</Value> computes the average of the /wBidPrice and /wAskPrice and places the result in the variable { {price} } .
[00143] The next step of the action is to publish a new message containing the { {symbol] } and the normalized { {price} ), along with a "source" field that indicates the source of the pricing data. This may be published to the AUPrices table and may be formatted as JSON.
[00144] At this point, the incoming market data has been normalized and transformed into a single AUPrices table. Conflation and replication may be used to minimize the total volume of messages that the
PortfolioPricing instance has to deal with, while giving PortfolioPricing a complete view of all of the security prices needed to price portfolios.
[00145] Next, the portfolio may be modeled as a table "Portfolio", with one row per portfolio per security. Each message in Portfolio may have a portfolio ID, a security, and the weight of that security in that portfolio. This table is essentially the input data for the portfolios of interest, and the system may dynamically re-compute portfolio prices as the contents of the portfolio are adjusted, added, and/or removed.
[00146] This computation is done in the next layer, the aggregations. This may be modeled as two Views. The first view is a join between
Portfolio and AlTPrices, called "PricedPortfolio." There may be one message here for ever message in Portfolio, but because of the Join, current price for each security may be included.
[00147] A top PricedPortfolio may be the 2nd view that computes the portfolio price, "PricedPortfolioAgg." This is in essence a translation of the pseudocode (above) over to a view, containing one message per portfolio with the portfolio ID, the count of securities in the portfolio, and the SUM of the /price's/weight of the securities in the portfolio.
[00148] Thus, in embodiments, the system may receive portfolio data (e.g., by push), and then be receive queries/subscriptions to PricedPortfio and PricedPortfolioAgg to see the dynamically updated portfolio prices as the underlying market data changes.
[00149] In embodiments, each portfolio's official start of day valuation may be submitted as part of a daily load into a PortfolioRef table. This permits the PricedPortfolioAgg view to include a calculation of an unrealized start of day gain/loss by comparing current valuation and the reference data submitted into the PortfolioRef table. In one embodiment, a percentage change from start of day portfolio valuation may be calculated.
[00150] In embodiments, an alerting engine may set up subscriptions into the CEP engine that maps human-identifiable concepts to an underlying data structure. Arithmetic expressions and formulas in SQL language may he used to define when the alerting engine should receive a notification of an event. Upon data being received on a specific subscription, the alerting engine will map it back to a desired action (e.g., sending a message, such as an email or SMS, popping up a window on Advisor's user interface, etc.) and may execute that action. It may be further be configured to remain as an active alert, or disable itself by dropping a subscription.
[00151] Non-limiting examples of alerts include portfolio gain/loss above/below dollar amount or percentage, specific holding's value exceeds a specific threshold, etc.
[00152] Advisors, representing multiple clients, may maintain alerts covering multiple portfolios of interest for a holistic view of their clients' holdings.
[00153] At an organizational level, all holdings may be analyzed holistically in a CEP engine to facilitate hedging, cross-account stock loan, risk management, etc.
[00154] In one embodiment, cash may be represented in a portfolio as a number of units of underlying currency, such as the U.S. Dollar. Market data for the currency may be static (such as for U.S. Dollar for Dollar- denominated accounts), or represent current foreign exchange rates.
[00155] In another embodiment, a similar mechanism may be used for hypothetical modeling of portfolio changes. A separate CEP engine or a side table may be used to store the final representation of portfolio's hypothetical contents, while the same aggregation mechanisms can be used to price the portfolio. [00156] In one embodiment, a separate table in the CEP engine may store stock to sector relationships that may be used to build dynamic risk models to help customers manage risk. Multiple risk models may be defined to facilitate different risk appetites.
[00157] In one embodiment, a separate subscription may be set up on the table containing aggregated views that may then be extracted and preserved in a database. This historical record may be used as an audit trail, or to train machine learning systems to generate automated investment advice to customers and advisors.
[00158] In one embodiment, the CEP engine may be used for real-time margin management, since it is aware of the current valuation status of portfolio lots as well as availability of monetary instruments.
[00159] Although several embodiments have been disclosed, it should be recognized that these embodiments are not exclusive to each other.
[00160] Hereinafter, general aspects of implementation of the systems and methods of the invention will be described.
[00161] The system of the invention or portions of the system of the invention may be in the form of a "processing machine," such as a general purpose computer, for example. As used herein, the term "processing machine" is to be understood to include at least one processor that uses at least one memory. The at least one memory stores a set of instructions. The instructions may be either permanently or temporarily stored in the memory or memories of the processing machine. The processor executes the instructions that are stored in the memory or memories in order to process data. The set of instructions may include various instructions that perform a particular task or tasks, such as those tasks described above. Such a set of instructions for performing a particular task may be characterized as a program, software program, or simply software,
[00162] In one embodiment, the processing machine may be a specialized processor,
[00163] As noted above, the processing machine executes the instructions that are stored in the memory or memories to process data. This processing of data may be in response to commands by a user or users of the processing machine, in response to previous processing, in response to a request by another processing machine and/or any other input, for example.
[00164] As noted above, the processing machine used to implement the invention may be a general purpose computer. However, the processing- machine described above may also utilize any of a wide variety of other technologies including a special purpose computer, a computer system including, for example, a microcomputer, mini-computer or mainframe, a programmed microprocessor, a micro-controller, a peripheral integrated circuit element, a CSIC (Customer Specific Integrated Circuit) or ASIC (Application Specific Integrated Circuit) or other integrated circuit, a logic circuit, a digital signal processor, a programmable logic device such as a FPGA, PLD, PL A or PAL, or any other device or arrangement of devices that is capable of implementing the steps of the processes of the invention,
[00165] The processing machine used to implement the invention may utilize a suitable operating system. Thus, embodiments of the invention may include a processing machine running the iOS operating system, the OS X operating system, the Android operating system, the Microsoft Windows™ operating systems, the Unix operating system, the Linux operating system, the Xenix operating system, the IBM ATX™ operating system, the Hewlett- Packard UX™ operating system, the Novell Netware™ operating system, the Sun Microsystems Solaris™ operating system, the OS/2™ operating- system, the BeOS™1 operating system, the Macintosh operating system, the Apache operating system, an OpenStep™ operating system or another operating system or platform.
[00166] It is appreciated that in order to practice the method of the invention as described above, it is not necessary that the processors and/or the memories of the processing machine be physically located in the same geographical place. That is, each of the processors and the memories used by the processing machine may be located in geographically distinct locations and connected so as to communicate in any suitable manner.
Additionally, it is appreciated that each of the processor and/or the memory may be composed of different physical pieces of equipment. Accordingly, it is not necessary that the processor be one single piece of equipment in one location and that the memory be another single piece of equipment in another location. That is, it is contemplated that the processor may be two pieces of equipment in two different physical locations. The two distinct pieces of equipment may be connected in any suitable manner. Additionally, the memory may include two or more portions of memory in two or more physical locations.
[00167] To explain further, processing, as described above, is performed by various components and various memories. However, it is appreciated that the processing performed by two distinct components as described above may, in accordance with a further embodiment of the invention, be performed by a single component. Further, the processing performed by one distinct component as described above may be performed by two distinct components, hi a similar manner, the memory storage performed by two distinct memory portions as described above may, in accordance with a further embodiment of the invention, be performed by a single memory portion. Further, the memory storage performed by one distinct memory portion as described above may be performed by two memory portions,
[00168] Further, various technologies may be used to provide
communication between the various processors and/or memories, as well as to allow the processors and/or the memories of the invention to communicate with any other entity; i.e., so as to obtain further instructions or to access and use remote memory stores, for example. Such technologies used to provide such communication mieht include a network, the Internet. Intranet.
Extranet, LAN, an Ethernet, wireless communication via cell tower or satellite, or any client server system that provides communication, for example. Such communications technologies may use any suitable protocol such as TCP/IP, UDP, or OSI, for example.
[00169] As described above, a set of instructions may be used in the processing of the invention. The set of instructions may be in the form of a program or software. The software may be in the form of system software or application software, for example. The software might also be in the form of a collection of separate programs, a program module within a larger program, or a portion of a program module, for example. The software used might also include modular programming in the form of object oriented programming. The software tells the processing machine what to do with the data being processed.
[00170] Further, it is appreciated that the instructions or set of instructions used in the implementation and operation of the invention may be in a suitable form such that the processing machine may read the instructions. For example, the instructions that form a program may be in the form of a suitable programming language, which is converted to machine language or object code to allow the processor or processors to read the instructions. That is, written lines of programming code or source code, in a particular programming language, are converted to machine language using a compiler, assembler or interpreter. The machine language is binary coded machine instructions that are specific to a particular type of processing machine, i.e., to a particular type of computer, for example. The computer understands the machine language.
[00171] Any suitable programming language may be used in accordance with the various embodiments of the invention. Illustratively, the
programming language used may include assembly language, Ada, APL, Basic, C, C++, COBOL, dBase, Forth, Fortran, Java, Modula-2, Pascal, Prolog, REXX, Visual Basic, and/or JavaScript, for example. Further, it is not necessary that a single type of instruction or single programming language be utilized in conjunction with the operation of the system and method of the invention. Rather, any number of different programming languages may be utilized as is necessary and/or desirable.
[00172] Also, the instructions and/or data used in the practice of the invention may utilize any compression or encryption technique or algorithm, as may be desired. An enciyption module might be used to encrypt data. Further, files or other data may be decrypted using a suitable decryption module, for example.
[00173] As described above, the invention may illustratively be embodied in the form of a processing machine, including a computer or computer system, for example, that includes at least one memory. It is to be appreciated that the set of instructions, i.e., the software for example, that enables the computer operating system to perform the operations described above may be contained on any of a wide variety of media or medium, as desired. Further, the data that is processed by the set of instructions might also be contained on any of a wide variety of media or medium. That is, the particular medium, i.e., the memory in the processing machine, utilized to hold the set of instructions and/or the data used in the invention may take on any of a variety of physical forms or transmissions, for example.
Illustratively, the medium may be in the form of paper, paper transparencies, a compact disk, a DVD, an integrated circuit, a hard disk, a floppy disk, an optical disk, a magnetic tape, a RAM, a ROM, a PROM, an EPROM, a wire, a cable, a fiber, a communications channel, a satellite transmission, a memory card, a SIM card, or other remote transmission, as well as any other medium or source of data that may be read by the processors of the invention.
[00174] Further, the memory or memories used in the processing machine that implements the invention may be in any of a wide variety of forms to allow the memory to hold instructions, data, or other information, as is desired. Thus, the memory might be in the form of a database to hold data. The database might use any desired arrangement of files such as a flat file arrangement or a relational database arrangement, for example.
[00175] In the system and method of the invention, a variety of "user interfaces" may be utilized to allow a user to interface with the processing machine or machines that are used to implement the invention. As used herein, a user interface includes any hardware, software, or combination of hardware and software used by the processing machine that allows a user to interact with the processing machine. A user interface may be in the form of a dialogue screen for example. A user interface may also include any of a mouse, touch screen, keyboard, keypad, voice reader, voice recognizer, dialogue screen, menu box, list, checkbox, toggle switch, a pushbutton or any other device that allows a user to receive information regarding the operation of the processing machine as it processes a set of instructions and/or provides the processing machine with information. Accordingly, the user interface is any device that provides communication between a user and a processing machine. The information provided by the user to the processing machine through the user interface may be in the form of a command, a selection of data, or some other input, for example.
[00176] As discussed above, a user interface is utilized by the processing machine that performs a set of instructions such that the processing machine processes data for a user. The user interface is typically used by the processing machine for interacting with a user either to convey information or receive information from the user. However, it should be appreciated thai in accordance with some embodiments of the system and method of the invention, it is not necessary that a human user actually interact with a user interface used by the processing machine of the invention. Rather, it is also contemplated that the user interface of the invention might interact, i.e., convey and receive information, with another processing machine, rather than a human user. Accordingly, the other processing machine might be characterized as a user. Further, it is contemplated that a user interface utilized in the system and method of the invention may interact partially with another processing machine or processing machines, while also interacting partially with a human user.
[00177] It will be readily understood by those persons skilled in the art that the present invention is susceptible to broad utility and application. Many embodiments and adaptations of the present invention other than those herein described, as well as many variations, modifications and equivalent arrangements, will be apparent from or reasonably suggested by the present invention and foregoing description thereof, without departing from the substance or scope of the invention.
[00178] Accordingly, while the present invention has been described here in detail in relation to its exemplary embodiments, it is to be understood that this disclosure is only illustrative and exemplaiy of the present invention and is made to provide an enabling disclosure of the invention. Accordingly, the foregoing disclosure is not intended to be construed or to limit the present invention or otherwise to exclude any other such embodiments, adaptations, variations, modifications or equivalent arrangements.

Claims

CLAIMS What s claimed is:
1. A method for presenting streaming data, comprising:
receiving, at web services layer for a server comprising at least one computer processor, a query from a client, wherein the query comprises a plurality of parameters;
a data caching layer for the server receiving streaming data from at least one predefined streaming data source;
the data caching layer conflating the streaming data for each of the plurality of parameters;
the data caching layer aggregating the conflated data:
the data caching layer generating a snapshot of the conflated data by simultaneously running the query against the conflated data; and
outputting the snapshot to the client.
2. The method of claim 1, wherein the parameters comprise a specific descriptor for at least one of a security and an investment.
3. The method of claim 1 , wherein the query further comprises an identification of a streaming data source.
4. The method of claim 1, wherein the streaming data comprises market data.
5. The method of claim 1 , wherein the web services layer outputs the snapshot that is delayed by a predetermined amount of time, wherein the period of time is based on one or more rale associated with the streaming data.
6. The method of claim 1, further comprising:
an entitlement services layer for the server verifying that the client is authorized to access information responsive to the query.
7. The method of claim 1 , further comprising:
authenticating the client based on at least one client credential that is received from the client.
8. The method of claim 2, wherein the snapshot is accurate for the at least one of the security and the investment for a certain time.
9. The method of claim 2, wherein the snapshot comprises a proper state for the at least one of the security and the investment for a certain time.
10. A system for presenting streaming data, comprising:
a plurality of streaming data sources;
a data loader for each streaming data source, the data loader receiving streaming data from the streaming data source;
a data caching layer the receives the streaming data from the data loaders; and
a web services layer comprising at least one computer processor in communication with the data caching layer;
wherein: the web services layer receiving a query from a client, wherein the query comprises a plurality of parameters;
the data caching layer conflates the streaming data for each of the plurality of parameters;
the data caching layer aggregating the conflated data;
the data caching layer generating a snapshot of the conflated data by simultaneously running the query against the conflated data; and
the snapshot is output to the client.
1 1. The system of claim 10, wherein the parameters comprise a specific descriptor for at least one of a security and an investment.
12. The system of claim 10, wherein the query further comprises an identification of a streaming data source.
13. The system claim 10, wherein the streaming data comprises market data.
14. The system of claim 10, wherein the web services layer outputs the snapshot that is delayed by a predetermined amount of time.
15. The system of claim 14, wherein the period of time is based on one or more rules associated with the streaming data.
16. The system of claim 10, wherein the query is received from at least one of a cloud application and a local application.
17. The system of claim 10, further comprising:
an entitlement services layer for the server verifying that the client is authorized to access information responsive to the query.
18. The system of claim 17, wherein the entitlement services layer further authenticates the client based on at least one client credential that is received from the client.
19. The system of claim 1 1 , wherein the snapshot is accurate for the at least one of the security and the investment for a certain time.
20. The system of claim 11 , wherein the snapshot comprises a proper state for the at least one of the security and the investment for a certain time.
PCT/US2017/066068 2016-12-14 2017-12-13 Systems and methods for aggregating, filtering, and presenting streaming data WO2018112023A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
AU2017378245A AU2017378245B2 (en) 2016-12-14 2017-12-13 Systems and methods for aggregating, filtering, and presenting streaming data
EP17881926.4A EP3555741A4 (en) 2016-12-14 2017-12-13 Systems and methods for aggregating, filtering, and presenting streaming data
CN201780085598.0A CN110249322B (en) 2016-12-14 2017-12-13 System and method for aggregating, filtering, and presenting streaming data
JP2019531996A JP7048614B2 (en) 2016-12-14 2017-12-13 Systems and methods for aggregating, filtering and presenting streaming data

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US15/378,501 2016-12-14
US15/378,501 US10657137B2 (en) 2016-12-14 2016-12-14 Systems and methods for aggregating, filtering, and presenting streaming data
US201762534749P 2017-07-20 2017-07-20
US62/534,749 2017-07-20

Publications (1)

Publication Number Publication Date
WO2018112023A1 true WO2018112023A1 (en) 2018-06-21

Family

ID=62559201

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2017/066068 WO2018112023A1 (en) 2016-12-14 2017-12-13 Systems and methods for aggregating, filtering, and presenting streaming data

Country Status (5)

Country Link
EP (1) EP3555741A4 (en)
JP (1) JP7048614B2 (en)
CN (1) CN110249322B (en)
AU (1) AU2017378245B2 (en)
WO (1) WO2018112023A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110187780A (en) * 2019-06-10 2019-08-30 北京百度网讯科技有限公司 Long text prediction technique, device, equipment and storage medium
EP4088248A4 (en) * 2020-12-10 2024-02-07 JPMorgan Chase Bank, N.A. System and method for cloud-first streaming and market data utility
US11922437B2 (en) 2018-04-12 2024-03-05 Jpmorgan Chase Bank, N.A. System and method for implementing a market data hub

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113779083A (en) * 2021-01-29 2021-12-10 北京京东拓先科技有限公司 Data processing method, device, server and storage medium
US20220358499A1 (en) * 2021-05-07 2022-11-10 Jpmorgan Chase Bank, N.A. Method and system for autonomous portfolio platform management

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090292677A1 (en) * 2008-02-15 2009-11-26 Wordstream, Inc. Integrated web analytics and actionable workbench tools for search engine optimization and marketing
US20140157370A1 (en) * 2012-05-22 2014-06-05 Hasso-Plattner-Institu für Softwaresystemtechnik GmbH Transparent Control of Access Invoking Real-time Analysis of the Query History
US20140351233A1 (en) * 2013-05-24 2014-11-27 Software AG USA Inc. System and method for continuous analytics run against a combination of static and real-time data

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6697806B1 (en) * 2000-04-24 2004-02-24 Sprint Communications Company, L.P. Access network authorization
JP2004521530A (en) 2000-10-25 2004-07-15 トムソン・フィナンシャル・インコーポレイテッド E-commerce system
US9870415B2 (en) * 2013-09-18 2018-01-16 Quintiles Ims Incorporated System and method for fast query response

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090292677A1 (en) * 2008-02-15 2009-11-26 Wordstream, Inc. Integrated web analytics and actionable workbench tools for search engine optimization and marketing
US20140157370A1 (en) * 2012-05-22 2014-06-05 Hasso-Plattner-Institu für Softwaresystemtechnik GmbH Transparent Control of Access Invoking Real-time Analysis of the Query History
US20140351233A1 (en) * 2013-05-24 2014-11-27 Software AG USA Inc. System and method for continuous analytics run against a combination of static and real-time data

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3555741A4 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11922437B2 (en) 2018-04-12 2024-03-05 Jpmorgan Chase Bank, N.A. System and method for implementing a market data hub
CN110187780A (en) * 2019-06-10 2019-08-30 北京百度网讯科技有限公司 Long text prediction technique, device, equipment and storage medium
CN110187780B (en) * 2019-06-10 2023-07-21 北京百度网讯科技有限公司 Long text prediction method, long text prediction device, long text prediction equipment and storage medium
EP4088248A4 (en) * 2020-12-10 2024-02-07 JPMorgan Chase Bank, N.A. System and method for cloud-first streaming and market data utility

Also Published As

Publication number Publication date
EP3555741A1 (en) 2019-10-23
AU2017378245A1 (en) 2019-07-11
JP2020504867A (en) 2020-02-13
AU2017378245B2 (en) 2022-10-13
CN110249322B (en) 2023-06-20
CN110249322A (en) 2019-09-17
JP7048614B2 (en) 2022-04-05
EP3555741A4 (en) 2020-05-06

Similar Documents

Publication Publication Date Title
AU2017378245B2 (en) Systems and methods for aggregating, filtering, and presenting streaming data
US10482534B2 (en) Method and system for aggregating and managing data from disparate sources in consolidated storage
US20170293980A1 (en) System and method for managing processing resources of a computing system
US8924274B2 (en) For and method of providing portfolio risk information to investors without revealing position information
US11949757B2 (en) Method, apparatus and system for subscription management
US12014290B2 (en) Projecting data trends using customized modeling
US10474692B2 (en) Data conversion and distribution systems
WO2018140484A1 (en) Systems and methods for intraday facility monitoring
US10915968B1 (en) System and method for proactively managing alerts
US20220207606A1 (en) Prediction of future occurrences of events using adaptively trained artificial-intelligence processes
US10007950B2 (en) Integrating multiple trading platforms with a central trade processing system
US20200042164A1 (en) System and Method for a Mobile Computing Device Having a User Interface and Options Selection in the User Interface
US11922437B2 (en) System and method for implementing a market data hub
US20110202452A1 (en) System and method for providing borrow coverage services to short sell securities
US20240338703A1 (en) Method, apparatus, system, and non-transitory computer readable medium for preserving trading time series
US20120323817A1 (en) Systems and Methods for Implementing a Separately Managed Account Platform
CN117078340A (en) Order processing method and device, electronic equipment and storage medium
US20150317738A1 (en) Computerized method and system for secure communication, and method and system for matching customers with options for investment
WO2014008573A1 (en) Banking server, networked system and method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17881926

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2019531996

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2017378245

Country of ref document: AU

Date of ref document: 20171213

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2017881926

Country of ref document: EP

Effective date: 20190715