BACKGROUND OF THE INVENTION
1 . Field of the Invention
[0001] The present disclosure generally relates to systems and methods for aggregating, filtering, and presenting streaming data.
2. Description of The Related Art
[0002] Both public and private clouds have become increasingly popular computing environments. These environments, however, do not support streaming data due to variable latency between the host system and the remote operating system. In addition, the cost of pushing terabytes of streaming market data into various clouds incurs very high "transit" fees from providers, decreasing the value proposition of using clouds for data- driven applications.
SUMMARY OF THE INVENTION
[0003] Systems and methods for aggregating, filtering, and presenting streaming data are disclosed. In one embodiment, a method for presenting streaming data may include (1 ) receiving, at web services layer for a server comprising at least one computer processor, a query from a client, wherein the query comprises a plurality of parameters; (2) a data caching layer for the server receiving streaming data from at least one predefined streaming data source; (3) the data caching layer conflating the streaming data for each of the plurality of parameters; (4) the data caching layer aggregating the conflated data; (5) the data caching layer generating a snapshot of the
conflated data by simultaneously running the query against the conflated data; and (6) outputting the snapshot to the client.
[0004] In one embodiment, the parameters may include a specific descriptor for at least one of a security and an investment.
[0005] In one embodiment, the query may further include an identification of a streaming data source.
[0006] In one embodiment, the streaming data may include market data.
[0007] In one embodiment, the web services layer may output the snapshot that is delayed by a predetermined amount of time, wherein the period of time is based on one or more rale associated with the streaming data.
[0008] In one embodiment, the method may further include an entitlement services layer for the server verifying that the client is authorized to access information responsive to the query.
[0009] In one embodiment, the method may further include
authenticating the client based on at least one client credential that is received from the client.
[0010] In one embodiment, the snapshot may be accurate for the at least one of the security and the investment for a certain time.
[0011] In one embodiment, the snapshot may include a proper state for the at least one of the security and the investment for a certain time.
[0012] According to another embodiment, a system for presenting streaming data may include a plurality of streaming data sources; a data
loader for each streaming data source, the data loader receiving streaming data from the streaming data source; a data caching layer the receives the streaming data from the data loaders; and a web services layer comprising at least one computer processor in communication with the data caching layer. The web services layer may receive a query from a client, wherein the query comprises a plurality of parameters; the data caching layer may conflate the streaming data for each of the plurality of parameters; the data caching layer may aggregate the conflated data; the data caching layer may generate a snapshot of the conflated data by simultaneously running the query against the conflated data; and the snapshot may be output to the client,
[0013] In one embodiment, the parameters may include a specific descriptor for at least one of a security and an investment.
[0014] In one embodiment, the query may include an identification of a streaming data source.
[0015] In one embodiment, the streaming data may include market data.
[0016] In one embodiment, the web services layer may output the snapshot that is delayed by a predetermined amount of time.
[0017] In one embodiment, the period of time may be based on one or more rules associated with the streaming data.
[0018] In one embodiment, the query may be received from at least one of a cloud application and a local application.
[0019] In one embodiment, the system may further include an entitlement services layer for the server verifying that the client is authorized to access information responsive to the query. The entitlement services layer
may further authenticate the client based on at least one client credential that is received from the client.
[0020] In one embodiment, the snapshot may be accurate for the at least one of the security and the investment for a certain time,
[0021] In one embodiment, the snapshot may represent a proper state for the at least one of the security and the investment for a certain time.
BRIEF DESCRIPTION OF THE DRAWINGS
[0022] For a more complete understanding of the present invention, the objects and advantages thereof, reference is now made to the following descriptions taken in connection with the accompanying drawings in which:
[0023] Figure 1 depicts a system for aggregating, filtering, and presenting streaming data according to one embodiment;
[0024] Figure 2 depicts a method for aggregating, filtering, and presenting streaming data according to one embodiment;
[0025] Figure 3 depicts a strategic flow is provided according to one embodiment; and
[0026] Figure 4 depicts an exemplary process flow according to one embodiment.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
[0027] Several embodiments of the present invention and their advantages may be understood by referring to Figures 1-4.
[0028] Embodiments are directed to systems and methods for aggregating, filtering, and presenting streaming data.
[0029] In one embodiment, data from internal and external sources may be pushed into a high-speed data aggregation engine. Clients may make a simple web services request, specifying their credentials, source of information, and list of requested information. For example, in a financial institution, clients may submit a list of securities and fields. The client then receives a single response with the information that the client needs. This reduces or eliminates high-speed stream management that all applications have to support as well as the required infrastructure and development.
[0030] In one embodiment, any suitable aggregation engine may be used as is necessary and/or desired. In one embodiment, the aggregation engine may permit Structured Query Language ("SQL")-style content filtering on individual fields within data payloads, allowing for near realtime analytics on a wide array of products.
[0031] Embodiments may provide some or all of the following: (1) increased price advantage of Cloud-based applications through lower data transit costs; (2) large scale reduction of technology costs through decreased infrastructure requirements and shorter development cycles for data-driven applications; (3) more fungible developer workforce due to simplification of API and integration with spreadsheets; (4) advanced content filtering across disparate sources permits development of real-time application features that were cost prohibitive in the past; (5) facilitated distributed analytics through merging of external and internal data; and (6) the ability to age, or delay, the delivery of real-time data for individuals that do not need the data in realtime to reduce the cost of that data. Other benefits may also be provided.
[0032] In the financial industry, because market data and messaging involves a large amount of data, clients are expected to do significant work associated with dealing with payload coming out of the ticker plants. They
often have to deal with hardware and software challenges related to large volumes of information, conflate it in order for users to derive benefits, aggregate across many threads and systems to support portfolio
management, etc. Embodiments address these and other challenges while giving end users and developers an interface with filtering.
[0033] Referring to Figure 1, a system for aggregating, filtering, and presenting streaming data is disclosed according to one embodiment.
System 100 may include a plurality of client access points 110, web services layer 120, data caching layer 130, data loaders 140l5 1402, ... 140n, and data sources 150i, 1502, ... 150n. Data sources 150 may receive data from one or more streaming data source, such as internal and external market data, industry news, etc.
[0034] Although three data loaders 140 and data sources 150 are illustrated in Figure 1 , it should be recognized that a greater or fewer number of data loaders 140 and data sources 150 may be provided as is necessary and/or desired.
[0035] Data loaders 140 may receive data from data sources 150. In one embodiment, each data loader 140 may "feed" streaming data from one or more data source 150 into data caching layer 130. Each data loader 140 may further be in communication with web services layer 120. For example, each data loader 140 may communicate with control plane sendee 126, which may instruct one or more data loader 140 to create a new subscription, retry a failure, etc.
[0036] Data caching layer 130 may service requests from, web services layer 120. In one embodiment, data caching layer 130 may function as both a database and a message bus.
[0037] In one embodiment, data caching layer 130 may also filter the data and may provide a "snapshot" of requested data.
[0038] An example of a suitable data caching layer 130 is the
Advanced Message Processing System (AMPS) from 60East Technologies.
[0039] Web services layer 120 may interface with clients using one or more client access points 1 10. In one embodiment, web services layer may provide sendees, such as configuration management sendee 122, monitoring service 124, control plane layer 126, and entitlement services 128. Other services may be provided as is necessary and/or desired.
[0040] In one embodiment, configuration management service 122 may provide configuration data for one or more data loader 140, such as what data sources 150 to connect to, how to connect to data sources 150, etc. In one embodiment, configuration management service 122 may provide run-time configuration information, for example, connection information (e.g., hosnport) for data loaders 140, web services layer 120, entitlement sendees 128, what data feeds to load, what web connections to support (http or https), SSL certificate locations, connection and thread pool sizing, logging intervals, environment (dev/test/prod), etc.
[0041] In one embodiment, monitoring service 124 may monitor a status of one or more data source 150, and may re-route requests if a data source becomes unavailable.
[0042] Control plane sendee 150 may interface with one or more data loaders 140. In one embodiment, control plane service 150 may pass directives between services, allowing both automated instructions (e.g., monitoring sendee 124 notices a failure of a data loader and instructs web
service layer 120 to switch servers within data aggregation layer 130), or manual directives (e.g., an operations team wants to take a service offline).
[0043] In one embodiment, entitlement services 128 may verify that the client is allowed to access the data that is requested.
[0044] In one embodiment, client access points 1 10 may include, for example, cloud applications, local application, user interfaces, APIs, etc. In one embodiment, client access point 1 10 may provide, for example, a spreadsheet as its output to a client.
[0045] In one embodiment, system 100 may include the ability to provide a "subscription" for a client to receive updates. In one embodiment, updates to a complex query may be smaller in size than re-submitting the query, and receiving a full result set several times a second.
[0046] Referring to Figure 2, a method for aggregating, filtering, and presenting streaming data is disclosed according to one embodiment.
[0047] In step 210, a client may request information of interest, via an interface, such as a cloud application, a local application, an upload, etc. In one embodiment, the client may also provide credentials and identify the source of information and a list of requested information. In another embodiment, machine learning may be used to identify the sources based on past, queries,
[0048] In one embodiment, the client may use a web service protocol to request the information. The web service protocol may use, for example, SQL-like syntax for filtering information. An example query that may be submitted is "(/BID / /ASK >= (0.05 * /TRDPRC....1) )", which means
"where value of BID divided by value of ASK is greater or equal to 5% of the value of "TRDPRC_1."
[0049] In a financial system, the query may include an identification of a stock, industry, an field/area of interest, etc. The query may be used in filters.
[0050] In one embodiment, the request may be a subscription (e.g., to a stock, industry, etc.). In one embodiment, the request may specify a period or condition for each snapshot to be provided.
[0051] In step 220, the client may be authorized to access the requested data. In one embodiment, this may involve an entitlement check to determine if the client is permitted access to the requested information,
[0052] In one embodiment, the client may be further authenticated based on credentials that may be provided as part of the request.
[0053] In step 230, the request may be provided to the data caching layer, which, in step 240, may filter the data received from streaming data sources. The data caching layer may also "snapshot" the data, that is, provide a data state at a particular instance in time.
[0054] In one embodiment, business logic may reduce the number of fields of data that are presented to the client.
[0055] In step 250, the filtered data may be conflated. For example, data from more than one source that has been filtered may be aggregated. For example, in a minute, a stock may change in price from 70.00 to 70.05, back to 69.90 and up to 70.01. With a conflated source, client would only receive 70.01 since that was the proper state of the price at the end of the minute. This value, however, will not be the last value as it is not always the
proper data point, in particular for a dataset with multiple fields, such as Bid and Ask,
[0056] In step 260, a data snapshot of the data may be generated. In one embodiment, a snapshot may be for a collection of securities, and may be accurate across that collection at that point of time. For example, the query may be r n simultaneously across the data thai has been aggregated (e.g., the securities of interest) rather than sequentially.
[0057] In step 270, the snapshot may be output to the client via a web application, local application, API, spreadsheet, etc.
[0058] In one embodiment, the delivery of the snapshot may be delayed by a predetermined amount of time in order to "age" the data. For example, the timing for data deliver}' may determine whether the data can be delivered for free, or if there is a fee for the data. Not all clients require realtime data; for example, an investment manager may not need real-time data when meeting prospective clients.
[0059] Thus, in one embodiment, the system may determine how long to delay the output of the snapshot, which may be based on rules associated with the data and/or the data source, and may delay the snapshot by at least this amount of time. In one embodiment, an auditable trail of the age of the data in the snapshot may be maintained.
[0060] Embodiments disclosed herein may separate the mechanics of the various processes performed by multiple lines of business from content that is relevant to solving the problem. That content, such as various lots in clients' portfolios, changes resulting from trades, money transfers, additions, withdrawals, etc. may be treated as generic internally published data, anonymized, and provided to one or more system/method for aggregating,
filtering, and presenting streaming data. Examples of such a system/method are disclosed in U.S. Patent Application Ser. No. 15/378,501, the disclosure of which is hereby incorporated, by reference, in its entirety.
[0061] In embodiments, market data from various venues may be conflated inside of one or more system/method for aggregating, filtering, and presenting streaming data, and may be merged with portfolio information. As data goes through several merging stages, each lot becomes priced against real-time market data, then aggregated into a real-time portfolio views.
[0062] In embodiments, customers may use a custom rales engine to subscribe to "alerts" relevant to the business. For example, the system may send a notice to the client when a particular stock reaches a price threshold, when a concentration risk of a particular holding breaches a predefined level, when a valuation of the lot or portfolio rises/falls by a certain percentage, etc.
[0063] Referring to Figure 3, a strategic flow is provided according to one embodiment. Strategic flow mav include customers, a financial institution interface, a gateway service, one or more graphical user interface, a perimeter gateway, a Complex Event Processor (CEP) engine, an internal data platform (e.g., an internal data platform that may provide secure reliable messaging between publishes and consumers. Access may be provided using Open MAM A API), one or more market, data feeds (e.g., NASDAQ Basic, which carries NASDAQ and N YSE market data. A commercial feed handler may be used; OPRA, Pink Sheets, Trade Web, etc.), an Anonymizer a data injector, a security master, a portfolio management system, intraday activity monitors (e.g., intraday activity, order management systems, and cash flow systems), an alerting engine, and a notification platform.
[0064] In one embodiment, the CEP engine may process data for applications that require market data and one or more of the following: (1) are in configurations that cannot support high-rate streaming data; (2) only require prices a few times a day but need those prices to have a very small hysteresis; (3) need to submit a large number of symbols at a single time (as opposed to streaming subscriptions that each take a single symbol); (4) aggregation of market and analytical data from a number of sources (e.g., Reuters RMDS, Direct Feeds, Internal Data, LOB market analytics); or (5) need the ability to filter the data on dynamic queries taking into account the values of one or more fields, or can use market data that has been delayed a determined time to lower its cost.
[0065] In one embodiment, the anonymizer component may strip identity of clients from portfolio contents prior to injection into the CEP engine. A reverse process allows processes to look up the ID associated with a specific client's portfolio they are interested in.
[0066] The data injector may provide a single interface to one or more technical assets. It may map reference data (e.g., symbology) between systems and communicate with both and an Internal Data Platform (e.g., to send portfolio information) and the CEP engine (e.g. ,to ensure instruments of interest are on the watch list).
[0067] In one embodiment, the portfolio management system may be the source of portfolio content data for clients' holdings. It may publish daily refreshes of portfolios of interest, publish portfolios on demand, as well as maintain the "official" start of day prices for all lots/portfolios.
[0068] In one embodiment, intraday activity monitors may monitor activity such as that originating from Order Management Systems, Cash
Flow engines, etc. It may publish updates into the CEP engine as portfolio changes since start of day to reflect changes in valuation in real-time.
[0069] The alerting engine may translate client directions from the financial institution website into subscriptions against the CEP engine and other back-end systems. Persists alerts across days, listens for signals from the CEP engine that conditions were satisfied, performs desired actions, disables aciioried alerts, etc.
[0070] Figure 4 depicts an exemplary process flow according to one embodiment. In one embodiment, the data source layer may receive market data (e.g., market data streams) and may provide the data to the data loader layer, where it may be loaded into the data caching layer using one or more data loader. The CEP Engine may access the data from the data caching layer, and may provide configuration management, data filter processing, monitoring services, control plane services, entitlement services, etc. In one embodiment, the CEP Engine may provide services to customers via the customer access layer.
[0071] EXAMPLE
[0072] A non-limiting example is provided below.
[0073] Data Set 1: Equity instrument from direct feed: [0074] <snapSymbol>JPM.N</snapSymbol> [0075] < snapS tatu s>ok</s n apS tatu s>
[0076] <wBidPrice>87 ,790Q00</wBidPrice>
[0077] <wBidSize> 18</wBidSize>
[0078] <wAskPrice>87.80Q00Q</wAskPrice>
[0079] <wAskSize>4</wAskSize>
[0080] <wInstrumentType>Stock</wIn strumentType>
[0081] < wOpenPrice>90.360000</wOpenPrice>
[0082] <wAdjPrevClose>90.070000</wAdjPrevClose>
[0083] <wPrevClosePrice>90.070000</wPrevClosePrice>
[0084] <wBidOpen>90.340000</wBidOpen>
[0085] <wAskOpen>90.400000</wAskOpen>
[0086] The instrument price is an average of wBidPrice and wAskPrice.
[0087] Data Set 2: Commodity future from direct feed:
[0088] <snapSymbol>CLJ7.NYM</snapSymbol>
[0089] < snapS tatu s>ok</ s napS tatu s>
[0090] <wTssueSymbol>CLJ7.NYM</wIssueSymbol>
[0091] <wBidHigh>48.620000</wBidHigh>
[0092] <wAskLow>47.400000</wAskLow>
[0093] < wAskPrice>47.4100G0</w AskPrice>
[0094] < wA skSize>5 </w AskSizo
< wB idPrice>47.390000</wB idPrice>
[0096] <wBidSize>52</wBidSize>
[0097] <wB idLo w>47.340000</wB idLo w>
[0098] <wAskHigh>48.750000</wAskHigh>
[0099] < wOpenPrice>47.930000</wOpenPrice>
[00100] <wOpenTime>2017-03-21
17:45:27,229603</wOpenTime>
[00101] < CfiCode>FCMXSX</wCfiCode>
[00102] <wCurrency>USD</wCurrency>
[00103] <wEntryStatus>0</wEntryStatus>
[00104] Same rule for price applies
[00105] Data Set 3: Bond instrument from vendor feed:
[00106] <snapSymbol>ISIN_USN82008AK46„MODEL.FUSE</ snapSymbol>
[00107] <snapStatus>ok</snapStatus>
[00108] <CLJENT_OFFER>0.000000</CLIENT_OFFER>
[00109] <CLIENT_BID>0.000000</CLJENT_BID>
[00110] <POSITION>0,000000</POSITION>
[00111] <COUPON_RATE>2.000000</COUPON„RATE>
[00112] <STRING2>HG -
Industrials/Defense/Chemicals</STRING2>
<MKT_PRICE_2>1000000.000000</MKT_PRICE_2>
<MKT .PRICE,J>1000000.000000</MKT .PRICE...1>
[00115] <MATURD ATE> 15 Sep 2023</MATURDATE>
[00116] <ISIN...,BENCHMARK>US912828 W556</1SIN.. BENC
HMARK>
7] <CURVE_TYPE>EMEA< CURVE_TYPE> [00118] <UPDATE_TM> 12:31 : 15.000</UPDATE_TM>
[00119] <lTEM...ID/>
[00120] Price is in field MKT_PRICE_1
Data sets do not need to exist in the same CEP instance. For example, Data sets 1 and 3 may exist on the same CEP instance where the portfolio view is created. Data set 2 may be on another instance due to volume demands of ingestion.
[00122] An example portfolio design is provided. The portfolio may be stored in its own table, called Portfolios:
[00123] :
[00124] {
[00125] "portfolio" : "12345", "version" : 3.2.6,
[00127] "holdings" : [
[00128] { "source" : "Data Set 1 ", "security" : "JPM.N", "weight"
: 150 },
[00129] { "source" : "Data Set 2", "security" : "CLJ7.NYM", " weight" : 200 ) ,
[00130] { "source" : "Data Set 3", "security" :
"ISIN_USN82008AK46_MODEL,FUSE", " weight" : 2 } j
[00133] For example, portfolio pricing may se the following pseud o code:
[00134] for specific portfolio ID
[00135] for each asset identifier
[00136] total price ~ total price + (asset weight *
as setjpirce [as set identifier J )
[00137] return total price
[00138] In other words, a single value representing the real-time price of the portfolio, which will change as data changes, may be returned.
[00139] In one embodiment, there may be four layers of tables defined for both instances,
[00140] The first layer may receive data from market data publishers. For example, for both PortfolioPricing and CommodityFutures, there may be one table per data feed/source. On PortfolioPricing, these may be named
"Equity" and "Bond" and on Commodity Futures, it may be named "Comm". These tables are not transaction logged nor replicated and the "shape" of the data in these tables is whatever that data source defines.
[00141] At the next layer is a replicated table that may be called
"AUPrices," which may act as the normalized union of price data from all sources. Data may be normalized and published to AUPrices from the Comm, Equity and Bond tables. In embodiments, this information may be conflated in order to reduce the maximum data rate.
[00142] The first step in the chain of actions may be to parse the underlying data and extract values of interest into variables. Extraction of values may be specified as a projection expression. For example when processing a message from Equity, the line
<Value>price=(/price/wBidPrice+/price/wAskPrice) / 2</Value> computes the average of the /wBidPrice and /wAskPrice and places the result in the variable { {price} } .
[00143] The next step of the action is to publish a new message containing the { {symbol] } and the normalized { {price} ), along with a "source" field that indicates the source of the pricing data. This may be published to the AUPrices table and may be formatted as JSON.
[00144] At this point, the incoming market data has been normalized and transformed into a single AUPrices table. Conflation and replication may be used to minimize the total volume of messages that the
PortfolioPricing instance has to deal with, while giving PortfolioPricing a complete view of all of the security prices needed to price portfolios.
[00145] Next, the portfolio may be modeled as a table "Portfolio", with one row per portfolio per security. Each message in Portfolio may have a
portfolio ID, a security, and the weight of that security in that portfolio. This table is essentially the input data for the portfolios of interest, and the system may dynamically re-compute portfolio prices as the contents of the portfolio are adjusted, added, and/or removed.
[00146] This computation is done in the next layer, the aggregations. This may be modeled as two Views. The first view is a join between
Portfolio and AlTPrices, called "PricedPortfolio." There may be one message here for ever message in Portfolio, but because of the Join, current price for each security may be included.
[00147] A top PricedPortfolio may be the 2nd view that computes the portfolio price, "PricedPortfolioAgg." This is in essence a translation of the pseudocode (above) over to a view, containing one message per portfolio with the portfolio ID, the count of securities in the portfolio, and the SUM of the /price's/weight of the securities in the portfolio.
[00148] Thus, in embodiments, the system may receive portfolio data (e.g., by push), and then be receive queries/subscriptions to PricedPortfio and PricedPortfolioAgg to see the dynamically updated portfolio prices as the underlying market data changes.
[00149] In embodiments, each portfolio's official start of day valuation may be submitted as part of a daily load into a PortfolioRef table. This permits the PricedPortfolioAgg view to include a calculation of an unrealized start of day gain/loss by comparing current valuation and the reference data submitted into the PortfolioRef table. In one embodiment, a percentage change from start of day portfolio valuation may be calculated.
[00150] In embodiments, an alerting engine may set up subscriptions into the CEP engine that maps human-identifiable concepts to an underlying
data structure. Arithmetic expressions and formulas in SQL language may he used to define when the alerting engine should receive a notification of an event. Upon data being received on a specific subscription, the alerting engine will map it back to a desired action (e.g., sending a message, such as an email or SMS, popping up a window on Advisor's user interface, etc.) and may execute that action. It may be further be configured to remain as an active alert, or disable itself by dropping a subscription.
[00151] Non-limiting examples of alerts include portfolio gain/loss above/below dollar amount or percentage, specific holding's value exceeds a specific threshold, etc.
[00152] Advisors, representing multiple clients, may maintain alerts covering multiple portfolios of interest for a holistic view of their clients' holdings.
[00153] At an organizational level, all holdings may be analyzed holistically in a CEP engine to facilitate hedging, cross-account stock loan, risk management, etc.
[00154] In one embodiment, cash may be represented in a portfolio as a number of units of underlying currency, such as the U.S. Dollar. Market data for the currency may be static (such as for U.S. Dollar for Dollar- denominated accounts), or represent current foreign exchange rates.
[00155] In another embodiment, a similar mechanism may be used for hypothetical modeling of portfolio changes. A separate CEP engine or a side table may be used to store the final representation of portfolio's hypothetical contents, while the same aggregation mechanisms can be used to price the portfolio.
[00156] In one embodiment, a separate table in the CEP engine may store stock to sector relationships that may be used to build dynamic risk models to help customers manage risk. Multiple risk models may be defined to facilitate different risk appetites.
[00157] In one embodiment, a separate subscription may be set up on the table containing aggregated views that may then be extracted and preserved in a database. This historical record may be used as an audit trail, or to train machine learning systems to generate automated investment advice to customers and advisors.
[00158] In one embodiment, the CEP engine may be used for real-time margin management, since it is aware of the current valuation status of portfolio lots as well as availability of monetary instruments.
[00159] Although several embodiments have been disclosed, it should be recognized that these embodiments are not exclusive to each other.
[00160] Hereinafter, general aspects of implementation of the systems and methods of the invention will be described.
[00161] The system of the invention or portions of the system of the invention may be in the form of a "processing machine," such as a general purpose computer, for example. As used herein, the term "processing machine" is to be understood to include at least one processor that uses at least one memory. The at least one memory stores a set of instructions. The instructions may be either permanently or temporarily stored in the memory or memories of the processing machine. The processor executes the instructions that are stored in the memory or memories in order to process data. The set of instructions may include various instructions that perform a particular task or tasks, such as those tasks described above. Such a set of
instructions for performing a particular task may be characterized as a program, software program, or simply software,
[00162] In one embodiment, the processing machine may be a specialized processor,
[00163] As noted above, the processing machine executes the instructions that are stored in the memory or memories to process data. This processing of data may be in response to commands by a user or users of the processing machine, in response to previous processing, in response to a request by another processing machine and/or any other input, for example.
[00164] As noted above, the processing machine used to implement the invention may be a general purpose computer. However, the processing- machine described above may also utilize any of a wide variety of other technologies including a special purpose computer, a computer system including, for example, a microcomputer, mini-computer or mainframe, a programmed microprocessor, a micro-controller, a peripheral integrated circuit element, a CSIC (Customer Specific Integrated Circuit) or ASIC (Application Specific Integrated Circuit) or other integrated circuit, a logic circuit, a digital signal processor, a programmable logic device such as a FPGA, PLD, PL A or PAL, or any other device or arrangement of devices that is capable of implementing the steps of the processes of the invention,
[00165] The processing machine used to implement the invention may utilize a suitable operating system. Thus, embodiments of the invention may include a processing machine running the iOS operating system, the OS X operating system, the Android operating system, the Microsoft Windows™ operating systems, the Unix operating system, the Linux operating system, the Xenix operating system, the IBM ATX™ operating system, the Hewlett-
Packard UX™ operating system, the Novell Netware™ operating system, the Sun Microsystems Solaris™ operating system, the OS/2™ operating- system, the BeOS™1 operating system, the Macintosh operating system, the Apache operating system, an OpenStep™ operating system or another operating system or platform.
[00166] It is appreciated that in order to practice the method of the invention as described above, it is not necessary that the processors and/or the memories of the processing machine be physically located in the same geographical place. That is, each of the processors and the memories used by the processing machine may be located in geographically distinct locations and connected so as to communicate in any suitable manner.
Additionally, it is appreciated that each of the processor and/or the memory may be composed of different physical pieces of equipment. Accordingly, it is not necessary that the processor be one single piece of equipment in one location and that the memory be another single piece of equipment in another location. That is, it is contemplated that the processor may be two pieces of equipment in two different physical locations. The two distinct pieces of equipment may be connected in any suitable manner. Additionally, the memory may include two or more portions of memory in two or more physical locations.
[00167] To explain further, processing, as described above, is performed by various components and various memories. However, it is appreciated that the processing performed by two distinct components as described above may, in accordance with a further embodiment of the invention, be performed by a single component. Further, the processing performed by one distinct component as described above may be performed by two distinct components, hi a similar manner, the memory storage performed by two
distinct memory portions as described above may, in accordance with a further embodiment of the invention, be performed by a single memory portion. Further, the memory storage performed by one distinct memory portion as described above may be performed by two memory portions,
[00168] Further, various technologies may be used to provide
communication between the various processors and/or memories, as well as to allow the processors and/or the memories of the invention to communicate with any other entity; i.e., so as to obtain further instructions or to access and use remote memory stores, for example. Such technologies used to provide such communication mieht include a network, the Internet. Intranet.
Extranet, LAN, an Ethernet, wireless communication via cell tower or satellite, or any client server system that provides communication, for example. Such communications technologies may use any suitable protocol such as TCP/IP, UDP, or OSI, for example.
[00169] As described above, a set of instructions may be used in the processing of the invention. The set of instructions may be in the form of a program or software. The software may be in the form of system software or application software, for example. The software might also be in the form of a collection of separate programs, a program module within a larger program, or a portion of a program module, for example. The software used might also include modular programming in the form of object oriented programming. The software tells the processing machine what to do with the data being processed.
[00170] Further, it is appreciated that the instructions or set of instructions used in the implementation and operation of the invention may be in a suitable form such that the processing machine may read the instructions. For example, the instructions that form a program may be in
the form of a suitable programming language, which is converted to machine language or object code to allow the processor or processors to read the instructions. That is, written lines of programming code or source code, in a particular programming language, are converted to machine language using a compiler, assembler or interpreter. The machine language is binary coded machine instructions that are specific to a particular type of processing machine, i.e., to a particular type of computer, for example. The computer understands the machine language.
[00171] Any suitable programming language may be used in accordance with the various embodiments of the invention. Illustratively, the
programming language used may include assembly language, Ada, APL, Basic, C, C++, COBOL, dBase, Forth, Fortran, Java, Modula-2, Pascal, Prolog, REXX, Visual Basic, and/or JavaScript, for example. Further, it is not necessary that a single type of instruction or single programming language be utilized in conjunction with the operation of the system and method of the invention. Rather, any number of different programming languages may be utilized as is necessary and/or desirable.
[00172] Also, the instructions and/or data used in the practice of the invention may utilize any compression or encryption technique or algorithm, as may be desired. An enciyption module might be used to encrypt data. Further, files or other data may be decrypted using a suitable decryption module, for example.
[00173] As described above, the invention may illustratively be embodied in the form of a processing machine, including a computer or computer system, for example, that includes at least one memory. It is to be appreciated that the set of instructions, i.e., the software for example, that enables the computer operating system to perform the operations described
above may be contained on any of a wide variety of media or medium, as desired. Further, the data that is processed by the set of instructions might also be contained on any of a wide variety of media or medium. That is, the particular medium, i.e., the memory in the processing machine, utilized to hold the set of instructions and/or the data used in the invention may take on any of a variety of physical forms or transmissions, for example.
Illustratively, the medium may be in the form of paper, paper transparencies, a compact disk, a DVD, an integrated circuit, a hard disk, a floppy disk, an optical disk, a magnetic tape, a RAM, a ROM, a PROM, an EPROM, a wire, a cable, a fiber, a communications channel, a satellite transmission, a memory card, a SIM card, or other remote transmission, as well as any other medium or source of data that may be read by the processors of the invention.
[00174] Further, the memory or memories used in the processing machine that implements the invention may be in any of a wide variety of forms to allow the memory to hold instructions, data, or other information, as is desired. Thus, the memory might be in the form of a database to hold data. The database might use any desired arrangement of files such as a flat file arrangement or a relational database arrangement, for example.
[00175] In the system and method of the invention, a variety of "user interfaces" may be utilized to allow a user to interface with the processing machine or machines that are used to implement the invention. As used herein, a user interface includes any hardware, software, or combination of hardware and software used by the processing machine that allows a user to interact with the processing machine. A user interface may be in the form of a dialogue screen for example. A user interface may also include any of a mouse, touch screen, keyboard, keypad, voice reader, voice recognizer,
dialogue screen, menu box, list, checkbox, toggle switch, a pushbutton or any other device that allows a user to receive information regarding the operation of the processing machine as it processes a set of instructions and/or provides the processing machine with information. Accordingly, the user interface is any device that provides communication between a user and a processing machine. The information provided by the user to the processing machine through the user interface may be in the form of a command, a selection of data, or some other input, for example.
[00176] As discussed above, a user interface is utilized by the processing machine that performs a set of instructions such that the processing machine processes data for a user. The user interface is typically used by the processing machine for interacting with a user either to convey information or receive information from the user. However, it should be appreciated thai in accordance with some embodiments of the system and method of the invention, it is not necessary that a human user actually interact with a user interface used by the processing machine of the invention. Rather, it is also contemplated that the user interface of the invention might interact, i.e., convey and receive information, with another processing machine, rather than a human user. Accordingly, the other processing machine might be characterized as a user. Further, it is contemplated that a user interface utilized in the system and method of the invention may interact partially with another processing machine or processing machines, while also interacting partially with a human user.
[00177] It will be readily understood by those persons skilled in the art that the present invention is susceptible to broad utility and application. Many embodiments and adaptations of the present invention other than those herein described, as well as many variations, modifications and equivalent
arrangements, will be apparent from or reasonably suggested by the present invention and foregoing description thereof, without departing from the substance or scope of the invention.
[00178] Accordingly, while the present invention has been described here in detail in relation to its exemplary embodiments, it is to be understood that this disclosure is only illustrative and exemplaiy of the present invention and is made to provide an enabling disclosure of the invention. Accordingly, the foregoing disclosure is not intended to be construed or to limit the present invention or otherwise to exclude any other such embodiments, adaptations, variations, modifications or equivalent arrangements.