WO2017171826A1 - Entropic classification of objects - Google Patents
Entropic classification of objects Download PDFInfo
- Publication number
- WO2017171826A1 WO2017171826A1 PCT/US2016/025466 US2016025466W WO2017171826A1 WO 2017171826 A1 WO2017171826 A1 WO 2017171826A1 US 2016025466 W US2016025466 W US 2016025466W WO 2017171826 A1 WO2017171826 A1 WO 2017171826A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- corpus
- gentropy
- document
- score
- entropy
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/16—Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/93—Document management systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/11—Complex mathematical operations for solving equations, e.g. nonlinear equations, general mathematical optimization problems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/50—Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
- G06F21/55—Detecting local intrusion or implementing counter-measures
- G06F21/552—Detecting local intrusion or implementing counter-measures involving long-term monitoring or reporting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/02—Knowledge representation; Symbolic representation
- G06N5/022—Knowledge engineering; Knowledge acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/10—Office automation; Time management
- G06Q10/105—Human resources
- G06Q10/1053—Employment or hiring
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/40—Document-oriented image-based pattern recognition
- G06V30/41—Analysis of document content
- G06V30/416—Extracting the logical structure, e.g. chapters, sections or page numbers; Identifying elements of the document, e.g. authors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2216/00—Indexing scheme relating to additional aspects of information retrieval not explicitly covered by G06F16/00 and subgroups
- G06F2216/11—Patent retrieval
Definitions
- This disclosure relates in general to the field of machine learning, and more particularly, though not exclusively to, a system and method for providing entropic classification of objects.
- Machine-driven classification of objects such as web pages, text documents, or multimedia content, is useful in contexts such as security and productivity.
- Classifiers are a category of tool in the analytics arsenal that help divide input data into one or more classes. For example, email spam filter classifiers decide what incoming email is classified as “spam” and what is classified as “ham .” A classifier with a binary result (“yes,” the object is in the target class, or "no,” the object is not in the target class) may be referred to as a binary classifier. A classifier that designates an object as belonging to one of several potential classes may be referred to as a multinominal classifier.
- FIGURE 1 is a block diagram of a security-enabled network according to one or more examples of the present specification.
- FIGURE 2 is a block diagram of a computing device according to one or more examples of the present specification .
- FIGURE 3 is a block diagram of a server according to one or more examples of the present specification.
- FIGURE 4 is a block diagram of a classification system according to one or more examples of the present specification.
- FIGURE 5 is an illustration of analysis of an issued patent according to one or more examples of the present specification.
- FIGURE 6 is a workflow diagram of document classification according to one or more examples of the present specification.
- FIGURE 7 is a message graph according to one or more examples of the present specification.
- FIGURE 8 is a graph of impact potential according to one or more examples of the present specification.
- FIGURE 9 is a block diagram of a mathematical theory of communication in information theory, according to one or more examples of the present specification.
- FIGURE 10 is a block diagram of a classification method according to one or more examples of the present specification.
- FIGURE 11 is an illustration of evolution of a message graph according to one or more examples of the present specification.
- FIGURE 12 is a message graph according to one or more examples of the present specification.
- FIGURE 13 is an illustration of evolution of a message graph according to one or more examples of the present specification
- FIGURE 14 is an illustration of evolution of a message graph according to one or more examples of the present specification
- FIGURE 15 a graph of a distribution of gentropy scores according to one or more examples of the present specification
- FIGURE 16 is a histogram of the percentile range of labeled data with respect to a whole data set according to one or more examples of the present specification.
- FIGURE 17 is a block diagram of a classification process according to one or more examples of the present specification.
- FIGURE 18 is a block diagram of a classification process according to one or more examples of the present specification.
- FIGURE 19 is a block diagram of a classification process according to one or more examples of the present specification.
- FIGURE 20 is a flow chart of a method of computing and applying gentropy scores according to one or more examples of the present specification .
- a computing apparatus for assigning an entropy score to a document to be added to a corpus in a first temporal state having a first corpus entropy, having one or more logic elements, including at least one hardware logic element, providing a classification engine to: receive the document to be added to the corpus; add the document to the corpus, creating a second temporal state of the corpus; compute a second corpus entropy for the second temporal state, based at least in part on a morphism; and assign the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
- Classification of documents and other objects is an important task for an enterprise. Such classification can have uses for security, privacy, litigation, and many other purposes.
- an enterprise may have firewall rules configured to treat certain classes of websites differently.
- Business-essential websites such as corporate websites, business information, partner sites, and similar
- Non-essential but harmless websites such as news, general information, and e-commerce
- Time- wasting websites such as sports, games, most social media, fashion, or other non-business-related subject matters
- Malicious websites (such as malware, or websites with illegal or unethical content) may be blocked and permissible only with special administrative dispensation.
- a machine classifier may be used to assign ownership or other attributes to business documents.
- an enterprise may deploy a database file system, in which files must be assigned attributes, such as which group owns each document, which client or job number a document belongs to, or similar attributes.
- the classifier may use existing metadata fields, file locations, and content to assign these attributes (each constituting a "class") to each document.
- classification is mutually exclusive.
- the classes are mutually exclusive (e.g., the owner can be either "Business” or "Marketing,” but not both).
- classes may be more analogous to "tags,” and thus may be non-exclusive.
- a popular sports website that also frequently features scantily-clad women and men in skimpy swimsuits, may be tagged as both "sports" and "soft pornography.” In the latter case, the site may inherit the permission set of its most- restricted tag.
- machine classification may be used to mitigate the expense of electronic production for litigation .
- a binary classifier may scan a large corpus of documents, and tag each one as “relevant” or “not relevant” to the litigation. Additional classifiers may be used, for example, to tag documents as "privileged,” “attorneys' eyes only,” or as potential trial exhibits. Even if a human user is required to verify the classifications, having the initial tagging in place can significantly speed the process.
- Embodiments of the present specification deal with the problem of information overload, which occurs when the amount of input to a system exceeds its processing capacity. For example, a web search might find millions of relevant results, or there might be tens of thousands of relevant technical articles in a given field. In those cases, it is implausible for a cognitive observer to review all the available data and to choose a best fit.
- a recommendation engine may at least partially offset the limitations of keyword searching.
- Recommendation engines are used by services like Pandora, last.fm, Google Music, Netflix, Youtube, Vimeo, and many others to recommend media such as movies, music, and books based on the user's previous selections and the selections of users with similar profiles, and stated preferences.
- these systems like keyword-based search systems, may lack information diversity in certain embodiments. This can become particularly important when the elements being searched contain highly redundant information. Taking search results as an example, higher diversity of returned results allows the system to present a wider span of information in the first few hits, or for recommender systems to present a broader selection of recommendations, some of them even surprising to the user but still useful .
- embodiments of the methods described in this specification can be used to rank a collection of items according to how similar or dissimilar they are compared to an existing collection, based at least in part on one or more morphisms, or a combination of morphisms.
- Morphisms are physical or conceptual relationship schemes between attributes of the items or the items themselves. In information theory, these items, their attributes, or their relationships could be thought of as the result of a transmitter operating on a message.
- each object or item in the data set is called a "document.”
- the term “document” should not be misconstrued to be limited merely to discrete units of paginated textual data. Rather, it is a message produced by an information source (from an information theoretic point of view).
- a document in this context includes any object or datum that may be represented or modeled digitally, including its attributes and metadata where appropriate.
- "documents” may be text documents, web pages, books, movies, music tracks, sounds, periodicals, articles, businesses, stocks, living organism (plants, animals, or people), network objects, network addresses, users of a network, executable files, astronomical objects, weather events, or weather patterns, to name just a few nonlimiting examples.
- a collection of documents being managed by the system may be referred to as the "corpus."
- the corpus In an initial “null corpus" state, the corpus has zero documents.
- the corpus When the first document is added to the corpus (including any characterization or classification performed on the document), the corpus enters a second state. Thereafter, with the addition of each new document or group of documents, the corpus moves into a new state.
- the corpus With each state, the corpus may have a total information entropy H, the computation of which is detailed below.
- the total information entropy of a particular state may be notated as H n , where n is the state number.
- a change in information entropy occurs.
- the change in information entropy caused by the addition of a new document d to the corpus may be represented as S d .
- identifying exemplars may be most desirable (e.g., review articles, where exemplars may represent the best summarizations of field), or identifying outliers may be most desirable (e.g . , network security, where outl iers may represent a security risk) .
- a system of the present specification may quantitatively measure the graph entropy (the "gentropy") of a mathematical graph where the nodes and edges are assigned according to a particu lar morphism or combination of morphisms and based on the documents of a corpus, and thus be able to calcu late the change in gentropy due to actual or potential new additions.
- the graph such constructed is a "message graph" since in an information theoretic framework, the information contained in each relationship between attributes is a message .
- a morphism is akin to a receiver that reconstructs certain aspects of the message .
- One interpretation of the gentropy is as a measure of diversity. For exam ple, in ecology, ecosystems with more an i mal species and about the same num ber of animals belonging to each species are more diverse than ecosystems with fewer species or those in wh ich most an imals belong to a few species.
- the value of S d may be conditioned to account for corpus size.
- i n a corpus of three documents
- the addition of a fourth document may be expected to have a very substantial effect on the overall entropy of the corpus, not because the document is very significant, but because the corpus is qu ite small .
- the addition of a one-mi ll ion-and-first document will be expected to have only a smal l effect on the total entropy, not because the document is unimportant, but because the corpus is very large. Additionally, certain attribute relationships m ight be more important that others.
- G d of document d is its conditioned S d .
- the methods disclosed herein may be appl ied in a variety of cases, and may be used to predict items that a user wi ll find the most interesting from a large corpus.
- One application of the present method is to aid in the identification of technical documents that a user might find interesting .
- This use case is not intended to be limiting, but was used as a validation test of the model disclosed herein.
- document metadata i.e., not document content
- This test case illustrates an advantage of the present method, because neither keywords (as used by a search engine), nor past user behavior or preferences (as used by recommendation engines) are necessary, though keywords and past behavior can be integrated with the present methods in appropriate embodiments.
- test implementation was based on classification of contents, and was intended to identify items similar to the collection, as well as items that are unusual but still relevant.
- the disclosed method was used to score each item in a collection of about 14,000 technical documents belonging broadly to the same industry and published periodically since 1976.
- the gentropy score for each document measured the technological diversity of each document with respect to the rest of the documents published up to that point, and is based on a classification assigned to it upon publication .
- the classification scheme in this case is a human-created taxonomy with several thousand unique categories.
- a publicly-available publication generally considered reputable, and maintained by a group of domain experts was consulted.
- the experts who maintain the website read newly-published documents in a particular industry, and provide a digest with commentary. They publish additional commentary about documents that may be of particular interests to industry groups.
- This specification provides a method to calculate the information entropy of the message graph of a corpus (i .e. , the gentropy) using the content or metadata of each document (i .e . , item of any type) along with a morphism (a structure-preserving map from one mathematical structure to another) .
- the combination of these methods measures the relative diversity of particular items of the same kind (i .e ., how much new information each provides. )
- the conditioned entropy, the gentropy score may be used to identify noteworthy documents in a corpus, and to make recom mendations to a user.
- ⁇ ( ⁇ ⁇ ) is the probabil ity distribution function and H is measured in bits.
- X can parameterize more than one scoring aspect as explained below. These aspects could include factors like content and user ranking .
- ⁇ ( ⁇ ;) is the stationary occupation probabi lity of each node based at least in part on the number of directed or undirected edges and the occupation probabi lities of neighboring nodes. This concept is i llustrated in FIGU RE 12, where node A has a large occupation probabil ity because almost all other nodes point to it. Node G has a relatively large occupation probabil ity because node A points to it (even though only node A points to it) .
- a morphism is a structure-preserving map from one mathematical structure to another, i .e. , a rule or set of rules that transforms data but preserves certain relationships or properties of the data .
- the critical properties to be preserved are growth and preferential attachment (or other non-trivial topological features), which results in a graph that has a degree distribution that follows a power law. Without preferential attachment or other non-trivial topological features, the results of the present method could be inconclusive.
- scale- freeness is a characteristic of complex systems, and most networks of interest or other systems that can be mapped into a network are scale-free and show power law behavior, such as social networks and the world wide web (WWW) .
- the rule that we used for the results illustrated herein is that for every document, a node was added to the graph (if the node did not previously exist), corresponding to each classification code assigned to the document, and edges were created between them (so it was an undirected multigraph based on the co-classification network of the documents whose stationary probability was calculated and updated after each step) .
- G d was calculated for each document.
- the power law behavior of the system is evidenced by the linearity in the log-log plot of the gentropy score distribution of the documents that increased diversity. Of these documents, about 4,000 contributed 0.1 bits or less while a handful contributed more than 10 bits.
- the method disclosed herein may be used with a collection of items constructed with the results of a keyword search, user behavior, or similar. The user may then elect whether to look at items that are representative of the collection or those that are more unusual or span more information, based on a gentropy score ranking .
- a user may explore which items, not already in a collection, would change the collection's diversity the most. This can be used to provide recommendations for movies, books, music, and other media. Exploration could be manual or automatic. Items that decrease diversity are useful in some situations, for example in a collection of books with two subgroups, one on Java and Scala, and the other one on C and C++, yet another book on Java and Scala might be exactly what the user is looking for. Items that increase diversity (and the degree to which they do) might be useful in other situations, for example a book on all four languages will increase diversity more than one on Java and C++, and the user might be interested in acquiring different knowledge.
- the method can be used to monitor compliance with anti-discrimination regulations or other hiring criteria.
- the method may be used to assign a gentropy score (i.e., a diversity score) to a hiring decision, which may be orthogonal to other hiring criteria. If hiring decisions group strongly around a single gender, ethnicity, skill set, work experience, age or other factor, the hiring may be very non-diverse.
- a hiring decision that involves a person from a technical or life background that is less commonly hired may be considered “more diverse.” This could be desirable, for example, when having an ethnically diverse team of product designers is required to penetrate new markets. Records can be used to measure diversity trends over time, and to identify "outliers" that are substantially different from "business as usual.”
- FIGURES A system and method for binary linear classification will now be described with more particular reference to the attached FIGURES. It should be noted that throughout the FIGURES, certain reference numerals may be repeated to indicate that a particular device or block is wholly or substantially consistent across the FIGURES. This is not, however, intended to imply any particular relationship between the various embodiments disclosed.
- a genus of elements may be referred to by a particular reference numeral ("widget 10"), while individual species or examples of the genus may be referred to by a hyphenated numeral ("first specific widget 10-1" and "second specific widget 10- 2").
- FIGURE 1 is a network-level diagram of an enterprise 100 receiving classification services according to one or more examples of the present specification. This figure illustrates one nonlimiting example of using classification to provide enterprise-class spam filtering on a network. This example is intended to be nonlimiting, and other examples are provided herein to illustrate the broad applicability of the methods disclosed herein.
- one or more users 120 operate one or more client devices 110.
- Each device may include an appropriate operating system, such as Microsoft Windows, Linux, Android, Mac OSX, Apple iOS, Unix, or similar. Some of the foregoing may be more often used on one type of device than another. For example, desktop computers or engineering workstation may be more likely to use one of Microsoft Windows, Linux, Unix, or Mac OSX. Laptop computers, which are usually a portable off-the-shelf device with fewer customization options, may be more likely to run Microsoft Windows or Mac OSX. Mobile devices may be more likely to run Android or iOS. However, these examples are not intended to be limiting.
- Client devices 110 may be communicatively coupled to one another and to other network resources via enterprise network 170.
- Enterprise network 170 may be any suitable network or combination of one or more networks operating on one or more suitable networking protocols, including for example, a local area network, an intranet, a virtual network, a wide area network, a wireless network, a cellular network, or the Internet (optionally accessed via a proxy, virtual machine, or other similar security mechanism) by way of nonlimiting example.
- Enterprise network 170 may also include one or more servers, firewalls, routers, switches, security appliances, antivirus servers, or other useful network devices, which in an example may be virtualized within workload cluster 142.
- enterprise network 170 is shown as a single network for simplicity, but in some embodiments, enterprise network 170 may include a large number of networks, such as one or more enterprise intranets connected to the internet. Enterprise network 170 may also provide access to an external network, such as the Internet, via external network 172. External network 172 may similarly be any suitable type of network.
- a workload cluster 142 may be provided, for example as a virtual cluster running in a hypervisor on a plurality of rack-mounted blade servers, or as a cluster of physical servers.
- Workload cluster 142 may provide one or more server functions, or one or more "microclouds" in one or more hypervisors.
- a virtualization environment such as vCenter may provide the ability to define a plurality of "tenants," with each tenant being functionally separate from each other tenant, and each tenant operating as a single-purpose microcloud.
- Each microcloud may serve a distinctive function, and may include a plurality of virtual machines (VMs) of many different flavors, including agentful and agentless VMs.
- VMs virtual machines
- endpoint devices 110 may also be provided via workload cluster 142.
- one microcloud may provide a remote desktop hypervisor such as a Citrix workspace, which allows users 120 operating endpoints 110 to remotely login to a remote enterprise desktop and access enterprise applications, workspaces, and data.
- endpoint 120 could be a "thin client" such as a Google Chromebook, running only a stripped-down operating system, and still provide user 120 useful access to enterprise resources.
- one or more virtual machines are provisioned within workload cluster 142 to act as a machine classifier 190.
- Machine classifier 190 may be tasked with classification of network objects, including static and/or dynamic objects, for any suitable purpose, such as identifying spam and malware. For example, classification may be provided for enforcement of enterprise security and network usage policies, such as by classifying websites into various categories, each with different security settings or permissions.
- machine classifier 190 may classify text documents for storage in a database file system.
- the machine classifier may be tasked with indexing content to ease searching and compilation, such as indexing patents by subject matter.
- network objects are classified for security purposes, and assigned to a classification such as "white” (permitted), “gray” (unknown), or “black” (barred) as part of an antivirus or antimalware scheme.
- the objects may be any network object as described herein, including static objects, active objects, and network-attached devices.
- classifications may be based, for example, on identifying objects that introduce high entropy into the system, or alternately, that group strongly around other known objects of the same type.
- These embodiments are provided as nonlimiting examples only, and should be understood to be non-exclusive. Many other uses for object classification are possible, and are within the intended scope of this specification.
- One or more computing devices configured as a management console 140 may also operate on enterprise network 170.
- Management console 140 may provide a user interface for a security administrator 150 to define enterprise security policies, which management console 140 may enforce on enterprise network 170 and across client devices 110 and workload cluster 142.
- management console 140 may run a server-class operating system, such as Linux, Unix, or Windows Server.
- management console 140 may be provided as a web interface, on a desktop-class machine, or via a VM provisioned within workload cluster 142.
- Security administrator 150 or user 120 may use either management console 140 or client device 110 to "train" a data set, either interactively, or by submitting a batch of pre-classified documents.
- Enterprise 100 may communicate across enterprise boundary 104 with external network 172.
- Enterprise boundary 104 may represent a physical, logical, or other boundary.
- External network 172 may include, for example, websites, servers, network protocols, and other network-based services.
- an attacker 180 (or other similar malicious or negligent actor) also connects to external network 172.
- attacker 180 is a malware or spam author whose goal or purpose is to cause malicious harm or mischief, for example by injecting malicious object 182 (e.g., spam or malware) into client device 110.
- malicious object 182 Once malicious object 182 gains access to client device 110, it may try to perform work such as social engineering of user 120, a hardware- based attack on client device 110, modifying storage, modifying client application 112, or gaining access to enterprise servers 142.
- Attacker 180 may also want to commit industrial or other espionage against secured enterprise 100, such as stealing classified or proprietary data, stealing identities, or gaining unauthorized access to enterprise resources.
- attacker 180's strategy may also include trying to gain physical access to one or more client devices 110 and operating them without authorization, so that an effective security policy may also include provisions for preventing such access.
- enterprise 100 may simply be a family, with parents assuming the role of security administrator 150.
- the parents may wish to protect their children from undesirable content, such as pornography, adware, spyware, age-inappropriate content, advocacy for certain political, religious, or social movements, or forums for discussing illegal or dangerous activities, by way of non-limiting example.
- the parent may perform some or all of the duties of security administrator 150.
- security policies may initially treat it as "gray” or "suspect.”
- machine classifier 190 may inspect the object and attempt to classify it. For example, if the object is an e-mail, machine classifier 190 may attempt to determine whether it is “ham” or “spam .” If the object is executable, machine classifier 190 may attempt to determine whether it is “malware” or “not malware,” or whether it is a "permissible application” or “impermissible application . Again, these classifications may be achieved because an object either has high entropy, or groups strongly with other similar objects. Advantageously, this classification may be achieved via metadata, without the need to inspect the content of the object.
- FIGURE 2 is a block diagram of client-class computing device 200 according to one or more examples of the present specification.
- Computing device 200 may be any suitable computing device.
- a "computing device” may be or comprise, by way of non-limiting example, a computer, workstation, server, mainframe, virtual machine (whether emulated or on a "bare- metal” hypervisor), embedded computer, embedded controller, embedded sensor, personal digital assistant, laptop computer, cellular telephone, IP telephone, smart phone, tablet computer, convertible tablet computer, computing appliance, network appliance, receiver, wearable computer, handheld calculator, or any other electronic, microelectronic, or microelectromechanical device for processing and communicating data.
- Any computing device may be designated as a host on the network. Each computing device may refer to itself as a "local host,” while any computing device external to it may be designated as a "remote host.”
- client device 110 may be an example of computing device 200.
- Users 120 may use computing device 200 to interact with a classified data set, or to perform tasks such as performing interactive training or submitting batches of pre-classified documents.
- Computing device 200 includes a processor 210 connected to a memory 220, having stored therein executable instructions for providing an operating system 222 and at least software portions of a security engine 224.
- Other components of client device 200 include a storage 250, network interface 260, and peripheral interface 240.
- This architecture is provided by way of example only, and is intended to be non-exclusive and non-limiting. Furthermore, the various parts disclosed are intended to be logical divisions only, and need not necessarily represent physically separate hardware and/or software components.
- Certain computing devices provide main memory 220 and storage 250, for example, in a single physical memory device, and in other cases, memory 220 and/or storage 250 are functionally distributed across many physical devices.
- each logical block disclosed herein is broadly intended to include one or more logic elements configured and operable for providing the disclosed logical operation of that block.
- logic elements may include hardware, external hardware (digital, analog, or mixed-signal), software, reciprocating software, services, drivers, interfaces, components, modules, algorithms, sensors, components, firmware, microcode, programmable logic, or objects that can coordinate to achieve a logical operation.
- processor 210 is communicatively coupled to memory 220 via memory bus 270-3, which may be for example a direct memory access (DMA) bus by way of example, though other memory architectures are possible, including ones in which memory 220 communicates with processor 210 via system bus 270-1 or some other bus.
- processor 210 may be communicatively coupled to other devices via a system bus 270-1.
- a "bus" includes any wired or wireless interconnection line, network, connection, bundle, single bus, multiple buses, crossbar network, single-stage network, multistage network or other conduction medium operable to carry data, signals, or power between parts of a computing device, or between computing devices. It should be noted that these uses are disclosed by way of non-limiting example only, and that some embodiments may omit one or more of the foregoing buses, while others may employ additional or different buses.
- a "processor” may include any combination of logic elements operable to execute instructions, whether loaded from memory, or implemented directly in hardware, including by way of non-limiting example a microprocessor, digital signal processor, field-programmable gate array, graphics processing unit, programmable logic array, application-specific integrated circuit, or virtual machine processor.
- a multi-core processor may be provided, in which case processor 210 may be treated as only one core of a multi-core processor, or may be treated as the entire multi-core processor, as appropriate.
- one or more co-processors may also be provided for specialized or support functions.
- Processor 210 may be connected to memory 220 in a DMA configuration via DMA bus 270-3.
- memory 220 is disclosed as a single logical block, but in a physical embodiment may include one or more blocks of any suitable volatile or non-volatile memory technology or technologies, including for example DDR RAM, SRAM, DRAM, cache, LI or L2 memory, on-chip memory, registers, flash, ROM, optical media, virtual memory regions, magnetic or tape memory, or similar.
- memory 220 may comprise a relatively low-latency volatile main memory, while storage 250 may comprise a relatively higher-latency non-volatile memory.
- memory 220 and storage 250 need not be physically separate devices, and in some examples may represent simply a logical separation of function. It should also be noted that although DMA is disclosed by way of non-limiting example, DMA is not the only protocol consistent with this specification, and that other memory architectures are available.
- Storage 250 may be any species of memory 220, or may be a separate device.
- Storage 250 may include one or more non-transitory computer- readable mediums, including by way of non-limiting example, a hard drive, solid- state drive, external storage, redundant array of independent disks (RAID), network-attached storage, optical storage, tape drive, backup system, cloud storage, or any combination of the foregoing.
- Storage 250 may be, or may include therein, a database or databases or data stored in other configurations, and may include a stored copy of operational software such as operating system 222 and software portions of security agent 224. Many other configurations are also possible, and are intended to be encompassed within the broad scope of this specification.
- Network interface 260 may be provided to communicatively couple client device 200 to a wired or wireless network.
- a "network,” as used throughout this specification, may include any communicative platform operable to exchange data or information within or between computing devices, including by way of non- limiting example, an ad-hoc local network, an internet architecture providing computing devices with the ability to electronically interact, a plain old telephone system (POTS), which computing devices could use to perform transactions in which they may be assisted by human operators or in which they may manually key data into a telephone or other suitable electronic equipment, any packet data network (PDN) offering a communications interface or exchange between any two nodes in a system, or any local area network (LAN), metropolitan area network (MAN), wide area network (WAN), wireless local area network (WLAN), virtual private network (VPN), intranet, or any other appropriate architecture or system that facilitates communications in a network or telephonic environment.
- POTS plain old telephone system
- PDN packet data network
- LAN local area network
- MAN metropolitan area network
- WAN wide
- Security agent 224 in one example, is operable to carry out computer-implemented methods as described in this specification.
- Security engine 224 may include one or more tangible non-transitory computer-readable mediums having stored thereon executable instructions operable to instruct a processor to provide a security engine 224.
- an "engine” includes any combination of one or more logic elements, of similar or dissimilar species, operable for and configured to perform one or more methods provided by the engine.
- security engine 224 may comprise one or more logic elements configured to provide methods as disclosed in this specification .
- security engine 224 may include a special integrated circuit designed to carry out a method or a part thereof, and may also include software instructions operable to instruct a processor to perform the method.
- security engine 224 may run as a "daemon” process.
- a “daemon” may include any program or series of executable instructions, whether implemented in hardware, software, firmware, or any combination thereof, that runs as a background process, a terminate-and- stay-resident program, a service, system extension, control panel, bootup procedure, BIOS subroutine, or any similar program that operates without direct user interaction.
- daemon processes may run with elevated privileges in a "driver space,” or in ring 0, 1, or 2 in a protection ring architecture.
- security engine 224 may also include other hardware and software, including configuration files, registry entries, and interactive or user-mode software by way of non-limiting example.
- security engine 224 includes executable instructions stored on a non-transitory medium operable to perform a method according to this specification.
- processor 210 may retrieve a copy of the instructions from storage 250 and load it into memory 220.
- Processor 210 may then iteratively execute the instructions of security engine 224 to provide the desired method.
- security engine 224 may be provisioned on client device 200 to enforce enterprise security, resource usage, or other policies. For example, when a user 120 directs client device 200 to a particular website, security engine 224 may query a server engine for an appropriate class and associated permissions. The machine may then locally enforce those permissions on client device 200. Enforcement may also be performed in whole or in part by server devices 300 (FIGURE 3). In another example, security engine 224 may mark e-mails as "spam" or "ham.” In yet another example, security engine 224 may simply provide an interface for accessing documents or objects that have been classified by classification engine 326.
- Peripheral interface 240 may be configured to interface with any auxiliary device that connects to client device 200 but that is not necessarily a part of the core architecture of client device 200.
- a peripheral may be operable to provide extended functionality to client device 200, and may or may not be wholly dependent on client device 200.
- a peripheral may be a computing device in its own right.
- Peripherals may include input and output devices such as displays, terminals, printers, keyboards, mice, modems, data ports (e.g., serial, parallel, USB, Firewire, or similar), network controllers, optical media, external storage, sensors, transducers, actuators, controllers, data acquisition buses, cameras, microphones, speakers, or external storage by way of non-limiting example.
- peripherals include display adapter 242, audio driver 244, and input/output (I/O) driver 246.
- Display adapter 242 may be configured to provide a human-readable visual output, such as a command-line interface (CLI) or graphical desktop such as Microsoft Windows, Apple OSX desktop, or a Unix/Linux X Window System-based desktop.
- Display adapter 242 may provide output in any suitable format, such as a coaxial output, composite video, component video, VGA, or digital outputs such as DVI or HDMI, by way of nonlimiting example.
- display adapter 242 may include a hardware graphics card, which may have its own memory and its own graphics processing unit (GPU).
- GPU graphics processing unit
- Audio driver 244 may provide an interface for audible sounds, and may include in some examples a hardware sound card. Sound output may be provided in analog (such as a 3.5mm stereo jack), component ("RCA") stereo, or in a digital audio format such as S/PDIF, AES3, AES47, HDMI, USB, Bluetooth or Wi-Fi audio, by way of non-limiting example.
- analog such as a 3.5mm stereo jack
- RCA component
- Digital audio format such as S/PDIF, AES3, AES47, HDMI, USB, Bluetooth or Wi-Fi audio, by way of non-limiting example.
- FIGURE 3 is a block diagram of a server-class device 300 according to one or more examples of the present specification.
- Server 300 may be any suitable computing device, as described in connection with FIGURE 2.
- the definitions and examples of FIGURE 2 may be considered as equally applicable to FIGURE 3, unless specifically stated otherwise.
- Server 300 is described herein separately to illustrate that in certain embodiments, logical operations according to this specification may be divided along a client-server model, wherein client device 200 provides certain localized tasks, while server 300 provides certain other centralized tasks. In contemporary practice, server 300 is more likely than client device 200 to be provided as a "headless" VM running on a computing cluster, or as a standalone appliance, though these configurations are not required.
- any of the devices in workload cluster 142 may be a server-class device 300.
- Server-class device 300 is disclosed as a separate device from client device 200 to illustrate one logical division of function between devices. But it should be understood that this example is not intended to be limiting, and that a classification engine 326 may be provided on a client device 200 in suitable circumstances.
- Server 300 includes a processor 310 connected to a memory 320, having stored therein executable instructions for providing an operating system 322 and at least software portions of a server engine 324.
- Other components of server 300 include a storage 350, network interface 360, and peripheral interface 340. As described in FIGURE 2, each logical block may be provided by one or more similar or dissimilar logic elements.
- processor 310 is communicatively coupled to memory 320 via memory bus 370-3, which may be for example a direct memory access (DMA) bus.
- processor 310 may be communicatively coupled to other devices via a system bus 370-1.
- DMA direct memory access
- Processor 310 may be connected to memory 320 in a DMA configuration via DMA bus 370-3, or via any other suitable memory configuration .
- memory 320 may include one or more logic elements of any suitable type.
- Storage 350 may be any species of memory 320, or may be a separate device, as described in connection with storage 250 of FIGURE 2.
- Storage 350 may be, or may include therein, a database or databases or data stored in other configurations, and may include a stored copy of operational software such as operating system 322 and software portions of server engine 324.
- Network interface 360 may be provided to communicatively couple server 140 to a wired or wireless network, and may include one or more logic elements as described in FIGURE 2.
- Server engine 324 is an engine as described in FIGURE 2 and, in one example, includes one or more logic elements operable to carry out computer- implemented methods as described in this specification. Software portions of server engine 324 may run as a daemon process.
- Server engine 324 may include one or more non-transitory computer-readable mediums having stored thereon executable instructions operable to instruct a processor to provide appropriate services.
- processor 310 may retrieve a copy of server engine 324 (or software portions thereof) from storage 350 and load it into memory 320. Processor 310 may then iteratively execute the instructions of server engine 324 to provide the desired method.
- Server engine 324 may provide services such as a webserver, database server, database file system server, network antivirus or antimalware engine, or other functions that may benefit from the classification methods disclosed herein .
- Classification engine 326 is also an engine as described herein, and may include one or more logic elements operable to carry out computer- implemented methods as described in this specification. Software portions of classification engine 326 may run as a daemon process. Note that "classification engine” is used throughout as a general term to include the engine that carries out the methods disclosed herein. This is true even in cases where classification engine 326 performs only a portion of its available function (e.g., even when it performs only scoring without classifying a document or message).
- Classification engine 326 may include one or more non-transitory computer-readable mediums having stored thereon executable instructions operable to instruct a processor to provide classification services. At an appropriate time, such as upon booting server 140 or upon a command from operating system 322 or a user 120 or security administrator 150, processor 310 may retrieve a copy of classification engine 326 (or software portions thereof) from storage 350 and load it into memory 320. Processor 310 may then iteratively execute the instructions of classification engine 326 to provide the desired method.
- classification engine 326 More specific functions of classification engine 326 are provided throughout this specification.
- FIGURE 4 is a block diagram of classification of a corpus of documents according to one or more examples of the present specification.
- Training set 410 may commonly be much smaller than corpus 430, and may include documents selected from corpus 430.
- corpus 430 is the corpus of published and issued U.S. patents and patent applications. Because this corpus comprises many millions of documents, it may be impractical for a human user to manually classify all of them . However, it may be useful for a user to manually or automatically classify some documents to give the corpus a starting point.
- classification engine 326 receives training set 410, along with the larger patent corpus 430. Employing method 2100 of FIGURE 20, or another suitable method, classification engine 326 assigns a class to each document based on its a individual gentropy score 460 and those of the training set.
- classification engine 326 may include a synonym engine 480 for data, metadata and other attributes, or any combination of these.
- Synonym engine 480 may be configured to identify words or phrases with identical or similar meanings, including variations on word forms (e.g., plurals and changes in tense), and true synonyms (e.g., "NIC” vs. "Ethernet card”) . It can also be configured to identify functionally or statistically equivalent attributes, e.g., patent classification codes in the US patent classification system and the cooperative classification system.
- Corpus 430 starts out in a first temporal state, having a first information entropy.
- Classification engine 326 receives a document to be added to corpus 430.
- Classification engine 326 adds the document to corpus 430, thus causing corpus 430 to assume a second temporal state, having a second information entropy.
- Classification engine 326 computes the corpus information entropy and gentropy after adding the document, and also computes an individual gentropy score 460 (SG d ) for the document that was added.
- FIGURE 5 is an illustration of feature sources that may be used in an embodiment where issued and published patents comprise corpus 430.
- the text of the abstract 590 or the content of the detailed description comprise the "data" of the document, which need not be considered for the present method.
- the other fields of a patent may be considered “metadata,” and may be used in embodiments of the present method.
- the patent classification codes are disclosed as an example that was used for validation of the method, but other metadata fields may be used either individually or in combination.
- words appearing in the abstract may stand in as a proxy for classification, as the abstract includes words most likely to be representative of the subject matter of the patent as a whole.
- words in the abstract may be used as classes.
- Certain embodiments may also employ the concept of word variations and synonyms.
- SoC is a common abbreviation for "system-on-a-chip,” and they may be treated as an identical word.
- system on a chip “systems on a chip,” and “systems-on-a-chip” may all be considered identical for purpose of classification engine 326.
- the terms “jelly bean” and "Easter egg” may be considered to be in related arts, though they are not strict synonyms.
- Non-identical words with similar meaning may also be useful. This may include a recognition that words can be both contemporaneously synonymous, and that words can change over time. Indeed, as technologies mature, they may organically acquire new and different nomenclatures.
- the first known patents on transistors did not refer to the devices as “transistors.” Rather, they referred to them as “amplifiers” or similar.
- patent date 510 (FIGURE 5) may be a relevant metadata factor. For a 2015 date, "transistor” may be a more predictive term, while for a 1925 date, “amplifier” may be a more predictive term .
- classification engine 326 may include a synonym engine 480 (FIGURE 4) to identify synonyms, including temporal spacing of synonyms.
- Additional metadata fields may also be used as attributes.
- classification codes 560 were used in a validation test, and both high and low entropy in this field was found to be predictive of "interesting" patents.
- Other fields that can be used in a classification scheme include title 504, inventor names 530, assignee 540, filing date 550, references cited 520 (similar to hyperlinks— note that additional connections may be formed when future patents cite back to this one), examiner names 570, attorney 580, and abstract 590, by way of nonlimiting example.
- FIGURE 6 is a workflow diagram of document classification according to one or more examples of the present specification.
- corpus 430 is any suitable corpus consistent with the present specification. From within corpus 430, interesting documents 620 are identified. A human user or a computer or automated system 630 interact with interesting documents 620 and can perform desired statistical analyses on important documents 620 which might be a subset of corpus 430 or the whole corpus 430.
- FIGURE 7 is a message graph created according to one or more examples of morphisms of the present specification that represents relationships between documents in a corpus.
- message graph 700 includes three clusters, namely cluster 1 710- 1, cluster 2 710-2, and cluster 3 710-3.
- Each cluster may represent, for example, a broad subject matter area.
- cluster 1 could represent the area of fishing lures.
- Cluster 2 could represent the area of semiconductor device manufacturing .
- Cluster 3 could represent the area of computer security.
- Each node in message graph 700 may represent an attribute or category, such as a classification code.
- Each edge of the graph represents a connection between attributes, such as a patent that was assigned several .
- Each time an "event" happens (in this case, the event is that a new patent is added to the corpus), its connections are added to the message graph . This represents a change in state.
- Message graph 700 was in a first state before the new patent was added. After the patent was added, some change in information entropy ⁇ ⁇ has occurred, and the corpus is now in a second state with a new information entropy.
- Incremental innovation 720 may occur when, for example, a new patent issues with two classification codes that already exist in cluster 3 710-3, plus one new code that was not previously linked to cluster 3 710-3. It should be understood that labeling the patent as "incremental" does not in any way imply that it is unimportant, nor is it a judgment of the breadth, scope, or monetary value of the newly-issued patent. Rather, the innovation is incremental because it has only a modest effect on the overall entropy of the graph, because it has introduced merely a new node to cluster 3 710-3 which already had numerous nodes.
- radical innovation 730 has a much more significant effect on the entropy of the corpus.
- radical innovation 730 may arise from a patent that, for example, identifies a novel material or novel use for an existing material that is applicable to both fishing lures and semiconductor manufacturing .
- labeling node 730 as a "radical" innovation does not represent a judgment on the importance, breadth, scope, or monetary value of the newly-issued patent. Rather, it indicates that radical innovation 730 has a relatively substantial impact on the overall entropy of the corpus.
- the method not only finds exemplars or outliers with respect to the corpus, but because a score is assigned, it can also quantify and rank the degree to which, e.g., outliers are outliers. For particular cases, the most interesting documents might be in a particular range as opposed to in the extremes. In a nonlimiting case, a user might like movie recommendations that are not very extreme outliers, but are close to the 80th percentile. In another nonlimiting case, certain categories of items that the user might be interested in, such as funny yet disturbing movies, might consistently fall in a particular range.
- Message graph 700 can be used to visually represent data in a way that is very useful for a human observer. For example, radical innovation 730 may be highlighted in a special color that draws particular attention to it. If a human user is searching for prior art patents that may anticipate a patent currently being litigated, radical innovation 730 may in fact be an outlier that is less important, and may be discarded. In this example, the searcher may be looking rather for patents that are more narrowly focused on the subject matter at hand. Similarly, a scientist reviewing literature in her art may prefer to focus on nodes that have many links within their clusters (e.g., frequently-cited publications).
- message graph 700 may color nodes with different colors or intensities to represent the nodes with the most links (citations) within cluster 710-3.
- a web user is searching for information about the Free Speech clause of the First Amendment.
- low-entropy nodes (those with many links) may be of more interest than high-entropy nodes (e.g., a Timecube rant that links free speech to (pseudo-) quantum mechanics and government conspiracy theories).
- radical innovation 730 may be a curiosity at best and of little practical value to the system .
- clusters 710 may be of little interest.
- cluster 3 710-3 may represent users in the accounting department accessing resources commonly accessed by accounting personnel.
- a radical innovation 730 occurs, that may represent a user or machine accessing a resource outside of his normal sphere of influence (for example, a user in accounting accesses a network management console generally used by the IT staff) . If the gentropy from a network object or event is above one or more thresholds, this may indicate that a user is behaving suspiciously, or that his machine has been compromised and may be trying to access restricted network resources.
- FIGURE 8 is a graph of the impact potential of newly acquired knowledge (with respect to, for example, anindividual, company, government, or human civilization) according to one or more examples of the present specification . This graph illustrates four possible scenarios, and their potential impact.
- Scenario 1 is completely new knowledge or a completely isolated event. Although this new knowledge or event may be a substantial or technological contribution to the body of knowledge of human thought, and indeed may be a seminal innovation in its field, its instantaneous impact on the entropy of the corpus as a whole is fairly modest, as it has few, if any connections. In the security field, it might be a user using an application for the first time on a machine or network.
- Scenario 2 is new knowledge that joins previously-isolated clusters. As illustrated above, this innovation may have a relatively significant impact (a "leap") on the entropy of the corpus, as in the case of a novel material with applications both in fishing lures and semiconductor manufacturing . Another example is realizing that the force that pulls an apple towards the ground and the force that keeps the Moon rotating around the Earth is one and the same, thus joining botany and astronomy and starting a scientific revolution .
- Scenario 3 is incremental knowledge in existing cluster. As illustrated by incremental innovation 720 above, this may be a relatively modest innovation that strengthens ties within a cluster, and may add a new node to the cluster. An example is the realization that phonon anharmonicity has thermodynamic importance in a material after it has been determined to be the case in a number of other materials.
- Scenario 4 is solidification of existing knowledge. This may take place, for example, after a subfield like phonon thermodynamics or hydrogen storage become mature enough that an expert can write a review article summarizing the findings over the previous decade or so. Another example is the publication of a textbook when a whole field becomes mature and widely-known, for example introductory physics or introductory chemistry. Documents like the American Constitution summarize a set of values generally held by the population of the country circa 1789.
- FIGURE 9 is a block diagram of a general communication system in the mathematical theory of communication according to one or more examples of the present specification . This figure illustrates the piece parts of the methods disclosed herein .
- Information source 990 produces a message (or sequence of messages) 920 to be communicated to the receiving terminal 940.
- the message 920 might be of various types, such as a sequence of letters in a telegraph of teletype system, or scientific knowledge in the scientific enterprise.
- Transmitter 910 operates on a message 920 in some way to produce a signal suitable for transmission over channel 930.
- patent attorneys encode technical and scientific knowledge into legal documents.
- Priests and clerics encode divinely inspired ethical and social guidelines into sacred texts and artifacts.
- the signal might be perturbed by transm itter 910, in which case the encoding of message 920 is not perfect.
- Channel 930 is the medium used to transmit the signal from transmitter to receiver. It might be a pair of wires, a coaxial cable, a patent or scientific paper, the Ark of the Covenant, gravitational waves, or any other medium . Transmission could be across time, space, or both . The signal might be perturbed by the channel 930, inducing noise.
- Receiver 940 performs the inverse operation of that done by the transmitter, reconstructing the message 920 from the signal .
- a patent attorney could reconstruct patents into technical or scientific knowledge.
- a priest could reconstruct sacred texts into ethical or social guidelines.
- a morphism reconstructs aspects of a document or relationships between documents or their attributes.
- the signal might be perturbed by receiver 940, in which case the decoding of message 920 is not perfect.
- Destination 980 is the person or thing for whom message 920 is intended .
- FIGURE 10 is a block diagram of a classification method according to one or more examples of the present specification .
- Block 1010 represents message generation and message encoding, such as when information source 990 generates a message 920 that is operated on by transmitter 910 and transmitted through channel 930. This could represent, for example, a person creating new knowledge, acquiring the services of a patent attorney to apply for and receive a patent, and the patent being loaded into a patent database such as corpus 430.
- one or more morphisms operate on a document. Morphisms act as receiver 940 and translate document data or metadata into a mathematical message graph that approximate destination 980.
- the individual message is added to the message graph of the corpus 430.
- the total information entropy of a state, H n may be computed after each addition to the corpus.
- the individual contributions might be saved for future reference in ledger 1030.
- Each addition represents a change of state in corpus 430, with each discrete state having its own entropy value. For example :
- states 1 - 3 one message each is added for each state, and S d for that individual message is computed.
- state 4 three additional documents are added . Because these documents are added as a group, and hence comprise a single message, the individual entropy for each document is not computed . Rather, S d for this state is the composite entropy contributed by the three documents together. As can be seen here, the total S d for the three documents is only 0.39 bits (whereas surrounding contributions are more on the order or 1.0 bits), so it may be decided that the individual entropy of any one of the three messages is not of any special interest.
- the message added in state 1 has an individual information entropy of 3.14 bits. This appears to be a radical innovation. However, it should be noted that the calculation of S d is not a linear calculation (as explored above). Thus, the first message in a corpus is expected to make a significant contribution to the overall entropy. Thus, this message may not be as important as it appears at first pass. Stated otherwise, the fist message to join two clusters may appear to be a radical innovation, but if that message is the third message in the corpus, and the two clusters each have only one node, and those nodes end up being densely packed in the center of a larger cluster as the corpus evolves, the contribution of message 3 may appear much less radical in hindsight.
- co-classification such as the case where the corpus includes classification codes for patents.
- This is a classification system that roughly spans the full human-knowledge space.
- Each classification code is a symbol and the universe of symbols is the alphabet of the corpus (in the information theoretical sense).
- the co-classification e.g., the connection between the set of codes assigned to a given document
- the co-classification is the message transmitted in that document.
- the message is analogous to a sentence in written language where the individual classification codes can be regarded as words with a semantic relationship, i.e., the codes are not assigned randomly but convey a fundamental attribute of the message. Relationships between primary and secondary classification codes can provide additional information.
- LSI latent semantic indexing
- each LSI topic vector is a symbol .
- the universe of topic vectors is the alphabet of the corpus.
- a co-topic (set of topics present in a document) is the message transmitted by that document.
- FIGURE 11 is an illustration of evolution of a message graph 1040 according to one or more examples of the present specification .
- the message graph starts with three nodes, named Al, Bl, and C3.
- Document 1 may be a published paper that cites three previous papers, namely Al, Bl, and C3.
- the resulting corpus graph (body of knowledge) is identical to that of the document, as this is the first document in the corpus.
- Document 2 is another academic paper, this one citing Al, B3, and D5.
- no new connections are made between Al, Bl, and C3, as Bl and C3 are not cited in this paper.
- two new nodes, B3 and D5 are introduced to the message graph, with one connection each to Al .
- This is a co- citation network.
- the weight of the edges can be conditioned, for example by an exponential decay factor that can account for the half-life of a particular message to account for, e.g ., obsoleteness. This would be particularly appropriate for a co-citation network.
- Document 3 is yet another academic paper, this one citing Al, Bl, and B3.
- Al, Bl, and B3 are all already in the message graph, and there is already an Al to B3 connection, and Bl to B3 connection .
- these links are strengthened, as one additional Al to B3 connection, and one additional Al to Bl connection is added .
- FIGURE 12 is a message graph according to one or more examples of the present specification .
- This graph shows more particularly the "occupation probability" of each node.
- This is an illustration of the stationary state of a graph in which the size of the nodes is approximately proportional to their occupation probabilities based on the number and directions of the edges.
- a probability density function with a categorical variable can be calculated from the occupation probabilities.
- the occupation probability ⁇ ( ⁇ ;) of node i in the message graph of a corpus depends on its number of edges and on its neighbors in a ripple effect. It can be calculated using any suitable algorithm, including repeated sampling taking advantage of the ergodicity of the graph, or other method that outputs a probability distribution related to the likelihood that a traverse of the graph visits a particular node.
- the set of occupation probabilities of the nodes is the probability distribution P( ) .
- the message graph information entropy (gentropy) is calculated as :
- the information entropy is maximized when the probabilities are equally distributed .
- the coin is fair, i .e., not biased towards either head or tails
- H 1 bit.
- the information entropy is zero when all the probabilities ⁇ ( ⁇ ;) are zero except for one which will have the value unity. In this case, there is no uncertainty and no information to be gained from the coin toss.
- the probabilities for each node can be calculated iteratively using an algorithm, such as brute force or PageRank, or via repeated sampling of the system .
- a node is chosen based on its occupation probability based on a previous estimation of the occupation probability density function of the graph (or completely at random if no previous estimation exists), and edges are selected at random to traverse other nodes until a cutoff number of nodes have been selected or a deadend is reached.
- the number of samples and cutoff number of traverse nodes can be optimized so that the probability density of function is approximated appropriately for the particular use with the minimum amount of computational power.
- the previous occupation probability is updated with the new statistics and the estimation is finalized. The calculation converges when the node occupation probabilities consistently change by less than a specified cutoff (the ergodic theorem ensures that the true stationary probability distribution is approximated by this empirical sampling) .
- the information entropy of an information source is the total information contained in the information source divided by the number of messages n, so it is the average amount of information per message.
- the precise nature of the information depends on the morphism .
- FIGURE 13 is an illustration of evolution of a message graph according to one or more examples of the present specification .
- the original graph 1300 includes two clusters, centered around nodes Zl and Al .
- the size of the nodes represents their occupation probabilities but are not necessarily drawn to scale.
- This graph is very lopsided (two very high probability nodes surrounded by several lower-probability nodes), so it has low entropy.
- the next message in the corpus is S1-I5.
- the result is that the graph is less lopsided, because nodes Zl and Al have decreased in probability, while SI and 15 have increased in probability.
- the overall entropy of the graph has increased by the joinder of the two nodes.
- FIGURE 14 is an illustration of evolution of a message graph according to one or more examples of the present specification .
- original graph 1400 is lopsided, with high-probability node Zl surrounded by lower probability nodes 13, SI, and V3.
- the next message in the corpus is Z1-V3-S1 as illustrated.
- the result is an even more lopsided graph, with Zl increasing in probability, and additional connections being made to V3 and S I .
- the entropy of the graph has decreased because the last message did not introduce any new connections, but rather merely reinforced existing connections.
- graph 1300 FIG. 13
- 1400 FIG. 14
- FIGURE 15 a graph of a distribution of the absolute value of gentropy scores according to one or more examples of the present specification .
- This graph illustrates a system with a power law distribution, which is common for systems with growth and with preferential attachment. These attributes are expected to be present in systems with processes in which some quantity is distributed among a number of objects according to how much of that quantity they already have. These systems are widely observed in natural and human made systems, including the Internet and the World Wide Web, social networks, citation networks, traffic systems, astronomical systems, income inequality, etc.
- This figure illustrates the technological diversity of patents in a particular industry.
- the morphism rules can be designed to score a collection of items relative to desired attributes and their relationships.
- the documents that increase diversity are those that explore unusual regions of design or knowledge space (unusual combinations of ideas) and have a positive gentropy score, while those that decrease diversity are representative of the collection of documents as a whole and have a negative gentropy score.
- Patent examiners tend to reference patents that are referenced often by other patents and other examiners because they are easier to find using traditional search methods, and because patent examiners (who may be considered experts in searching their particular fields) get to know them very well. These patents may even end up in personal bookmarks. The result is that patents that are highly referenced and more visible are more likely to be referenced by new patents than patents with few other patents referencing them. In many cases though, less well-known patents could be important or relevant, but fall into accidental obscurantism because nobody has read them or cited them in the first place, sometimes by arbitrary or random reasons, possibly due to the sheer volume of new knowledge being created.
- FIGURE 16 is a histogram of the percentile range of 99 technical documents in a particular industry that were of interest to domain experts with respect to the whole corpus of about 2,000 documents and a co-classification morphism . If the method disclosed herein lacked predictive power, or if the domain experts reviewed documents randomly, each range in the histogram would be expected to have roughly the same number of documents. However, as seen, the histogram is weighted heavily to the extreme ends, illustrating that in this embodiment, the documents with large and positive (top percentile) and large and negative (bottom percentile) gentropy scores are of particular interest. In other words, in a corpus of patents, those of most interest to a group of human experts are those that are highly innovative or matures technologies that can be readily be applied to current processes in the industry.
- FIGURE 17 is a block diagram of a classification of documents according to one or more examples of the present specification.
- a morphism 1740 is used to build a message graph 1720 according to the attributes of one or more documents and the relationships between their attributes.
- This message graph 1720 might represent knowledge space when the documents are patents.
- the gentropy score is calculated as described before and used for identification of interesting documents 1730.
- the documents identified as interesting become a new collection of documents 1710 and further refinement is possible.
- FIGURE 18 is a block diagram of the classification of documents according to one or more examples of the present specification. As this FIGURE illustrates, starting with a collection of documents 1810, a plurality of morphisms 1840 is used to build a message graph 1820 with different kinds of edges corresponding to the relationships between different kinds of attributes. This message graph might represent knowledge space when the documents are patents. The gentropy score is calculated as described before and used for identification of interesting documents 1830. The documents identified as interesting become a new collection of documents 1810 and further refinement is possible.
- FIGURE 19 is a block diagram of application of the classification of documents according to one or more examples of the present specification . As this FIGURE illustrates, starting with two or more collections of document, e.g.
- one or more morphisms e.g ., 1941, 1940, which might or might not be the same, are used to build a message graph for each collection, e.g., 1921, 1929.
- the message graph of each collection might or might not contain different kinds of edges.
- the nodes and edges of the message graphs convey relationships between different attributes of the documents.
- the gentropy score is calculated as described before for each collection of items and can be used for identification of interesting documents.
- the gentropy scores according to different morphisms can be recorded, e.g., 1961, 1969, and machine learning algorithms can be then be applied for identification of interesting documents 1930.
- FIGURE 20 is a flow chart of a method 2000 of computing and applying entropy according to one or more examples of the present specification .
- the message graph of corpus 430 in the state n is built by the application of one or more morphisms to the documents of the corpus or their attributes and the information entropy of corpus 430 in the state n, (H n ) is calculated. This is stored in ledger 2006.
- the information entropy of the message graph of corpus 430 in state n is compared to the information entropy of message graph of the corpus 430 in state n- 1 to calculate a value for S d , the information entropy contribution of the present message (document), also referred to as the gentropy.
- S d may be conditioned or otherwise adjusted based on an appropriate function, functional, or set of instructions.
- S d is multiplied by the number of documents in the corpus.
- the conditioned quantity is the exponentiation of the number of documents as the exponent and Euler's number as the base.
- the conditioned quantity is the number of documents times the inverse of the exponentiation of the difference in the number of documents between the initial state and the current state times a constant and Euler's constant as the base.
- the conditioned quantity is the gentropy score G d . As described above, this is to ensure that G d is not unfairly penalized simply because corpus 430 has grown relatively large, and that G d is not unfairly inflated simply because corpus 430 is relatively small, but it can serve other purposes as well .
- a conditioned incremental entropy score, G d is assigned to the document under consideration and stored in ledger 2020 for reference.
- the magnitude of this entropy score may indicate whether the document is of particular interest. For example, in a web search, a low G d may indicate a particularly relevant search result. In a security application, a high G d may indicate a particularly interesting event.
- the system acts on the gentropy score G d .
- acting on the score could include returning "hits" for those documents in corpus 430 that have the lowest G d .
- acting on the score could include returning "hits" for documents that have both the highest and lowest values for G d l meaning that they may be of the greatest interest to subject matter experts.
- acting on the score may include alerting a security administrator 150 that an event with a particularly high entropy has occurred, thus giving security administrator 150 an opportunity to take appropriate action . In appropriate circumstances, many other species of acting on the score are possible.
- SOC system on chip
- CPU central processing unit
- An SOC represents an integrated circuit (IC) that integrates components of a computer or other electronic system into a single chip. It may contain digital, analog, mixed- signal, and radio frequency functions: all of which may be provided on a single chip substrate.
- Other embodiments may include a multi-chip-module (MCM), with a plurality of chips located within a single electronic package and configured to interact closely with each other through the electronic package.
- MCM multi-chip-module
- the digital signal processing functionalities may be implemented in one or more silicon cores in Application Specific Integrated Circuits (ASICs), Field Programmable Gate Arrays (FPGAs), and other semiconductor chips.
- ASICs Application Specific Integrated Circuits
- FPGAs Field Programmable Gate Arrays
- Any suitably-configured processor component can execute any type of instructions associated with the data to achieve the operations detailed herein .
- Any processor disclosed herein could transform an element or an article (for example, data) from one state or thing to another state or thing.
- some activities outlined herein may be implemented with fixed logic or programmable logic (for example, software and/or computer instructions executed by a processor) and the elements identified herein could be some type of a programmable processor, programmable digital logic (for example, a field programmable gate array (FPGA), an erasable programmable read only memory (EPROM), an electrically erasable programmable read only memory (EEPROM)), an ASIC that includes digital logic, software, code, electronic instructions, flash memory, optical disks, CD-ROMs, DVD ROMs, magnetic or optical cards, other types of machine-readable mediums suitable for storing electronic instructions, or any suitable combination thereof.
- FPGA field programmable gate array
- EPROM erasable programmable read only memory
- EEPROM electrically
- processors may store information in any suitable type of non-transitory storage medium (for example, random access memory (RAM), read only memory (ROM), field programmable gate array (FPGA), erasable programmable read only memory (EPROM), electrically erasable programmable ROM (EEPROM), etc.), software, hardware, or in any other suitable component, device, element, or object where appropriate and based on particular needs.
- RAM random access memory
- ROM read only memory
- FPGA field programmable gate array
- EPROM erasable programmable read only memory
- EEPROM electrically erasable programmable ROM
- Computer program logic implementing all or part of the functionality described herein is embodied in various forms, including, but in no way limited to, a source code form, a computer executable form, and various intermediate forms (for example, forms generated by an assembler, compiler, linker, or locator).
- source code includes a series of computer program instructions implemented in various programming languages, such as an object code, an assembly language, or a high-level language such as OpenCL, Fortran, C, C++, JAVA, or HTML for use with various operating systems or operating environments.
- the source code may define and use various data structures and communication messages.
- the source code may be in a computer executable form (e.g., via an interpreter), or the source code may be converted (e.g., via a translator, assembler, or compiler) into a computer executable form.
- any number of electrical circuits of the FIGURES may be implemented on a board of an associated electronic device.
- the board can be a general circuit board that can hold various components of the internal electronic system of the electronic device and, further, provide connectors for other peripherals. More specifically, the board can provide the electrical connections by which the other components of the system can communicate electrically.
- Any suitable processors (inclusive of digital signal processors, microprocessors, supporting chipsets, etc.), memory elements, etc. can be suitably coupled to the board based on particular configuration needs, processing demands, computer designs, etc.
- Other components such as external storage, additional sensors, controllers for audio/video display, and peripheral devices may be attached to the board as plug-in cards, via cables, or integrated into the board itself.
- the electrical circuits of the FIGURES may be implemented as stand-alone modules (e.g., a device with associated components and circuitry configured to perform a specific application or function) or implemented as plug-in modules into application specific hardware of electronic devices.
- a computing apparatus for assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy, comprising : one or more logic elements, including at least one hardware logic element, comprising a classification engine to : receive the document to be added to the corpus; add the document to the corpus, creating a second temporal state of the corpus; compute a second corpus entropy for the second temporal state, based at least in part on a morphism ; and assign the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
- assigning the document a gentropy score comprises conditioning the gentropy score based at least in part on a function or functional of a number of documents in the corpus in the first temporal state.
- computing the second corpus entropy is based at least in part on a plurality of morphisms.
- classification engine is further to add a second document to the corpus, and compute a gentropy score for the second document.
- classification engine is further to act on the gentropy score.
- the corpus comprises patents, and wherein acting on the gentropy score comprises determining that the gentropy score falls within a range designated a region of interest, and identifying the document as a patent of interest.
- the corpus comprises network resources, and wherein acting on the gentropy score comprises determining that the gentropy score is higher than a threshold, and marking the document as suspicious.
- the corpus comprises employee or contract diversity data
- acting on the gentropy score comprises determining that a hiring or contracting decision with a high entropy represents a hiring of or contracting with a person or enterprise with high diversity.
- one or more tangible, non- transitory, computer-readable storage mediums for instructing a processor to provide a classification engine for assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy, the classification engine to: receive the document to be added to the corpus; add the document to the corpus, creating a second temporal state of the corpus; compute a second corpus entropy for the second temporal state, based at least in part on a morphism; and assign the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
- assigning the document a gentropy score comprises conditioning the gentropy score based at least in part on a function or functional of a number of documents in the corpus in the first temporal state.
- computing the second corpus entropy is based at least in part on a plurality of morphisms.
- classification engine is further to add a second document to the corpus, and compute a gentropy score for the second document.
- the classification engine is further to act on the gentropy score.
- the corpus comprises patents, and wherein acting on the gentropy score comprises determining that the gentropy score falls within a range designated a region of interest, and identifying the document as a patent of interest.
- the corpus comprises network resources
- acting on the gentropy score comprises determining that the gentropy score is higher than a threshold, and marking the document as suspicious.
- the corpus comprises employee or contract diversity data
- acting on the gentropy score comprises determining that a hiring or contracting decision with a high entropy represents a hiring of or contracting with a person or enterprise with high diversity.
- a method of assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy comprising : receiving the document to be added to the corpus; adding the document to the corpus, creating a second temporal state of the corpus; computing a second corpus entropy for the second temporal state, based at least in part on a morphism; and assigning the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
- assigning the document a gentropy score comprises conditioning the gentropy score based at least in part on a function or functional of a number of documents in the corpus in the first temporal state.
- classification engine is further to add a second document to the corpus, and compute a gentropy score for the second document.
- the classification engine is further to act on the gentropy score.
- the corpus comprises patents
- acting on the gentropy score comprises determining that the gentropy score falls within a range designated a region of interest, and identifying the document as a patent of interest.
- a computing apparatus for assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy comprising : one or more logic elements, including at least one hardware logic element, comprising a classification engine to: receive the document to be added to the corpus; create a message graph of the first corpus using at least a first morphism including a conditioning of weight of individual edges of the message graph; calculate a gentropy of the corpus in the first temporal state comprising calculating a first node occupation probability density function of the message graph; receiving the document to be added to the first corpus, creating a second temporal state of the corpus having a second corpus information entropy; create a message graph of the second temporal state using at least a second morphism including a conditioning of weight of individual edges of the message graph; calculate a second gentropy of the corpus in the second temporal state comprising calculating
- classification engine is further to apply machine learning techniques to classify the document according to the gentropy score.
- a method of assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy comprising : receiving the document to be added to the corpus; creating a message graph of the first corpus using at least a first morphism including a conditioning of weight of individual edges of the message graph; calculating a gentropy of the corpus in the first temporal state comprising calculating a first node occupation probability density function of the message graph; receiving the document to be added to the first corpus, creating a second temporal state of the corpus having a second corpus information entropy; creating a message graph of the second temporal state using at least a second morphism including a conditioning of weight of individual edges of the message graph; calculating a second gentropy of the corpus in the second temporal state comprising calculating a second node occupation probability density function of the message graph; and assigning the gentrop
- classification engine is further to apply machine learning techniques to classify the document according to the gentropy score.
- An apparatus comprising means for performing the method of any of the preceding examples.
- the means comprise a processor and a memory.
- the means comprise a machine readable medium having stored thereon executable instructions to instruct a processor to perform the method.
- the apparatus is a computing system.
- the means comprise a processor and a memory.
- the means comprise one or more tangible, non-transitory computer-readable storage mediums.
- the apparatus is a computing device.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Business, Economics & Management (AREA)
- Software Systems (AREA)
- Human Resources & Organizations (AREA)
- Databases & Information Systems (AREA)
- Mathematical Physics (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Strategic Management (AREA)
- Entrepreneurship & Innovation (AREA)
- General Business, Economics & Management (AREA)
- Computer Hardware Design (AREA)
- Computing Systems (AREA)
- Computer Security & Cryptography (AREA)
- Operations Research (AREA)
- Mathematical Analysis (AREA)
- Pure & Applied Mathematics (AREA)
- Mathematical Optimization (AREA)
- Computational Mathematics (AREA)
- Economics (AREA)
- Marketing (AREA)
- Quality & Reliability (AREA)
- Tourism & Hospitality (AREA)
- Medical Informatics (AREA)
- Computational Linguistics (AREA)
- Algebra (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Multimedia (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
There is disclosed in an example a computing apparatus for assigning an entropy score to a document to be added to a corpus in a first temporal state having a first corpus entropy, having one or more logic elements, including at least one hardware logic element, providing a classification engine to: receive the document to be added to the corpus; add the document to the corpus, creating a second temporal state of the corpus; compute a second corpus entropy for the second temporal state, based at least in part on a morphism; and assign the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
Description
ENTROPIC CLASSIFICATION OF OBJECTS
Field of the specification
[0001] This disclosure relates in general to the field of machine learning, and more particularly, though not exclusively to, a system and method for providing entropic classification of objects.
Background
[0002] Machine-driven classification of objects, such as web pages, text documents, or multimedia content, is useful in contexts such as security and productivity.
[0003] Classifiers are a category of tool in the analytics arsenal that help divide input data into one or more classes. For example, email spam filter classifiers decide what incoming email is classified as "spam" and what is classified as "ham ." A classifier with a binary result ("yes," the object is in the target class, or "no," the object is not in the target class) may be referred to as a binary classifier. A classifier that designates an object as belonging to one of several potential classes may be referred to as a multinominal classifier.
Brief Description of the Drawings
[0004] The present disclosure is best understood from the following detailed description when read with the accompanying figures. It is emphasized that, in accordance with the standard practice in the industry, various features are not necessarily drawn to scale, and are used for illustration purposes only. Where a scale is shown, explicitly or implicitly, it provides only one illustrative example. In other embodiments, the dimensions of the various features may be arbitrarily increased or reduced for clarity of discussion.
[0005] FIGURE 1 is a block diagram of a security-enabled network according to one or more examples of the present specification.
[0006] FIGURE 2 is a block diagram of a computing device according to one or more examples of the present specification .
[0007] FIGURE 3 is a block diagram of a server according to one or more examples of the present specification.
[0008] FIGURE 4 is a block diagram of a classification system according to one or more examples of the present specification.
[0009] FIGURE 5 is an illustration of analysis of an issued patent according to one or more examples of the present specification.
[0010] FIGURE 6 is a workflow diagram of document classification according to one or more examples of the present specification.
[0011] FIGURE 7 is a message graph according to one or more examples of the present specification.
[0012] FIGURE 8 is a graph of impact potential according to one or more examples of the present specification.
[0013] FIGURE 9 is a block diagram of a mathematical theory of communication in information theory, according to one or more examples of the present specification.
[0014] FIGURE 10 is a block diagram of a classification method according to one or more examples of the present specification.
[0015] FIGURE 11 is an illustration of evolution of a message graph according to one or more examples of the present specification.
[0016] FIGURE 12 is a message graph according to one or more examples of the present specification.
[0017] FIGURE 13 is an illustration of evolution of a message graph according to one or more examples of the present specification
[0018] FIGURE 14 is an illustration of evolution of a message graph according to one or more examples of the present specification
[0019] FIGURE 15 a graph of a distribution of gentropy scores according to one or more examples of the present specification
[0020] FIGURE 16 is a histogram of the percentile range of labeled data with respect to a whole data set according to one or more examples of the present specification.
[0021] FIGURE 17 is a block diagram of a classification process according to one or more examples of the present specification.
[0022] FIGURE 18 is a block diagram of a classification process according to one or more examples of the present specification.
[0023] FIGURE 19 is a block diagram of a classification process according to one or more examples of the present specification.
[0024] FIGURE 20 is a flow chart of a method of computing and applying gentropy scores according to one or more examples of the present specification .
Summary
[0025] There is disclosed in an example a computing apparatus for assigning an entropy score to a document to be added to a corpus in a first temporal state having a first corpus entropy, having one or more logic elements, including at least one hardware logic element, providing a classification engine to: receive the document to be added to the corpus; add the document to the corpus, creating a second temporal state of the corpus; compute a second corpus entropy for the second temporal state, based at least in part on a morphism; and assign the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
Embodiments of the Disclosure
[0026] The following disclosure provides many different embodiments, or examples, for implementing different features of the present disclosure. Specific examples of components and arrangements are described below to simplify the present disclosure. These are, of course, merely examples and are not intended to be limiting. Further, the present disclosure may repeat reference numerals and/or letters in the various examples. This repetition is for the purpose of simplicity and clarity and does not in itself dictate a relationship between the various embodiments and/or configurations discussed. Different embodiments may have different advantages, and no particular advantage is necessarily required of any embodiment.
[0027] Classification of documents and other objects is an important task for an enterprise. Such classification can have uses for security, privacy, litigation, and many other purposes.
[0028] In one nonlimiting example, an enterprise may have firewall rules configured to treat certain classes of websites differently. Business-essential websites (such as corporate websites, business information, partner sites, and similar) may be always allowed. Non-essential but harmless websites (such as news, general information, and e-commerce) may be allowed but logged. Time- wasting websites (such as sports, games, most social media, fashion, or other non-business-related subject matters) may be allowed only with explicit user input. Malicious websites (such as malware, or websites with illegal or unethical content) may be blocked and permissible only with special administrative dispensation.
[0029] Because there are many millions of websites, it may be impractical to have a human user manually classify each website. Rather, one or more machine classifiers may be employed to classify each website, and then permissions may be applied en grosse to each class. Thus, machine learning can be used to help deploy and enforce enterprise security and usage policies.
[0030] Similarly, a machine classifier may be used to assign ownership or other attributes to business documents. For example, an enterprise may deploy a database file system, in which files must be assigned attributes, such as which group owns each document, which client or job number a document belongs to, or similar attributes. The classifier may use existing metadata fields, file locations, and content to assign these attributes (each constituting a "class") to each document.
[0031] In some cases, classification is mutually exclusive. For example, when assigning ownership to a document, enterprise constraints may dictate that only one group can own the document. Thus, the classes are mutually exclusive (e.g., the owner can be either "Business" or "Marketing," but not both). In other cases, classes may be more analogous to "tags," and thus may be non-exclusive. For example, a popular sports website that also frequently features scantily-clad women and men in skimpy swimsuits, may be tagged as both "sports" and "soft pornography." In the latter case, the site may inherit the permission set of its most- restricted tag.
[0032] In another example, machine classification may be used to mitigate the expense of electronic production for litigation . A binary classifier may scan a
large corpus of documents, and tag each one as "relevant" or "not relevant" to the litigation. Additional classifiers may be used, for example, to tag documents as "privileged," "attorneys' eyes only," or as potential trial exhibits. Even if a human user is required to verify the classifications, having the initial tagging in place can significantly speed the process.
[0033] Embodiments of the present specification deal with the problem of information overload, which occurs when the amount of input to a system exceeds its processing capacity. For example, a web search might find millions of relevant results, or there might be tens of thousands of relevant technical articles in a given field. In those cases, it is implausible for a cognitive observer to review all the available data and to choose a best fit.
[0034] In the field of web searches, first-generation search engines such as the "World Wide Web Worm" relied on raw keyword matching to find information, but they typically returned too many low quality matches to be very useful to general users. Later, Google combined keyword matching with the "PageRank" algorithm, which ranks websites, inter alia based on the number of links to and from the page. These improvements in ranking results have resulted in significantly better results returned to users, though this approach is still limited by the user's knowledge of the appropriate keywords to use.
[0035] In the field of "recommenders," a recommendation engine may at least partially offset the limitations of keyword searching. Recommendation engines are used by services like Pandora, last.fm, Google Music, Netflix, Youtube, Vimeo, and many others to recommend media such as movies, music, and books based on the user's previous selections and the selections of users with similar profiles, and stated preferences. But even these systems, like keyword-based search systems, may lack information diversity in certain embodiments. This can become particularly important when the elements being searched contain highly redundant information. Taking search results as an example, higher diversity of returned results allows the system to present a wider span of information in the first few hits, or for recommender systems to present a broader selection of recommendations, some of them even surprising to the user but still useful .
[0036] Thus, embodiments of the methods described in this specification can be used to rank a collection of items according to how similar or dissimilar they
are compared to an existing collection, based at least in part on one or more morphisms, or a combination of morphisms. Morphisms are physical or conceptual relationship schemes between attributes of the items or the items themselves. In information theory, these items, their attributes, or their relationships could be thought of as the result of a transmitter operating on a message.
[0037] Throughout this specification, each object or item in the data set is called a "document." However, the term "document" should not be misconstrued to be limited merely to discrete units of paginated textual data. Rather, it is a message produced by an information source (from an information theoretic point of view). A document in this context includes any object or datum that may be represented or modeled digitally, including its attributes and metadata where appropriate. Thus, "documents" may be text documents, web pages, books, movies, music tracks, sounds, periodicals, articles, businesses, stocks, living organism (plants, animals, or people), network objects, network addresses, users of a network, executable files, astronomical objects, weather events, or weather patterns, to name just a few nonlimiting examples.
[0038] A collection of documents being managed by the system may be referred to as the "corpus." In an initial "null corpus" state, the corpus has zero documents. When the first document is added to the corpus (including any characterization or classification performed on the document), the corpus enters a second state. Thereafter, with the addition of each new document or group of documents, the corpus moves into a new state. With each state, the corpus may have a total information entropy H, the computation of which is detailed below. The total information entropy of a particular state may be notated as Hn, where n is the state number. Furthermore, with the addition of each new document, a change in information entropy occurs. The change in information entropy caused by the addition of a new document d to the corpus may be represented as Sd.
[0039] In evaluating the corpus, it is straightforward to include user rankings or multiple scoring criteria. The methods of the present specification may be used to identify items that a user will likely find useful, either because they are representative of the collection (e.g., exemplars), or because they are unusual combinations (e.g., outliers). Depending on the use case, identifying exemplars may be most desirable (e.g., review articles, where exemplars may represent the
best summarizations of field), or identifying outliers may be most desirable (e.g . , network security, where outl iers may represent a security risk) .
[0040] A system of the present specification may quantitatively measure the graph entropy (the "gentropy") of a mathematical graph where the nodes and edges are assigned according to a particu lar morphism or combination of morphisms and based on the documents of a corpus, and thus be able to calcu late the change in gentropy due to actual or potential new additions. The graph such constructed is a "message graph" since in an information theoretic framework, the information contained in each relationship between attributes is a message . In the same framework, a morphism is akin to a receiver that reconstructs certain aspects of the message . One interpretation of the gentropy is as a measure of diversity. For exam ple, in ecology, ecosystems with more an i mal species and about the same num ber of animals belonging to each species are more diverse than ecosystems with fewer species or those in wh ich most an imals belong to a few species.
[0041] In appropriate em bodiments, the value of Sd may be conditioned to account for corpus size. In other words, i n a corpus of three documents, the addition of a fourth document may be expected to have a very substantial effect on the overall entropy of the corpus, not because the document is very significant, but because the corpus is qu ite small . On the other hand, in a corpus of one m il lion documents, the addition of a one-mi ll ion-and-first document will be expected to have only a smal l effect on the total entropy, not because the document is unimportant, but because the corpus is very large. Additionally, certain attribute relationships m ight be more important that others. Thus, there are also disclosed herein methods for conditioning the value of Sd to more accurately reflect the importance of a particular document to the corpus. The "gentropy score" Gd of document d is its conditioned Sd . The methods disclosed herein may be appl ied in a variety of cases, and may be used to predict items that a user wi ll find the most interesting from a large corpus.
[0042] In certain em bodiments, different rules may be designed for different purposes, and then com bined . For example, it is observed herein that documents considered interesting by subject matter experts tended to significantly increase
the information entropy (they were unusual), or to significantly decrease it (they were strongly representative of the collection).
[0043] One application of the present method is to aid in the identification of technical documents that a user might find interesting . This use case is not intended to be limiting, but was used as a validation test of the model disclosed herein. In the test case, document metadata (i.e., not document content) were used to flag potentially relevant materials from a large collection of similar documents. This test case illustrates an advantage of the present method, because neither keywords (as used by a search engine), nor past user behavior or preferences (as used by recommendation engines) are necessary, though keywords and past behavior can be integrated with the present methods in appropriate embodiments.
[0044] The test implementation was based on classification of contents, and was intended to identify items similar to the collection, as well as items that are unusual but still relevant. The disclosed method was used to score each item in a collection of about 14,000 technical documents belonging broadly to the same industry and published periodically since 1976. The gentropy score for each document measured the technological diversity of each document with respect to the rest of the documents published up to that point, and is based on a classification assigned to it upon publication .
[0045] The classification scheme in this case is a human-created taxonomy with several thousand unique categories. A publicly-available publication, generally considered reputable, and maintained by a group of domain experts was consulted. The experts who maintain the website read newly-published documents in a particular industry, and provide a digest with commentary. They publish additional commentary about documents that may be of particular interests to industry groups.
[0046] For example, in a period of about 10 months, the experts wrote special commentaries on approximately 100 newly-published documents, out of about 2,000 total (i.e., about 5%). As illustrated in FIGURE 16 below, 61 of those documents were either in the bottom or top 14th percentile. The documents in the bottom 14th percentile had large and negative gentropy scores, decreasing the overall content diversity because their content was representative of the collection
as a whole, whereas those in the top 14th percentile had large and positive gentropy scores, increasing diversity because their content is new or is an unusual combination of existing content. The domain experts wrote disproportionally about those documents that were outliers accordi ng to the present method (i .e ., either in the upper or lower percenti les) .
[0047] This specification provides a method to calculate the information entropy of the message graph of a corpus (i .e. , the gentropy) using the content or metadata of each document (i .e . , item of any type) along with a morphism (a structure-preserving map from one mathematical structure to another) . The combination of these methods measures the relative diversity of particular items of the same kind (i .e ., how much new information each provides. ) The conditioned entropy, the gentropy score, may be used to identify noteworthy documents in a corpus, and to make recom mendations to a user.
[0048] In an em bodiment, the "Shannon information entropy" (named for Claude Shannon, considered "the father of information theory") is used . The Shannon entropy H of a discreet random variable X taken from a finite sam ple is
[0049] Where Ρ(χέ) is the probabil ity distribution function and H is measured in bits. With the appropriate morphism or combination of morphisms, X can parameterize more than one scoring aspect as explained below. These aspects could include factors like content and user ranking . I n this embodiment, Ρ(χ;) is the stationary occupation probabi lity of each node based at least in part on the number of directed or undirected edges and the occupation probabi lities of neighboring nodes. This concept is i llustrated in FIGU RE 12, where node A has a large occupation probabil ity because almost all other nodes point to it. Node G has a relatively large occupation probabil ity because node A points to it (even though only node A points to it) .
[0050] A morphism is a structure-preserving map from one mathematical structure to another, i .e. , a rule or set of rules that transforms data but preserves certain relationships or properties of the data . The critical properties to be preserved are growth and preferential attachment (or other non-trivial topological features), which results in a graph that has a degree distribution that follows a
power law. Without preferential attachment or other non-trivial topological features, the results of the present method could be inconclusive. However, scale- freeness is a characteristic of complex systems, and most networks of interest or other systems that can be mapped into a network are scale-free and show power law behavior, such as social networks and the world wide web (WWW) . The rule that we used for the results illustrated herein is that for every document, a node was added to the graph (if the node did not previously exist), corresponding to each classification code assigned to the document, and edges were created between them (so it was an undirected multigraph based on the co-classification network of the documents whose stationary probability was calculated and updated after each step) . Gd was calculated for each document. As illustrated in FIGURE 15, the power law behavior of the system is evidenced by the linearity in the log-log plot of the gentropy score distribution of the documents that increased diversity. Of these documents, about 4,000 contributed 0.1 bits or less while a handful contributed more than 10 bits.
[0051] In an example application, the method disclosed herein may be used with a collection of items constructed with the results of a keyword search, user behavior, or similar. The user may then elect whether to look at items that are representative of the collection or those that are more unusual or span more information, based on a gentropy score ranking .
[0052] In another example, a user may explore which items, not already in a collection, would change the collection's diversity the most. This can be used to provide recommendations for movies, books, music, and other media. Exploration could be manual or automatic. Items that decrease diversity are useful in some situations, for example in a collection of books with two subgroups, one on Java and Scala, and the other one on C and C++, yet another book on Java and Scala might be exactly what the user is looking for. Items that increase diversity (and the degree to which they do) might be useful in other situations, for example a book on all four languages will increase diversity more than one on Java and C++, and the user might be interested in acquiring different knowledge. A book on an esoteric programing language such as Brainfuck or LOLCODE that is not part of the collection increases its diversity, but to a lesser (distinguishably) degree, and the user might select books in this range.
[0053] In yet another example application, the method can be used to monitor compliance with anti-discrimination regulations or other hiring criteria. For example, the method may be used to assign a gentropy score (i.e., a diversity score) to a hiring decision, which may be orthogonal to other hiring criteria. If hiring decisions group strongly around a single gender, ethnicity, skill set, work experience, age or other factor, the hiring may be very non-diverse. This is desirable in some cases, such as when more computer programmers versed in a particular computer language are required to complete a project on time, but undesirable in others. A hiring decision that involves a person from a technical or life background that is less commonly hired may be considered "more diverse." This could be desirable, for example, when having an ethnically diverse team of product designers is required to penetrate new markets. Records can be used to measure diversity trends over time, and to identify "outliers" that are substantially different from "business as usual."
[0054] A system and method for binary linear classification will now be described with more particular reference to the attached FIGURES. It should be noted that throughout the FIGURES, certain reference numerals may be repeated to indicate that a particular device or block is wholly or substantially consistent across the FIGURES. This is not, however, intended to imply any particular relationship between the various embodiments disclosed. In certain examples, a genus of elements may be referred to by a particular reference numeral ("widget 10"), while individual species or examples of the genus may be referred to by a hyphenated numeral ("first specific widget 10-1" and "second specific widget 10- 2").
[0055] FIGURE 1 is a network-level diagram of an enterprise 100 receiving classification services according to one or more examples of the present specification. This figure illustrates one nonlimiting example of using classification to provide enterprise-class spam filtering on a network. This example is intended to be nonlimiting, and other examples are provided herein to illustrate the broad applicability of the methods disclosed herein.
[0056] In the example of FIGURE 1, one or more users 120 operate one or more client devices 110. Each device may include an appropriate operating system, such as Microsoft Windows, Linux, Android, Mac OSX, Apple iOS, Unix, or
similar. Some of the foregoing may be more often used on one type of device than another. For example, desktop computers or engineering workstation may be more likely to use one of Microsoft Windows, Linux, Unix, or Mac OSX. Laptop computers, which are usually a portable off-the-shelf device with fewer customization options, may be more likely to run Microsoft Windows or Mac OSX. Mobile devices may be more likely to run Android or iOS. However, these examples are not intended to be limiting.
[0057] Client devices 110 may be communicatively coupled to one another and to other network resources via enterprise network 170. Enterprise network 170 may be any suitable network or combination of one or more networks operating on one or more suitable networking protocols, including for example, a local area network, an intranet, a virtual network, a wide area network, a wireless network, a cellular network, or the Internet (optionally accessed via a proxy, virtual machine, or other similar security mechanism) by way of nonlimiting example. Enterprise network 170 may also include one or more servers, firewalls, routers, switches, security appliances, antivirus servers, or other useful network devices, which in an example may be virtualized within workload cluster 142. In this illustration, enterprise network 170 is shown as a single network for simplicity, but in some embodiments, enterprise network 170 may include a large number of networks, such as one or more enterprise intranets connected to the internet. Enterprise network 170 may also provide access to an external network, such as the Internet, via external network 172. External network 172 may similarly be any suitable type of network.
[0058] A workload cluster 142 may be provided, for example as a virtual cluster running in a hypervisor on a plurality of rack-mounted blade servers, or as a cluster of physical servers. Workload cluster 142 may provide one or more server functions, or one or more "microclouds" in one or more hypervisors. For example, a virtualization environment such as vCenter may provide the ability to define a plurality of "tenants," with each tenant being functionally separate from each other tenant, and each tenant operating as a single-purpose microcloud. Each microcloud may serve a distinctive function, and may include a plurality of virtual machines (VMs) of many different flavors, including agentful and agentless VMs.
[0059] It should also be noted that some functionality of endpoint devices 110 may also be provided via workload cluster 142. For example, one microcloud may provide a remote desktop hypervisor such as a Citrix workspace, which allows users 120 operating endpoints 110 to remotely login to a remote enterprise desktop and access enterprise applications, workspaces, and data. In that case, endpoint 120 could be a "thin client" such as a Google Chromebook, running only a stripped-down operating system, and still provide user 120 useful access to enterprise resources.
[0060] In an example, one or more virtual machines are provisioned within workload cluster 142 to act as a machine classifier 190. Machine classifier 190 may be tasked with classification of network objects, including static and/or dynamic objects, for any suitable purpose, such as identifying spam and malware. For example, classification may be provided for enforcement of enterprise security and network usage policies, such as by classifying websites into various categories, each with different security settings or permissions. Similarly, machine classifier 190 may classify text documents for storage in a database file system. In another nonlimiting example, the machine classifier may be tasked with indexing content to ease searching and compilation, such as indexing patents by subject matter. In yet another example, network objects are classified for security purposes, and assigned to a classification such as "white" (permitted), "gray" (unknown), or "black" (barred) as part of an antivirus or antimalware scheme. In the latter case, the objects may be any network object as described herein, including static objects, active objects, and network-attached devices. These classifications may be based, for example, on identifying objects that introduce high entropy into the system, or alternately, that group strongly around other known objects of the same type. These embodiments are provided as nonlimiting examples only, and should be understood to be non-exclusive. Many other uses for object classification are possible, and are within the intended scope of this specification.
[0061] One or more computing devices configured as a management console 140 may also operate on enterprise network 170. Management console 140 may provide a user interface for a security administrator 150 to define enterprise security policies, which management console 140 may enforce on enterprise
network 170 and across client devices 110 and workload cluster 142. In an example, management console 140 may run a server-class operating system, such as Linux, Unix, or Windows Server. In another case, management console 140 may be provided as a web interface, on a desktop-class machine, or via a VM provisioned within workload cluster 142. Security administrator 150 or user 120 may use either management console 140 or client device 110 to "train" a data set, either interactively, or by submitting a batch of pre-classified documents.
[0062] Enterprise 100 may communicate across enterprise boundary 104 with external network 172. Enterprise boundary 104 may represent a physical, logical, or other boundary. External network 172 may include, for example, websites, servers, network protocols, and other network-based services. In one example, an attacker 180 (or other similar malicious or negligent actor) also connects to external network 172.
[0063] It may be a goal of users 120 and enterprise 100 to successfully operate client devices 110 and workload cluster 142 without interference from attacker 180 or from unwanted security objects. In one example, attacker 180 is a malware or spam author whose goal or purpose is to cause malicious harm or mischief, for example by injecting malicious object 182 (e.g., spam or malware) into client device 110. Once malicious object 182 gains access to client device 110, it may try to perform work such as social engineering of user 120, a hardware- based attack on client device 110, modifying storage, modifying client application 112, or gaining access to enterprise servers 142.
[0064] Attacker 180 may also want to commit industrial or other espionage against secured enterprise 100, such as stealing classified or proprietary data, stealing identities, or gaining unauthorized access to enterprise resources. Thus, attacker 180's strategy may also include trying to gain physical access to one or more client devices 110 and operating them without authorization, so that an effective security policy may also include provisions for preventing such access.
[0065] In another example, enterprise 100 may simply be a family, with parents assuming the role of security administrator 150. The parents may wish to protect their children from undesirable content, such as pornography, adware, spyware, age-inappropriate content, advocacy for certain political, religious, or social movements, or forums for discussing illegal or dangerous activities, by way
of non-limiting example. In this case, the parent may perform some or all of the duties of security administrator 150.
[0066] When a new object is first encountered on the network, security policies may initially treat it as "gray" or "suspect." As a first line of defense, machine classifier 190 may inspect the object and attempt to classify it. For example, if the object is an e-mail, machine classifier 190 may attempt to determine whether it is "ham" or "spam ." If the object is executable, machine classifier 190 may attempt to determine whether it is "malware" or "not malware," or whether it is a "permissible application" or "impermissible application . Again, these classifications may be achieved because an object either has high entropy, or groups strongly with other similar objects. Advantageously, this classification may be achieved via metadata, without the need to inspect the content of the object.
[0067] FIGURE 2 is a block diagram of client-class computing device 200 according to one or more examples of the present specification. Computing device 200 may be any suitable computing device. In various embodiments, a "computing device" may be or comprise, by way of non-limiting example, a computer, workstation, server, mainframe, virtual machine (whether emulated or on a "bare- metal" hypervisor), embedded computer, embedded controller, embedded sensor, personal digital assistant, laptop computer, cellular telephone, IP telephone, smart phone, tablet computer, convertible tablet computer, computing appliance, network appliance, receiver, wearable computer, handheld calculator, or any other electronic, microelectronic, or microelectromechanical device for processing and communicating data. Any computing device may be designated as a host on the network. Each computing device may refer to itself as a "local host," while any computing device external to it may be designated as a "remote host."
[0068] In certain embodiments, client device 110 may be an example of computing device 200. Users 120 may use computing device 200 to interact with a classified data set, or to perform tasks such as performing interactive training or submitting batches of pre-classified documents.
[0069] Computing device 200 includes a processor 210 connected to a memory 220, having stored therein executable instructions for providing an operating system 222 and at least software portions of a security engine 224.
Other components of client device 200 include a storage 250, network interface 260, and peripheral interface 240. This architecture is provided by way of example only, and is intended to be non-exclusive and non-limiting. Furthermore, the various parts disclosed are intended to be logical divisions only, and need not necessarily represent physically separate hardware and/or software components. Certain computing devices provide main memory 220 and storage 250, for example, in a single physical memory device, and in other cases, memory 220 and/or storage 250 are functionally distributed across many physical devices. In the case of virtual machines or hypervisors, all or part of a function may be provided in the form of software or firmware running over a virtualization layer to provide the disclosed logical function. In other examples, a device such as a network interface 260 may provide only the minimum hardware interfaces necessary to perform its logical operation, and may rely on a software driver to provide additional necessary logic. Thus, each logical block disclosed herein is broadly intended to include one or more logic elements configured and operable for providing the disclosed logical operation of that block. As used throughout this specification, "logic elements" may include hardware, external hardware (digital, analog, or mixed-signal), software, reciprocating software, services, drivers, interfaces, components, modules, algorithms, sensors, components, firmware, microcode, programmable logic, or objects that can coordinate to achieve a logical operation.
[0070] In an example, processor 210 is communicatively coupled to memory 220 via memory bus 270-3, which may be for example a direct memory access (DMA) bus by way of example, though other memory architectures are possible, including ones in which memory 220 communicates with processor 210 via system bus 270-1 or some other bus. Processor 210 may be communicatively coupled to other devices via a system bus 270-1. As used throughout this specification, a "bus" includes any wired or wireless interconnection line, network, connection, bundle, single bus, multiple buses, crossbar network, single-stage network, multistage network or other conduction medium operable to carry data, signals, or power between parts of a computing device, or between computing devices. It should be noted that these uses are disclosed by way of non-limiting example
only, and that some embodiments may omit one or more of the foregoing buses, while others may employ additional or different buses.
[0071] In various examples, a "processor" may include any combination of logic elements operable to execute instructions, whether loaded from memory, or implemented directly in hardware, including by way of non-limiting example a microprocessor, digital signal processor, field-programmable gate array, graphics processing unit, programmable logic array, application-specific integrated circuit, or virtual machine processor. In certain architectures, a multi-core processor may be provided, in which case processor 210 may be treated as only one core of a multi-core processor, or may be treated as the entire multi-core processor, as appropriate. In some embodiments, one or more co-processors may also be provided for specialized or support functions.
[0072] Processor 210 may be connected to memory 220 in a DMA configuration via DMA bus 270-3. To simplify this disclosure, memory 220 is disclosed as a single logical block, but in a physical embodiment may include one or more blocks of any suitable volatile or non-volatile memory technology or technologies, including for example DDR RAM, SRAM, DRAM, cache, LI or L2 memory, on-chip memory, registers, flash, ROM, optical media, virtual memory regions, magnetic or tape memory, or similar. In certain embodiments, memory 220 may comprise a relatively low-latency volatile main memory, while storage 250 may comprise a relatively higher-latency non-volatile memory. However, memory 220 and storage 250 need not be physically separate devices, and in some examples may represent simply a logical separation of function. It should also be noted that although DMA is disclosed by way of non-limiting example, DMA is not the only protocol consistent with this specification, and that other memory architectures are available.
[0073] Storage 250 may be any species of memory 220, or may be a separate device. Storage 250 may include one or more non-transitory computer- readable mediums, including by way of non-limiting example, a hard drive, solid- state drive, external storage, redundant array of independent disks (RAID), network-attached storage, optical storage, tape drive, backup system, cloud storage, or any combination of the foregoing. Storage 250 may be, or may include therein, a database or databases or data stored in other configurations, and may
include a stored copy of operational software such as operating system 222 and software portions of security agent 224. Many other configurations are also possible, and are intended to be encompassed within the broad scope of this specification.
[0074] Network interface 260 may be provided to communicatively couple client device 200 to a wired or wireless network. A "network," as used throughout this specification, may include any communicative platform operable to exchange data or information within or between computing devices, including by way of non- limiting example, an ad-hoc local network, an internet architecture providing computing devices with the ability to electronically interact, a plain old telephone system (POTS), which computing devices could use to perform transactions in which they may be assisted by human operators or in which they may manually key data into a telephone or other suitable electronic equipment, any packet data network (PDN) offering a communications interface or exchange between any two nodes in a system, or any local area network (LAN), metropolitan area network (MAN), wide area network (WAN), wireless local area network (WLAN), virtual private network (VPN), intranet, or any other appropriate architecture or system that facilitates communications in a network or telephonic environment.
[0075] Security agent 224, in one example, is operable to carry out computer-implemented methods as described in this specification. Security engine 224 may include one or more tangible non-transitory computer-readable mediums having stored thereon executable instructions operable to instruct a processor to provide a security engine 224. As used throughout this specification, an "engine" includes any combination of one or more logic elements, of similar or dissimilar species, operable for and configured to perform one or more methods provided by the engine. Thus, security engine 224 may comprise one or more logic elements configured to provide methods as disclosed in this specification . In some cases, security engine 224 may include a special integrated circuit designed to carry out a method or a part thereof, and may also include software instructions operable to instruct a processor to perform the method. In some cases, security engine 224 may run as a "daemon" process. A "daemon" may include any program or series of executable instructions, whether implemented in hardware, software, firmware, or any combination thereof, that runs as a background process, a terminate-and-
stay-resident program, a service, system extension, control panel, bootup procedure, BIOS subroutine, or any similar program that operates without direct user interaction. In certain embodiments, daemon processes may run with elevated privileges in a "driver space," or in ring 0, 1, or 2 in a protection ring architecture. It should also be noted that security engine 224 may also include other hardware and software, including configuration files, registry entries, and interactive or user-mode software by way of non-limiting example.
[0076] In one example, security engine 224 includes executable instructions stored on a non-transitory medium operable to perform a method according to this specification. At an appropriate time, such as upon booting client device 200 or upon a command from operating system 222 or a user 120, processor 210 may retrieve a copy of the instructions from storage 250 and load it into memory 220. Processor 210 may then iteratively execute the instructions of security engine 224 to provide the desired method.
[0077] In certain embodiments, security engine 224 may be provisioned on client device 200 to enforce enterprise security, resource usage, or other policies. For example, when a user 120 directs client device 200 to a particular website, security engine 224 may query a server engine for an appropriate class and associated permissions. The machine may then locally enforce those permissions on client device 200. Enforcement may also be performed in whole or in part by server devices 300 (FIGURE 3). In another example, security engine 224 may mark e-mails as "spam" or "ham." In yet another example, security engine 224 may simply provide an interface for accessing documents or objects that have been classified by classification engine 326.
[0078] Peripheral interface 240 may be configured to interface with any auxiliary device that connects to client device 200 but that is not necessarily a part of the core architecture of client device 200. A peripheral may be operable to provide extended functionality to client device 200, and may or may not be wholly dependent on client device 200. In some cases, a peripheral may be a computing device in its own right. Peripherals may include input and output devices such as displays, terminals, printers, keyboards, mice, modems, data ports (e.g., serial, parallel, USB, Firewire, or similar), network controllers, optical media, external storage, sensors, transducers, actuators, controllers, data acquisition buses,
cameras, microphones, speakers, or external storage by way of non-limiting example.
[0079] In one example, peripherals include display adapter 242, audio driver 244, and input/output (I/O) driver 246. Display adapter 242 may be configured to provide a human-readable visual output, such as a command-line interface (CLI) or graphical desktop such as Microsoft Windows, Apple OSX desktop, or a Unix/Linux X Window System-based desktop. Display adapter 242 may provide output in any suitable format, such as a coaxial output, composite video, component video, VGA, or digital outputs such as DVI or HDMI, by way of nonlimiting example. In some examples, display adapter 242 may include a hardware graphics card, which may have its own memory and its own graphics processing unit (GPU). Audio driver 244 may provide an interface for audible sounds, and may include in some examples a hardware sound card. Sound output may be provided in analog (such as a 3.5mm stereo jack), component ("RCA") stereo, or in a digital audio format such as S/PDIF, AES3, AES47, HDMI, USB, Bluetooth or Wi-Fi audio, by way of non-limiting example.
[0080] FIGURE 3 is a block diagram of a server-class device 300 according to one or more examples of the present specification. Server 300 may be any suitable computing device, as described in connection with FIGURE 2. In general, the definitions and examples of FIGURE 2 may be considered as equally applicable to FIGURE 3, unless specifically stated otherwise. Server 300 is described herein separately to illustrate that in certain embodiments, logical operations according to this specification may be divided along a client-server model, wherein client device 200 provides certain localized tasks, while server 300 provides certain other centralized tasks. In contemporary practice, server 300 is more likely than client device 200 to be provided as a "headless" VM running on a computing cluster, or as a standalone appliance, though these configurations are not required.
[0081] In an example, any of the devices in workload cluster 142, including machine classifier 190, may be a server-class device 300. Server-class device 300 is disclosed as a separate device from client device 200 to illustrate one logical division of function between devices. But it should be understood that this example is not intended to be limiting, and that a classification engine 326 may be provided on a client device 200 in suitable circumstances.
[0082] Server 300 includes a processor 310 connected to a memory 320, having stored therein executable instructions for providing an operating system 322 and at least software portions of a server engine 324. Other components of server 300 include a storage 350, network interface 360, and peripheral interface 340. As described in FIGURE 2, each logical block may be provided by one or more similar or dissimilar logic elements.
[0083] In an example, processor 310 is communicatively coupled to memory 320 via memory bus 370-3, which may be for example a direct memory access (DMA) bus. Processor 310 may be communicatively coupled to other devices via a system bus 370-1.
[0084] Processor 310 may be connected to memory 320 in a DMA configuration via DMA bus 370-3, or via any other suitable memory configuration . As discussed in FIGURE 2, memory 320 may include one or more logic elements of any suitable type.
[0085] Storage 350 may be any species of memory 320, or may be a separate device, as described in connection with storage 250 of FIGURE 2. Storage 350 may be, or may include therein, a database or databases or data stored in other configurations, and may include a stored copy of operational software such as operating system 322 and software portions of server engine 324.
[0086] Network interface 360 may be provided to communicatively couple server 140 to a wired or wireless network, and may include one or more logic elements as described in FIGURE 2.
[0087] Server engine 324 is an engine as described in FIGURE 2 and, in one example, includes one or more logic elements operable to carry out computer- implemented methods as described in this specification. Software portions of server engine 324 may run as a daemon process.
[0088] Server engine 324 may include one or more non-transitory computer-readable mediums having stored thereon executable instructions operable to instruct a processor to provide appropriate services. At an appropriate time, such as upon booting server 140 or upon a command from operating system 322 or a user 120 or security administrator 150, processor 310 may retrieve a copy of server engine 324 (or software portions thereof) from storage 350 and
load it into memory 320. Processor 310 may then iteratively execute the instructions of server engine 324 to provide the desired method.
[0089] Server engine 324 may provide services such as a webserver, database server, database file system server, network antivirus or antimalware engine, or other functions that may benefit from the classification methods disclosed herein .
[0090] Classification engine 326 is also an engine as described herein, and may include one or more logic elements operable to carry out computer- implemented methods as described in this specification. Software portions of classification engine 326 may run as a daemon process. Note that "classification engine" is used throughout as a general term to include the engine that carries out the methods disclosed herein. This is true even in cases where classification engine 326 performs only a portion of its available function (e.g., even when it performs only scoring without classifying a document or message).
[0091] Classification engine 326 may include one or more non-transitory computer-readable mediums having stored thereon executable instructions operable to instruct a processor to provide classification services. At an appropriate time, such as upon booting server 140 or upon a command from operating system 322 or a user 120 or security administrator 150, processor 310 may retrieve a copy of classification engine 326 (or software portions thereof) from storage 350 and load it into memory 320. Processor 310 may then iteratively execute the instructions of classification engine 326 to provide the desired method.
[0092] More specific functions of classification engine 326 are provided throughout this specification.
[0093] FIGURE 4 is a block diagram of classification of a corpus of documents according to one or more examples of the present specification.
[0094] In this example, a training set 410 is first defined. Training set 410 may commonly be much smaller than corpus 430, and may include documents selected from corpus 430. In one example, corpus 430 is the corpus of published and issued U.S. patents and patent applications. Because this corpus comprises many millions of documents, it may be impractical for a human user to manually
classify all of them . However, it may be useful for a user to manually or automatically classify some documents to give the corpus a starting point.
[0095] In an example, classification engine 326 receives training set 410, along with the larger patent corpus 430. Employing method 2100 of FIGURE 20, or another suitable method, classification engine 326 assigns a class to each document based on its a individual gentropy score 460 and those of the training set.
[0096] Optionally, classification engine 326 may include a synonym engine 480 for data, metadata and other attributes, or any combination of these. Synonym engine 480 may be configured to identify words or phrases with identical or similar meanings, including variations on word forms (e.g., plurals and changes in tense), and true synonyms (e.g., "NIC" vs. "Ethernet card") . It can also be configured to identify functionally or statistically equivalent attributes, e.g., patent classification codes in the US patent classification system and the cooperative classification system.
[0097] Corpus 430 starts out in a first temporal state, having a first information entropy. Classification engine 326 then receives a document to be added to corpus 430. Classification engine 326 adds the document to corpus 430, thus causing corpus 430 to assume a second temporal state, having a second information entropy. Classification engine 326 computes the corpus information entropy and gentropy after adding the document, and also computes an individual gentropy score 460 (SGd) for the document that was added.
[0098] FIGURE 5 is an illustration of feature sources that may be used in an embodiment where issued and published patents comprise corpus 430. In this example, the text of the abstract 590 or the content of the detailed description comprise the "data" of the document, which need not be considered for the present method.
[0099] The other fields of a patent may be considered "metadata," and may be used in embodiments of the present method. The patent classification codes are disclosed as an example that was used for validation of the method, but other metadata fields may be used either individually or in combination.
[0100] For example, words appearing in the abstract may stand in as a proxy for classification, as the abstract includes words most likely to be representative
of the subject matter of the patent as a whole. Thus, words in the abstract may be used as classes. Certain embodiments may also employ the concept of word variations and synonyms. For example, "SoC" is a common abbreviation for "system-on-a-chip," and they may be treated as an identical word. Similarly, "system on a chip," "systems on a chip," and "systems-on-a-chip" may all be considered identical for purpose of classification engine 326. Similarly, the terms "jelly bean" and "Easter egg" may be considered to be in related arts, though they are not strict synonyms.
[0101] Non-identical words with similar meaning may also be useful. This may include a recognition that words can be both contemporaneously synonymous, and that words can change over time. Indeed, as technologies mature, they may organically acquire new and different nomenclatures. To provide just one example, the first known patents on transistors did not refer to the devices as "transistors." Rather, they referred to them as "amplifiers" or similar. Thus, patent date 510 (FIGURE 5) may be a relevant metadata factor. For a 2015 date, "transistor" may be a more predictive term, while for a 1925 date, "amplifier" may be a more predictive term . Thus, classification engine 326 may include a synonym engine 480 (FIGURE 4) to identify synonyms, including temporal spacing of synonyms.
[0102] Additional metadata fields may also be used as attributes. As described herein, classification codes 560 were used in a validation test, and both high and low entropy in this field was found to be predictive of "interesting" patents. Other fields that can be used in a classification scheme include title 504, inventor names 530, assignee 540, filing date 550, references cited 520 (similar to hyperlinks— note that additional connections may be formed when future patents cite back to this one), examiner names 570, attorney 580, and abstract 590, by way of nonlimiting example.
[0103] FIGURE 6 is a workflow diagram of document classification according to one or more examples of the present specification. In this example, corpus 430 is any suitable corpus consistent with the present specification. From within corpus 430, interesting documents 620 are identified. A human user or a computer or automated system 630 interact with interesting documents 620 and can perform
desired statistical analyses on important documents 620 which might be a subset of corpus 430 or the whole corpus 430.
[0104] FIGURE 7 is a message graph created according to one or more examples of morphisms of the present specification that represents relationships between documents in a corpus. In the example of FIGURE 7, message graph 700 includes three clusters, namely cluster 1 710- 1, cluster 2 710-2, and cluster 3 710-3. Each cluster may represent, for example, a broad subject matter area. By way of illustration, cluster 1 could represent the area of fishing lures. Cluster 2 could represent the area of semiconductor device manufacturing . Cluster 3 could represent the area of computer security.
[0105] Each node in message graph 700 may represent an attribute or category, such as a classification code. Each edge of the graph represents a connection between attributes, such as a patent that was assigned several . Each time an "event" happens (in this case, the event is that a new patent is added to the corpus), its connections are added to the message graph . This represents a change in state. Message graph 700 was in a first state before the new patent was added. After the patent was added, some change in information entropy Δ Η has occurred, and the corpus is now in a second state with a new information entropy.
[0106] Incremental innovation 720 may occur when, for example, a new patent issues with two classification codes that already exist in cluster 3 710-3, plus one new code that was not previously linked to cluster 3 710-3. It should be understood that labeling the patent as "incremental" does not in any way imply that it is unimportant, nor is it a judgment of the breadth, scope, or monetary value of the newly-issued patent. Rather, the innovation is incremental because it has only a modest effect on the overall entropy of the graph, because it has introduced merely a new node to cluster 3 710-3 which already had numerous nodes.
[0107] In contrast, radical innovation 730 has a much more significant effect on the entropy of the corpus. In this case, radical innovation 730 may arise from a patent that, for example, identifies a novel material or novel use for an existing material that is applicable to both fishing lures and semiconductor manufacturing . Again, labeling node 730 as a "radical" innovation does not represent a judgment on the importance, breadth, scope, or monetary value of the newly-issued patent.
Rather, it indicates that radical innovation 730 has a relatively substantial impact on the overall entropy of the corpus.
[0108] The method not only finds exemplars or outliers with respect to the corpus, but because a score is assigned, it can also quantify and rank the degree to which, e.g., outliers are outliers. For particular cases, the most interesting documents might be in a particular range as opposed to in the extremes. In a nonlimiting case, a user might like movie recommendations that are not very extreme outliers, but are close to the 80th percentile. In another nonlimiting case, certain categories of items that the user might be interested in, such as funny yet disturbing movies, might consistently fall in a particular range.
[0109] Message graph 700 can be used to visually represent data in a way that is very useful for a human observer. For example, radical innovation 730 may be highlighted in a special color that draws particular attention to it. If a human user is searching for prior art patents that may anticipate a patent currently being litigated, radical innovation 730 may in fact be an outlier that is less important, and may be discarded. In this example, the searcher may be looking rather for patents that are more narrowly focused on the subject matter at hand. Similarly, a scientist reviewing literature in her art may prefer to focus on nodes that have many links within their clusters (e.g., frequently-cited publications). In that case, message graph 700 may color nodes with different colors or intensities to represent the nodes with the most links (citations) within cluster 710-3. In another example, a web user is searching for information about the Free Speech clause of the First Amendment. In that case, low-entropy nodes (those with many links) may be of more interest than high-entropy nodes (e.g., a Timecube rant that links free speech to (pseudo-) quantum mechanics and government conspiracy theories). In these cases, radical innovation 730 may be a curiosity at best and of little practical value to the system .
[0110] On the other hand, if message graph 700 is being used for security purposes, nodes heavily within clusters 710 may be of little interest. For example, if each link represents an access event by a user (or machine) to a particular network resource, cluster 3 710-3 may represent users in the accounting department accessing resources commonly accessed by accounting personnel. Thus, there may be many, many links between nodes within cluster 3 710-3.
However, if a radical innovation 730 occurs, that may represent a user or machine accessing a resource outside of his normal sphere of influence (for example, a user in accounting accesses a network management console generally used by the IT staff) . If the gentropy from a network object or event is above one or more thresholds, this may indicate that a user is behaving suspiciously, or that his machine has been compromised and may be trying to access restricted network resources.
[0111] FIGURE 8 is a graph of the impact potential of newly acquired knowledge (with respect to, for example, anindividual, company, government, or human civilization) according to one or more examples of the present specification . This graph illustrates four possible scenarios, and their potential impact.
[0112] Scenario 1 is completely new knowledge or a completely isolated event. Although this new knowledge or event may be a substantial or groundbreaking contribution to the body of knowledge of human thought, and indeed may be a seminal innovation in its field, its instantaneous impact on the entropy of the corpus as a whole is fairly modest, as it has few, if any connections. In the security field, it might be a user using an application for the first time on a machine or network.
[0113] Scenario 2 is new knowledge that joins previously-isolated clusters. As illustrated above, this innovation may have a relatively significant impact (a "leap") on the entropy of the corpus, as in the case of a novel material with applications both in fishing lures and semiconductor manufacturing . Another example is realizing that the force that pulls an apple towards the ground and the force that keeps the Moon rotating around the Earth is one and the same, thus joining botany and astronomy and starting a scientific revolution .
[0114] Scenario 3 is incremental knowledge in existing cluster. As illustrated by incremental innovation 720 above, this may be a relatively modest innovation that strengthens ties within a cluster, and may add a new node to the cluster. An example is the realization that phonon anharmonicity has thermodynamic importance in a material after it has been determined to be the case in a number of other materials.
[0115] Scenario 4 is solidification of existing knowledge. This may take place, for example, after a subfield like phonon thermodynamics or hydrogen storage become mature enough that an expert can write a review article summarizing the findings over the previous decade or so. Another example is the publication of a textbook when a whole field becomes mature and widely-known, for example introductory physics or introductory chemistry. Documents like the American Constitution summarize a set of values generally held by the population of the country circa 1789.
[0116] FIGURE 9 is a block diagram of a general communication system in the mathematical theory of communication according to one or more examples of the present specification . This figure illustrates the piece parts of the methods disclosed herein .
[0117] Information source 990 produces a message (or sequence of messages) 920 to be communicated to the receiving terminal 940. The message 920 might be of various types, such as a sequence of letters in a telegraph of teletype system, or scientific knowledge in the scientific enterprise.
[0118] Transmitter 910 operates on a message 920 in some way to produce a signal suitable for transmission over channel 930. For example, patent attorneys encode technical and scientific knowledge into legal documents. Priests and clerics encode divinely inspired ethical and social guidelines into sacred texts and artifacts. The signal might be perturbed by transm itter 910, in which case the encoding of message 920 is not perfect.
[0119] Channel 930 is the medium used to transmit the signal from transmitter to receiver. It might be a pair of wires, a coaxial cable, a patent or scientific paper, the Ark of the Covenant, gravitational waves, or any other medium . Transmission could be across time, space, or both . The signal might be perturbed by the channel 930, inducing noise.
[0120] Receiver 940 performs the inverse operation of that done by the transmitter, reconstructing the message 920 from the signal . A patent attorney could reconstruct patents into technical or scientific knowledge. A priest could reconstruct sacred texts into ethical or social guidelines. A morphism reconstructs aspects of a document or relationships between documents or their attributes. The
signal might be perturbed by receiver 940, in which case the decoding of message 920 is not perfect.
[0121] Destination 980 is the person or thing for whom message 920 is intended .
[0122] FIGURE 10 is a block diagram of a classification method according to one or more examples of the present specification . Block 1010 represents message generation and message encoding, such as when information source 990 generates a message 920 that is operated on by transmitter 910 and transmitted through channel 930. This could represent, for example, a person creating new knowledge, acquiring the services of a patent attorney to apply for and receive a patent, and the patent being loaded into a patent database such as corpus 430.
[0123] In block 1020, one or more morphisms operate on a document. Morphisms act as receiver 940 and translate document data or metadata into a mathematical message graph that approximate destination 980.
[0124] Finally, the individual message is added to the message graph of the corpus 430. As illustrated in this figure, the total information entropy of a state, Hn, as well as the entropic contribution of the new message Sd, may be computed after each addition to the corpus. The individual contributions might be saved for future reference in ledger 1030. Each addition represents a change of state in corpus 430, with each discrete state having its own entropy value. For example :
[0125] Note that in states 1 - 3, one message each is added for each state, and Sd for that individual message is computed. In state 4, three additional documents are added . Because these documents are added as a group, and hence comprise a single message, the individual entropy for each document is not computed . Rather, Sd for this state is the composite entropy contributed by the three documents together. As can be seen here, the total Sd for the three
documents is only 0.39 bits (whereas surrounding contributions are more on the order or 1.0 bits), so it may be decided that the individual entropy of any one of the three messages is not of any special interest.
[0126] In contrast, the message added in state 1 has an individual information entropy of 3.14 bits. This appears to be a radical innovation. However, it should be noted that the calculation of Sd is not a linear calculation (as explored above). Thus, the first message in a corpus is expected to make a significant contribution to the overall entropy. Thus, this message may not be as important as it appears at first pass. Stated otherwise, the fist message to join two clusters may appear to be a radical innovation, but if that message is the third message in the corpus, and the two clusters each have only one node, and those nodes end up being densely packed in the center of a larger cluster as the corpus evolves, the contribution of message 3 may appear much less radical in hindsight.
[0127] On the other hand, messages added much later (such as state 1459) may be expected to contribute a much smaller Sd to the overall information entropy of the corpus. Thus, while message 1482 contributes an Sd of only 0.21 bits, which is the smallest raw sd in the table, weighted against the size of the corpus at the time (1,459 messages), 0.21 bits may in fact be a substantially entropic contribution. Thus, methods are disclosed herein for conditioning and weighting values of Sd so that entropy calculations are more fair. The conditioned sd is the gentropy score Gd of the message.
[0128] One example of a morphism is co-classification, such as the case where the corpus includes classification codes for patents. This is a classification system that roughly spans the full human-knowledge space. Each classification code is a symbol and the universe of symbols is the alphabet of the corpus (in the information theoretical sense). The co-classification (e.g., the connection between the set of codes assigned to a given document) is the message transmitted in that document. Thus, the message is analogous to a sentence in written language where the individual classification codes can be regarded as words with a semantic relationship, i.e., the codes are not assigned randomly but convey a fundamental attribute of the message. Relationships between primary and secondary classification codes can provide additional information.
[0129] Another example of a morphism is a co-topic with latent semantic indexing (LSI) . In this case, each LSI topic vector is a symbol . The universe of topic vectors is the alphabet of the corpus. A co-topic (set of topics present in a document) is the message transmitted by that document.
[0130] Note that these morphisms need not rely on the content of the documents themselves, such as "key words" or web scrapers. Rather, they rely on metadata attached to or associated with the document (patent classifications, citations in a paper, LSI topics, inventors, bags of words, links to and from a web page, semantic web, internet hits, "likes," etc.) .
[0131] FIGURE 11 is an illustration of evolution of a message graph 1040 according to one or more examples of the present specification . In this example, the message graph starts with three nodes, named Al, Bl, and C3. Consider the example where the graph represents citations in academic literature. Document 1 may be a published paper that cites three previous papers, namely Al, Bl, and C3. The resulting corpus graph (body of knowledge) is identical to that of the document, as this is the first document in the corpus.
[0132] Document 2 is another academic paper, this one citing Al, B3, and D5. In this case, no new connections are made between Al, Bl, and C3, as Bl and C3 are not cited in this paper. However, two new nodes, B3 and D5, are introduced to the message graph, with one connection each to Al . This is a co- citation network. Generally, the weight of the edges can be conditioned, for example by an exponential decay factor that can account for the half-life of a particular message to account for, e.g ., obsoleteness. This would be particularly appropriate for a co-citation network.
[0133] Document 3 is yet another academic paper, this one citing Al, Bl, and B3. Al, Bl, and B3 are all already in the message graph, and there is already an Al to B3 connection, and Bl to B3 connection . Thus, these links are strengthened, as one additional Al to B3 connection, and one additional Al to Bl connection is added .
[0134] FIGURE 12 is a message graph according to one or more examples of the present specification . This graph shows more particularly the "occupation probability" of each node. This is an illustration of the stationary state of a graph in which the size of the nodes is approximately proportional to their occupation
probabilities based on the number and directions of the edges. A probability density function with a categorical variable can be calculated from the occupation probabilities.
[0135] The occupation probability Ρ(χ;) of node i in the message graph of a corpus depends on its number of edges and on its neighbors in a ripple effect. It can be calculated using any suitable algorithm, including repeated sampling taking advantage of the ergodicity of the graph, or other method that outputs a probability distribution related to the likelihood that a traverse of the graph visits a particular node. The set of occupation probabilities of the nodes is the probability distribution P( ) . The message graph information entropy (gentropy) is calculated as :
[0136] The information entropy is maximized when the probabilities are equally distributed . For example, the information entropy in the case of two possibilities, such as a coin toss, with probabilities p and q = 1 - p for heads and tails respectively is H = -(p log2 p + q log2 q) . When the coin is fair, i .e., not biased towards either head or tails, p = q = 0.5 and H = 1 bit. By definition, the information entropy is zero when all the probabilities Ρ(χ;) are zero except for one which will have the value unity. In this case, there is no uncertainty and no information to be gained from the coin toss.
[0137] The probabilities for each node can be calculated iteratively using an algorithm, such as brute force or PageRank, or via repeated sampling of the system . Using repeated sampling to calculate the node occupation probabilities, a node is chosen based on its occupation probability based on a previous estimation of the occupation probability density function of the graph (or completely at random if no previous estimation exists), and edges are selected at random to traverse other nodes until a cutoff number of nodes have been selected or a deadend is reached. The number of samples and cutoff number of traverse nodes can be optimized so that the probability density of function is approximated appropriately for the particular use with the minimum amount of computational power. The previous occupation probability is updated with the new statistics and the estimation is finalized. The calculation converges when the node occupation
probabilities consistently change by less than a specified cutoff (the ergodic theorem ensures that the true stationary probability distribution is approximated by this empirical sampling) .
[0138] The information entropy of an information source is the total information contained in the information source divided by the number of messages n, so it is the average amount of information per message. The amount of information Δ 7 that a new item adds to the information source (the collection of items) can be measured by calculating the change in entropy when the nodes and edges related to that new item are added to the graph, A/ = (n - l) A H . The precise nature of the information depends on the morphism .
[0139] FIGURE 13 is an illustration of evolution of a message graph according to one or more examples of the present specification . In this case, the original graph 1300 includes two clusters, centered around nodes Zl and Al . The size of the nodes represents their occupation probabilities but are not necessarily drawn to scale. This graph is very lopsided (two very high probability nodes surrounded by several lower-probability nodes), so it has low entropy.
[0140] The next message in the corpus is S1-I5. The result is that the graph is less lopsided, because nodes Zl and Al have decreased in probability, while SI and 15 have increased in probability. The overall entropy of the graph has increased by the joinder of the two nodes.
[0141] FIGURE 14 is an illustration of evolution of a message graph according to one or more examples of the present specification . In this example, as in the previous example, original graph 1400 is lopsided, with high-probability node Zl surrounded by lower probability nodes 13, SI, and V3.
[0142] The next message in the corpus is Z1-V3-S1 as illustrated. The result is an even more lopsided graph, with Zl increasing in probability, and additional connections being made to V3 and S I . In this case, the entropy of the graph has decreased because the last message did not introduce any new connections, but rather merely reinforced existing connections. As discussed above, depending on the application, either one of graph 1300 (FIG. 13) or 1400 (FIG. 14) could be of greater interest.
[0143] FIGURE 15 a graph of a distribution of the absolute value of gentropy scores according to one or more examples of the present specification . This graph
illustrates a system with a power law distribution, which is common for systems with growth and with preferential attachment. These attributes are expected to be present in systems with processes in which some quantity is distributed among a number of objects according to how much of that quantity they already have. These systems are widely observed in natural and human made systems, including the Internet and the World Wide Web, social networks, citation networks, traffic systems, astronomical systems, income inequality, etc.
[0144] This figure illustrates the technological diversity of patents in a particular industry. The morphism rules can be designed to score a collection of items relative to desired attributes and their relationships. In the co-classification scheme used in the embodiment illustrated here, the documents that increase diversity are those that explore unusual regions of design or knowledge space (unusual combinations of ideas) and have a positive gentropy score, while those that decrease diversity are representative of the collection of documents as a whole and have a negative gentropy score.
[0145] The effect of decreasing diversity can be achieved by, for instance, a review article or a text book. These summarize a field, and are important references, and are located in well-explored regions of the design space. These typically do not advance novel ideas. Seminal works, on the other hand, may be expect to increase diversity (although not all documents that increase diversity are seminal).
[0146] One example of a system with growth and preferential attachment is the patent reference network. In this case, growth comes from the issuing of new patents and the quantity that gets preferentially distributed is the number of references to a patent. Patent examiners tend to reference patents that are referenced often by other patents and other examiners because they are easier to find using traditional search methods, and because patent examiners (who may be considered experts in searching their particular fields) get to know them very well. These patents may even end up in personal bookmarks. The result is that patents that are highly referenced and more visible are more likely to be referenced by new patents than patents with few other patents referencing them. In many cases though, less well-known patents could be important or relevant, but fall into accidental obscurantism because nobody has read them or cited them
in the first place, sometimes by arbitrary or random reasons, possibly due to the sheer volume of new knowledge being created.
[0147] FIGURE 16 is a histogram of the percentile range of 99 technical documents in a particular industry that were of interest to domain experts with respect to the whole corpus of about 2,000 documents and a co-classification morphism . If the method disclosed herein lacked predictive power, or if the domain experts reviewed documents randomly, each range in the histogram would be expected to have roughly the same number of documents. However, as seen, the histogram is weighted heavily to the extreme ends, illustrating that in this embodiment, the documents with large and positive (top percentile) and large and negative (bottom percentile) gentropy scores are of particular interest. In other words, in a corpus of patents, those of most interest to a group of human experts are those that are highly innovative or matures technologies that can be readily be applied to current processes in the industry.
[0148] FIGURE 17 is a block diagram of a classification of documents according to one or more examples of the present specification. As this FIGURE illustrates, starting with a collection of documents 1710, a morphism 1740 is used to build a message graph 1720 according to the attributes of one or more documents and the relationships between their attributes. This message graph 1720 might represent knowledge space when the documents are patents. The gentropy score is calculated as described before and used for identification of interesting documents 1730. The documents identified as interesting become a new collection of documents 1710 and further refinement is possible.
[0149] FIGURE 18 is a block diagram of the classification of documents according to one or more examples of the present specification. As this FIGURE illustrates, starting with a collection of documents 1810, a plurality of morphisms 1840 is used to build a message graph 1820 with different kinds of edges corresponding to the relationships between different kinds of attributes. This message graph might represent knowledge space when the documents are patents. The gentropy score is calculated as described before and used for identification of interesting documents 1830. The documents identified as interesting become a new collection of documents 1810 and further refinement is possible.
[0150] FIGURE 19 is a block diagram of application of the classification of documents according to one or more examples of the present specification . As this FIGURE illustrates, starting with two or more collections of document, e.g. 1911, 1919, which might or might not contain the same documents, one or more morphisms, e.g ., 1941, 1940, which might or might not be the same, are used to build a message graph for each collection, e.g., 1921, 1929. The message graph of each collection might or might not contain different kinds of edges. The nodes and edges of the message graphs convey relationships between different attributes of the documents. The gentropy score is calculated as described before for each collection of items and can be used for identification of interesting documents. The gentropy scores according to different morphisms can be recorded, e.g., 1961, 1969, and machine learning algorithms can be then be applied for identification of interesting documents 1930.
[0151] FIGURE 20 is a flow chart of a method 2000 of computing and applying entropy according to one or more examples of the present specification .
[0152] In block 2002, a document is added to corpus 430 and state n is created . Corpus 430 in state n - 1 already has an existing total information entropy Ηη_! stored in total entropy ledger 2006. This is true even if corpus 430 starts out as an empty corpus, with an information entropy of zero.
[0153] In block 2004, the message graph of corpus 430 in the state n is built by the application of one or more morphisms to the documents of the corpus or their attributes and the information entropy of corpus 430 in the state n, (Hn) is calculated. This is stored in ledger 2006.
[0154] In block 2108, the information entropy of the message graph of corpus 430 in state n is compared to the information entropy of message graph of the corpus 430 in state n- 1 to calculate a value for Sd, the information entropy contribution of the present message (document), also referred to as the gentropy.
[0155] In block 2010, Sd may be conditioned or otherwise adjusted based on an appropriate function, functional, or set of instructions. In a nonlimiting example, Sd is multiplied by the number of documents in the corpus. In another nonlimiting example, the conditioned quantity is the exponentiation of the number of documents as the exponent and Euler's number as the base. In yet another nonlimiting example, the conditioned quantity is the number of documents times
the inverse of the exponentiation of the difference in the number of documents between the initial state and the current state times a constant and Euler's constant as the base. The conditioned quantity is the gentropy score Gd . As described above, this is to ensure that Gd is not unfairly penalized simply because corpus 430 has grown relatively large, and that Gd is not unfairly inflated simply because corpus 430 is relatively small, but it can serve other purposes as well .
[0156] In block 2012, a conditioned incremental entropy score, Gd is assigned to the document under consideration and stored in ledger 2020 for reference. Depending on the application, the magnitude of this entropy score may indicate whether the document is of particular interest. For example, in a web search, a low Gd may indicate a particularly relevant search result. In a security application, a high Gd may indicate a particularly interesting event.
[0157] In block 2014, the system acts on the gentropy score Gd . This could be any of a large number of potential actions. For example, in a search engine for web pages or scientific publications, acting on the score could include returning "hits" for those documents in corpus 430 that have the lowest Gd . In a patent search, acting on the score could include returning "hits" for documents that have both the highest and lowest values for Gd l meaning that they may be of the greatest interest to subject matter experts. In a security application, acting on the score may include alerting a security administrator 150 that an event with a particularly high entropy has occurred, thus giving security administrator 150 an opportunity to take appropriate action . In appropriate circumstances, many other species of acting on the score are possible.
[0158] In block 2099, the method is done.
[0159] The foregoing outlines features of several embodiments so that those skilled in the art may better understand the aspects of the present disclosure. Those skilled in the art should appreciate that they may readily use the present disclosure as a basis for designing or modifying other processes and structures for carrying out the same purposes and/or achieving the same advantages of the embodiments introduced herein . Those skilled in the art should also realize that such equivalent constructions do not depart from the spirit and scope of the present disclosure, and that they may make various changes, substitutions, and
alterations herein without departing from the spirit and scope of the present disclosure.
[0160] The particular embodiments of the present disclosure may readily include a system on chip (SOC) central processing unit (CPU) package. An SOC represents an integrated circuit (IC) that integrates components of a computer or other electronic system into a single chip. It may contain digital, analog, mixed- signal, and radio frequency functions: all of which may be provided on a single chip substrate. Other embodiments may include a multi-chip-module (MCM), with a plurality of chips located within a single electronic package and configured to interact closely with each other through the electronic package. In various other embodiments, the digital signal processing functionalities may be implemented in one or more silicon cores in Application Specific Integrated Circuits (ASICs), Field Programmable Gate Arrays (FPGAs), and other semiconductor chips.
[0161] Additionally, some of the components associated with described microprocessors may be removed, or otherwise consolidated. In a general sense, the arrangements depicted in the figures may be more logical in their representations, whereas a physical architecture may include various permutations, combinations, and/or hybrids of these elements. It is imperative to note that countless possible design configurations can be used to achieve the operational objectives outlined herein. Accordingly, the associated infrastructure has a myriad of substitute arrangements, design choices, device possibilities, hardware configurations, software implementations, equipment options, etc.
[0162] Any suitably-configured processor component can execute any type of instructions associated with the data to achieve the operations detailed herein . Any processor disclosed herein could transform an element or an article (for example, data) from one state or thing to another state or thing. In another example, some activities outlined herein may be implemented with fixed logic or programmable logic (for example, software and/or computer instructions executed by a processor) and the elements identified herein could be some type of a programmable processor, programmable digital logic (for example, a field programmable gate array (FPGA), an erasable programmable read only memory (EPROM), an electrically erasable programmable read only memory (EEPROM)), an ASIC that includes digital logic, software, code, electronic instructions, flash
memory, optical disks, CD-ROMs, DVD ROMs, magnetic or optical cards, other types of machine-readable mediums suitable for storing electronic instructions, or any suitable combination thereof. In operation, processors may store information in any suitable type of non-transitory storage medium (for example, random access memory (RAM), read only memory (ROM), field programmable gate array (FPGA), erasable programmable read only memory (EPROM), electrically erasable programmable ROM (EEPROM), etc.), software, hardware, or in any other suitable component, device, element, or object where appropriate and based on particular needs. Further, the information being tracked, sent, received, or stored in a processor could be provided in any database, register, table, cache, queue, control list, or storage structure, based on particular needs and implementations, all of which could be referenced in any suitable timeframe. Any of the memory items discussed herein should be construed as being encompassed within the broad term 'memory.'
[0163] Computer program logic implementing all or part of the functionality described herein is embodied in various forms, including, but in no way limited to, a source code form, a computer executable form, and various intermediate forms (for example, forms generated by an assembler, compiler, linker, or locator). In an example, source code includes a series of computer program instructions implemented in various programming languages, such as an object code, an assembly language, or a high-level language such as OpenCL, Fortran, C, C++, JAVA, or HTML for use with various operating systems or operating environments. The source code may define and use various data structures and communication messages. The source code may be in a computer executable form (e.g., via an interpreter), or the source code may be converted (e.g., via a translator, assembler, or compiler) into a computer executable form.
[0164] In one example embodiment, any number of electrical circuits of the FIGURES may be implemented on a board of an associated electronic device. The board can be a general circuit board that can hold various components of the internal electronic system of the electronic device and, further, provide connectors for other peripherals. More specifically, the board can provide the electrical connections by which the other components of the system can communicate electrically. Any suitable processors (inclusive of digital signal processors,
microprocessors, supporting chipsets, etc.), memory elements, etc. can be suitably coupled to the board based on particular configuration needs, processing demands, computer designs, etc. Other components such as external storage, additional sensors, controllers for audio/video display, and peripheral devices may be attached to the board as plug-in cards, via cables, or integrated into the board itself. In another example embodiment, the electrical circuits of the FIGURES may be implemented as stand-alone modules (e.g., a device with associated components and circuitry configured to perform a specific application or function) or implemented as plug-in modules into application specific hardware of electronic devices.
[0165] Note that with the numerous examples provided herein, interaction may be described in terms of two, three, four, or more electrical components. However, this has been done for purposes of clarity and example only. It should be appreciated that the system can be consolidated in any suitable manner. Along similar design alternatives, any of the illustrated components, modules, and elements of the FIGURES may be combined in various possible configurations, all of which are clearly within the broad scope of this specification. In certain cases, it may be easier to describe one or more of the functionalities of a given set of flows by only referencing a limited number of electrical elements. It should be appreciated that the electrical circuits of the FIGURES and its teachings are readily scalable and can accommodate a large number of components, as well as more complicated/sophisticated arrangements and configurations. Accordingly, the examples provided should not limit the scope or inhibit the broad teachings of the electrical circuits as potentially applied to a myriad of other architectures.
[0166] Numerous other changes, substitutions, variations, alterations, and modifications may be ascertained to one skilled in the art and it is intended that the present disclosure encompass all such changes, substitutions, variations, alterations, and modifications as falling within the scope of the appended claims. In order to assist the United States Patent and Trademark Office (USPTO) and, additionally, any readers of any patent issued on this application in interpreting the claims appended hereto, Applicant wishes to note that the Applicant: (a) does not intend any of the appended claims to invoke paragraph six (6) of 35 U.S.C. section 112 (pre-AIA) or paragraph (f) of the same section (post-AIA), as it exists
on the date of the filing hereof unless the words "means for" or "steps for" are specifically used in the particular claims; and (b) does not intend, by any statement in the specification, to limit this disclosure in any way that is not otherwise reflected in the appended claims.
Example Implementations
[0167] There is disclosed in an example, a computing apparatus for assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy, comprising : one or more logic elements, including at least one hardware logic element, comprising a classification engine to : receive the document to be added to the corpus; add the document to the corpus, creating a second temporal state of the corpus; compute a second corpus entropy for the second temporal state, based at least in part on a morphism ; and assign the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
[0168] There is further disclosed an example, wherein assigning the document a gentropy score comprises conditioning the gentropy score based at least in part on a function or functional of a number of documents in the corpus in the first temporal state.
[0169] There is further disclosed an example, wherein computing the second corpus entropy comprises computing an entropy of a message graph of the form Η ) = -∑ί Ρ(χ ί) 1ο§2 Ρ(χ ί) .
[0170] There is further disclosed an example, wherein computing the second corpus entropy is based at least in part on a plurality of morphisms.
[0171] There is further disclosed an example, wherein the classification engine is further to add a second document to the corpus, and compute a gentropy score for the second document.
[0172] There is further disclosed an example, wherein the classification engine is further to act on the gentropy score.
[0173] There is further disclosed an example, wherein the corpus comprises patents, and wherein acting on the gentropy score comprises determining that the gentropy score falls within a range designated a region of interest, and identifying the document as a patent of interest.
[0174] There is further disclosed an example, wherein the corpus comprises network resources, and wherein acting on the gentropy score comprises determining that the gentropy score is higher than a threshold, and marking the document as suspicious.
[0175] There is further disclosed an example, wherein the corpus comprises employee or contract diversity data, and wherein acting on the gentropy score comprises determining that a hiring or contracting decision with a high entropy represents a hiring of or contracting with a person or enterprise with high diversity.
[0176] There is further disclosed in an example, one or more tangible, non- transitory, computer-readable storage mediums for instructing a processor to provide a classification engine for assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy, the classification engine to: receive the document to be added to the corpus; add the document to the corpus, creating a second temporal state of the corpus; compute a second corpus entropy for the second temporal state, based at least in part on a morphism; and assign the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
[0177] There is further disclosed an example, wherein assigning the document a gentropy score comprises conditioning the gentropy score based at least in part on a function or functional of a number of documents in the corpus in the first temporal state.
[0178] There is further disclosed an example, wherein computing the second corpus entropy comprises computing an entropy of a message graph of the form H( ) = -∑i P(xi) log2 P(xi).
[0179] There is further disclosed an example, wherein computing the second corpus entropy is based at least in part on a plurality of morphisms.
[0180] There is further disclosed an example, wherein the classification engine is further to add a second document to the corpus, and compute a gentropy score for the second document.
[0181] There is further disclosed an example, wherein the classification engine is further to act on the gentropy score.
[0182] There is further disclosed an example, wherein the corpus comprises patents, and wherein acting on the gentropy score comprises determining that the gentropy score falls within a range designated a region of interest, and identifying the document as a patent of interest.
[0183] There is further disclosed an example, wherein the corpus comprises network resources, and wherein acting on the gentropy score comprises determining that the gentropy score is higher than a threshold, and marking the document as suspicious.
[0184] There is further disclosed an example, wherein the corpus comprises employee or contract diversity data, and wherein acting on the gentropy score comprises determining that a hiring or contracting decision with a high entropy represents a hiring of or contracting with a person or enterprise with high diversity.
[0185] A method of assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy, comprising : receiving the document to be added to the corpus; adding the document to the corpus, creating a second temporal state of the corpus; computing a second corpus entropy for the second temporal state, based at least in part on a morphism; and assigning the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
[0186] There is further disclosed an example, wherein assigning the document a gentropy score comprises conditioning the gentropy score based at least in part on a function or functional of a number of documents in the corpus in the first temporal state.
[0187] There is further disclosed an example, wherein computing the second corpus gentropy comprises computing an entropy of a message graph of the form Η ) = -∑ί Ρ(χ ί) 1ο§2 Ρ(χ ί) .
[0188] There is further disclosed an example, wherein computing the second corpus entropy is based on a plurality of morphisms.
[0189] There is further disclosed an example, wherein the classification engine is further to add a second document to the corpus, and compute a gentropy score for the second document.
[0190] There is further disclosed an example, wherein the classification engine is further to act on the gentropy score.
[0191] There is further disclosed an example, wherein the corpus comprises patents, and wherein acting on the gentropy score comprises determining that the gentropy score falls within a range designated a region of interest, and identifying the document as a patent of interest.
[0192] There is further disclosed in an example, a computing apparatus for assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy, comprising : one or more logic elements, including at least one hardware logic element, comprising a classification engine to: receive the document to be added to the corpus; create a message graph of the first corpus using at least a first morphism including a conditioning of weight of individual edges of the message graph; calculate a gentropy of the corpus in the first temporal state comprising calculating a first node occupation probability density function of the message graph; receiving the document to be added to the first corpus, creating a second temporal state of the corpus having a second corpus information entropy; create a message graph of the second temporal state using at least a second morphism including a conditioning of weight of individual edges of the message graph; calculate a second gentropy of the corpus in the second temporal state comprising calculating a second node occupation probability density function of the message graph; assigning the gentropy score to the document comprising computing a gentropy difference between the first gentropy and the second gentropy.
[0193] There is further disclosed an example, wherein the classification engine is further to apply machine learning techniques to classify the document according to the gentropy score.
[0194] There is further disclosed in an example, a method of assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy, comprising : receiving the document to be added to the corpus; creating a message graph of the first corpus using at least a first morphism including a conditioning of weight of individual edges of the message graph; calculating a gentropy of the corpus in the first temporal state comprising calculating a first node occupation probability density function of the
message graph; receiving the document to be added to the first corpus, creating a second temporal state of the corpus having a second corpus information entropy; creating a message graph of the second temporal state using at least a second morphism including a conditioning of weight of individual edges of the message graph; calculating a second gentropy of the corpus in the second temporal state comprising calculating a second node occupation probability density function of the message graph; and assigning the gentropy score to the document comprising computing a gentropy difference between the first gentropy and the second gentropy.
[0195] There is further disclosed an example, wherein the classification engine is further to apply machine learning techniques to classify the document according to the gentropy score.
[0196] An apparatus comprising means for performing the method of any of the preceding examples.
[0197] There is further disclosed an example, wherein the means comprise a processor and a memory.
[0198] There is further disclosed an example, wherein the means comprise a machine readable medium having stored thereon executable instructions to instruct a processor to perform the method.
[0199] There is further disclosed an example, wherein the apparatus is a computing system.
[0200] There is further disclosed an example of one or more tangible, non- transitory computer-readable storage mediums having stored thereon executable instructions for instructing one or more processors for providing a classification engine operable for performing any or all of the operations of any of the preceding examples.
[0201] There is further disclosed an example of a method of providing a classification engine comprising performing any or all of the operations of the preceding examples.
[0202] There is further disclosed an example of an apparatus comprising means for performing the method.
[0203] There is further disclosed an example wherein the means comprise a processor and a memory.
[0204] There is further disclosed an example wherein the means comprise one or more tangible, non-transitory computer-readable storage mediums.
[0205] There is further disclosed an example wherein the apparatus is a computing device.
Claims
1. A computing apparatus for assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy, comprising : one or more logic elements, including at least one hardware logic element, comprising a classification engine to: receive the document to be added to the corpus; add the document to the corpus, creating a second temporal state of the corpus; compute a second corpus entropy for the second temporal state, based at least in part on a morphism; and assign the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
2. The computing apparatus of claim 1, wherein assigning the document a gentropy score comprises conditioning the gentropy score based at least in part on a function or functional of a number of documents in the corpus in the first temporal state.
3. The computing apparatus of claim 1, wherein computing the second corpus entropy comprises computing an entropy of a message graph of the form
Η ) = -∑ί Ρ(χ ί) 1ο§2 Ρ(χ ί) .
4. The computing apparatus of claim 1, wherein computing the second corpus entropy is based at least in part on a plurality of morphisms.
5. The computing apparatus of claim 1, wherein the classification engine is further to add a second document to the corpus, and compute a gentropy score for the second document.
6. The computing apparatus of any of claims 1 - 5, wherein the classification engine is further to act on the gentropy score.
7. The computing apparatus of claim 6, wherein the corpus comprises patents, and wherein acting on the gentropy score comprises determining that the gentropy score falls within a range designated a region of interest, and identifying the document as a patent of interest.
8. The computing apparatus of claim 6, wherein the corpus comprises network resources, and wherein acting on the gentropy score comprises determining that the gentropy score is higher than a threshold, and marking the document as suspicious.
9. The computing apparatus of claim 6, wherein the corpus comprises employee or contract diversity data, and wherein acting on the gentropy score comprises determining that a hiring or contracting decision with a high entropy represents a hiring of or contracting with a person or enterprise with high diversity.
10. One or more tangible, non-transitory, computer-readable storage mediums for instructing a processor to provide a classification engine for assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy, the classification engine to: receive the document to be added to the corpus; add the document to the corpus, creating a second temporal state of the corpus; compute a second corpus entropy for the second temporal state, based at least in part on a morphism; and assign the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
11. The one or more tangible, non-transitory computer-readable storage mediums of claim 10, wherein assigning the document a gentropy score comprises conditioning the gentropy score based at least in part on a function or functional of a number of documents in the corpus in the first temporal state.
12. The one or more tangible, non-transitory computer-readable storage mediums of claim 10, wherein computing the second corpus entropy comprises computing an entropy of a message graph of the form H( ) = -∑i P(x;) log2 P(x;) -
13. The one or more tangible, non-transitory computer-readable storage mediums of claim 10, wherein computing the second corpus entropy is based at least in part on a plurality of morphisms.
14. The one or more tangible, non-transitory computer-readable storage mediums of claim 10, wherein the classification engine is further to add a second document to the corpus, and compute a gentropy score for the second
document.
15. The one or more tangible, non-transitory computer-readable storage mediums of any of claims 10 - 14, wherein the classification engine is further to act on the gentropy score.
16. The one or more tangible, non-transitory computer-readable storage mediums of claim 15, wherein the corpus comprises patents, and wherein acting on the gentropy score comprises determining that the gentropy score falls within a range designated a region of interest, and identifying the document as a patent of interest.
17. The one or more tangible, non-transitory computer-readable storage mediums of claim 15, wherein the corpus comprises network resources, and wherein acting on the gentropy score comprises determining that the gentropy score is higher than a threshold, and marking the document as suspicious.
18. The one or more tangible, non-transitory computer-readable storage mediums of claim 15, wherein the corpus comprises employee or contract diversity data, and wherein acting on the gentropy score comprises determining
that a hiring or contracting decision with a high entropy represents a hiring of or contracting with a person or enterprise with high diversity.
19. A method of assigning a gentropy score to a document to be added to a corpus in a first temporal state having a first corpus information entropy, comprising : receiving the document to be added to the corpus; adding the document to the corpus, creating a second temporal state of the corpus; computing a second corpus entropy for the second temporal state, based at least in part on a morphism; and assigning the document a gentropy score based at least in part on a difference between the first corpus entropy and the second corpus entropy.
20. The method of claim 19, wherein assigning the document a gentropy score comprises conditioning the gentropy score based at least in part on a function or functional of a number of documents in the corpus in the first temporal state.
21. The method of claim 19, wherein computing the second corpus gentropy comprises computing an entropy of a message graph of the form H( ) =
-∑i P(*i) log2 P(*i) -
22. The method of claim 19, wherein computing the second corpus entropy is based on a plurality of morphisms.
23. The method of claim 19, wherein the classification engine is further to add a second document to the corpus, and compute a gentropy score for the second document.
24. The method of any of claims 19 - 23, wherein the classification engine is further to act on the gentropy score.
25. The method of claim 24, wherein the corpus comprises patents, and wherein acting on the gentropy score comprises determining that the gentropy score falls within a range designated a region of interest, and identifying the document as a patent of interest.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/US2016/025466 WO2017171826A1 (en) | 2016-04-01 | 2016-04-01 | Entropic classification of objects |
EP16897366.7A EP3436966A4 (en) | 2016-04-01 | 2016-04-01 | Entropic classification of objects |
US16/083,108 US10956476B2 (en) | 2016-04-01 | 2016-04-01 | Entropic classification of objects |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/US2016/025466 WO2017171826A1 (en) | 2016-04-01 | 2016-04-01 | Entropic classification of objects |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017171826A1 true WO2017171826A1 (en) | 2017-10-05 |
Family
ID=59966266
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2016/025466 WO2017171826A1 (en) | 2016-04-01 | 2016-04-01 | Entropic classification of objects |
Country Status (3)
Country | Link |
---|---|
US (1) | US10956476B2 (en) |
EP (1) | EP3436966A4 (en) |
WO (1) | WO2017171826A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10436615B2 (en) | 2017-04-24 | 2019-10-08 | Carnegie Mellon University | Virtual sensor system |
JP2020004393A (en) * | 2018-06-26 | 2020-01-09 | インテル・コーポレーション | Entropic clustering of objects |
US11321625B2 (en) | 2019-04-25 | 2022-05-03 | International Business Machines Corporation | Quantum circuit optimization using machine learning |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101742258B1 (en) * | 2016-02-29 | 2017-05-31 | 경희대학교 산학협력단 | Apparatus and method for providing contents using web-based virtual desktop protocol |
US11165694B2 (en) * | 2018-07-31 | 2021-11-02 | Mcafee, Llc | Methods, systems, articles of manufacture and apparatus to identify applications |
US11182800B2 (en) | 2019-04-08 | 2021-11-23 | Bank Of America Corporation | Controlling enterprise software policy compliance assessment processes based on quantum combinations of assessment elements |
CN112016582B (en) * | 2019-05-31 | 2023-11-24 | 口口相传(北京)网络技术有限公司 | Dish recommending method and device |
US11580415B2 (en) * | 2019-07-09 | 2023-02-14 | Baidu Usa Llc | Hierarchical multi-task term embedding learning for synonym prediction |
CN111209363B (en) * | 2019-12-25 | 2024-02-09 | 华为技术有限公司 | Corpus data processing method, corpus data processing device, server and storage medium |
US11522837B2 (en) * | 2020-06-24 | 2022-12-06 | Microsoft Technology Licensing, Llc | Systems and methods for web tracker classification and mitigation |
US20230018387A1 (en) * | 2021-07-06 | 2023-01-19 | metacluster lt, UAB | Dynamic web page classification in web data collection |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7320000B2 (en) * | 2002-12-04 | 2008-01-15 | International Business Machines Corporation | Method and apparatus for populating a predefined concept hierarchy or other hierarchical set of classified data items by minimizing system entrophy |
US20110270845A1 (en) * | 2010-04-29 | 2011-11-03 | International Business Machines Corporation | Ranking Information Content Based on Performance Data of Prior Users of the Information Content |
US20150348216A1 (en) * | 2014-05-29 | 2015-12-03 | General Electric Company | Influencer analyzer platform for social and traditional media document authors |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6397205B1 (en) * | 1998-11-24 | 2002-05-28 | Duquesne University Of The Holy Ghost | Document categorization and evaluation via cross-entrophy |
US20030221163A1 (en) * | 2002-02-22 | 2003-11-27 | Nec Laboratories America, Inc. | Using web structure for classifying and describing web pages |
US7711747B2 (en) * | 2007-04-06 | 2010-05-04 | Xerox Corporation | Interactive cleaning for automatic document clustering and categorization |
US8166051B1 (en) * | 2009-02-03 | 2012-04-24 | Sandia Corporation | Computation of term dominance in text documents |
US20160307113A1 (en) * | 2015-04-20 | 2016-10-20 | Xerox Corporation | Large-scale batch active learning using locality sensitive hashing |
US9798818B2 (en) * | 2015-09-22 | 2017-10-24 | International Business Machines Corporation | Analyzing concepts over time |
-
2016
- 2016-04-01 US US16/083,108 patent/US10956476B2/en active Active
- 2016-04-01 EP EP16897366.7A patent/EP3436966A4/en not_active Ceased
- 2016-04-01 WO PCT/US2016/025466 patent/WO2017171826A1/en active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7320000B2 (en) * | 2002-12-04 | 2008-01-15 | International Business Machines Corporation | Method and apparatus for populating a predefined concept hierarchy or other hierarchical set of classified data items by minimizing system entrophy |
US20110270845A1 (en) * | 2010-04-29 | 2011-11-03 | International Business Machines Corporation | Ranking Information Content Based on Performance Data of Prior Users of the Information Content |
US20150348216A1 (en) * | 2014-05-29 | 2015-12-03 | General Electric Company | Influencer analyzer platform for social and traditional media document authors |
Non-Patent Citations (3)
Title |
---|
IOANNIS KORKONTZELOS ET AL.: "Graph Connectivity Measures for Unsupervised Parameter Tuning of Graph-Based Sense Induction Systems", PROCEEDINGS OF THE NAACL HLT WORKSHOP ON UNSUPERVISED AND MINIMALLY SUPERVISED LEARNING OF LEXICAL SEMANTICS, June 2009 (2009-06-01), pages 36 - 44, Retrieved from the Internet <URL:http://www.aclweb.org/anthology/WO9-1705> * |
ROBERTO NAVIGLI ET AL.: "An Experimental Study of Graph Connectivity for Unsupervised Word Sense Disambiguation", IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, vol. 32, no. 4, 13 February 2010 (2010-02-13), pages 678 - 692, XP011280707, Retrieved from the Internet <URL:https://www.cl.cam.ac.uk/teaching/1314/L114/PAMI.pdf> * |
See also references of EP3436966A4 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10436615B2 (en) | 2017-04-24 | 2019-10-08 | Carnegie Mellon University | Virtual sensor system |
JP2020004393A (en) * | 2018-06-26 | 2020-01-09 | インテル・コーポレーション | Entropic clustering of objects |
US11238308B2 (en) | 2018-06-26 | 2022-02-01 | Intel Corporation | Entropic clustering of objects |
JP7404599B2 (en) | 2018-06-26 | 2023-12-26 | インテル・コーポレーション | Entropy clustering of objects |
US11321625B2 (en) | 2019-04-25 | 2022-05-03 | International Business Machines Corporation | Quantum circuit optimization using machine learning |
Also Published As
Publication number | Publication date |
---|---|
EP3436966A4 (en) | 2019-11-13 |
EP3436966A1 (en) | 2019-02-06 |
US20190080000A1 (en) | 2019-03-14 |
US10956476B2 (en) | 2021-03-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10956476B2 (en) | Entropic classification of objects | |
EP2866421B1 (en) | Method and apparatus for identifying a same user in multiple social networks | |
Kang et al. | Detecting and classifying android malware using static analysis along with creator information | |
US11250256B2 (en) | Binary linear classification | |
US20170177627A1 (en) | Multi-label content recategorization | |
US8078642B1 (en) | Concurrent traversal of multiple binary trees | |
US20150356091A1 (en) | Method and system for identifying microblog user identity | |
CN107408115B (en) | Web site filter, method and medium for controlling access to content | |
US8346774B1 (en) | Protecting network entity data while preserving network properties | |
CN109697641A (en) | The method and apparatus for calculating commodity similarity | |
US10417579B2 (en) | Multi-label classification for overlapping classes | |
WO2021004124A1 (en) | Data comparison-based information recommendation method and device, and storage medium | |
CN109582967B (en) | Public opinion abstract extraction method, device, equipment and computer readable storage medium | |
US20200117738A1 (en) | Semantic concept discovery over event databases | |
Kahvedžić et al. | DIALOG: A framework for modeling, analysis and reuse of digital forensic knowledge | |
Li et al. | MapReduce-based web mining for prediction of web-user navigation | |
CN110851582A (en) | Text processing method and system, computer system and computer readable storage medium | |
US11847412B2 (en) | Enforcing data ownership at gateway registration using natural language processing | |
CN111538930A (en) | Home page display method and related equipment | |
Youn | SPONGY (SPam ONtoloGY): Email Classification Using Two‐Level Dynamic Ontology | |
US9563845B1 (en) | Rule evaluation based on precomputed results | |
Huth et al. | Accessibility of covariance information creates vulnerability in Federated Learning frameworks | |
CN111209403A (en) | Data processing method, device, medium and electronic equipment | |
Hu et al. | Stochastic stability of a class of unbounded delay neutral stochastic differential equations with general decay rate | |
Gao et al. | Twitter hyperlink recommendation with user-tweet-hyperlink three-way clustering |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16897366 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2016897366 Country of ref document: EP |