US20160335234A1 - Systems and Methods for Generating Summaries of Documents - Google Patents

Systems and Methods for Generating Summaries of Documents Download PDF

Info

Publication number
US20160335234A1
US20160335234A1 US15/131,997 US201615131997A US2016335234A1 US 20160335234 A1 US20160335234 A1 US 20160335234A1 US 201615131997 A US201615131997 A US 201615131997A US 2016335234 A1 US2016335234 A1 US 2016335234A1
Authority
US
United States
Prior art keywords
article
articles
sentences
sentence
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/131,997
Inventor
Douglas Dane Baker
Paulo Malvar Fernández
Brian Fernandes
Rodrigo Alarcon Martinez
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CODEQ LLC
Original Assignee
CODEQ LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CODEQ LLC filed Critical CODEQ LLC
Priority to US15/131,997 priority Critical patent/US20160335234A1/en
Assigned to CODEQ, LLC reassignment CODEQ, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: FERNANDES, BRIAN, MARTINEZ, RODRIGO ALARCON, BAKER, DOUGLAS DANE, FERNÁNDEZ, PAULO MALVAR
Publication of US20160335234A1 publication Critical patent/US20160335234A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/14Tree-structured documents
    • G06F40/143Markup, e.g. Standard Generalized Markup Language [SGML] or Document Type Definition [DTD]
    • G06F17/24
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/34Browsing; Visualisation therefor
    • G06F16/345Summarisation for human users
    • G06F17/2247
    • G06F17/2705
    • G06F17/278
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/194Calculation of difference between files
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • G06F40/295Named entity recognition

Definitions

  • FIG. 1 is a schematic diagram illustrating a system according to a particular embodiment
  • FIG. 2 is a block diagram of a hardware component of the monitoring system according to a particular embodiment
  • FIG. 3 is a flowchart illustrating the functionality of a method according to a particular embodiment
  • FIG. 4A is a screenshot of a capsule screen according to a particular embodiment
  • FIG. 4B is a screenshot of a particular embodiment
  • FIG. 4C is a screenshot of a particular embodiment
  • FIG. 4D is a screenshot of a particular embodiment.
  • FIG. 5 is a block diagram of a hardware component according to a particular embodiment.
  • the disclosure herein describes systems and methods for summarizing single document articles for consumption on a user device, specifically a mobile device (such as a smartphone or tablet).
  • the systems and methods disclosed herein feature an innovative single document summarization engine that combines both unsupervised Machine Learning techniques and linguistically motivated rules to significantly improve the quality of generated summaries.
  • the systems and methods may improve the quality of the generated summaries by reducing the risk of including sentences that will impact the summary coherence.
  • Coherence is a linguistic concept that can be succinctly defined as the logical connections that humans perceive in a text (written or oral).
  • the systems and methods disclosed herein block sentences that contain linguistic markers that build the internal coherence of articles while at the same time taking advantage of the content of these blocked sentences to select the set sentences of an article (such as a news article) that most relevantly summarize it.
  • an article may include several sentences with discourse markers and/or connectors like “however”, “in addition”, “meanwhile”, etc. Including these sentences in the generated summary could break the flow of ideas of the original article. For instance, a sentence that starts with the adversative “however” is contradicting an idea expressed before this sentence. Including this sentence in the summary without including the sentences that includes the idea that it is contradicting would lead to an incomprehensible summary from the point of view of its internal coherence. The summary is designed to fit on the screen of most tablets and smartphones.
  • the system may initially extract the main body of text from the HTML code of an article.
  • the system may then classify the extracted article into one of several different categories.
  • the system also identifies duplicate articles and removes them.
  • the system breaks down each article into its component sentences, and each sentence is classified into one of three categories: (1) potential candidate sentences that may be included in the generated summary; (2) weakly rejected sentences that will not be included in the summary, but may be used to help generate the summary; and (3) strongly rejected sentences that are not included in the summary or used to generate the summary.
  • the system applies a document summarizer, which features an innovative summarization algorithm that leverages linguistic information to reduce the risk of producing incoherent summaries. In this way, the system generates quickly readable article summaries, for viewing on the screen of a smartphone or tablet computer, using relevant sentences from the article that maintain the coherence of the article.
  • module may be understood to refer to computing software, firmware, hardware, and/or various combinations thereof. It is noted that the modules are examples. The modules may be combined, integrated, separated, and/or duplicated to support various applications. Also, a function described herein as being performed at a particular module may be performed at one or more other modules and/or by one or more other devices instead of or in addition to the function performed at the particular module. Further, the modules may be implemented across multiple devices and/or other components local or remote to one another. Additionally, the modules may be moved from one device and added to another device, and/or may be included in both devices.
  • software described herein may be tangibly embodied in one or more physical media, such as, but not limited to, a compact disc (“CD”), a digital versatile disc (“DVD”), a floppy disk, a hard drive, read only memory (“ROM”), random access memory (“RAM”), as well as other physical media capable of storing software, and/or combinations thereof.
  • CD compact disc
  • DVD digital versatile disc
  • ROM read only memory
  • RAM random access memory
  • the functions described as being performed at various components may be performed at other components, and the various components may be combined and/or separated. Other modifications also may be made
  • FIG. 1 is a schematic diagram illustrating a system according to particular embodiments.
  • system 100 may include a user device 102 a , content provider 104 , network 108 , and summarization system 110 .
  • User device 102 a may be, for example, but not limited to, a cellular telephone, Session Initiation Protocol (“SIP”) phone, software client/phone, a desktop computer, a laptop/notebook, a server, a module, a satellite phone, a personal digital assistant (“PDA”), a tablet computer, a smartphone, a remote controller, a personal computer (“PC”), a workstation, a handheld PC, a handheld MP3 player, a handheld video player, a personal media player, a gaming device, a thin system, a fat system, a network appliance, and/or other mobile communication device that may be capable of transmitting and/or receiving data.
  • SIP Session Initiation Protocol
  • PDA personal digital assistant
  • PC personal computer
  • workstation a
  • user device 102 a may include one or more transmitters, receivers, and/or transceivers to transmit and/or receive one or more signals to and/or from other components depicted in FIG. 1 , including, for example, content provider 104 , network 108 , and/or summarization system 110 .
  • Network 108 may be a wireless network, a wired network, or any combination of wireless network and wired network.
  • network 108 may include one or more of a fiber optics network, a passive optical network, a cable network, an Internet network, a satellite network (e.g., operating in Band C, Band Ku or Band Ka), a wireless LAN, a Global System for Mobile Communication (“GSM”), a Personal Communication Service (“PCS”), a Personal Area Network (“PAN”), D-AMPS, Wi-Fi, Fixed Wireless Data, Long Term Evolution (LTE) network, LTE Advanced (LTE-A) network, Mobile WiMAX (IEEE 802.16e), Mobile WiMAX Release 2 (IEEE 802.16m) network, IEEE 802.11a, 802.11b, 802.15.1, 802.11n and 802.11g or any other wired or wireless network for transmitting and/or receiving a data signal.
  • GSM Global System for Mobile Communication
  • PCS Personal Communication Service
  • PAN Personal Area Network
  • D-AMPS
  • network 108 may include, without limitation, telephone line, fiber optics, IEEE Ethernet 802.3, a wide area network (“WAN”), a local area network (“LAN”), or a global network such as the Internet.
  • network 108 may support, an Internet network, a wireless communication network, a cellular network, or the like, or any combination thereof.
  • Networks 108 may further include one, or any number of the exemplary types of networks mentioned above operating as a stand-alone network or in cooperation with each other.
  • Network 108 may utilize one or more protocols of one or more network elements to which it is communicatively coupled.
  • Network 108 may translate to or from other protocols to one or more protocols of network devices.
  • network 108 may comprise a plurality of interconnected networks, such as, for example, a service provider network, the Internet, a broadcaster's network, a cable television network, corporate networks, and home networks.
  • the components depicted in FIG. 1 may transmit and receive data to and from network 108 representing broadcast content, user request content, parallel search queries, parallel search responses, and other data.
  • the data may be transmitted and received utilizing a standard telecommunications protocol or a standard networking protocol.
  • a standard telecommunications protocol For example, one embodiment may utilize Session Initiation Protocol (“SIP”).
  • SIP Session Initiation Protocol
  • VOIP Voice Over IP
  • data may also be transmitted and/or received using Wireless Application Protocol (“WAP”), Multimedia Messaging Service (“MMS”), Enhanced Messaging Service (“EMS”), Short Message Service (“SMS”), Global System for Mobile Communications (“GSM”) based systems, Code Division Multiple Access (“CDMA”) based systems, Transmission Control Protocol/Internet (“TCP/IP”) Protocols, or other protocols and systems suitable for transmitting and receiving broadcast or parallel search data.
  • WAP Wireless Application Protocol
  • MMS Multimedia Messaging Service
  • EMS Enhanced Messaging Service
  • SMS Short Message Service
  • GSM Global System for Mobile Communications
  • CDMA Code Division Multiple Access
  • TCP/IP Transmission Control Protocol/Internet Protocols
  • Data may be transmitted and received wirelessly or may utilize cabled network or telecom connections such as an Ethernet RJ45/Category 5 Ethernet connection, a fiber connection, a traditional phone wireline connection, a cable connection or other wired network connection.
  • Network 108 may use standard wireless protocols including IEEE 802.11a, 802.11b and 802.11g.
  • Network 108
  • Content provider 104 may be one or more web-based media sources. Content provider 104 may host one or more websites that include news articles, commentary, blogs, videos, graphics, images, and other media content. Content provider 104 may aggregate media content from a plurality of other sources. User device 102 a and/or summarization system 110 may access content provider 104 via network 108 .
  • User device 102 a may include summarization application 102 b .
  • Summarization application 102 b may be a combination of software and/or hardware configured to provide one or more GUIs for user device 102 a to transmit data to and from summarization system 110 .
  • Summarization application 102 b may be a local application on user device 102 a .
  • Summarization application 102 b may be a web-based application hosted by summarization system 110 .
  • summarization system 110 may include a user interface module 202 , a parsing module 204 , an extraction module 206 , a classification module 208 , a deduplication module 210 , a preprocessing module 212 , a rules module 214 , a marker module 216 , and a summarization module 218 .
  • the modules 202 , 204 , 206 , 208 , 210 , 212 , 214 , 216 , and 218 are exemplary and the functions performed by one or more of the modules may be combined with that performed by other modules.
  • summarization system 110 may be a resolution server or may be a module of, or communicatively coupled to, a Domain Name System (“DNS”) server, such as a BIND server, for converting host names and domain names into IP addresses over the Internet.
  • DNS Domain Name System
  • Summarization system 110 may comprise one or more network enabled computers.
  • a network-enabled computer system and/or device may include, but is not limited to: e.g., any computer device, or communications device including, e.g., a server, a network appliance, a personal computer (PC), a workstation, a mobile device, a phone, a handheld PC, a personal digital assistant (PDA), a thin client, a fat client, an Internet browser, or other device.
  • PC personal computer
  • PDA personal digital assistant
  • Interface module 202 may be configured to interface with user device 102 a and/or content provider 104 .
  • Interface module 202 may receive data from user device 102 a and/or content provider 104 and provide data to user device 102 a and/or content provider 104 .
  • the user of user device 102 a may interact with interface module 202 using manual input (e.g., typing into a keyboard or keypad, etc.), voice input, touch screen input, graphical input (e.g., camera or camcorder) and/or any other method for inputting information or data to user device 102 a.
  • manual input e.g., typing into a keyboard or keypad, etc.
  • voice input e.g., touch screen input
  • graphical input e.g., camera or camcorder
  • Parsing module 204 is configured to parse one or more Rich Site Summary (RSS) feeds.
  • the RSS feeds may be provided by content provider 104 .
  • the RSS feed may include one or more entries identifying an article published by content provider 104 .
  • the RSS feeds may be monitored by interface module 202 and/or parsing module 204 .
  • the RSS feeds may be a precompiled list of RSS feeds that are published and constantly updated by content provider 104 .
  • Parser module 204 stores a precompiled list of RSS feeds in data storage 120 and retrieves them at regular intervals. Parser module 204 may parse the RSS feeds using a third-party open source Python module called Feedparser, for example.
  • Parser module 204 is configured to parse the RSS feeds to retrieve one or more URLs. Each URL corresponds to a specific article. For each URL retrieved, Parser module 204 also extracts a title and a publication date for that article.
  • Parser module 204 may maintain a list of previously-retrieved URLs in data storage 120 and compare the latest URL with the URLs in the list to ensure that the URL has not already been retrieved. Parser module 204 may delete URLs from the list of previously-retrieved URLs that were created before a certain date. For example, parser module 204 may delete URL entries whose creation date is older than one month.
  • Parser module 204 may prepare a set of articles based on the information extracted from the RSS feeds. Each article in the set of articles is associated with a title, a URL, and a publication date. Once parser module 204 has parsed a group of RSS feeds and extracted the URL, title, and publication date for each article, for each article in the set of articles, extraction module 206 may retrieve the textual content embedded in the HTML code of the article, by, for example, scraping the textual content embedded in the article's HTML code. Extraction module 206 may use one or more third-party open source scraping programs, such as, for example, Goose Extractor, Justtext, or some combination of the two.
  • third-party open source scraping programs such as, for example, Goose Extractor, Justtext, or some combination of the two.
  • extraction module 206 Once extraction module 206 has extracted the body of text from the HTML code for each article, extraction module 206 then removes boilerplate text snippets that do not belong to the body of text of the reported story. Examples of boilerplate text may include copyright notices, and error messages generated during the scraping and image captioning text. Extraction module 206 may also extract the author of the article, which is used as an anchor to remove automatically generated paragraphs that describe the background and experience of the author reporting the story.
  • Extraction module 206 then analyzes the base URL of the article to reduce it to its fully qualified domain name (FQDN).
  • Extraction module 206 may store the FQDN of the news articles' URLs in order to keep track of news stories published by the same news outlet.
  • classification module 208 receives the body of text for each article from extraction module 206 and classifies the article into one or more categories. Categories may include for example, and without limitation, U.S. News, World News, Business, Technology, Politics, Sports, Entertainment, Science, and Health.
  • Preprocessing module 212 is configured to implement one or more preprocessing methods to the body of text and/or the title of each article in the set of articles.
  • Preprocessing module 212 may first implement a tokenization step. During tokenization, the body of text for an article is broken up into words, phrases, symbols and other elements known as tokens.
  • preprocessing module 212 may implement a lowercasing step. During lowercasing, the body of text tokens are normalized by lowercasing them.
  • preprocessing module 212 may implement a stopwords removal step. During stopwords removal, grammatical words, known as stopwords, are removed from the tokenized and lowercased bodies of text.
  • Stopwords include, without limitation, “a”, “about”, “after”, “because”, “between”, “the”, “for”, “or”, etc.
  • preprocessing module 212 may implement a stemming step. During stemming, remaining body of text tokens are stemmed, that is, reduced to their root form by removing inflectional word endings.
  • classification module 208 may vectorize the body of text by using a bag-of-words representation using a mixture of unigrams and bigrams.
  • Classification module vectorizes the text by converting the text from words into a sequence of numbers in the form of a one dimensional array.
  • a vector encodes the textual information in the form of numbers that encapsulate information about the text. For example, the sentence “The dog chased the cat”, has 4 word types (the, dog, chased, cat).
  • a vector of this sentence this sentence may represent the number of times each word type appears in the sentence: [2, 1, 1, 1], where dimension 1 means the number of times “the” appears, dimension 2 means the number of times dog appears, and so on.
  • the bag-of-words representation is a form of textual vectorization that does not take into account the position in which word types appear in texts. This means that a sentence like “The dog chased the cat” has the same bag-of-words representation as the sentence “The cat chased the dog”. Both sentences have the very same 4 word types (the, dog, chased, cat). The sentence “The dog chased the cat” has 5 unigrams: “the”, “dog”, “chased”, “the”, “cat”, that is each individual word token. The same sentence has 3 bigrams: “the dog”, “dog chased”, “chased cat”, that is sequences of immediately adjacent words.
  • the generated vector space is first transformed by using the TF-IDF algorithm and, second, its dimensionality is reduced using the Truncated SVD algorithm.
  • TF-IDF term frequency-inverse document frequency
  • the tf-idf value increases proportionally to the number of times a word appears in the document, but is offset by the frequency of the word in the corpus, which helps to control for the fact that some words are generally more common than others.
  • the Truncated SVD algorithm is an algorithm specialized in reducing the dimensionality of vector representations.
  • the sentence “The dog chased the cat” may be represented as a vector of the following form: [2, 1, 1, 1]. This vector has 4 dimensions.
  • the Truncated SVD algorithm applies a set of mathematical functions that analyze each of the dimensions and convert a vector to another vector with a lower dimensionality that encodes the same information in a compacted/condensed manner.
  • classification module 208 is classifying each article into one of a plurality of news categories.
  • categorization module 208 may use the following 9 categories: U.S. News, World News, Business, Technology, Politics, Sports, Entertainment, Science, and Health. In some instances, a single article may be assignable to more than one category.
  • classification module 208 includes a classifier that distinguishes between US News and World News, US News and Business News, US News and Technology News, Technology News and Sports News, Technology News and World News, etc. Each of these is a binary classification where a news item is classified as one category or the other. The category that receives more votes, that is, that is selected more times for a particular article, is the category that it is finally assigned to that article.
  • Deduplication module 210 is configured to determine whether the body of text for each article in the set of articles represents an article or story that is a duplicate of another article or story that has already been processed and/or summarized. Deduplication module 210 will only compare articles that have been placed in the same category by classification module 208 . Deduplication module 210 may use a two stage approach. First, deduplication module 210 may compare new articles within the same category to one another to find duplicate stories. Next, deduplication module may compare the new articles to articles that were published and previously stored in data storage 120 within a certain time period. For example, deduplication module 210 may compare new articles to stored articles form the past 48 hours.
  • Deduplication module 210 may receive the set of articles that have gone through the previous processing steps described above. Each article may include the body text, title, date, URL, and other information. Deduplication module 210 may first use the titles of each article in the set as a filtering stage to reduce the search space to be explored for article deduplication. Only articles whose titles have a degree similarity higher than a predefined threshold are compared to each other to find duplicates.
  • Preprocessing module 212 may implement the one or more preprocessing steps on the title of each article using the same steps described previously: (1) tokenization; (2) lowercasing; (3) stopwords removal; and (4) stemming.
  • deduplication module 210 may vectorize the titles using a bag-of-words vectorization approach and the vectors are weighted using the algorithm known as TF-IDF. Finally vectors are reduced in dimensionality using the Latent Semantic Indexing (LSI) algorithm.
  • Latent semantic indexing (LSI) is an indexing and retrieval method that uses a mathematical technique called singular value decomposition (SVD) to identify patterns in the relationships between the terms and concepts contained in an unstructured collection of text. LSI is based on the principle that words that are used in the same contexts tend to have similar meanings. A key feature of LSI is its ability to extract the conceptual content of a body of text by establishing associations between those terms that occur in similar contexts.
  • Deduplication module 210 compares each title vector to the rest of the title vectors in the set using the cosine similarity algorithm, which outputs a similarity measure for each pair of compared vectors.
  • Cosine similarity is a measure of similarity between two vectors of an inner product space that measures the cosine of the angle between them. The cosine of 0° is 1, and it is less than 1 for any other angle.
  • Cosine similarity is thus a measure of orientation and not magnitude: two vectors with the same orientation have a Cosine similarity of 1, two vectors at 90° have a similarity of 0, and two vectors diametrically opposed have a similarity of ⁇ 1, independent of their magnitude.
  • Cosine similarity is particularly used in positive space, where the outcome is neatly bounded in [0,1].
  • the predefined threshold is a constant that is checked against the cosine similarity value that each pair of articles is assigned. So, in one embodiment, if a pair of article titles have a cosine similarity of 0.5, but the predefined threshold is 0.6, deduplication module 210 deem this pair of article titles as not similar enough. But if a pair of news items have a cosine similarity of 0.7 and the threshold is 0.6 then those two items are deemed as similar enough.
  • the value of this predefined threshold may be empirically set by inspecting the results obtained using different values and choosing the value that renders better results. The set of pairs that do not pass this filtering are removed from search space of potential duplicate pairs of articles.
  • deduplication module 210 is configured to further analyze each pair of articles based on the notion of “containment” as their similarity comparison measure.
  • Containment is a measure that defines similarity as set a intersection problem as described in the following, labeled Formula 1:
  • preprocessing module 212 may preprocess the body of text of each article in the pair using the following preprocessing steps; (1) tokenization, (2) lowercasing, and (3) stopwords removal. Once this has been done, for each pair of articles, deduplication module 210 may extract the unigrams, bigrams and trigrams from each pair of preprocessed bodies of text and converted into sets of tokens.
  • deduplication module 210 calculates the following two measures of containment according to Formula 1:
  • Deduplication module 210 then compares c(A,B) with c(B,A) to determine the direction of containment. Thus, if Containment of A in B (c(B,A)) is higher than Containment of B in A (c(A,B)), A is marked as a superset of B, and vice versa.
  • the containment measure of the winning superset is higher than a predefined containment threshold, the article that is contained by the superset article is classified as a duplicate and removed.
  • the predefined containment threshold may be empirically set, similar to the threshold used for the cosine similarity algorithm that compared title vectors for two articles.
  • the winning article superset is retained for future comparisons with other articles.
  • the final result of repeating this process for all pairs of articles is a set of unique stories from which duplicate articles have been removed. This set of articles may be referred to as the final set of articles.
  • Preprocessing module 212 is configured to preprocess the final set of articles prior to summarization. For each article, preprocessing module 212 may first perform tokenization to break up the body of text into words, phrases, symbols and other elements known as “tokens”. Preprocessing module 212 may then perform sentence splitting on the tokenized article text by splitting the text into individual sentences by using punctuation as anchors to determine sentence boundaries. Preprocessing module 212 may then perform truecasing on the tokens of each individual sentence to normalize the tokens to their true canonical casing. Preprocessing module 212 may then perform stopwords removal to remove stopwords from individual tokenized and truecased sentences. Finally, preprocessing module 212 may perform stemming by reducing each remaining token to its root form by removing inflexional word endings.
  • rules module 214 applies a series of rules to each article to accomplish the following steps: 1) extract a series of preliminary numeric measures to be reused by summarization module 218 to score and rank potential candidate sentences; and 2) to identify linguistic markers that can potentially impact the coherence of the generated summary.
  • Potential candidate sentences are sentences within an article that may be used as part of the generated summary.
  • rules module 214 is configured to extract a number of preliminary statistics in order to provide a first process for sentence scoring and ranking within each article.
  • the first type are positive scoring factors, which contribute to positively score sentences higher in the ranking.
  • the second type are negative scoring factors, which diminish a sentence score, thus, lowering its ranking.
  • Rules module 214 is configured to scale all negative and positive scoring factors are using the following formula, labeled as Formula 2:
  • alpha is constant that smooths the scale into which factor is converted.
  • Alpha is an empirically set constant that affects the effect that a smoothing function will have on a particular data series. In this equation higher values of alpha may intensify the effect of the smoothing function, whereas lower values of alpha may reduce the effect of the smoothing function.
  • the factor_score may the score of the scoring factor being considered.
  • Rules module 214 is configured to collect positive scoring factors for each individual sentence in an article.
  • the first positive scoring factor is the sentence position in the article. This factor is used to favor sentences that are closer to the beginning of the article as these sentences tend to include the most relevant pieces of information for each particular article.
  • rules module 214 may use a scaling formula, Formula 3, which is slightly different than Formula 2:
  • the next positive scoring factor is the number of frequent words present in the sentence.
  • rules module 214 may define a “frequent word” as a word that appears a minimum number of times in an article.
  • rules module 214 may define a frequent word as a word that appears at least 3 times in an article. This factor (document frequent words) is used to favor sentences that contain instances of the most frequently used words in the document.
  • Another positive scoring factor may be the number of document Named Entities present in the sentence. This factor is used to favor sentences that contain instances of the Named Entities (e.g., a proper noun) present in the document.
  • Another positive scoring factor may be the number of sentiment words, that is, words with either positive or negative polarity, present in the sentence.
  • Rules module 214 may maintain a lexicon of sentiment terms that are used to favor sentences that are positively or negatively subjective. For each sentence in the article, rules module 214 may compare the words in the sentence with the lexicon of sentiment terms to determine how many sentiment words are present in that sentence.
  • Another positive scoring factor may be the number of money expressions present in the sentence (e.g., “$20”, “$2”, “$1,000”).
  • a simple regular expression is used to identify and favor sentences that contain money-related expressions.
  • the regular expression may be: ⁇ $ ⁇ d+,? ⁇ d* ⁇ .? ⁇ d*. This regular expression accepts strings that start with a dollar sign, followed by one of more digits, followed by an optional comma, followed by 0 or more digits, followed by an optional decimal point, followed by 0 or more digits.
  • Rules module 214 is configured to collect negative scoring factors for each individual sentence in the article.
  • One negative scoring factor may be the length of the sentence. This factor is used to favor sentences that are not too long nor too short.
  • Formula 4 a slightly modified version of Formula 2 is used:
  • sent_length represents the length of the sentence being considered.
  • Doc_sent_length mean represents the mean of the sentence lengths for the entire article.
  • Another negative scoring factor may be the number of sentence negative features. This factor is used to favor sentences that don't have any of the following features present: (1) last character is not a question mark (?); (2) last character is not a colon (:); (3) a demonstrative (this, that, etc.) is not present in the first positions of the sentence; and (4) opening quotations, parentheses or square brackets are not present in the first positions of the sentence.
  • rules module 214 will combine the positive and negative scoring factors of each sentence using Formula 5, resulting into the final preliminary score that will be assigned to that sentence:
  • rule_score sum(positive_scoring_factors) ⁇ sum(negative_scoring_factors).
  • marker module 216 is configured to split each sentence of the article into one of three different groups: Potential candidates, weakly rejected candidates, and strongly rejected candidates.
  • Potential candidates are sentences for which no potentially impact in summary coherence has been detected, therefore, they are signaled as fully-qualified candidates for the summary to be generated by summarization module 218 .
  • Weakly rejected candidates are sentences for which linguistic markers that could negatively impact the summary coherence have been identified. These sentences are banned from being included in the generated summary but they're kept for further processing by summarization module 218 as they contain valuable information about the article content that needs to be explored.
  • Strongly rejected candidates are sentences that, because they contain some kind of boilerplate element, completely discarded as they are deemed as not containing valuable content for the article.
  • Marker module 216 is configured to designate a sentence as a weakly rejected sentence based on one or more linguistic markers found in the sentence. These linguistic markers represent local, sentence-to-sentence transition of ideas and concepts within the article discourse. Weakly rejected sentences break the sentence-to-sentence transition of ideas and concepts that the linguistic markers articulate locally. With their local coherence broken, these sentences negatively impact the global coherence of the summary by introducing non-contextualized idea transitions. Marker module 216 may evaluate each sentence based on a predetermined list of weak linguistic markers that indicate that a sentence should be designated as a weakly rejected sentence.
  • weak linguistic markers may include personal pronouns, demonstrative pronouns, adversative conjunctions, concessive conjunctions, coordinating conjunctions, adverbial discourse connectors, deictic metadiscourse connectors, and temporal adverbs.
  • marker module 216 may designate a sentence as a weakly rejected sentence that doesn't include a verb, as these sentences are normally section titles that contribute to the visual understanding of the article without contributing to its meaning. Marker module 216 may require a predetermined number of these weak linguistic markers to be found in a sentence before marker module 216 will designate that sentence as a weakly rejected sentence. In various embodiments, all weakly linguistic markers are weighted equally.
  • Marker module 216 is configured to designate a sentence as a strongly rejected sentence based on one or more strong linguistic markers. Strongly rejected sentences are sentences that do not contribute to the concepts and ideas of that the article elaborates. Thus, these sentences could be considered discourse boilerplate. Marker module 216 may evaluate each sentence based on a predetermined list of strong linguistic markers that indicate that a sentence should be designated as a strongly rejected sentence. These strong linguistic markers may include gratitude nouns and verbs, greetings, and/or exclamations.
  • marker module 216 may designate a sentence as a strongly rejected sentence based on the presence of lexical terms that belong to the jargon of hypertext, like the verb ‘click’ in “Click here to find out more”, are also used as clues that signal strongly rejected sentences. If marker module 216 evaluates a sentence and determines that the sentence is neither a weakly rejected sentence nor a strongly rejected sentence, marker module 216 may classify that sentence as a potential candidate sentence. At the end of this process, marker module 216 may have created three sets of sentences for each article—a set of potential candidate sentences, a set of weakly rejected sentences, and a set of strongly rejected sentences.
  • Summarization module 218 is configured to generate an article summary for each article in the final set of articles.
  • the article summary is comprised of a series of one or more sentences from the set of potential candidate sentences, as designated by marker module 216 .
  • Summarization module 218 may generate the article summary by implementing a modified version of the algorithm known as Maximal Marginal Relevance to iteratively and dynamically rank potential candidate sentences.
  • Maximal Marginal Relevance is an Information Retrieval (IR) algorithm that measures the relevant novelty of a document.
  • IR Information Retrieval
  • a first approximation to measuring relevant novelty is to measure relevance and novelty independently and provide a linear combination as the metric. This linear combination may be described as “marginal relevance”, i.e., a document has high marginal relevance if it is both relevant to the query and contains minimal similarity to previously selected documents.
  • MMR's basic formula can be derived as Formula 6:
  • lambda is a weighting constant
  • sim is the cosine similarity algorithm
  • Di is a candidate document
  • Q is the query
  • DP is the pool of documents already selected as relevant in the previous iterations of the algorithm.
  • summarization module 218 applies this IR paradigm to generate the article summary by treating the title of the article as the query and the set of potential candidate sentences as the collection of documents searched by the query.
  • summarization module 218 uses the following modified version, Formula 7, of the original MMR algorithm.
  • lambda is a weighting constant
  • sim is the cosine similarity algorithm
  • Si is a potential candidate sentence
  • T is the title of the article
  • DC is the centroid of the article vector space resulting from the combination of both the potential candidate sentences and the weakly rejected sentences sets
  • Se_score is the score calculated for the potential candidate sentence by rules module 214
  • Su is the dynamically generated summary, that is, the set of sentences picked to be included in the summary in previous iterations of the algorithm.
  • lambda may be empirically set value of 0.6.
  • Summarization module 218 may select sentences from the potential candidate sentences set to be included in the generated summary, whereas weakly rejected sentences and strongly rejected sentences cannot be used in the generated summary.
  • summarization module 218 uses term statistics for sentences from both sets (potential candidate sentences and weakly rejected sentences) to generate a combined vector space, which is first transformed by using the TF-IDF algorithm.
  • the TF-IDF algorithm is a numerical statistic that transforms a vector by measuring the importance that each dimension has across the collection of all vectors. This algorithm intensifies the importance of dimensions that are infrequent across all vectors to boost their discriminative power.
  • summarization module 218 reduces the TF-IDF transformed vectors in dimensionality by applying the Truncated SVD algorithm.
  • the Truncated SVD algorithm performs a factorization of a vector matrix to project it into a lower dimensionality space that encodes the information contained in the higher dimensionality space by combining the value of interdependent dimensions.
  • Summarization module 218 then calculates the centroid of the resulting combined and modified vector space by averaging the values of each feature across the entire vector space.
  • the centroid of a vector space represents the point that lays in the center of that high dimensional space.
  • the centroid is a vector with the same dimensionality as the vectors that compose the vector space.
  • Summarization module 218 then reduces the vector space by removing the vectors that represent the sentences signaled as weakly rejected sentences, preventing them this way from being processed by Formula 7 and, therefore, from being included in the generated summary.
  • summarization module 218 For a given article, summarization module 218 generates the article summary by iteratively running each of the sentences from the set of potential candidate sentences through Formula 7. In the first iteration, summarization module 218 determines the MMR for each potential candidate sentence in the article, then selects the sentence with the highest MMR and adds it to the article summary. Once a sentence has been added to the article summary, summarization module 218 removes that sentence from the set of potential candidate sentences for that article in order to prevent it from being repeatedly selected.
  • summarization module 218 performs the same process (only this time the set of potential candidate sentences has been modified to remove the previously selected sentence), and selects the sentence having the highest MMR in that iteration, adds that sentence to the article summary, and removes it from the set of potential candidate sentences.
  • Summarization module 218 repeats this process until the number of characters in the article summary reaches a maximum threshold.
  • the maximum threshold may have been preprogrammed into summarization module 218 .
  • the maximum threshold may be 325 characters.
  • a user may be able to provide the maximum threshold via interface module 202 .
  • summarization module 218 may reorder the sentences in the article summary so that their position in the summary follows the same occurrence order of the sentences in the source article.
  • FIG. 3 is a flowchart illustrating the functionality for generating an article summary based on an article title and body of text.
  • This exemplary method 300 may be provided by way of example, as there are a variety of ways to carry out the method.
  • the method 300 shown in FIG. 3 can be executed or otherwise performed by one or a combination of various systems.
  • the method 300 is described below may be carried out by the systems and networks shown in FIGS. 1 and 2 , by way of example, and various elements of the systems and networks are referenced in explaining the example method of FIG. 3 .
  • Each block shown in FIG. 3 represents one or more processes, methods or subroutines carried out in exemplary method 300 .
  • exemplary method 300 may begin at block 302 .
  • summarization system 110 retrieves an RSS feed.
  • the RSS feed may include one or more entries identifying articles from content provider 104 .
  • Summarization system 110 may regularly retrieve one or more RSS feeds and store them in data storage 120 .
  • Method 300 may proceed to block 306 .
  • summarization system 110 generates an initial set of articles from the RSS feed.
  • Summarization system 110 may first parse the RSS feeds to retrieve a list of URLs, where each URL is associated with an article.
  • Summarization system 110 may compare the list of URLs to a list of previously retrieved URLs in data storage 120 and remove any matching URLs.
  • Summarization system 110 may also parse the RSS feed to retrieve the title and publication date of each article in the feed.
  • Summarization system 110 may then retrieve the textual content embedded in the HTML code of the article from the URL.
  • Summarization system 110 may remove boilerplate text snippets that do not belong to the body of text of the reported story.
  • boilerplate text may include copyright notices, and error messages generated during the scraping and image captioning text.
  • Extraction module 206 may also extract the author of the article, which is used as an anchor to remove automatically generated paragraphs that describe the background and experience of the author reporting the story. Summarization system may then generate an initial set of articles, where each article in the set includes the title, publication date, source, and body text of the article. Method 300 may proceed to block 308 .
  • summarization system 110 classifies each article in the initial set of articles.
  • Summarization system 110 may classify each article into one or more news categories.
  • News categories may include for example, and without limitation, U.S. News, World News, Business, Technology, Politics, Sports, Entertainment, Science, and Health.
  • Summarization system 110 may apply one or more preprocessing steps to the body of text and/or title of each article in the initial set of articles. The steps may include tokenization, lowercasing, stopwords removal, and stemming. Following these preprocessing steps, summarization system 110 may generate a vectorized document for each body of text and categorize the article text into one of the news categories. Method 300 may proceed to block 310 .
  • summarization system 110 removes duplicate articles from the initial set of articles to generate a final set of articles.
  • This process may involve a two stage approach. First, summarization system 110 may compare articles with other articles in the same category to find duplicate articles. Next, summarization system 110 may compare the articles in the initial set of articles to articles that were published and previously stored in data storage 120 within a certain time period. For example, summarization system 110 may compare new articles to stored articles form the past 48 hours. Using the titles of each article in the initial set of articles, summarization system 110 may implement one or more algorithms to quantify the degree of similarity between a pair of articles. A pair of articles with titles that meet a predetermined degree of similarity will then be compared using a containment algorithm, described above as Formula 1. Summarization system 110 may use this to remove duplicates from the initial set of articles to generate a final set of articles. Method 300 may proceed to block 312 .
  • summarization system 110 For each article in the final set of articles, summarization system 110 generates a preliminary score for each sentence in that article.
  • Summarization system 110 may first preprocess each article in the final set of articles by performing tokenization, sentence splitting, truecasing, stopwords removal, and stemming. Summarization system 110 then evaluates each sentence in an article and assigns it a score. The score may be based on a series of positive and negative scoring factors. Positive scoring factors lead to a higher score and a higher rank for the sentence in the article. Negative scoring factors lower the score and lead to a lower rank for the sentence within the article. The positive and negative scoring factors may be scaled using Formula 2, as described previously.
  • Positive scoring factors may include the position of the sentence in the article, the number of frequent words present in the sentence (where a frequent word is a word that appears a minimum number of times in the article), the number of Named Entities present in the sentence, the number of sentiment words in the sentence, and/or the number of money-related expressions in the sentence.
  • Negative scoring factors may include the length of the sentence, and the presence or absence of different negative features. The positive and negative scoring factors for each sentence may be determined and combined using Formula 2, Formula 3, Formula 4, and/or Formula 5, as described above. Method 300 may proceed to block 314 .
  • summarization system 110 may classify each sentence in an article into one of three categories: (1) potential candidate sentences; (2) weakly rejected sentences; and (3) strongly rejected sentences.
  • Potential candidate sentences are sentences for which no potentially impact in summary coherence has been detected, therefore, they are signaled as fully-qualified candidates for the article summary.
  • Weakly rejected candidates are sentences for which linguistic markers that could negatively impact the summary coherence have been identified. These sentences are banned from being included in the generated summary but they're kept for further processing as they contain valuable information about the article content that needs to be explored.
  • Strongly rejected candidates are sentences that, because they contain some kind of boilerplate element, completely discarded as they are deemed as not containing valuable content for the article.
  • Summarization system 110 may classify a sentence as a weakly rejected sentence based on the presence of one or more weak linguistic markers.
  • Weak linguistic markers may include personal pronouns, demonstrative pronouns, adversative conjunctions, concessive conjunctions, coordinating conjunctions, adverbial discourse connectors, deictic metadiscourse connectors, and/or temporal adverbs.
  • Summarization system 110 may classify a sentence as a weakly rejected sentence if the sentence does not include a verb.
  • Summarization system 110 may classify a sentence as a strongly rejected sentence based on the presence of one or more strong linguistic markers. Strong linguistic markers may include gratitude nouns and verbs, greetings, and/or exclamations.
  • Summarization system 110 may classify a sentence as a strongly rejected sentence based on the presence of lexical terms that belong to the jargon of hypertext, like the verb ‘click’ in “Click here to find out more”, are also used as clues that signal strongly rejected sentences. If a sentence does not contain a minimum number of strong linguistic markers and/or weak linguistic markers, summarization system 110 may classify the sentence as a potential candidate sentence. Method 300 may proceed to block 316 .
  • summarization system generates an article summary for each article in the final set of articles.
  • the article summary is comprised of a series of one or more sentences from the set of potential candidate sentences for that article.
  • Summarization system 110 uses an iterative algorithm, Formula 7, as described above, to construct the article summary from the set of potential candidate sentences for that article.
  • Summarization system 110 may then provide the article summary to be viewed on an interactive interface on a user device. Method 300 may proceed to block 318 .
  • method 300 may end.
  • FIGS. 4A-4D show different interface screens generated by summarization system 110 , according to an embodiment.
  • the screens may be provided to a user via an application on a mobile device (such as summarization application 102 b ) and/or a web interface.
  • FIG. 4A shows an initial screen for summarization application 102 b .
  • FIG. 4B shows a list of article titles from a final set of articles. The article titles also include the source for the article, listed below each title.
  • the screen shown in FIG. 4B may be interactive and receive inputs from the user. For example, the user may be able to select a given article title and be presented with the article summary, as shown in FIG. 4D .
  • FIG. 4C shows a list of articles that have been categorized under the category of “U.S. News.”
  • the list includes the article title and the source.
  • the screen may receive one or more selections from the user, as with the screen in FIG. 4B .
  • FIG. 4D depicts an article summary generated by summarization system 110 as described above.
  • the article summary is designed to be viewed on a single screen of user device 102 a .
  • the summary may include the article's title, source, date, author name, and/or byline. The user may be able to scroll through the summary and view the entire text of the article.
  • certain types of articles may require modified processing, based on the structure of the article.
  • These types of articles have an internal discourse structure does not fit the assumption that a unique discourse coherence articulates the paragraphs and sentences of the entire article from the beginning to the end.
  • many articles are a “Top X” list articles (e.g., “Top 10 Smartphones on the Market”, “15 best beaches in California”, “Top 50 NFL Players of all Time”).
  • the structure of these “Top X” list articles typically features a preliminary set of paragraphs that introduce the general topic discussed in the article, followed by a series of sections that examine this general topic from different angles. Each of these sections can be considered as a self-focused mini document with its own internal coherence.
  • any one of parser module 204 , extraction module 206 , and/or classification module 208 may check the title of an article using a set of functions to determine whether the article belongs in the category of “Top X” list articles.
  • these functions check for the existence of numbers of up to three digits that are not followed by time expressions (such as “years”, “months” or “days”, etc.) or the existence of numbers in word form (such as “four”, “ten”, “two”, etc.).
  • time expressions such as “years”, “months” or “days”, etc.
  • These modules also check the body text of the article to find the number of items listed in the article.
  • One or more modules may explore the HTML code of the article to identify a set of HTML tags that enclose the title of each section of the independent list items in the article. If the number of section titles in the article body text matches the number expressed in the title, the article is classified as a “Top X” list type article and the section titles are stored in a database for later use as structural anchors during summarization.
  • Preprocessing module 212 may compare the number of section titles with the number of text blocks in the articles. If the number of section titles matches the number of text blocks it is assumed that no introduction is present, whereas if the number of section titles is lower than the number of text blocks, one or more modules mark the first section as the article introduction.
  • preprocessing module 212 may generate a summary of this introductory section, as described above. Summarization module 218 may then append the list of section titles to the end of the generated summary, and this may be displayed on user device 102 . If no introduction is identified by preprocessing module 212 , then a summary is generated by simply appending the list of section titles found during scraping.
  • clustering module 220 may clustering articles to help avid news readers to follow particular stories without having to hunt for related reports in the continuous stream of the news cycle.
  • Clustering module 220 receives article documents that have been pre-processed by pre-processing module 212 (which applied lowercasing, tokenization, stopwords removal, and token stemming as described above).
  • Unigram tokens left after applying these pre-processing routines are used to update the TF and IDF indices. Subsequently, unigram tokens with a term frequency lower than 2 in the global TF index are filtered out.
  • f(t,d) is the number of times term t occurs in the article document d
  • N is the total number of documents in the collection of documents D and
  • is the number of documents in which term t occurs. If either tf(t,d) or idf(t,D) result in non-numerical or infinite values they are converted to 0.
  • This optimization consists in computing a correction cluster age factor that is subtracted from the cosine similarity:
  • cacf ⁇ ( caf /(60*60*24*30))
  • Cluster module 220 may filter out cluster centroids that have a similarity lower than an empirically defined similarity threshold with respect to the vectorized article document. Cluster module 220 may then assign the article document to the cluster with the greatest similarity among all the remaining clusters, that is, among all the clusters with a similarity greater than the predefined similarity threshold. If no cluster has a similarity greater than the predefined similarity threshold, the article document defines its own cluster and will be subsequently compared to each new document that is processed by the cluster module 220 .
  • Cluster module 220 may cause a cluster to be “frozen” (i.e., no new articles are added to the cluster) if the cluster's age becomes greater than a predefined age threshold or no new documents have been assigned to a cluster after a predefined period of time. Clusters may be presented to the user via the application interfaces shown in FIGS. 4 a - 4 d.
  • FIG. 5 is a schematic diagram illustrating a system according to particular embodiments.
  • system 500 may include a named entity recognition and classification (NERC) module 502 and a named entity disambiguation module 504 .
  • NERC named entity recognition and classification
  • modules 502 and 504 may be integrated into summarization system 110 (e.g., as part of named entities module 222 ).
  • NERC module 502 may extract Named Entities (NEs) mentions from an article. These inputs may then be provided to Named Entity disambiguation module 504 which may be configured to disambiguate the NEs, and identify the most salient NEs for each article.
  • named entities module 222 may link articles that share unique salient NEs so that users can follow particular entities they are interesting in.
  • NERC module 502 may first extract NEs from an article using a two-step process: (1) Named Entity Recognition, and (2) Named Entity Classification.
  • Named Entity Recognition is defined as the process of identifying the boundaries of NEs in sentences so that all tokens belonging to a NE are grouped together and all tokens outside the boundaries of a NE are excluded.
  • Named Entity classification is defined as the process of assigning a unique classification label to all NEs identified by during the Named Entity Recognition phase (e.g., Person, Organization, Location, or Miscellaneous). Both processes use machine learning.
  • NERC module 502 may then be provided to Named Entity Disambiguation module 504 , which may first map an entity mention extracted from documents to unique entries of a knowledge base.
  • An entity mention is a sequence of tokens in a text which refers to a specific NE and has been previously identified by NERC module 502 .
  • NERC module 502 Given the high ambiguity of language, a named entity can have multiple names and a name can be denoted by different named entities.
  • the main goal of the named entity disambiguation process is to disambiguate the entity mentions in their textual context and identify their concrete NE referent.
  • the knowledge base may include a Named Entity index 508 and various third party databases 510 such as Wikipedia (https://www.wikipedia.org/), Freebase (https://www.freebase.com/) or DBpedia (http://dbpedia.org/) or other data sources.
  • third party databases 510 such as Wikipedia (https://www.wikipedia.org/), Freebase (https://www.freebase.com/) or DBpedia (http://dbpedia.org/) or other data sources.
  • NERC module 502 may analyze an article that includes the sentence “With knowledge of such widespread ferocity, I recently saw Mel Gibson's movie ‘Apocalypto’ which deals with the gore of the Mayan civilization.” NERC module 502 may first extract the mentions “Mel Gibson” and “Apocalypto”. After that, NE disambiguation module 504 would retrieve from a knowledge base ( 508 , 510 ) a list of possible NE candidates that refer to each of those mentions (for example an actor and a basketball player for the mention “Mel Gibson”). The goal of the disambiguation module 504 is to identify the correct referents, in this case an actor and a film. The NE disambiguation process consists of three steps to retrieve, score and rank NE candidates for each entity mention previously extracted from articles during the NE recognition process. The output is a single disambiguated NE for each entity mention.
  • NE disambiguation module 504 gathers a list of NE candidates for each of the entity mentions extracted by the NERC module 502 . This may be extracted from one or more databases, such as index 508 and/or third party database 510 .
  • the index 508 may include NEs with the following information: title, alias (other names of the NE), links to other NEs, in-degree score (the count of links pointing to the entity), out-degree score (the count of links to other NEs).
  • NE disambiguation module 504 may use the Named Entity Index 508 to retrieve candidates for a given entity mention by searching on the title and alias fields.
  • the index For each candidate, the index provides a popularity score computed by the entity in-degree and out-degree link counts in the Wikipedia graph.
  • the popularity score is also enriched with the search relevance score provided by ElasticSearch, which is based on Lucene's practical scoring function (http://www.elasticsearch.org/guide/en/elasticsearch/guide/current/practical-scoring-function.html).
  • ElasticSearch is based on Lucene's practical scoring function (http://www.elasticsearch.org/guide/en/elasticsearch/guide/current/practical-scoring-function.html).
  • the second step is a set of scoring functions that aim at obtaining two main representations of each NE candidate.
  • the first is a confidence score that is an independent representation of each single NE candidate without considering any other candidates of the article. This score includes two dimensions: (1) Local NE confidence, and (2) Global NE confidence.
  • Local NE confidence is the sum of: a) cosine similarity of the NE candidate name and alias against the sequence of the entity mention as found in the article (e.g. “mention: Mel Gibson” vs.
  • the Global NE confidence refers to the popularity of the candidate in the knowledge base where it was obtained.
  • the second score is a collective score.
  • Named entity disambiguation module 504 may generate a graph where it stores all NE candidates as nodes and existing relations among them as edges. The edges are drawn only between candidates of different entity mentions and only if a given relation is found using the links of the entity page from Wikipedia. Using this graph, the collective score is computed by generating the PageRank of each node.
  • Named entity disambiguation module 504 may generate different ranks for each entity mention and the dynamic selection of the best top rank.
  • a rank is an ordered list of the NE candidates generated by some combinations of the scores described above. The main ranks used are Rank by Local Confidence+PageRank, and Rank by Combined Confidence (Local+Global Confidence)+PageRank.
  • Named Entity Disambiguation module 504 selects the best top rank by normalizing the scores of each list to sum to 1 and computing the difference between the scores of the top two candidates in the given list. The idea behind this is that a bigger difference in the scores would indicate more confidence for the discrimination of candidates. In this way, the final NE candidate is selected from the rank with a highest difference score. This method equilibrates the cases where a false candidate is selected as the top one because it has a significantly higher popularity score (retrieved by the global confidence); in this case, the local confidence provides a higher difference between the top candidates and avoids the selection of the wrong one.
  • Named entity disambiguation module 504 may analyze and classify the NEs to determine what NEs are most salient. As pointed out by (Gamon et al., 2013a: 1), “salience and importance/prominence are not the same”. Whereas importance/prominence is a function of the reader needs, that is, what the reader considers to be most important/prominent; salience is a function of the author, that is, the entities that the author considers to be the most relevant for a particular article. In other words, the most salient NEs are the most prominent entities from a content and tendive point of view, better summarizing the information encapsulated in an article.
  • NESI short for Named Entity Salience Identification
  • nyt-salience is a Machine Learning module trained using a subset of the New York Times Salience corpus (known as nyt-salience; it can be downloaded from https://code.google.com/p/nyt-salience/), which accompanied the paper Dunietz & Gillick, 2014.
  • the New York Times Salience corpus extends the 1.8 million articles compiled in the New York Times Annotated Corpus (https://catalog.ldc.upenn.edu/LDC2008T19) with NEs automatically extracted from those articles that are primarily annotated with their Freebase mid, that is their unique id, and their automatically inferred salience.
  • NESI was trained using a subset of the New York Times Salience corpus. In particular we used the first 30000 articles, but discarded all articles for which no salient entities were automatically identified by the algorithm that Dunietz & Gillick (2014) used. Finally, in order to use a more balanced dataset, non-salient NEs were randomized and reduced to the same number of salient NEs. During training this proved to be beneficial as it reduced the bias of the trained algorithm towards non-salient NEs.
  • NESI was trained using a supervised Support Vector Machines algorithm with RBF kernel that learned from the following features: The complete Named Entity is in the article's title; The complete Named Entity is in the article's first line; Named Entity count; Named Entity start byte-offset; Number of Named Entity words that are in the article's title; Length in words of the Named Entity; Length in words of the article's title; and the Number of related entities that are present in the article's list of extracted Named Entities after their disambiguation and linking to a Knowledge Base (Wikipedia).
  • Wikipedia Knowledge Base
  • NESI's with respect to Dunietz & Gillick's are that: 1) NESI is lighter in terms of running time and processing power as it doesn't need to run a PoS tagger, a dependency parser and a NP extractor; and 2) The feature set NESI uses is not lexicalized, and therefore not dependent on the corpus used for training and evaluation.
  • Heuristic 1 If no NE has been found to be salient, all NEs that are either present in the title or the first line of the article and their count is greater than 1 it is classified as salient.
  • Heuristic 2 If after applying heuristic 1 no NE has yet been found to be salient, the NE with the lowest start offset and the greatest count is classified as salient.
  • the various computing devices above generally include computer-executable instructions, where the instructions may be executable by one or more processors.
  • Computer-executable instructions may be compiled or interpreted from computer programs created using a variety of programming languages and/or technologies, including, without limitation, and either alone or in combination, JavaTM, C, C++, Visual Basic, Java Script, Perl, etc.
  • a processor or microprocessor receives instructions, e.g., from a memory, a computer-readable medium, etc., and executes these instructions, thereby performing one or more processes, including one or more of the processes described herein.
  • Such instructions and other data may be stored and transmitted using a variety of computer-readable media.
  • Databases, data repositories or other data stores described herein may include various kinds of mechanisms for storing, accessing, and retrieving various kinds of data, including a hierarchical database, a set of files in a file system, an application database in a proprietary format, a relational database management system (RDBMS), etc.
  • Each such data store is generally included within a computing device employing a computer operating system such as one of those mentioned above, and are accessed via a network in any one or more of a variety of manners.
  • a file system may be accessible from a computer operating system, and may include files stored in various formats.
  • An RDBMS generally employs the Structured Query Language (SQL) in addition to a language for creating, storing, editing, and executing stored procedures, such as the PL/SQL language mentioned above.
  • SQL Structured Query Language

Abstract

Systems and methods for summarizing online articles for consumption on a user device are disclosed herein. The system extracts the main body of an article's text from the HTML code of an online article. The system may then classify the extracted article into one of several different categories and removes duplicate articles. The system breaks down the article into its component sentences, and each sentence is classified into one of three categories: (1) potential candidate sentences that may be included in the generated summary; (2) weakly rejected sentences that will not be included in the summary but may be used to generate the summary; and (3) strongly rejected sentences that are not included in the summary. Finally, the system applies a document summarizer to generate quickly readable article summaries, for viewing on the user device, using relevant sentences from the article while maintaining the coherence of the article.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application contains subject matter related to and claims the benefit of U.S. Provisional Patent Application No. 62/002,350, filed on May 23, 2014, the entire contents of which is incorporated herein by reference.
  • BACKGROUND INFORMATION
  • With the growth of online and digital media, today, more and more consumers are accessing media content on mobile platforms. Many consumers receive all of their news-related content on mobile platforms, instead of in print form. Many mobile platforms, such as smartphones, are not ideally suited for presenting long-form news content, due to the limited screen size. These and other drawbacks exist.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention, together with further objects and advantages, may best be understood by reference to the following description taken in conjunction with the accompanying drawings, in the several figures of which like reference numerals identify like elements, and in which:
  • FIG. 1 is a schematic diagram illustrating a system according to a particular embodiment;
  • FIG. 2 is a block diagram of a hardware component of the monitoring system according to a particular embodiment;
  • FIG. 3 is a flowchart illustrating the functionality of a method according to a particular embodiment;
  • FIG. 4A is a screenshot of a capsule screen according to a particular embodiment;
  • FIG. 4B is a screenshot of a particular embodiment;
  • FIG. 4C is a screenshot of a particular embodiment;
  • FIG. 4D is a screenshot of a particular embodiment; and
  • FIG. 5 is a block diagram of a hardware component according to a particular embodiment.
  • DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
  • The disclosure herein describes systems and methods for summarizing single document articles for consumption on a user device, specifically a mobile device (such as a smartphone or tablet). The systems and methods disclosed herein feature an innovative single document summarization engine that combines both unsupervised Machine Learning techniques and linguistically motivated rules to significantly improve the quality of generated summaries. The systems and methods may improve the quality of the generated summaries by reducing the risk of including sentences that will impact the summary coherence. Coherence is a linguistic concept that can be succinctly defined as the logical connections that humans perceive in a text (written or oral). The systems and methods disclosed herein block sentences that contain linguistic markers that build the internal coherence of articles while at the same time taking advantage of the content of these blocked sentences to select the set sentences of an article (such as a news article) that most relevantly summarize it. For example, an article may include several sentences with discourse markers and/or connectors like “however”, “in addition”, “meanwhile”, etc. Including these sentences in the generated summary could break the flow of ideas of the original article. For instance, a sentence that starts with the adversative “however” is contradicting an idea expressed before this sentence. Including this sentence in the summary without including the sentences that includes the idea that it is contradicting would lead to an incomprehensible summary from the point of view of its internal coherence. The summary is designed to fit on the screen of most tablets and smartphones.
  • The system may initially extract the main body of text from the HTML code of an article. The system may then classify the extracted article into one of several different categories. The system also identifies duplicate articles and removes them. The system breaks down each article into its component sentences, and each sentence is classified into one of three categories: (1) potential candidate sentences that may be included in the generated summary; (2) weakly rejected sentences that will not be included in the summary, but may be used to help generate the summary; and (3) strongly rejected sentences that are not included in the summary or used to generate the summary. Finally, the system applies a document summarizer, which features an innovative summarization algorithm that leverages linguistic information to reduce the risk of producing incoherent summaries. In this way, the system generates quickly readable article summaries, for viewing on the screen of a smartphone or tablet computer, using relevant sentences from the article that maintain the coherence of the article.
  • The description below describes user interface modules, parsing modules, extraction modules, classification modules, deduplication modules, preprocessing modules, rules modules, marker modules, summarization modules, user devices, content providers, computer systems, and networks that may include one or more modules, some of which are explicitly shown while others are not. As used herein, the term “module” may be understood to refer to computing software, firmware, hardware, and/or various combinations thereof. It is noted that the modules are examples. The modules may be combined, integrated, separated, and/or duplicated to support various applications. Also, a function described herein as being performed at a particular module may be performed at one or more other modules and/or by one or more other devices instead of or in addition to the function performed at the particular module. Further, the modules may be implemented across multiple devices and/or other components local or remote to one another. Additionally, the modules may be moved from one device and added to another device, and/or may be included in both devices.
  • It is further noted that software described herein may be tangibly embodied in one or more physical media, such as, but not limited to, a compact disc (“CD”), a digital versatile disc (“DVD”), a floppy disk, a hard drive, read only memory (“ROM”), random access memory (“RAM”), as well as other physical media capable of storing software, and/or combinations thereof. The functions described as being performed at various components may be performed at other components, and the various components may be combined and/or separated. Other modifications also may be made
  • FIG. 1 is a schematic diagram illustrating a system according to particular embodiments. As illustrated in FIG. 1, system 100 may include a user device 102 a, content provider 104, network 108, and summarization system 110. User device 102 a may be, for example, but not limited to, a cellular telephone, Session Initiation Protocol (“SIP”) phone, software client/phone, a desktop computer, a laptop/notebook, a server, a module, a satellite phone, a personal digital assistant (“PDA”), a tablet computer, a smartphone, a remote controller, a personal computer (“PC”), a workstation, a handheld PC, a handheld MP3 player, a handheld video player, a personal media player, a gaming device, a thin system, a fat system, a network appliance, and/or other mobile communication device that may be capable of transmitting and/or receiving data. Also, user device 102 a may include one or more transmitters, receivers, and/or transceivers to transmit and/or receive one or more signals to and/or from other components depicted in FIG. 1, including, for example, content provider 104, network 108, and/or summarization system 110.
  • Network 108 may be a wireless network, a wired network, or any combination of wireless network and wired network. For example, network 108 may include one or more of a fiber optics network, a passive optical network, a cable network, an Internet network, a satellite network (e.g., operating in Band C, Band Ku or Band Ka), a wireless LAN, a Global System for Mobile Communication (“GSM”), a Personal Communication Service (“PCS”), a Personal Area Network (“PAN”), D-AMPS, Wi-Fi, Fixed Wireless Data, Long Term Evolution (LTE) network, LTE Advanced (LTE-A) network, Mobile WiMAX (IEEE 802.16e), Mobile WiMAX Release 2 (IEEE 802.16m) network, IEEE 802.11a, 802.11b, 802.15.1, 802.11n and 802.11g or any other wired or wireless network for transmitting and/or receiving a data signal. In addition, network 108 may include, without limitation, telephone line, fiber optics, IEEE Ethernet 802.3, a wide area network (“WAN”), a local area network (“LAN”), or a global network such as the Internet. Also, network 108 may support, an Internet network, a wireless communication network, a cellular network, or the like, or any combination thereof. Networks 108 may further include one, or any number of the exemplary types of networks mentioned above operating as a stand-alone network or in cooperation with each other. Network 108 may utilize one or more protocols of one or more network elements to which it is communicatively coupled. Network 108 may translate to or from other protocols to one or more protocols of network devices. Although network 108 is depicted as one network, it should be appreciated that according to one or more embodiments, network 108 may comprise a plurality of interconnected networks, such as, for example, a service provider network, the Internet, a broadcaster's network, a cable television network, corporate networks, and home networks.
  • The components depicted in FIG. 1 may transmit and receive data to and from network 108 representing broadcast content, user request content, parallel search queries, parallel search responses, and other data. The data may be transmitted and received utilizing a standard telecommunications protocol or a standard networking protocol. For example, one embodiment may utilize Session Initiation Protocol (“SIP”). In other embodiments, the data may be transmitted and/or received utilizing other Voice Over IP (“VOIP”) or messaging protocols. For example, data may also be transmitted and/or received using Wireless Application Protocol (“WAP”), Multimedia Messaging Service (“MMS”), Enhanced Messaging Service (“EMS”), Short Message Service (“SMS”), Global System for Mobile Communications (“GSM”) based systems, Code Division Multiple Access (“CDMA”) based systems, Transmission Control Protocol/Internet (“TCP/IP”) Protocols, or other protocols and systems suitable for transmitting and receiving broadcast or parallel search data. Data may be transmitted and received wirelessly or may utilize cabled network or telecom connections such as an Ethernet RJ45/Category 5 Ethernet connection, a fiber connection, a traditional phone wireline connection, a cable connection or other wired network connection. Network 108 may use standard wireless protocols including IEEE 802.11a, 802.11b and 802.11g. Network 108 may also use protocols for a wired connection, such as an IEEE Ethernet 802.3.
  • Content provider 104 may be one or more web-based media sources. Content provider 104 may host one or more websites that include news articles, commentary, blogs, videos, graphics, images, and other media content. Content provider 104 may aggregate media content from a plurality of other sources. User device 102 a and/or summarization system 110 may access content provider 104 via network 108.
  • User device 102 a may include summarization application 102 b. Summarization application 102 b may be a combination of software and/or hardware configured to provide one or more GUIs for user device 102 a to transmit data to and from summarization system 110. Summarization application 102 b may be a local application on user device 102 a. Summarization application 102 b may be a web-based application hosted by summarization system 110.
  • As shown in FIG. 2, summarization system 110 may include a user interface module 202, a parsing module 204, an extraction module 206, a classification module 208, a deduplication module 210, a preprocessing module 212, a rules module 214, a marker module 216, and a summarization module 218. It is noted that the modules 202, 204, 206, 208, 210, 212, 214, 216, and 218 are exemplary and the functions performed by one or more of the modules may be combined with that performed by other modules. The functions described herein as being performed by the modules 202, 204, 206, 208, 210, 212, 214, 216, and 218 also may be separated and may be located or performed by other modules. Moreover, the modules 202, 204, 206, 208, 210, 212, 214, 216, and 218 may be implemented at other devices of the system 100 (e.g., content provider 104). Also, in various embodiments, summarization system 110 may be a resolution server or may be a module of, or communicatively coupled to, a Domain Name System (“DNS”) server, such as a BIND server, for converting host names and domain names into IP addresses over the Internet. Summarization system 110 may comprise one or more network enabled computers. As referred to herein, a network-enabled computer system and/or device may include, but is not limited to: e.g., any computer device, or communications device including, e.g., a server, a network appliance, a personal computer (PC), a workstation, a mobile device, a phone, a handheld PC, a personal digital assistant (PDA), a thin client, a fat client, an Internet browser, or other device.
  • Interface module 202 may be configured to interface with user device 102 a and/or content provider 104. Interface module 202 may receive data from user device 102 a and/or content provider 104 and provide data to user device 102 a and/or content provider 104. The user of user device 102 a may interact with interface module 202 using manual input (e.g., typing into a keyboard or keypad, etc.), voice input, touch screen input, graphical input (e.g., camera or camcorder) and/or any other method for inputting information or data to user device 102 a.
  • Parsing module 204 is configured to parse one or more Rich Site Summary (RSS) feeds. The RSS feeds may be provided by content provider 104. The RSS feed may include one or more entries identifying an article published by content provider 104. The RSS feeds may be monitored by interface module 202 and/or parsing module 204. The RSS feeds may be a precompiled list of RSS feeds that are published and constantly updated by content provider 104. Parser module 204 stores a precompiled list of RSS feeds in data storage 120 and retrieves them at regular intervals. Parser module 204 may parse the RSS feeds using a third-party open source Python module called Feedparser, for example.
  • Parser module 204 is configured to parse the RSS feeds to retrieve one or more URLs. Each URL corresponds to a specific article. For each URL retrieved, Parser module 204 also extracts a title and a publication date for that article.
  • Parser module 204 may maintain a list of previously-retrieved URLs in data storage 120 and compare the latest URL with the URLs in the list to ensure that the URL has not already been retrieved. Parser module 204 may delete URLs from the list of previously-retrieved URLs that were created before a certain date. For example, parser module 204 may delete URL entries whose creation date is older than one month.
  • Parser module 204 may prepare a set of articles based on the information extracted from the RSS feeds. Each article in the set of articles is associated with a title, a URL, and a publication date. Once parser module 204 has parsed a group of RSS feeds and extracted the URL, title, and publication date for each article, for each article in the set of articles, extraction module 206 may retrieve the textual content embedded in the HTML code of the article, by, for example, scraping the textual content embedded in the article's HTML code. Extraction module 206 may use one or more third-party open source scraping programs, such as, for example, Goose Extractor, Justtext, or some combination of the two.
  • Once extraction module 206 has extracted the body of text from the HTML code for each article, extraction module 206 then removes boilerplate text snippets that do not belong to the body of text of the reported story. Examples of boilerplate text may include copyright notices, and error messages generated during the scraping and image captioning text. Extraction module 206 may also extract the author of the article, which is used as an anchor to remove automatically generated paragraphs that describe the background and experience of the author reporting the story.
  • Extraction module 206 then analyzes the base URL of the article to reduce it to its fully qualified domain name (FQDN). The FQDN is the domain name that specifies its exact location in the tree hierarchy of the Domain Name System (DNS). For example, the FQDN of the news article located at the URL http://www.nytimes.com/2014/03/05/world/europe/ukraine.html?hp&_r=0 is “www.nytimes.com”. Extraction module 206 may store the FQDN of the news articles' URLs in order to keep track of news stories published by the same news outlet.
  • For a set of articles, classification module 208 receives the body of text for each article from extraction module 206 and classifies the article into one or more categories. Categories may include for example, and without limitation, U.S. News, World News, Business, Technology, Politics, Sports, Entertainment, Science, and Health.
  • Preprocessing module 212 is configured to implement one or more preprocessing methods to the body of text and/or the title of each article in the set of articles. Preprocessing module 212 may first implement a tokenization step. During tokenization, the body of text for an article is broken up into words, phrases, symbols and other elements known as tokens. Next, preprocessing module 212 may implement a lowercasing step. During lowercasing, the body of text tokens are normalized by lowercasing them. Next, preprocessing module 212 may implement a stopwords removal step. During stopwords removal, grammatical words, known as stopwords, are removed from the tokenized and lowercased bodies of text. Stopwords include, without limitation, “a”, “about”, “after”, “because”, “between”, “the”, “for”, “or”, etc. Next, preprocessing module 212 may implement a stemming step. During stemming, remaining body of text tokens are stemmed, that is, reduced to their root form by removing inflectional word endings.
  • Once the all the previous preprocessing steps are applied, classification module 208 may vectorize the body of text by using a bag-of-words representation using a mixture of unigrams and bigrams. Classification module vectorizes the text by converting the text from words into a sequence of numbers in the form of a one dimensional array. A vector encodes the textual information in the form of numbers that encapsulate information about the text. For example, the sentence “The dog chased the cat”, has 4 word types (the, dog, chased, cat). A vector of this sentence this sentence may represent the number of times each word type appears in the sentence: [2, 1, 1, 1], where dimension 1 means the number of times “the” appears, dimension 2 means the number of times dog appears, and so on. The bag-of-words representation is a form of textual vectorization that does not take into account the position in which word types appear in texts. This means that a sentence like “The dog chased the cat” has the same bag-of-words representation as the sentence “The cat chased the dog”. Both sentences have the very same 4 word types (the, dog, chased, cat). The sentence “The dog chased the cat” has 5 unigrams: “the”, “dog”, “chased”, “the”, “cat”, that is each individual word token. The same sentence has 3 bigrams: “the dog”, “dog chased”, “chased cat”, that is sequences of immediately adjacent words.
  • The generated vector space is first transformed by using the TF-IDF algorithm and, second, its dimensionality is reduced using the Truncated SVD algorithm. TF-IDF (term frequency-inverse document frequency) is a numerical statistic that is intended to reflect how important a word is to a document in a collection or corpus. It is often used as a weighting factor in information retrieval and text mining. The tf-idf value increases proportionally to the number of times a word appears in the document, but is offset by the frequency of the word in the corpus, which helps to control for the fact that some words are generally more common than others. The Truncated SVD algorithm is an algorithm specialized in reducing the dimensionality of vector representations. From the previous example, the sentence “The dog chased the cat” may be represented as a vector of the following form: [2, 1, 1, 1]. This vector has 4 dimensions. The Truncated SVD algorithm applies a set of mathematical functions that analyze each of the dimensions and convert a vector to another vector with a lower dimensionality that encodes the same information in a compacted/condensed manner.
  • Finally, the individual vectorized document is classified into one of the above-mentioned categories by applying a cascade of One vs One pre-trained Support Vector Machine models. The One vs One classification strategy is applied to classification problems where items need to be classified into more than two categories. This algorithm converts a multi-category problem into a series of simpler binary classification problems by creating pairs of binary categories and training a single classifier for each pair. As previously noted, in this step classification module 208 is classifying each article into one of a plurality of news categories. For example, categorization module 208 may use the following 9 categories: U.S. News, World News, Business, Technology, Politics, Sports, Entertainment, Science, and Health. In some instances, a single article may be assignable to more than one category. The One vs One classification strategy simplifies the multi-category problem by creating pairs of categories and training a single classifier for each pair. Thus, classification module 208 includes a classifier that distinguishes between US News and World News, US News and Business News, US News and Technology News, Technology News and Sports News, Technology News and World News, etc. Each of these is a binary classification where a news item is classified as one category or the other. The category that receives more votes, that is, that is selected more times for a particular article, is the category that it is finally assigned to that article.
  • Deduplication module 210 is configured to determine whether the body of text for each article in the set of articles represents an article or story that is a duplicate of another article or story that has already been processed and/or summarized. Deduplication module 210 will only compare articles that have been placed in the same category by classification module 208. Deduplication module 210 may use a two stage approach. First, deduplication module 210 may compare new articles within the same category to one another to find duplicate stories. Next, deduplication module may compare the new articles to articles that were published and previously stored in data storage 120 within a certain time period. For example, deduplication module 210 may compare new articles to stored articles form the past 48 hours.
  • Deduplication module 210 may receive the set of articles that have gone through the previous processing steps described above. Each article may include the body text, title, date, URL, and other information. Deduplication module 210 may first use the titles of each article in the set as a filtering stage to reduce the search space to be explored for article deduplication. Only articles whose titles have a degree similarity higher than a predefined threshold are compared to each other to find duplicates.
  • Preprocessing module 212 may implement the one or more preprocessing steps on the title of each article using the same steps described previously: (1) tokenization; (2) lowercasing; (3) stopwords removal; and (4) stemming. Once preprocessing module 212 has preprocessed the titles in the set of articles, deduplication module 210 may vectorize the titles using a bag-of-words vectorization approach and the vectors are weighted using the algorithm known as TF-IDF. Finally vectors are reduced in dimensionality using the Latent Semantic Indexing (LSI) algorithm. Latent semantic indexing (LSI) is an indexing and retrieval method that uses a mathematical technique called singular value decomposition (SVD) to identify patterns in the relationships between the terms and concepts contained in an unstructured collection of text. LSI is based on the principle that words that are used in the same contexts tend to have similar meanings. A key feature of LSI is its ability to extract the conceptual content of a body of text by establishing associations between those terms that occur in similar contexts.
  • Deduplication module 210 then compares each title vector to the rest of the title vectors in the set using the cosine similarity algorithm, which outputs a similarity measure for each pair of compared vectors. Cosine similarity is a measure of similarity between two vectors of an inner product space that measures the cosine of the angle between them. The cosine of 0° is 1, and it is less than 1 for any other angle. Cosine similarity is thus a measure of orientation and not magnitude: two vectors with the same orientation have a Cosine similarity of 1, two vectors at 90° have a similarity of 0, and two vectors diametrically opposed have a similarity of −1, independent of their magnitude. Cosine similarity is particularly used in positive space, where the outcome is neatly bounded in [0,1].
  • Only pairs of articles whose titles have a cosine similarity above a predefined threshold are allow for further comparison. The predefined threshold is a constant that is checked against the cosine similarity value that each pair of articles is assigned. So, in one embodiment, if a pair of article titles have a cosine similarity of 0.5, but the predefined threshold is 0.6, deduplication module 210 deem this pair of article titles as not similar enough. But if a pair of news items have a cosine similarity of 0.7 and the threshold is 0.6 then those two items are deemed as similar enough. The value of this predefined threshold may be empirically set by inspecting the results obtained using different values and choosing the value that renders better results. The set of pairs that do not pass this filtering are removed from search space of potential duplicate pairs of articles.
  • For the remaining pairs of articles that pass the first title-based filtering step, deduplication module 210 is configured to further analyze each pair of articles based on the notion of “containment” as their similarity comparison measure. Containment is a measure that defines similarity as set a intersection problem as described in the following, labeled Formula 1:

  • c(A,B)=|set(A)∩set(B)|/set(A)  Formula 1:
  • In Formula 1, “A” is the first article in a pair of articles that met the title-based filtering step, and “B” is the second article. For a given pair (e.g., articles A and B), preprocessing module 212 may preprocess the body of text of each article in the pair using the following preprocessing steps; (1) tokenization, (2) lowercasing, and (3) stopwords removal. Once this has been done, for each pair of articles, deduplication module 210 may extract the unigrams, bigrams and trigrams from each pair of preprocessed bodies of text and converted into sets of tokens.
  • Extracted sets of tokens are subsequently intersected as follows:
      • unigrams_itersec=set(A_unigrams)∩set(B_unigrams),
      • bigrams_itersec=set(A_bigrams)∩set(B_bigrams), and
      • trigrams_itersec=set(A_trigrams)∩set(B_trigrams)
  • Using these n-gram intersections, deduplication module 210 calculates the following two measures of containment according to Formula 1:

  • Containment of B in A: c(A,B)=(|unigrams_itersec|+|bigrams_itersec|+|trigrams_itersec|)/set(A); and  (1)

  • Containment of A in B: c(B,A)=(|unigrams_itersec|+|bigrams_itersec|+|trigrams_itersec|)/set(B).  (2)
  • In the previous formulas, set(A) equals |A_unigrams|+|A_bigrams|+|A_trigrams|; and set(B) equals |B_unigrams|+|B_bigrams|+|B_trigrams|. Deduplication module 210 then compares c(A,B) with c(B,A) to determine the direction of containment. Thus, if Containment of A in B (c(B,A)) is higher than Containment of B in A (c(A,B)), A is marked as a superset of B, and vice versa.
  • Furthermore, if the containment measure of the winning superset is higher than a predefined containment threshold, the article that is contained by the superset article is classified as a duplicate and removed. The predefined containment threshold may be empirically set, similar to the threshold used for the cosine similarity algorithm that compared title vectors for two articles. The winning article superset is retained for future comparisons with other articles. The final result of repeating this process for all pairs of articles is a set of unique stories from which duplicate articles have been removed. This set of articles may be referred to as the final set of articles.
  • Preprocessing module 212 is configured to preprocess the final set of articles prior to summarization. For each article, preprocessing module 212 may first perform tokenization to break up the body of text into words, phrases, symbols and other elements known as “tokens”. Preprocessing module 212 may then perform sentence splitting on the tokenized article text by splitting the text into individual sentences by using punctuation as anchors to determine sentence boundaries. Preprocessing module 212 may then perform truecasing on the tokens of each individual sentence to normalize the tokens to their true canonical casing. Preprocessing module 212 may then perform stopwords removal to remove stopwords from individual tokenized and truecased sentences. Finally, preprocessing module 212 may perform stemming by reducing each remaining token to its root form by removing inflexional word endings.
  • Once individual sentences within each article (from the final set of articles) have been identified and pre-processed, rules module 214 applies a series of rules to each article to accomplish the following steps: 1) extract a series of preliminary numeric measures to be reused by summarization module 218 to score and rank potential candidate sentences; and 2) to identify linguistic markers that can potentially impact the coherence of the generated summary. Potential candidate sentences are sentences within an article that may be used as part of the generated summary.
  • For each individual sentence of an article, rules module 214 is configured to extract a number of preliminary statistics in order to provide a first process for sentence scoring and ranking within each article. There are two types of statistics that contribute to this score. The first type are positive scoring factors, which contribute to positively score sentences higher in the ranking. The second type are negative scoring factors, which diminish a sentence score, thus, lowering its ranking.
  • Rules module 214 is configured to scale all negative and positive scoring factors are using the following formula, labeled as Formula 2:

  • score=factor_score/sqrt(factor_score+alpha)  Formula 2:
  • In Formula 2, alpha is constant that smooths the scale into which factor is converted. Alpha is an empirically set constant that affects the effect that a smoothing function will have on a particular data series. In this equation higher values of alpha may intensify the effect of the smoothing function, whereas lower values of alpha may reduce the effect of the smoothing function. The factor_score may the score of the scoring factor being considered.
  • Rules module 214 is configured to collect positive scoring factors for each individual sentence in an article. The first positive scoring factor is the sentence position in the article. This factor is used to favor sentences that are closer to the beginning of the article as these sentences tend to include the most relevant pieces of information for each particular article. For this factor, rules module 214 may use a scaling formula, Formula 3, which is slightly different than Formula 2:

  • position_factor=(last_doc_position/(sentence_position+1))/sqrt((last_doc_position/(sentence_position+1))+alpha)  Formula 3:
      • In Formula 3, last_doc_position is the position of the last sentence in the news item. So, for example, if a article has 20 sentences, the last_doc_position for that article would be 20.
  • The next positive scoring factor is the number of frequent words present in the sentence. For a given article, rules module 214 may define a “frequent word” as a word that appears a minimum number of times in an article. For example, rules module 214 may define a frequent word as a word that appears at least 3 times in an article. This factor (document frequent words) is used to favor sentences that contain instances of the most frequently used words in the document.
  • Another positive scoring factor may be the number of document Named Entities present in the sentence. This factor is used to favor sentences that contain instances of the Named Entities (e.g., a proper noun) present in the document. Another positive scoring factor may be the number of sentiment words, that is, words with either positive or negative polarity, present in the sentence. Rules module 214 may maintain a lexicon of sentiment terms that are used to favor sentences that are positively or negatively subjective. For each sentence in the article, rules module 214 may compare the words in the sentence with the lexicon of sentiment terms to determine how many sentiment words are present in that sentence. Another positive scoring factor may be the number of money expressions present in the sentence (e.g., “$20”, “$2”, “$1,000”). A simple regular expression is used to identify and favor sentences that contain money-related expressions. The regular expression may be: \$\d+,?\d*\.?\d*. This regular expression accepts strings that start with a dollar sign, followed by one of more digits, followed by an optional comma, followed by 0 or more digits, followed by an optional decimal point, followed by 0 or more digits.
  • Rules module 214 is configured to collect negative scoring factors for each individual sentence in the article. One negative scoring factor may be the length of the sentence. This factor is used to favor sentences that are not too long nor too short. For this factor, Formula 4, a slightly modified version of Formula 2 is used:

  • length_factor=abs(sent_length−doc_sent_length_mean)/sqrt(abs(sent_length−doc_sent_length_mean)+alpha)  Formula 4:
  • In Formula 4, sent_length represents the length of the sentence being considered. Doc_sent_length mean represents the mean of the sentence lengths for the entire article.
  • Another negative scoring factor may be the number of sentence negative features. This factor is used to favor sentences that don't have any of the following features present: (1) last character is not a question mark (?); (2) last character is not a colon (:); (3) a demonstrative (this, that, etc.) is not present in the first positions of the sentence; and (4) opening quotations, parentheses or square brackets are not present in the first positions of the sentence.
  • Once all the positive and negative scoring factors are determined for each sentence, rules module 214 will combine the positive and negative scoring factors of each sentence using Formula 5, resulting into the final preliminary score that will be assigned to that sentence:

  • rule_score=sum(positive_scoring_factors)−sum(negative_scoring_factors).  Formula 5:
  • For each article in the final set of articles (i.e., the set of articles that remains after duplicate articles have been removed), marker module 216 is configured to split each sentence of the article into one of three different groups: Potential candidates, weakly rejected candidates, and strongly rejected candidates. Potential candidates are sentences for which no potentially impact in summary coherence has been detected, therefore, they are signaled as fully-qualified candidates for the summary to be generated by summarization module 218. Weakly rejected candidates are sentences for which linguistic markers that could negatively impact the summary coherence have been identified. These sentences are banned from being included in the generated summary but they're kept for further processing by summarization module 218 as they contain valuable information about the article content that needs to be explored. Strongly rejected candidates are sentences that, because they contain some kind of boilerplate element, completely discarded as they are deemed as not containing valuable content for the article.
  • Marker module 216 is configured to designate a sentence as a weakly rejected sentence based on one or more linguistic markers found in the sentence. These linguistic markers represent local, sentence-to-sentence transition of ideas and concepts within the article discourse. Weakly rejected sentences break the sentence-to-sentence transition of ideas and concepts that the linguistic markers articulate locally. With their local coherence broken, these sentences negatively impact the global coherence of the summary by introducing non-contextualized idea transitions. Marker module 216 may evaluate each sentence based on a predetermined list of weak linguistic markers that indicate that a sentence should be designated as a weakly rejected sentence. These weak linguistic markers may include personal pronouns, demonstrative pronouns, adversative conjunctions, concessive conjunctions, coordinating conjunctions, adverbial discourse connectors, deictic metadiscourse connectors, and temporal adverbs. Furthermore, marker module 216 may designate a sentence as a weakly rejected sentence that doesn't include a verb, as these sentences are normally section titles that contribute to the visual understanding of the article without contributing to its meaning. Marker module 216 may require a predetermined number of these weak linguistic markers to be found in a sentence before marker module 216 will designate that sentence as a weakly rejected sentence. In various embodiments, all weakly linguistic markers are weighted equally.
  • Marker module 216 is configured to designate a sentence as a strongly rejected sentence based on one or more strong linguistic markers. Strongly rejected sentences are sentences that do not contribute to the concepts and ideas of that the article elaborates. Thus, these sentences could be considered discourse boilerplate. Marker module 216 may evaluate each sentence based on a predetermined list of strong linguistic markers that indicate that a sentence should be designated as a strongly rejected sentence. These strong linguistic markers may include gratitude nouns and verbs, greetings, and/or exclamations. Furthermore, marker module 216 may designate a sentence as a strongly rejected sentence based on the presence of lexical terms that belong to the jargon of hypertext, like the verb ‘click’ in “Click here to find out more”, are also used as clues that signal strongly rejected sentences. If marker module 216 evaluates a sentence and determines that the sentence is neither a weakly rejected sentence nor a strongly rejected sentence, marker module 216 may classify that sentence as a potential candidate sentence. At the end of this process, marker module 216 may have created three sets of sentences for each article—a set of potential candidate sentences, a set of weakly rejected sentences, and a set of strongly rejected sentences.
  • Summarization module 218 is configured to generate an article summary for each article in the final set of articles. The article summary is comprised of a series of one or more sentences from the set of potential candidate sentences, as designated by marker module 216. Summarization module 218 may generate the article summary by implementing a modified version of the algorithm known as Maximal Marginal Relevance to iteratively and dynamically rank potential candidate sentences. Maximal Marginal Relevance (MMR) is an Information Retrieval (IR) algorithm that measures the relevant novelty of a document. A first approximation to measuring relevant novelty is to measure relevance and novelty independently and provide a linear combination as the metric. This linear combination may be described as “marginal relevance”, i.e., a document has high marginal relevance if it is both relevant to the query and contains minimal similarity to previously selected documents. MMR's basic formula can be derived as Formula 6:

  • MMR=argmax(lambda*sim(Di,Q)−(1−lambda)max_sim(Di,DP))  Formula 6:
  • In Formula 6, lambda is a weighting constant, sim is the cosine similarity algorithm, Di is a candidate document, Q is the query and DP is the pool of documents already selected as relevant in the previous iterations of the algorithm.
  • For a given article from the final set of articles, summarization module 218 applies this IR paradigm to generate the article summary by treating the title of the article as the query and the set of potential candidate sentences as the collection of documents searched by the query. Working within this IR paradigm, summarization module 218 uses the following modified version, Formula 7, of the original MMR algorithm.

  • MMR=argmax(lambda*(sim(Si,T)+sim(Si,DC)+Se_score)−(1−lambda)max_sim(Si,Su))  Formula 7:
  • In Formula 7, lambda is a weighting constant, sim is the cosine similarity algorithm, Si is a potential candidate sentence, T is the title of the article, DC is the centroid of the article vector space resulting from the combination of both the potential candidate sentences and the weakly rejected sentences sets, Se_score is the score calculated for the potential candidate sentence by rules module 214, and Su is the dynamically generated summary, that is, the set of sentences picked to be included in the summary in previous iterations of the algorithm. In various embodiments, lambda may be empirically set value of 0.6. Summarization module 218 may select sentences from the potential candidate sentences set to be included in the generated summary, whereas weakly rejected sentences and strongly rejected sentences cannot be used in the generated summary.
  • For each article, summarization module 218 uses term statistics for sentences from both sets (potential candidate sentences and weakly rejected sentences) to generate a combined vector space, which is first transformed by using the TF-IDF algorithm. The TF-IDF algorithm is a numerical statistic that transforms a vector by measuring the importance that each dimension has across the collection of all vectors. This algorithm intensifies the importance of dimensions that are infrequent across all vectors to boost their discriminative power. Next, summarization module 218 reduces the TF-IDF transformed vectors in dimensionality by applying the Truncated SVD algorithm. The Truncated SVD algorithm performs a factorization of a vector matrix to project it into a lower dimensionality space that encodes the information contained in the higher dimensionality space by combining the value of interdependent dimensions. Summarization module 218 then calculates the centroid of the resulting combined and modified vector space by averaging the values of each feature across the entire vector space. The centroid of a vector space represents the point that lays in the center of that high dimensional space. The centroid is a vector with the same dimensionality as the vectors that compose the vector space.
  • Summarization module 218 then reduces the vector space by removing the vectors that represent the sentences signaled as weakly rejected sentences, preventing them this way from being processed by Formula 7 and, therefore, from being included in the generated summary.
  • For a given article, summarization module 218 generates the article summary by iteratively running each of the sentences from the set of potential candidate sentences through Formula 7. In the first iteration, summarization module 218 determines the MMR for each potential candidate sentence in the article, then selects the sentence with the highest MMR and adds it to the article summary. Once a sentence has been added to the article summary, summarization module 218 removes that sentence from the set of potential candidate sentences for that article in order to prevent it from being repeatedly selected. In the next iteration, summarization module 218 performs the same process (only this time the set of potential candidate sentences has been modified to remove the previously selected sentence), and selects the sentence having the highest MMR in that iteration, adds that sentence to the article summary, and removes it from the set of potential candidate sentences.
  • Summarization module 218 repeats this process until the number of characters in the article summary reaches a maximum threshold. The maximum threshold may have been preprogrammed into summarization module 218. For example, in one embodiment, the maximum threshold may be 325 characters. In another embodiment, a user may be able to provide the maximum threshold via interface module 202.
  • Once the maximum threshold has been reached, summarization module 218 may reorder the sentences in the article summary so that their position in the summary follows the same occurrence order of the sentences in the source article.
  • FIG. 3 is a flowchart illustrating the functionality for generating an article summary based on an article title and body of text. This exemplary method 300 may be provided by way of example, as there are a variety of ways to carry out the method. The method 300 shown in FIG. 3 can be executed or otherwise performed by one or a combination of various systems. The method 300 is described below may be carried out by the systems and networks shown in FIGS. 1 and 2, by way of example, and various elements of the systems and networks are referenced in explaining the example method of FIG. 3. Each block shown in FIG. 3 represents one or more processes, methods or subroutines carried out in exemplary method 300. Referring to FIG. 3, exemplary method 300 may begin at block 302.
  • At block 304, summarization system 110 retrieves an RSS feed. The RSS feed may include one or more entries identifying articles from content provider 104. Summarization system 110 may regularly retrieve one or more RSS feeds and store them in data storage 120. Method 300 may proceed to block 306.
  • At block 306, summarization system 110 generates an initial set of articles from the RSS feed. Summarization system 110 may first parse the RSS feeds to retrieve a list of URLs, where each URL is associated with an article. Summarization system 110 may compare the list of URLs to a list of previously retrieved URLs in data storage 120 and remove any matching URLs. Summarization system 110 may also parse the RSS feed to retrieve the title and publication date of each article in the feed. Summarization system 110 may then retrieve the textual content embedded in the HTML code of the article from the URL. Summarization system 110 may remove boilerplate text snippets that do not belong to the body of text of the reported story. Examples of boilerplate text may include copyright notices, and error messages generated during the scraping and image captioning text. Extraction module 206 may also extract the author of the article, which is used as an anchor to remove automatically generated paragraphs that describe the background and experience of the author reporting the story. Summarization system may then generate an initial set of articles, where each article in the set includes the title, publication date, source, and body text of the article. Method 300 may proceed to block 308.
  • At block 308, summarization system 110 classifies each article in the initial set of articles. Summarization system 110 may classify each article into one or more news categories. News categories may include for example, and without limitation, U.S. News, World News, Business, Technology, Politics, Sports, Entertainment, Science, and Health. Summarization system 110 may apply one or more preprocessing steps to the body of text and/or title of each article in the initial set of articles. The steps may include tokenization, lowercasing, stopwords removal, and stemming. Following these preprocessing steps, summarization system 110 may generate a vectorized document for each body of text and categorize the article text into one of the news categories. Method 300 may proceed to block 310.
  • At block 310, summarization system 110 removes duplicate articles from the initial set of articles to generate a final set of articles. This process may involve a two stage approach. First, summarization system 110 may compare articles with other articles in the same category to find duplicate articles. Next, summarization system 110 may compare the articles in the initial set of articles to articles that were published and previously stored in data storage 120 within a certain time period. For example, summarization system 110 may compare new articles to stored articles form the past 48 hours. Using the titles of each article in the initial set of articles, summarization system 110 may implement one or more algorithms to quantify the degree of similarity between a pair of articles. A pair of articles with titles that meet a predetermined degree of similarity will then be compared using a containment algorithm, described above as Formula 1. Summarization system 110 may use this to remove duplicates from the initial set of articles to generate a final set of articles. Method 300 may proceed to block 312.
  • At block 312, for each article in the final set of articles, summarization system 110 generates a preliminary score for each sentence in that article. Summarization system 110 may first preprocess each article in the final set of articles by performing tokenization, sentence splitting, truecasing, stopwords removal, and stemming. Summarization system 110 then evaluates each sentence in an article and assigns it a score. The score may be based on a series of positive and negative scoring factors. Positive scoring factors lead to a higher score and a higher rank for the sentence in the article. Negative scoring factors lower the score and lead to a lower rank for the sentence within the article. The positive and negative scoring factors may be scaled using Formula 2, as described previously. Positive scoring factors may include the position of the sentence in the article, the number of frequent words present in the sentence (where a frequent word is a word that appears a minimum number of times in the article), the number of Named Entities present in the sentence, the number of sentiment words in the sentence, and/or the number of money-related expressions in the sentence. Negative scoring factors may include the length of the sentence, and the presence or absence of different negative features. The positive and negative scoring factors for each sentence may be determined and combined using Formula 2, Formula 3, Formula 4, and/or Formula 5, as described above. Method 300 may proceed to block 314.
  • At block 314, summarization system 110 may classify each sentence in an article into one of three categories: (1) potential candidate sentences; (2) weakly rejected sentences; and (3) strongly rejected sentences. For a given article, only sentences that have been classified as potential candidate sentences can be used as part of the article summary. Potential candidate sentences are sentences for which no potentially impact in summary coherence has been detected, therefore, they are signaled as fully-qualified candidates for the article summary. Weakly rejected candidates are sentences for which linguistic markers that could negatively impact the summary coherence have been identified. These sentences are banned from being included in the generated summary but they're kept for further processing as they contain valuable information about the article content that needs to be explored. Strongly rejected candidates are sentences that, because they contain some kind of boilerplate element, completely discarded as they are deemed as not containing valuable content for the article.
  • Summarization system 110 may classify a sentence as a weakly rejected sentence based on the presence of one or more weak linguistic markers. Weak linguistic markers may include personal pronouns, demonstrative pronouns, adversative conjunctions, concessive conjunctions, coordinating conjunctions, adverbial discourse connectors, deictic metadiscourse connectors, and/or temporal adverbs. Summarization system 110 may classify a sentence as a weakly rejected sentence if the sentence does not include a verb. Summarization system 110 may classify a sentence as a strongly rejected sentence based on the presence of one or more strong linguistic markers. Strong linguistic markers may include gratitude nouns and verbs, greetings, and/or exclamations. Summarization system 110 may classify a sentence as a strongly rejected sentence based on the presence of lexical terms that belong to the jargon of hypertext, like the verb ‘click’ in “Click here to find out more”, are also used as clues that signal strongly rejected sentences. If a sentence does not contain a minimum number of strong linguistic markers and/or weak linguistic markers, summarization system 110 may classify the sentence as a potential candidate sentence. Method 300 may proceed to block 316.
  • At block 316, summarization system generates an article summary for each article in the final set of articles. The article summary is comprised of a series of one or more sentences from the set of potential candidate sentences for that article. Summarization system 110 uses an iterative algorithm, Formula 7, as described above, to construct the article summary from the set of potential candidate sentences for that article.
  • As a sentence is added to the article summary, it is removed from the set of potential candidate sentences for that article. This process will continue until the length of the article summary has reached a predetermined number of characters. Summarization system 110 may then provide the article summary to be viewed on an interactive interface on a user device. Method 300 may proceed to block 318.
  • At block 318, method 300 may end.
  • FIGS. 4A-4D show different interface screens generated by summarization system 110, according to an embodiment. The screens may be provided to a user via an application on a mobile device (such as summarization application 102 b) and/or a web interface. FIG. 4A shows an initial screen for summarization application 102 b. FIG. 4B shows a list of article titles from a final set of articles. The article titles also include the source for the article, listed below each title. The screen shown in FIG. 4B may be interactive and receive inputs from the user. For example, the user may be able to select a given article title and be presented with the article summary, as shown in FIG. 4D.
  • FIG. 4C shows a list of articles that have been categorized under the category of “U.S. News.” The list includes the article title and the source. The screen may receive one or more selections from the user, as with the screen in FIG. 4B. FIG. 4D depicts an article summary generated by summarization system 110 as described above. The article summary is designed to be viewed on a single screen of user device 102 a. The summary may include the article's title, source, date, author name, and/or byline. The user may be able to scroll through the summary and view the entire text of the article.
  • In some embodiments, certain types of articles may require modified processing, based on the structure of the article. These types of articles have an internal discourse structure does not fit the assumption that a unique discourse coherence articulates the paragraphs and sentences of the entire article from the beginning to the end. For example, many articles are a “Top X” list articles (e.g., “Top 10 Smartphones on the Market”, “15 best beaches in California”, “Top 50 NFL Players of all Time”). The structure of these “Top X” list articles typically features a preliminary set of paragraphs that introduce the general topic discussed in the article, followed by a series of sections that examine this general topic from different angles. Each of these sections can be considered as a self-focused mini document with its own internal coherence. Using the single-document summarization process described above to access the sentences from each independent self-focused section would greatly increase the risk of generating non-coherent sentence summaries. Therefore, a special summarization strategy needs to be implemented in order to acknowledge their singularity in terms of their internal structure and coherence.
  • The modified summarization strategy would work using the following steps. First, any one of parser module 204, extraction module 206, and/or classification module 208 may check the title of an article using a set of functions to determine whether the article belongs in the category of “Top X” list articles. In particular, these functions check for the existence of numbers of up to three digits that are not followed by time expressions (such as “years”, “months” or “days”, etc.) or the existence of numbers in word form (such as “four”, “ten”, “two”, etc.). These modules also check the body text of the article to find the number of items listed in the article. One or more modules may explore the HTML code of the article to identify a set of HTML tags that enclose the title of each section of the independent list items in the article. If the number of section titles in the article body text matches the number expressed in the title, the article is classified as a “Top X” list type article and the section titles are stored in a database for later use as structural anchors during summarization.
  • During summarization articles marked as “Top X” list articles are decomposed in text blocks using the identified section title anchors. The result of this procedure is a list of sections in which articles can broken down. Preprocessing module 212 may compare the number of section titles with the number of text blocks in the articles. If the number of section titles matches the number of text blocks it is assumed that no introduction is present, whereas if the number of section titles is lower than the number of text blocks, one or more modules mark the first section as the article introduction.
  • If an introductory text block is found during the previous structure decomposition step, preprocessing module 212, rules module 214, marker module 216, and summarization module 218 may generate a summary of this introductory section, as described above. Summarization module 218 may then append the list of section titles to the end of the generated summary, and this may be displayed on user device 102. If no introduction is identified by preprocessing module 212, then a summary is generated by simply appending the list of section titles found during scraping.
  • In various embodiments, clustering module 220 may clustering articles to help avid news readers to follow particular stories without having to hunt for related reports in the continuous stream of the news cycle. Clustering module 220 receives article documents that have been pre-processed by pre-processing module 212 (which applied lowercasing, tokenization, stopwords removal, and token stemming as described above). Unigram tokens left after applying these pre-processing routines are used to update the TF and IDF indices. Subsequently, unigram tokens with a term frequency lower than 2 in the global TF index are filtered out.
  • As mentioned above, articles are vectorized by applying the BoW approach. For scalability and performance reasons each news event cluster keeps track of its own BoW; in other words, each cluster defines its own vector space. Using this optimization prevents both document vectors and cluster centroids from becoming too large and sparse when vectorized, thus minimizing the use of RAM during execution and lowering the processing power needed.
  • However, this approach requires centroid clusters to be recomputed for each new document. This has the positive effect of keeping cluster centroids accurate at all times and, since cluster vector spaces always use a small portion of the tokens of the TF and IDF indices, cluster centroid recomputations have a minimum performance impact.
  • Vectorized article documents and recomputed cluster centroids are subsequently weighted using the following TF/IDF algorithm version:
  • TF ( t , d ) = 1 + log ( f ( t , d ) ) idf ( t , D ) = log ( N d D ; t d ) tfidf ( f , d ) = tf ( t , d ) * idf * ( t , D )
  • where f(t,d) is the number of times term t occurs in the article document d, N is the total number of documents in the collection of documents D and |dεD; tεd| is the number of documents in which term t occurs. If either tf(t,d) or idf(t,D) result in non-numerical or infinite values they are converted to 0.
  • Vectorized and weighted article documents and recomputed cluster centroids are compared using the Cosine Similarity algorithm as a similarity measure:
  • similarity=(v·c/∥v∥*∥c∥), where v is the vectorized article document, c is the recomputed cluster centroid, v·c is the dot product of v and c and ∥v∥ and ∥c∥ are v and c vector norms, respectively.
  • In order to reduce the “attraction” power of older clusters, whose vector space tends to be larger than more recently created clusters, a second optimization is applied. This optimization consists in computing a correction cluster age factor that is subtracted from the cosine similarity:

  • cacf=dt−ca(if cacf<0 then cacf=0)

  • cacf=√(caf/(60*60*24*30))
  • similarity=similarity−(alpha*caf c), where cacf is the cluster age correction factor, dt is the article document time stamp and alpha is a weighting constant empirically set to 0.15. Cluster module 220 may filter out cluster centroids that have a similarity lower than an empirically defined similarity threshold with respect to the vectorized article document. Cluster module 220 may then assign the article document to the cluster with the greatest similarity among all the remaining clusters, that is, among all the clusters with a similarity greater than the predefined similarity threshold. If no cluster has a similarity greater than the predefined similarity threshold, the article document defines its own cluster and will be subsequently compared to each new document that is processed by the cluster module 220.
  • Cluster module 220 may cause a cluster to be “frozen” (i.e., no new articles are added to the cluster) if the cluster's age becomes greater than a predefined age threshold or no new documents have been assigned to a cluster after a predefined period of time. Clusters may be presented to the user via the application interfaces shown in FIGS. 4a -4 d.
  • FIG. 5 is a schematic diagram illustrating a system according to particular embodiments. As illustrated in FIG. 5, system 500 may include a named entity recognition and classification (NERC) module 502 and a named entity disambiguation module 504. These modules may comprise one or more network-enabled computers. Modules 502 and 504 may be integrated into summarization system 110 (e.g., as part of named entities module 222).
  • In various embodiments, NERC module 502 may extract Named Entities (NEs) mentions from an article. These inputs may then be provided to Named Entity disambiguation module 504 which may be configured to disambiguate the NEs, and identify the most salient NEs for each article. At the end of this process, named entities module 222 may link articles that share unique salient NEs so that users can follow particular entities they are interesting in.
  • NERC module 502 may first extract NEs from an article using a two-step process: (1) Named Entity Recognition, and (2) Named Entity Classification. First, Named Entity Recognition is defined as the process of identifying the boundaries of NEs in sentences so that all tokens belonging to a NE are grouped together and all tokens outside the boundaries of a NE are excluded. Second, Named Entity classification is defined as the process of assigning a unique classification label to all NEs identified by during the Named Entity Recognition phase (e.g., Person, Organization, Location, or Miscellaneous). Both processes use machine learning.
  • The outputs from NERC module 502 may then be provided to Named Entity Disambiguation module 504, which may first map an entity mention extracted from documents to unique entries of a knowledge base. An entity mention is a sequence of tokens in a text which refers to a specific NE and has been previously identified by NERC module 502. Given the high ambiguity of language, a named entity can have multiple names and a name can be denoted by different named entities. Hence, the main goal of the named entity disambiguation process is to disambiguate the entity mentions in their textual context and identify their concrete NE referent. The knowledge base may include a Named Entity index 508 and various third party databases 510 such as Wikipedia (https://www.wikipedia.org/), Freebase (https://www.freebase.com/) or DBpedia (http://dbpedia.org/) or other data sources.
  • For example, NERC module 502 may analyze an article that includes the sentence “With knowledge of such widespread ferocity, I recently saw Mel Gibson's movie ‘Apocalypto’ which deals with the gore of the Mayan civilization.” NERC module 502 may first extract the mentions “Mel Gibson” and “Apocalypto”. After that, NE disambiguation module 504 would retrieve from a knowledge base (508, 510) a list of possible NE candidates that refer to each of those mentions (for example an actor and a basketball player for the mention “Mel Gibson”). The goal of the disambiguation module 504 is to identify the correct referents, in this case an actor and a film. The NE disambiguation process consists of three steps to retrieve, score and rank NE candidates for each entity mention previously extracted from articles during the NE recognition process. The output is a single disambiguated NE for each entity mention.
  • First, NE disambiguation module 504 gathers a list of NE candidates for each of the entity mentions extracted by the NERC module 502. This may be extracted from one or more databases, such as index 508 and/or third party database 510. The index 508 may include NEs with the following information: title, alias (other names of the NE), links to other NEs, in-degree score (the count of links pointing to the entity), out-degree score (the count of links to other NEs). NE disambiguation module 504 may use the Named Entity Index 508 to retrieve candidates for a given entity mention by searching on the title and alias fields. For each candidate, the index provides a popularity score computed by the entity in-degree and out-degree link counts in the Wikipedia graph. The popularity score is also enriched with the search relevance score provided by ElasticSearch, which is based on Lucene's practical scoring function (http://www.elasticsearch.org/guide/en/elasticsearch/guide/current/practical-scoring-function.html). Finally, the output of this process is a dictionary of entity mentions as keys, and a list of NE candidates as values of each key.
  • The second step is a set of scoring functions that aim at obtaining two main representations of each NE candidate. The first is a confidence score that is an independent representation of each single NE candidate without considering any other candidates of the article. This score includes two dimensions: (1) Local NE confidence, and (2) Global NE confidence. Local NE confidence is the sum of: a) cosine similarity of the NE candidate name and alias against the sequence of the entity mention as found in the article (e.g. “mention: Mel Gibson” vs. “entity_name: Mel Gibson”, “entity_alias: Mel Columcille Gerard Gibson”, etc.), and b) cosine similarity of the candidate linked entities against the textual context surrounding the mention, which can be the single sentence, paragraph or complete article where the mention occur. The Global NE confidence refers to the popularity of the candidate in the knowledge base where it was obtained.
  • The second score is a collective score. Named entity disambiguation module 504 may generate a graph where it stores all NE candidates as nodes and existing relations among them as edges. The edges are drawn only between candidates of different entity mentions and only if a given relation is found using the links of the entity page from Wikipedia. Using this graph, the collective score is computed by generating the PageRank of each node.
  • Next, Named entity disambiguation module 504 may generate different ranks for each entity mention and the dynamic selection of the best top rank. A rank is an ordered list of the NE candidates generated by some combinations of the scores described above. The main ranks used are Rank by Local Confidence+PageRank, and Rank by Combined Confidence (Local+Global Confidence)+PageRank. Named Entity Disambiguation module 504 then selects the best top rank by normalizing the scores of each list to sum to 1 and computing the difference between the scores of the top two candidates in the given list. The idea behind this is that a bigger difference in the scores would indicate more confidence for the discrimination of candidates. In this way, the final NE candidate is selected from the rank with a highest difference score. This method equilibrates the cases where a false candidate is selected as the top one because it has a significantly higher popularity score (retrieved by the global confidence); in this case, the local confidence provides a higher difference between the top candidates and avoids the selection of the wrong one.
  • After NEs have been disambiguated, Named entity disambiguation module 504 may analyze and classify the NEs to determine what NEs are most salient. As pointed out by (Gamon et al., 2013a: 1), “salience and importance/prominence are not the same”. Whereas importance/prominence is a function of the reader needs, that is, what the reader considers to be most important/prominent; salience is a function of the author, that is, the entities that the author considers to be the most relevant for a particular article. In other words, the most salient NEs are the most prominent entities from a content and discursive point of view, better summarizing the information encapsulated in an article.
  • NESI, short for Named Entity Salience Identification, is a Machine Learning module trained using a subset of the New York Times Salience corpus (known as nyt-salience; it can be downloaded from https://code.google.com/p/nyt-salience/), which accompanied the paper Dunietz & Gillick, 2014. The New York Times Salience corpus extends the 1.8 million articles compiled in the New York Times Annotated Corpus (https://catalog.ldc.upenn.edu/LDC2008T19) with NEs automatically extracted from those articles that are primarily annotated with their Freebase mid, that is their unique id, and their automatically inferred salience.
  • As previously mentioned, NESI was trained using a subset of the New York Times Salience corpus. In particular we used the first 30000 articles, but discarded all articles for which no salient entities were automatically identified by the algorithm that Dunietz & Gillick (2014) used. Finally, in order to use a more balanced dataset, non-salient NEs were randomized and reduced to the same number of salient NEs. During training this proved to be beneficial as it reduced the bias of the trained algorithm towards non-salient NEs.
  • NESI was trained using a supervised Support Vector Machines algorithm with RBF kernel that learned from the following features: The complete Named Entity is in the article's title; The complete Named Entity is in the article's first line; Named Entity count; Named Entity start byte-offset; Number of Named Entity words that are in the article's title; Length in words of the Named Entity; Length in words of the article's title; and the Number of related entities that are present in the article's list of extracted Named Entities after their disambiguation and linking to a Knowledge Base (Wikipedia). In order to improve precision of the learned models, at the cost of some recall, all NEs identified as salient that have a prediction probability lower than 0.875 are deemed as non salient.
  • The most important advantages of NESI's with respect to Dunietz & Gillick's are that: 1) NESI is lighter in terms of running time and processing power as it doesn't need to run a PoS tagger, a dependency parser and a NP extractor; and 2) The feature set NESI uses is not lexicalized, and therefore not dependent on the corpus used for training and evaluation.
  • Since it is possible that for any particular article no NE is found to be salient by NESI's core ML learning algorithm, we applied a set of backup heuristics to ensure that at least one NE is classified as salient. In order to prioritize the set of two backup heuristics to be applied in case no NE is found to salient we used the results obtained by training NESI's ML algorithm using only one of the isolated binary features listed above. The reason why we chose only binary features is that results obtained using this type of features are easier to understand in contrast to continuous numerical features. Thus, the following heuristics are applied: Heuristic 1—If no NE has been found to be salient, all NEs that are either present in the title or the first line of the article and their count is greater than 1 it is classified as salient. Heuristic 2—If after applying heuristic 1 no NE has yet been found to be salient, the NE with the lowest start offset and the greatest count is classified as salient.
  • The various computing devices above (including phones and network equipment), generally include computer-executable instructions, where the instructions may be executable by one or more processors. Computer-executable instructions may be compiled or interpreted from computer programs created using a variety of programming languages and/or technologies, including, without limitation, and either alone or in combination, Java™, C, C++, Visual Basic, Java Script, Perl, etc. In general, a processor or microprocessor receives instructions, e.g., from a memory, a computer-readable medium, etc., and executes these instructions, thereby performing one or more processes, including one or more of the processes described herein. Such instructions and other data may be stored and transmitted using a variety of computer-readable media.
  • Databases, data repositories or other data stores described herein may include various kinds of mechanisms for storing, accessing, and retrieving various kinds of data, including a hierarchical database, a set of files in a file system, an application database in a proprietary format, a relational database management system (RDBMS), etc. Each such data store is generally included within a computing device employing a computer operating system such as one of those mentioned above, and are accessed via a network in any one or more of a variety of manners. A file system may be accessible from a computer operating system, and may include files stored in various formats. An RDBMS generally employs the Structured Query Language (SQL) in addition to a language for creating, storing, editing, and executing stored procedures, such as the PL/SQL language mentioned above.
  • In the preceding specification, various preferred embodiments have been described with references to the accompanying drawings. It will, however, be evident that various modifications and changes may be made thereto, and additional embodiments may be implemented, without departing from the broader scope of invention as set forth in the claims that follow. The specification and drawings are accordingly to be regarded in an illustrative rather than restrictive sense.
  • With regard to the processes, systems, methods, heuristics, etc. described herein, it should be understood that, although the steps of such processes, etc. have been described as occurring according to a certain ordered sequence, such processes could be practiced with the described steps performed in an order other than the order described herein. It further should be understood that certain steps could be performed simultaneously, that other steps could be added, or that certain steps described herein could be omitted. In other words, the descriptions of processes herein are provided for the purpose of illustrating certain embodiments, and should in no way be construed so as to limit the claims.
  • Accordingly, it is to be understood that the above description is intended to be illustrative and not restrictive. Many embodiments and applications other than the examples provided would be apparent upon reading the above description. The scope should be determined, not with reference to the above description, but should instead be determined with reference to the appended claims, along with the full scope of equivalents to which such claims are entitled. It is anticipated and intended that future developments will occur in the technologies discussed herein, and that the disclosed systems and methods will be incorporated into such future embodiments. In sum, it should be understood that the application is capable of modification and variation.
  • All terms used in the claims are intended to be given their broadest reasonable constructions and their ordinary meanings as understood by those knowledgeable in the technologies described herein unless an explicit indication to the contrary in made herein. In particular, use of the singular articles such as “a,” “the,” “said,” etc. should be read to recite one or more of the indicated elements unless a claim recites an explicit limitation to the contrary. The Abstract of the Disclosure is provided to allow the reader to quickly ascertain the nature of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. In addition, in the foregoing Detailed Description, it can be seen that various features are grouped together in various embodiments for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claimed embodiments require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separately claimed subject matter.

Claims (4)

1-20. (canceled)
21. A system, comprising:
an interface processor that provides an interactive graphical user interface to a user device over a network;
a parser processor that retrieves an RSS feed over the network, and generate an initial set of articles based on the RSS feed;
a categorization processor that categorizes each article in the initial set of articles into one of a plurality of subject matter categories;
a named entity recognizer that extracts entity mentions from the initial set of articles;
a named entity disambiguation module that identifies a named entity for each entity mention from at least one knowledge base;
a named entity salience identifier that identifies a salience for each named entity;
a deduplication processor that generates a final set of articles by removing duplicate articles from the initial set of articles, wherein each article in the final set of articles comprises a plurality of sentences and a title; and
a summarization processor that, for each article in the final set of articles:
generates a preliminary score for each sentence in the plurality of sentences,
assigns each sentence in the plurality of sentences to one of three categories,
generates an article summary, wherein the article summary comprises one or more sentences from one of the three categories, wherein the article summary is based at least in part on the preliminary score for each sentence, and
provides the article summary to the user device over the network via the interface processor.
22. The system of claim 21, wherein the named entity recognizer extracts entity mentions from the initial set of articles by identifying a plurality of entity boundaries in sentences in each of the initial set of articles and assigning a classification to each unique entity in the plurality of entity boundaries.
23. The system of claim 21, wherein the named entities comprise at least one of a person, a place, and a thing.
US15/131,997 2014-05-23 2016-04-18 Systems and Methods for Generating Summaries of Documents Abandoned US20160335234A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/131,997 US20160335234A1 (en) 2014-05-23 2016-04-18 Systems and Methods for Generating Summaries of Documents

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201462002350P 2014-05-23 2014-05-23
US14/681,612 US9317498B2 (en) 2014-05-23 2015-04-08 Systems and methods for generating summaries of documents
US15/131,997 US20160335234A1 (en) 2014-05-23 2016-04-18 Systems and Methods for Generating Summaries of Documents

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/681,612 Continuation US9317498B2 (en) 2014-05-23 2015-04-08 Systems and methods for generating summaries of documents

Publications (1)

Publication Number Publication Date
US20160335234A1 true US20160335234A1 (en) 2016-11-17

Family

ID=54554765

Family Applications (2)

Application Number Title Priority Date Filing Date
US14/681,612 Active US9317498B2 (en) 2014-05-23 2015-04-08 Systems and methods for generating summaries of documents
US15/131,997 Abandoned US20160335234A1 (en) 2014-05-23 2016-04-18 Systems and Methods for Generating Summaries of Documents

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US14/681,612 Active US9317498B2 (en) 2014-05-23 2015-04-08 Systems and methods for generating summaries of documents

Country Status (2)

Country Link
US (2) US9317498B2 (en)
WO (1) WO2015179643A1 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106844341A (en) * 2017-01-10 2017-06-13 北京百度网讯科技有限公司 News in brief extracting method and device based on artificial intelligence
CN109284357A (en) * 2018-08-29 2019-01-29 腾讯科技(深圳)有限公司 Interactive method, device, electronic equipment and computer-readable medium
CN110287280A (en) * 2019-06-24 2019-09-27 腾讯科技(深圳)有限公司 Method and apparatus, storage medium and the electronic equipment of word in a kind of analysis article
US10628474B2 (en) * 2016-07-06 2020-04-21 Adobe Inc. Probabalistic generation of diverse summaries
US10652592B2 (en) 2017-07-02 2020-05-12 Comigo Ltd. Named entity disambiguation for providing TV content enrichment
WO2020243846A1 (en) * 2019-06-06 2020-12-10 Bear Health Technologies Inc. System and method for automated file reporting
CN113420545A (en) * 2021-08-24 2021-09-21 平安科技(深圳)有限公司 Abstract generation method, device, equipment and storage medium
US11205103B2 (en) 2016-12-09 2021-12-21 The Research Foundation for the State University Semisupervised autoencoder for sentiment analysis
WO2022005188A1 (en) * 2020-07-01 2022-01-06 Samsung Electronics Co., Ltd. Entity recognition method, apparatus, electronic device and computer readable storage medium
US20220405343A1 (en) * 2021-06-17 2022-12-22 Verizon Media Inc. Generation and presentation of summary list based upon article

Families Citing this family (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150347390A1 (en) * 2014-05-30 2015-12-03 Vavni, Inc. Compliance Standards Metadata Generation
US9773499B2 (en) * 2014-06-18 2017-09-26 Google Inc. Entity name recognition based on entity type
US9672279B1 (en) * 2014-09-30 2017-06-06 EMC IP Holding Company LLC Cluster labeling system for documents comprising unstructured text data
US10430805B2 (en) * 2014-12-10 2019-10-01 Samsung Electronics Co., Ltd. Semantic enrichment of trajectory data
US10157178B2 (en) * 2015-02-06 2018-12-18 International Business Machines Corporation Identifying categories within textual data
US20160299881A1 (en) * 2015-04-07 2016-10-13 Xerox Corporation Method and system for summarizing a document
US10685289B2 (en) * 2015-06-05 2020-06-16 Intel Corporation Techniques for improving classification performance in supervised learning
US10515150B2 (en) * 2015-07-14 2019-12-24 Genesys Telecommunications Laboratories, Inc. Data driven speech enabled self-help systems and methods of operating thereof
US20170270096A1 (en) * 2015-08-04 2017-09-21 Yissum Research Development Company Of The Hebrew University Of Jerusalem Ltd. Method and system for generating large coded data set of text from textual documents using high resolution labeling
US10455088B2 (en) 2015-10-21 2019-10-22 Genesys Telecommunications Laboratories, Inc. Dialogue flow optimization and personalization
US10382623B2 (en) 2015-10-21 2019-08-13 Genesys Telecommunications Laboratories, Inc. Data-driven dialogue enabled self-help systems
US20170161372A1 (en) * 2015-12-04 2017-06-08 Codeq Llc Method and system for summarizing emails and extracting tasks
US10042880B1 (en) * 2016-01-06 2018-08-07 Amazon Technologies, Inc. Automated identification of start-of-reading location for ebooks
US10055489B2 (en) * 2016-02-08 2018-08-21 Ebay Inc. System and method for content-based media analysis
CN105808526B (en) * 2016-03-30 2019-07-30 北京京东尚科信息技术有限公司 Commodity short text core word extracting method and device
US20180039927A1 (en) * 2016-08-05 2018-02-08 General Electric Company Automatic summarization of employee performance
US10482086B2 (en) * 2016-11-30 2019-11-19 Salesforce.Com, Inc. Identifying similar database queries
US10706113B2 (en) 2017-01-06 2020-07-07 Microsoft Technology Licensing, Llc Domain review system for identifying entity relationships and corresponding insights
EP3382575A1 (en) 2017-03-27 2018-10-03 Skim It Ltd Electronic document file analysis
US10789281B2 (en) * 2017-06-29 2020-09-29 Xerox Corporation Regularities and trends discovery in a flow of business documents
JP6902945B2 (en) * 2017-07-07 2021-07-14 株式会社野村総合研究所 Text summarization system
CN109299454A (en) * 2017-07-24 2019-02-01 北京京东尚科信息技术有限公司 Abstraction generating method and device, storage medium and electric terminal based on chat log
US10127323B1 (en) * 2017-07-26 2018-11-13 International Business Machines Corporation Extractive query-focused multi-document summarization
US20190095438A1 (en) * 2017-09-22 2019-03-28 NewsRx, LLC Automating content generation
US10915707B2 (en) * 2017-10-20 2021-02-09 MachineVantage, Inc. Word replaceability through word vectors
CN109783795B (en) * 2017-11-14 2022-05-06 深圳市腾讯计算机系统有限公司 Method, device and equipment for obtaining abstract and computer readable storage medium
US10990602B2 (en) 2018-06-18 2021-04-27 Wipro Limited Method and system for generating optimized response to user input
CN110634172A (en) * 2018-06-25 2019-12-31 微软技术许可有限责任公司 Generating slides for presentation
US10929453B2 (en) * 2018-08-09 2021-02-23 Nec Corporation Verifying textual claims with a document corpus
RU2720899C2 (en) 2018-09-14 2020-05-14 Общество С Ограниченной Ответственностью "Яндекс" Method and system for determining user-specific content proportions for recommendation
RU2720952C2 (en) 2018-09-14 2020-05-15 Общество С Ограниченной Ответственностью "Яндекс" Method and system for generating digital content recommendation
US11526670B2 (en) * 2018-09-28 2022-12-13 The Mitre Corporation Machine learning of colloquial place names
RU2725659C2 (en) * 2018-10-08 2020-07-03 Общество С Ограниченной Ответственностью "Яндекс" Method and system for evaluating data on user-element interactions
CN109933787B (en) * 2019-02-14 2023-07-14 安徽省泰岳祥升软件有限公司 Text key information extraction method, device and medium
US11501073B2 (en) * 2019-02-26 2022-11-15 Greyb Research Private Limited Method, system, and device for creating patent document summaries
US11170169B2 (en) * 2019-03-29 2021-11-09 Innoplexus Ag System and method for language-independent contextual embedding
US11520971B2 (en) * 2019-03-30 2022-12-06 The Regents Of The University Of California System and method for artificial intelligence story generation allowing content introduction
CN110287491B (en) * 2019-06-25 2024-01-12 北京百度网讯科技有限公司 Event name generation method and device
US11080317B2 (en) 2019-07-09 2021-08-03 International Business Machines Corporation Context-aware sentence compression
RU2757406C1 (en) 2019-09-09 2021-10-15 Общество С Ограниченной Ответственностью «Яндекс» Method and system for providing a level of service when advertising content element
US11061951B2 (en) * 2019-11-21 2021-07-13 International Business Machines Corporation Unsupervised attention based scientific document summarization
CN111061860A (en) * 2019-12-12 2020-04-24 北京百度网讯科技有限公司 Abstract generation method and device
CN111046168B (en) * 2019-12-19 2023-12-01 智慧芽信息科技(苏州)有限公司 Method, apparatus, electronic device and medium for generating patent summary information
CN111104488B (en) * 2019-12-30 2023-10-24 广州广电运通信息科技有限公司 Method, device and storage medium for integrating retrieval and similarity analysis
US11423042B2 (en) * 2020-02-07 2022-08-23 International Business Machines Corporation Extracting information from unstructured documents using natural language processing and conversion of unstructured documents into structured documents
US11392753B2 (en) 2020-02-07 2022-07-19 International Business Machines Corporation Navigating unstructured documents using structured documents including information extracted from unstructured documents
KR20210102617A (en) * 2020-02-12 2021-08-20 삼성전자주식회사 Electronic apparatus and control method thereof
US11270061B2 (en) 2020-02-25 2022-03-08 International Business Machines Corporation Automatic generation of training data for scientific paper summarization using videos
US11157537B2 (en) * 2020-03-26 2021-10-26 Wipro Limited Method and system for summarizing content based on user preferences
CN111639175B (en) * 2020-05-29 2023-05-02 电子科技大学 Self-supervision dialogue text abstract method and system
CN111680152B (en) * 2020-06-10 2023-04-18 创新奇智(成都)科技有限公司 Method and device for extracting abstract of target text, electronic equipment and storage medium
US11626108B2 (en) * 2020-09-25 2023-04-11 Td Ameritrade Ip Company, Inc. Machine learning system for customer utterance intent prediction
CN112732898A (en) * 2020-12-30 2021-04-30 平安科技(深圳)有限公司 Document abstract generation method and device, computer equipment and storage medium
CN112686025B (en) * 2021-01-27 2023-09-19 浙江工商大学 Chinese choice question interference item generation method based on free text
CN113011152B (en) * 2021-03-04 2022-08-26 腾讯科技(深圳)有限公司 Text processing method, device and equipment and computer readable storage medium
US20220309245A1 (en) * 2021-03-24 2022-09-29 International Business Machines Corporation Automated natural language generation and selection
US11620320B1 (en) * 2021-10-13 2023-04-04 Dell Products L.P. Document summarization through iterative filtering of unstructured text data of documents
CN114201601B (en) * 2021-12-10 2023-03-28 北京金堤科技有限公司 Public opinion text abstract extraction method, device, equipment and computer storage medium
US11450124B1 (en) * 2022-04-21 2022-09-20 Morgan Stanley Services Group Inc. Scoring sentiment in documents using machine learning and fuzzy matching
CN117708545A (en) * 2024-02-01 2024-03-15 华中师范大学 Viewpoint contribution degree evaluation method and system integrating theme extraction and cosine similarity

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070260586A1 (en) * 2006-05-03 2007-11-08 Antonio Savona Systems and methods for selecting and organizing information using temporal clustering
US7496567B1 (en) * 2004-10-01 2009-02-24 Terril John Steichen System and method for document categorization
US20110055209A1 (en) * 2007-02-23 2011-03-03 Anthony Novac System and method for delivering content and advertisments
US20110107197A1 (en) * 2009-11-04 2011-05-05 International Business Machines Corporation Rendering sections of content in a document
US20110320543A1 (en) * 2010-06-28 2011-12-29 Bank Of America Corporation Analyzing News Content Information
US8423587B2 (en) * 2005-05-12 2013-04-16 Sybase, Inc. System and method for real-time content aggregation and syndication
US8849809B1 (en) * 2010-03-11 2014-09-30 Google Inc. Automated identification of news events

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7831597B2 (en) 2005-11-18 2010-11-09 The Boeing Company Text summarization method and apparatus using a multidimensional subspace
US10762080B2 (en) 2007-08-14 2020-09-01 John Nicholas and Kristin Gross Trust Temporal document sorter and method
CA2851772C (en) 2011-10-14 2017-03-28 Yahoo! Inc. Method and apparatus for automatically summarizing the contents of electronic documents
WO2014042612A1 (en) 2012-09-11 2014-03-20 Summly Ltd. Graphical user interface for presenting a menu of options

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7496567B1 (en) * 2004-10-01 2009-02-24 Terril John Steichen System and method for document categorization
US8423587B2 (en) * 2005-05-12 2013-04-16 Sybase, Inc. System and method for real-time content aggregation and syndication
US20070260586A1 (en) * 2006-05-03 2007-11-08 Antonio Savona Systems and methods for selecting and organizing information using temporal clustering
US20110055209A1 (en) * 2007-02-23 2011-03-03 Anthony Novac System and method for delivering content and advertisments
US20110107197A1 (en) * 2009-11-04 2011-05-05 International Business Machines Corporation Rendering sections of content in a document
US8849809B1 (en) * 2010-03-11 2014-09-30 Google Inc. Automated identification of news events
US20110320543A1 (en) * 2010-06-28 2011-12-29 Bank Of America Corporation Analyzing News Content Information

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Masum et al. "Emotion Sensitive News Agent: An Approach Towards User Centric Emotion Sensing from the News". 2007 IEEE/WIC/ACM International Conference in Web Intelligence. Pages 614-620 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10628474B2 (en) * 2016-07-06 2020-04-21 Adobe Inc. Probabalistic generation of diverse summaries
US11205103B2 (en) 2016-12-09 2021-12-21 The Research Foundation for the State University Semisupervised autoencoder for sentiment analysis
CN106844341A (en) * 2017-01-10 2017-06-13 北京百度网讯科技有限公司 News in brief extracting method and device based on artificial intelligence
US10652592B2 (en) 2017-07-02 2020-05-12 Comigo Ltd. Named entity disambiguation for providing TV content enrichment
CN109284357A (en) * 2018-08-29 2019-01-29 腾讯科技(深圳)有限公司 Interactive method, device, electronic equipment and computer-readable medium
WO2020243846A1 (en) * 2019-06-06 2020-12-10 Bear Health Technologies Inc. System and method for automated file reporting
CN110287280A (en) * 2019-06-24 2019-09-27 腾讯科技(深圳)有限公司 Method and apparatus, storage medium and the electronic equipment of word in a kind of analysis article
WO2022005188A1 (en) * 2020-07-01 2022-01-06 Samsung Electronics Co., Ltd. Entity recognition method, apparatus, electronic device and computer readable storage medium
US20220405343A1 (en) * 2021-06-17 2022-12-22 Verizon Media Inc. Generation and presentation of summary list based upon article
CN113420545A (en) * 2021-08-24 2021-09-21 平安科技(深圳)有限公司 Abstract generation method, device, equipment and storage medium

Also Published As

Publication number Publication date
WO2015179643A1 (en) 2015-11-26
US9317498B2 (en) 2016-04-19
US20150339288A1 (en) 2015-11-26

Similar Documents

Publication Publication Date Title
US9317498B2 (en) Systems and methods for generating summaries of documents
Mishra et al. A survey on question answering systems with classification
US9727637B2 (en) Retrieving text from a corpus of documents in an information handling system
US8751218B2 (en) Indexing content at semantic level
US20180300315A1 (en) Systems and methods for document processing using machine learning
US9645993B2 (en) Method and system for semantic searching
Moussa et al. A survey on opinion summarization techniques for social media
CA2774278C (en) Methods and systems for extracting keyphrases from natural text for search engine indexing
US20160299955A1 (en) Text mining system and tool
US20150120738A1 (en) System and method for document classification based on semantic analysis of the document
WO2010014082A1 (en) Method and apparatus for relating datasets by using semantic vectors and keyword analyses
Korayem et al. Sentiment/subjectivity analysis survey for languages other than English
US20160055155A1 (en) Answering Superlative Questions with a Question and Answer System
KR20160042896A (en) Browsing images via mined hyperlinked text snippets
Yalcin et al. An external plagiarism detection system based on part-of-speech (POS) tag n-grams and word embedding
Konchady Building Search Applications: Lucene, LingPipe, and Gate
Jia et al. A Chinese unknown word recognition method for micro-blog short text based on improved FP-growth
Fauzi et al. Image understanding and the web: a state-of-the-art review
Phan et al. Applying skip-gram word estimation and SVM-based classification for opinion mining Vietnamese food places text reviews
US20220365956A1 (en) Method and apparatus for generating patent summary information, and electronic device and medium
Wolyn et al. Summarization assessment methodology for multiple corpora using queries and classification for functional evaluation
WO2019173860A1 (en) Method and system for data curation
Wunnasri et al. Solving unbalanced data for Thai sentiment analysis
Baishya et al. Present state and future scope of Assamese text processing
Selvadurai A natural language processing based web mining system for social media analysis

Legal Events

Date Code Title Description
AS Assignment

Owner name: CODEQ, LLC, NORTH CAROLINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BAKER, DOUGLAS DANE;FERNANDEZ, PAULO MALVAR;FERNANDES, BRIAN;AND OTHERS;SIGNING DATES FROM 20160908 TO 20160919;REEL/FRAME:039925/0829

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION