WO2017023742A1 - Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system - Google Patents

Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system Download PDF

Info

Publication number
WO2017023742A1
WO2017023742A1 PCT/US2016/044687 US2016044687W WO2017023742A1 WO 2017023742 A1 WO2017023742 A1 WO 2017023742A1 US 2016044687 W US2016044687 W US 2016044687W WO 2017023742 A1 WO2017023742 A1 WO 2017023742A1
Authority
WO
WIPO (PCT)
Prior art keywords
content
submission
question
customer support
user
Prior art date
Application number
PCT/US2016/044687
Other languages
French (fr)
Inventor
Igor A. Podgorny
Benjamin John Koonse
Original Assignee
Intuit Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intuit Inc. filed Critical Intuit Inc.
Priority to CA2992563A priority Critical patent/CA2992563C/en
Priority to AU2016303436A priority patent/AU2016303436B2/en
Priority to EP16833601.4A priority patent/EP3329401A4/en
Publication of WO2017023742A1 publication Critical patent/WO2017023742A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/01Customer relationship services
    • G06Q30/015Providing customer assistance, e.g. assisting a customer within a business location or via helpdesk
    • G06Q30/016After-sales
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q40/00Finance; Insurance; Tax strategies; Processing of corporate or income taxes
    • G06Q40/12Accounting
    • G06Q40/123Tax preparation or submission

Definitions

  • a question and answer based customer support system includes a hosted forum through which a user can direct their specific questions, typically in a text format, to a support community that often includes other users and/or professional support personal.
  • customer support question and answer databases has numerous advantages including a self-help element whereby a searching user, i.e., a user accessing the resulting question and answer pair, can find an answer to their particular question by simply searching the customer support question and answer database for topics, questions, and answers related to their issue.
  • a searching user i.e., a user accessing the resulting question and answer pair
  • the user can then become an asking user by submitting their question to the question and answer based customer support system, typically through the same web-site and/or user interface. Consequently, by using a question and answer based customer support system that includes a customer support question and answer database, potentially millions of user questions can be answered in an efficient and effective manner, and with minimal duplicative effort.
  • the content that users are exposed to, within the question and answer based customer support system may affect the reputation of the service provider of the question and answer based customer support system. If users' queries consistently are directed to answers and/or questions that are unsatisfactory, the users will communicate dissatisfaction by, for example, using competitor question and answer systems, providing negative reviews in forums, and avoiding other products that are available from the service provider. Furthermore, the more dissatisfied users are with answers found in the question and answer based customer support system, the more likely the users are to request live customer support.
  • Embodiments of the present disclosure address some of the shortcomings associated with traditional question and answer based customer support systems for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, according to one embodiment.
  • the production environment assists the service provider in achieving business objectives such as, but not limited to, converting potential customers into paying customers of other services; reducing costs associated with user requests for live customer assistance; and attracting/directing/introducing new potential customers to products offered by the service provider, according to one embodiment.
  • the question and answer customer support system uses the probabilistic topic models to analyze various types of submission content, which can originate from different types of users, in order to generate various types of customer support content to facilitate and/or improve the user experience in the question and answer customer support system, according to one embodiment.
  • Probabilistic topic models extract hidden topics or summaries from content objects (e.g., database entries, webpages, in documents), without requiring the training of the model with known (e.g., manually verified) data sets, according to one embodiment.
  • the submission content can include question content (e.g., a question), response content (e.g., a comment or an answer to a question), and search query content (e.g., from a searching user), according to one embodiment.
  • the question and answer customer support system applies a probabilistic topic model to the submission content to generate customer support content such as, but not limited to, recommendations for improving question content, recommendations for improving response content, question quality indicators, question popularity indicators, answer quality indicators, answer popularity indicators, categorization of question content as product- related or tax -related, topically categorized navigation interfaces, and topically categorized search results, according to various embodiments.
  • customer support content such as, but not limited to, recommendations for improving question content, recommendations for improving response content, question quality indicators, question popularity indicators, answer quality indicators, answer popularity indicators, categorization of question content as product- related or tax -related, topically categorized navigation interfaces, and topically categorized search results, according to various embodiments.
  • the question and answer customer support system facilitates/enables: asking users to submit high-quality questions that result in more satisfying responses; asking users to submit popular questions, which can increase the likelihood that searching users will be directed to high-quality content; responding users to receive question content that is related to the responding users' particular areas of expertise (e.g., product-related topics or tax-related topics); and searching users to receive topically categorized and/or relevance-sorted search results in response to submission of search query content, according to various embodiments.
  • the probabilistic topic model that is applied to content in the tax environment is a Latent Dirichlet allocation ("LDA") algorithm or another version of a probabilistic topic model.
  • LDA Latent Dirichlet allocation
  • the disclosed method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system provides for significant improvements to the technical fields of customer support, information dissemination, software implementation, and user experience.
  • using the disclosed method and system for applying probabilistic topic models to content in a tax environment results in more efficient use of human and non-human resources, fewer processor cycles being utilized, reduced memory utilization, and less communications bandwidth being utilized to relay data to and from backend systems because users are less likely to request live customer support and because improving question quality and popularity results in users being directed to questions that are likely to result in user satisfaction and result in fewer additional question submissions.
  • computing systems are transformed into faster, more efficient, and more effective computing systems by implementing the method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system.
  • FIG. 1 is a block diagram representing one example of a hardware system and production environment for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system in accordance with one embodiment;
  • FIGs. 2A, 2B, and 2C are illustrative example graphs showing relationships between question quality and popularity for various topics that are determined by applying a probabilistic topic model to a sample data set in accordance with one embodiment;
  • FIGs. 3A, 3B, and 3C are illustrative example tables and a graph showing relationships between topic content, question popularity, and question quality, determined by applying a probabilistic topic model to a sample data set in accordance with one embodiment
  • FIGs. 4A and 4B are illustrate example user experience displays having topics and search results that are selected and sorted based on the application of a probabilistic topic model to a sample data set in accordance with one embodiment
  • FIG. 5 is a flow diagram of a method for applying probabilistic topic models to content in a tax environment in accordance with one embodiment.
  • FIG. 6 is a flow diagram of a method for applying probabilistic topic models to content in a tax environment in accordance with one embodiment.
  • a software system can be, but is not limited to, any data management system implemented on a computing system, accessed through one or more servers, accessed through a network, accessed through a cloud, and/or provided through any system or by any means, as discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing, that gathers/obtains data, from one or more sources and/or has the capability to analyze at least part of the data.
  • the term software system includes, but is not limited to the following: computing system implemented, and/or online, and/or web-based, personal and/or business tax preparation systems; computing system implemented, and/or online, and/or web- based, personal and/or business financial management systems, services, packages, programs, modules, or applications; computing system implemented, and/or online, and/or web-based, personal and/or business management systems, services, packages, programs, modules, or applications; computing system implemented, and/or online, and/or web-based, personal and/or business accounting and/or invoicing systems, services, packages, programs, modules, or applications; and various other personal and/or business electronic data management systems, services, packages, programs, modules, or applications, whether known at the time of filling or as developed later.
  • Specific examples of software systems include, but are not limited to the following: TurboTaxTM available from Intuit, Inc. of Mountain View, California; TurboTax OnlineTM available from Intuit, Inc. of Mountain View, California; QuickenTM, available from Intuit, Inc. of Mountain View, California; Quicken OnlineTM, available from Intuit, Inc. of Mountain View, California; QuickBooksTM, available from Intuit, Inc. of Mountain View, California; QuickBooks OnlineTM, available from Intuit, Inc. of Mountain View, California; MintTM, available from Intuit, Inc. of Mountain View, California; Mint OnlineTM, available from Intuit, Inc. of Mountain View, California; and/or various other software systems discussed herein, and/or known to those of skill in the art at the time of filing, and/or as developed after the time of filing.
  • computing entity include, but are not limited to, the following: a server computing system; a workstation; a desktop computing system; a mobile computing system, including, but not limited to, smart phones, portable devices, and/or devices worn or carried by a user; a database system or storage cluster; a virtual asset; a switching system; a router; any hardware system; any communications system; any form of proxy system; a gateway system; a firewall system; a load balancing system; or any device, subsystem, or mechanism that includes components that can execute all, or part, of any one of the processes and/or operations as described herein.
  • computing system and “computing entity,” can denote, but are not limited to the following: systems made up of multiple virtual assets, server computing systems, workstations, desktop computing systems, mobile computing systems, database systems or storage clusters, switching systems, routers, hardware systems, communications systems, proxy systems, gateway systems, firewall systems, load balancing systems, or any devices that can be used to perform the processes and/or operations as described herein.
  • production environment includes the various components, or assets, used to deploy, implement, access, and use, a given software system as that software system is intended to be used.
  • production environments include multiple computing systems and/or assets that are combined, communicatively coupled, virtually and/or physically connected, and/or associated with one another, to provide the production environment implementing the application.
  • the assets making up a given production environment can include, but are not limited to, the following: one or more computing environments used to implement at least part of the software system in the production environment such as a data center, a cloud computing environment, a dedicated hosting environment, and/or one or more other computing environments in which one or more assets used by the application in the production environment are implemented; one or more computing systems or computing entities used to implement at least part of the software system in the production environment; one or more virtual assets used to implement at least part of the software system in the production environment; one or more supervisory or control systems, such as hypervisors, or other monitoring and management systems used to monitor and control assets and/or components of the production environment; one or more communications channels for sending and receiving data used to implement at least part of the software system in the production environment; one or more access control systems for limiting access to various components of the production environment, such as firewalls and gateways; one or more traffic and/or routing systems used to direct, control, and/or buffer data traffic to components of the
  • computing environment includes, but is not limited to, a logical or physical grouping of connected or networked computing systems and/or virtual assets using the same infrastructure and systems such as, but not limited to, hardware systems, software systems, and networking/communications systems.
  • computing environments are either known, “trusted” environments or unknown, “untrusted” environments.
  • trusted computing environments are those where the assets, infrastructure, communication and networking systems, and security systems associated with the computing systems and/or virtual assets making up the trusted computing environment, are either under the control of, or known to, a party.
  • each computing environment includes allocated assets and virtual assets associated with, and controlled or used to create, and/or deploy, and/or operate at least part of the software system.
  • one or more cloud computing environments are used to create, and/or deploy, and/or operate at least part of the software system that can be any form of cloud computing environment, such as, but not limited to, a public cloud; a private cloud; a virtual private network (VPN); a subnet; a Virtual Private Cloud (VPC); a sub-net or any security/communications grouping; or any other cloud-based infrastructure, sub-structure, or architecture, as discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing.
  • VPN virtual private network
  • VPC Virtual Private Cloud
  • a given software system or service may utilize, and interface with, multiple cloud computing environments, such as multiple VPCs, in the course of being created, and/or deployed, and/or operated.
  • the term "virtual asset” includes any virtualized entity or resource, and/or virtualized part of an actual, or "bare metal” entity.
  • the virtual assets can be, but are not limited to, the following: virtual machines, virtual servers, and instances implemented in a cloud computing environment; databases associated with a cloud computing environment, and/or implemented in a cloud computing environment; services associated with, and/or delivered through, a cloud computing environment; communications systems used with, part of, or provided through a cloud computing environment; and/or any other virtualized assets and/or sub-systems of "bare metal" physical devices such as mobile devices, remote sensors, laptops, desktops, point-of-sale devices, etc., located within a data center, within a cloud computing environment, and/or any other physical or logical location, as discussed herein, and/or as known/available in the art at the time of filing, and/or as
  • any, or all, of the assets making up a given production environment discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing can be implemented as one or more virtual assets.
  • two or more assets such as computing systems and/or virtual assets, and/or two or more computing environments are connected by one or more
  • communications channels including but not limited to, Secure Sockets Layer (SSL)
  • SSL Secure Sockets Layer
  • communications channels and various other secure communications channels, and/or distributed computing system networks such as, but not limited to the following: a public cloud; a private cloud; a virtual private network (VPN); a subnet; any general network, communications network, or general network/communications network system; a combination of different network types; a public network; a private network; a satellite network; a cable network; or any other network capable of allowing communication between two or more assets, computing systems, and/or virtual assets, as discussed herein, and/or available or known at the time of filing, and/or as developed after the time of filing.
  • VPN virtual private network
  • the term "network” includes, but is not limited to, any network or network system such as, but not limited to, the following: a peer-to-peer network; a hybrid peer- to-peer network; a Local Area Network (LAN); a Wide Area Network (WAN); a public network, such as the Internet; a private network; a cellular network; any general network, communications network, or general network/communications network system; a wireless network; a wired network; a wireless and wired combination network; a satellite network; a cable network; any combination of different network types; or any other system capable of allowing communication between two or more assets, virtual assets, and/or computing systems, whether available or known at the time of filing or as later developed.
  • a peer-to-peer network such as, but not limited to, the following: a peer-to-peer network; a hybrid peer- to-peer network; a Local Area Network (LAN); a Wide Area Network (WAN); a public network, such as the Internet; a private network;
  • user experience display includes not only data entry and question submission user interfaces, but also other user experience features provided or displayed to the user such as, but not limited to the following: data entry fields; question quality indicators; images; backgrounds; avatars; highlighting mechanisms; icons; and any other features that individually, or in combination, create a user experience, as discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing.
  • the term "question quality indicator” includes any mechanism, means, or feature/function provided to indicate to a user a determined quality of a question being provided by the user.
  • Specific examples of question quality indicators include, but are not limited to, meter displays; line displays; score displays; audio content; visual content; images; backgrounds; avatars; highlighting mechanisms; icons; and any other features that individually, or in combination, relay to a user a determined quality of a question being submitted.
  • the term "question popularity indicator” includes any mechanism, means, or feature/function provided to indicate to a user a determined/estimated popularity of a question being provided by the user.
  • Specific examples of question popularity indicators include, but are not limited to, meter displays; line displays; score displays; audio content; visual content; images; backgrounds; avatars; highlighting mechanisms; icons; and any other features that individually, or in combination, relay to a user a determined, estimated, or predicted quality of a question being submitted.
  • the method and system for determining a level of popularity of submission content prior to publicizing the submission content with a question and answer support system as described herein, and/or a person and/or entity that interfaces with, and/or to whom information is provided by, the method and system for determining a level of popularity of submission content, prior to publicizing the submission content with a question and answer support system as described herein, and/or a legal guardian of person and/or entity that interfaces with, and/or to whom information is provided by, the method and system for determining a level of popularity of submission content, prior to publicizing the submission content with a question and answer support system as described herein, and/or an authorized agent of any party and/or person and/or entity that interfaces with, and/or to whom information is provided by, the method and system for determining a level of popularity of submission content, prior to publicizing the submission content with a question and answer support system as described herein.
  • a user can be, but is not limited to, a person, a commercial entity, an application, a service, and/or a computing system.
  • the term "asking user” includes a user of a software system submitting submission content (e.g., a question) to a question and answer based customer support system.
  • searching user includes a user of a software system submitting a search query to a customer support question and answer database associated with a question and answer customer support system.
  • An “answer recipient” includes one or both of the “asking user” and the “searching user", according to one embodiment.
  • the term “responding user” includes a user of a software system who submits a response to submission content for the question and answer based customer support system.
  • the term “response” is interchangeably used with the term “reply”
  • the term “responding user” is interchangeably used with the term(s) "replying user” and/or “answering user”.
  • a “responding user”, “replying user”, and/or “answering user” is a user who submits an answer to submission content (e.g., a question) and/or one who submits a comment to submission content in the question and answer based customer support system.
  • submission content includes a question content (inclusive of question summary and question details), response content, and search query content and the term “submission content” is used interchangeably with the term “question”.
  • a question summary is a character limited summary or articulation of the question, and the question details are additional information about the user or the circumstances surrounding the question summary.
  • a "post" is used to refer to a publicized or published version of the submission content, and may include comments and/or answers submitted by users in response to publicizing, publishing, hosting, and/or posting the submission content. Although submission content and a post may include similar information, one references content that has not been made publically available by a question and answer customer support system and the other has been made available for review, response, and comment by the public.
  • the term "probabilistic topic model” or “probabilistic model” denotes one or more individual or combined algorithms or sets of equations that describe, determine, or predict characteristics of or the performance of a datum, a data set, multiple data sets, data objects, a computing system, and/or multiple computing system.
  • the probabilistic topic model include algorithms configured to discover the hidden thematic (or semantic) structure in large data objects, text, and archives of documents. DETAILED DISCLOSURE
  • FIG.s depict one or more exemplary embodiments.
  • Embodiments may be implemented in many different forms and should not be construed as limited to the embodiments set forth herein, shown in the FIG.s, and/or described below. Rather, these exemplary embodiments are provided to allow a complete disclosure that conveys the principles of the invention, as set forth in the claims, to those of skill in the art.
  • FIG. 1 illustrates a block diagram of a production environment 100 for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, according to one embodiment.
  • the production environment assists the service provider in achieving business objectives such as, but not limited to, converting potential customers into paying customers of other services; reducing user requests for live customer assistance; and attracting/directing/introducing new potential customers to products offered by the service provider, according to one embodiment.
  • Probabilistic topic models extract hidden topics or extract content summaries from content objects (e.g., database entries, webpages, in documents), without requiring the training of the model with known (e.g., manually verified) data sets, according to one embodiment.
  • the hidden topics or content summaries can then be labeled by system administrators or other users, based on the terms returned from the model for the hidden topics or content summaries, according to one embodiment.
  • the production environment 100 uses the probabilistic topic models to analyze various types of submission content, which can originate from different types of users, in order to generate various types of customer support content to facilitate and/or improve the user experience in the question and answer customer support system, according to one embodiment.
  • the submission content can include question content (e.g., a question), response content (e.g., a comment or an answer to a question), and search query content, according to one embodiment.
  • the production environment 100 applies a probabilistic topic model to the submission content to generate a customer support content such as, but not limited to, recommendations for improving question content, recommendations for improving response content, question quality indicators, question popularity indicators, answer quality indicators, answer popularity indicators, categorization of question content as product-related or tax-related, topically categorized navigation interfaces, topically categorized search results, and recommendations for improving a response, according to various embodiments.
  • the production environment 100 facilitates/enables: asking users to submit high-quality questions that result in more satisfying responses; asking users to submit popular questions, which can increase the likelihood that searching users will be directed to high-quality content; responding users to receive question content that is related to the responding users' particular areas of expertise (e.g., product-related topics or tax-related topics); and searching users to receive topically categorized and/or relevance-sorted search results in response to submission of search query content, according to various embodiments.
  • the probabilistic topic model that is applied to content in the tax environment is a Latent Dirichlet allocation ("LDA”) algorithm or another version of a probabilistic topic model.
  • LDA Latent Dirichlet allocation
  • the production environment 100 includes a service provider computing environment 110, an asking user computing environment 140, a responding user computing environment 145, and a searching user computing environment 150, according to one embodiment.
  • the service provider computing environment 110 includes a question and answer customer support system 111 that is associated with and/or configured to support a tax return preparation system 112 and/or one or more additional service provider systems 113, according to one embodiment.
  • the question and answer customer support system 111, the tax return preparation system 112, and the one or more additional service provider systems 113 are software systems, according to one embodiment.
  • software system includes, but is not limited to the following: computing system implemented, and/or online, and/or web-based, personal and/or business tax preparation systems; computing system implemented, and/or online, and/or web-based, personal and/or business financial management systems, services, packages, programs, modules, or applications; computing system
  • the question and answer customer support system 111 e.g., a social question and answer (Q&A) system
  • Q&A social question and answer
  • the software system e.g., the tax return preparation system 112 and/or one or more additional service provider systems 113.
  • the question and answer customer support system 111 includes a customer support engine 114, an analytics module 115, and a customer support content database 116 for applying probabilistic topic models to content in a tax environment to improve/maintain customer satisfaction with the question and answer customer support system 111, according to one embodiment.
  • the customer support engine 114 includes a user interface 117 for providing a user interface display that receives new submission content 118 from a user and that delivers new customer support content 119 to the user, according to one embodiment.
  • the user interface 117 includes, but is not limited to one or more data entry fields; question quality indicators; images; backgrounds; avatars; highlighting mechanisms; icons; boxes; slides; buttons; and any other user interface elements or features that individually, or in combination, create a user experience, as discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing.
  • the customer support engine 114 uses the user interface 117 to receive different types of new submission content 118 from different types of users, according to one
  • the new submission content 118 includes question content 120, response content 121, and search query content 122, according to one embodiment.
  • the question content 120 is received from an asking user through the asking user computing environment 140
  • the response content 121 is received from a responding user through the responding user computing environment 145
  • the search query content 122 is received from a searching user through the searching user computing environment 150, according to one embodiment.
  • question and answer customer support system 111 is provided to support the tax return preparation system 112 and therefore the question content 120, the response content 121, and the search query content 122 are related to tax -related questions (e.g., federal and state taxation and tax preparation) and/or product-related questions (e.g., the installation and/or operations of the tax return preparation system 112), according to one embodiment.
  • tax -related questions e.g., federal and state taxation and tax preparation
  • product-related questions e.g., the installation and/or operations of the tax return preparation system 112
  • the question and answer customer support system 111 uses the customer support engine 114 and/or the user interface 117 to enable: asking users to submit question content 120 (e.g., questions); responding users to submit response content 121 (e.g., answers to questions); and searching users to submit search query content 122 (e.g., one or more search terms) to find answers to their questions that are already stored/maintained by the question and answer customer support system 111, according to one embodiment.
  • question content 120 e.g., questions
  • responding users to submit response content 121 e.g., answers to questions
  • searching users to submit search query content 122 e.g., one or more search terms
  • the customer support engine 114 uses the user interface 117 to provide different types of new customer support content 119 to the relevant type of user, e.g., asking user, responding user, searching user, etc., according to one embodiment.
  • the new customer support content 119 is generated by the analytics module 115 and includes, but is not limited to, recommendations, indicators, user interface ("UI") elements, topics, and/or search results that facilitate/improve users' experience/interactions with the question and answer customer support system 111, according to one embodiment. Additional example embodiments of the new customer support content 119 will be described below after a discussion of embodiments of applying a probabilistic topic model to the new submission content 118 with the analytics module 115.
  • the customer support engine 114 updates the customer support content database 116 in response to receiving the new submission content 118 and/or the new customer support content 119, according to one embodiment.
  • the customer support content database 116 includes existing submission content 123 (e.g., question content, response content, and search query content) and existing customer support content 124 (e.g., a history of recommendations and customer service provided to users), according to one embodiment.
  • the customer support content database 116 stores and maintains one or more tables or other data structures of previously received questions, responses, comments, and search queries received from users of the question and answer customer support system 111, according to one embodiment.
  • the customer support engine 114 updates the existing submission content 123 to reflect the new submission content 118, and the customer support engine 114 updates the existing customer support content 124 to reflect the new customer support content 119, according to one embodiment.
  • the question and answer customer support system 111 publicizes the contents of the customer support content database 116 to enable users to submit question content, submit/review response content associated with the question content, submit search query content to find response content that is relevant to the user's current needs, and otherwise view and/or interact with content hosted by the question and answer customer support system 111, according to one embodiment.
  • the question content 120 submitted to the question and answer customer support system 111 can be related to very different broad categories, be of various question types, have varying predicted answer lengths, and be formatted in various different ways, according to one embodiment.
  • the question content 120 includes a question summary (which provide a character-limited overview or description of the question), question details (which provide a non-character-limited description of the circumstances and/or background and/or context for the question summary, and user click-stream data (e.g., user IP address, web browsing history, geographical location, click speeds, hover durations, hardware identifier(s), and the like), according to one embodiment.
  • the response content 121 is provided by responding users who include paid support personnel in the employ of the service provider and volunteer experts, according to one embodiment.
  • the response content 121 can include comments and answers to questions in the question content 120, according to one embodiment.
  • the search query content 122 includes one or more terms, phrases, or sentences used by searching users to search the question and answer customer support system for answers to product-related and/or substantive questions about the tax return preparation system 112 and/or the one or more additional service provider systems 113, according to one embodiment.
  • the search query content 122 also includes other miscellaneous information about the searching user such as, but not limited to, the browsing history of the user, how the user arrived at the interface for the question and answer customer support system 111 (e.g., the landing page), the relationship of the user with the tax return preparation system 112 (e.g., system authenticated user, guest user, redirected user from search engine) and/or the one or more additional service provider systems 113, the IP address and/or geographic location of the user, and the like.
  • the browsing history of the user e.g., how the user arrived at the interface for the question and answer customer support system 111 (e.g., the landing page), the relationship of the user with the tax return preparation system 112 (e.g., system authenticated user, guest user, redirected user from search engine) and/or the one or more additional service provider systems 113, the IP address and/or geographic location of the user, and the like.
  • the tax return preparation system 112 e.g., system authenticated user, guest user,
  • the customer support engine 114 receives the new customer support content 119, in response to providing the new submission content 118 to the analytics module 115 for processing with a probabilistic topic model 125, according to one embodiment.
  • Probabilistic topic models are algorithms that processes text from a database elements, documents, files, and other content objects to determine topics, themes, and/or subjects of text-based content.
  • the topic is a term/phrase for which the text-based content includes words having a high probability of relevance to the term/phrase, according to one embodiment.
  • the probabilistic topic models populate a list, array, or other data structure with the topics of the text-based content, along with statistical information associated with the topics, according to one embodiment.
  • the statistical information associated with the topics include, but are not limited to, quantity of occurrences, distribution of topics in the content, distribution of words in the topic, probability of a word occurring in a topic, probability of a topic occurring in the text-based content, and the like, according to various embodiments.
  • the probabilistic topic model 125 is the Latent Dirichlet allocation algorithm or another version of a probabilistic topic model.
  • the probabilistic topic model 125 receives the new submission content 118 and model parameters 126, to generate model output 127, according to one embodiment.
  • the model parameters 126 include, but are not limited to, the number of topics for the probabilistic topic model 125 to generate and the number of iterations for the probabilistic topic model 125 to execute while processing the new submission content 118, according to one embodiment.
  • the probabilistic topic model 125 can be configured to process hundreds, thousands, or tens of thousands of data samples, e.g., question and answer pairs with corresponding view and vote data, in a matter of hours, when a similar manual processing of the data samples might take weeks or months of manual human processing, according to one embodiment.
  • the model output 127 includes submission content topics 128 and submission content topic statistics 129, according to one embodiment.
  • the submission content topics 128 are the terms/phrases for which the new submission content 118 includes words having a high probability of relevance to the terms/phrases, according to one embodiment.
  • submission content topics 128 are discrete portions of the new submission content 118 that provide quantifiable summaries of the submission content.
  • the submission content topics 128 include the identity of a topic of a word in the new submission content 118, the identity of a topic of all words in the new submission content 118, the identity of a word in the new submission content 118, and/or the identity of all words and all more than one instance of submission content, according to one embodiment.
  • the submission content topic statistics 129 include, but are not limited to, the probability of a word occurring in a topic, the distribution of words in a topic, the probability of a topic occurring in the new submission content 118 or in another text-based content, and/or the distribution of topics in the new submission content 118, according to one embodiment.
  • the analytics module 115 uses the customer support content generator 130 to generate different types of new customer support content 119, based on the model output 127 for the new submission content 118 and based on the content generator database 131, according to one embodiment.
  • the customer support content generator 130 is configured to index, search, and/or rank the submission content topics 128 based at least partially on the submission content topic statistics 129 to determine which content from the content generator database 131 to use to populate the new customer support content 119, according to one embodiment.
  • the customer support content generator 130 may apply one or more thresholds 132 to the model output 127 to determine whether the question content 120 is estimated or predicted to receive enough votes to be deemed "popular".
  • the customer support content generator 130 may display one or more user interface elements 133, such as meters, slides, digital displays, and the like to indicate a level of popularity and/or a level of quality of question content 120, according to one embodiment.
  • the customer support content generator 130 may use topics 134 and/or phrases 135 to populate templates for recommendations, suggestions, and/or encouragement for a user to alter the question summary and/or the question details used in question content, according to one embodiment.
  • the customer support content generator 130 may use the thresholds 132, the user interface elements 133, the topics 134, and/or the phrases 135, to encourage or recommend that a responding user alter response content 121, according to one embodiment.
  • the customer support content generator 130 may also use the contents of the content generator database 131 to display the submission content topics 128, based on the relevance of the search query content 122 to facilitate navigation of the user experience display for the user interface 117, according to one embodiment.
  • the analytics module 115 uses the customer support content generator 130 to generate new customer support content 119 that provides guidance to an asking user while the user is creating/generating the question content 120, according to one embodiment.
  • the analytics module 115 can be configured to populate the new customer support content 119 with real-time recommendations for improving the quality and/or the popularity of the question content 120 that the asking user is creating.
  • the customer support content generator 130 is configured to analyze model output 127 for the question content 120 and populate the new customer support content with one or more of a question quality indicator, a question popularity indicator, and/or recommendations for improving the quality and/or the popularity of the question content 120, according to one embodiment.
  • FIG. 2A illustrates a question quality graph 200
  • FIG. 2B illustrates a question popularity graph 220
  • FIG. 2C illustrates a question popularity and quality correlation graph 230, which are each generated by applying Latent Dirichlet allocation algorithm (an embodiment of the probabilistic topic model 125) to approximately 62,000 question content samples from 2013 (an embodiment of existing submission content 123).
  • Latent Dirichlet allocation algorithm was applied to question subjects and question details of the 62,000 question content to rank 50 topics within the question content samples based on the percentages of up votes received by each of the 50 topics and based on the quantity of posts made for each of the 50 topics.
  • the question quality graph 200 includes an x-axis that identifies each one of 50 topics and a y-axis that identifies the percentage of up vote received for each of the 50 topics.
  • the plot 201 includes the dark circles in the graph.
  • the question popularity graph 220 includes an x-axis that identifies each one of 50 topics and a y-axis that identifies the number of posts, i.e., question submissions, made for each of the 50 topics.
  • the plot 221 illustrates the relationship between the 50 topics and the number of posts made to a question and answer customer support system for each of the 50 topics.
  • Each of the 50 topics (derived from the 62,000 questions) were manually evaluated, and the Latent Dirichlet allocation algorithm proved to be consistent with the findings that were manually determined by people.
  • the question popularity and quality correlation graph 230 includes an x- axis that identifies the percentage of up vote for each of the 50 topics and a y-axis that identifies the views per post (i.e., an indication of popularity) for the 50 topics. As illustrated, there is very little correlation between the topics that users tend to up vote and the topics that receive above average views per post. Accordingly, the probabilistic topic model 125 can be used to predict question quality and question popularity separately, and can be used by the analytics module 115 to provide indicators and recommendations to the asking user to assist the asking user in strengthening the quality and/or the popularity of the question content formulated by the asking user, according to one embodiment.
  • the question and answer customer support system 111 can be configured to guide users to modify their question content so that it is likely to be of high-quality (e.g., receive customer satisfaction) and be popular (e.g., receive average or above average quantity of views), according to one embodiment.
  • the analytics module 115 also uses the model output 127 to determine whether the question content 120 contains a predominantly product-related question or a predominantly tax -related question, to properly route the question content 120 to the type of responding user who can effectively generate satisfactory response content associated with the question content 120, according to one embodiment.
  • some questions submitted to the question and answer based customer support system are product-related questions, e.g., questions related to pricing, installation, version choice, etc. for the software systems that often have little or no relation to the subject matter/job of the software system, i.e., the endeavor supported by the software system.
  • some questions submitted to the question and answer based customer support system are subject matter related, or substantive, questions directly related to the subject matter/endeavor of the software system.
  • the analytics module 115 uses the probabilistic topic model 125 to identify broad category/subject matter of the questions, e.g., product-related questions and subject matter related questions, to facilitate appropriately routing the questions to support personnel or volunteer expert users of the software system.
  • FIGs. 3A, 3B, and 3C illustrate example tables and a graph that demonstrate the effectiveness of applying the Latent Dirichlet allocation algorithm (i.e., an embodiment of the probabilistic topic model 125) to characterize the popularity and quality of topics based on whether the topic is predominantly a product-related question or predominantly a tax -related question.
  • FIG. 3A includes a content type popularity table 300 illustrating results of applying the Latent Dirichlet allocation algorithm to a data set, and shows that topic 18 received more views than topic 27.
  • the content type popularity table 300 includes columns of terms and frequency of the terms for each of the topics 18 and 27.
  • topic 18 The terms for topic 18 indicate that topic 18 predominantly includes product-related question content and is almost 7 times more popular than topic 27, which has terms that indicate the topic 27 permanently includes tax-related question content.
  • One conclusion that can be drawn from the content type popularity table 300 is that procedure or product-related question content may be more popular than tax- related question content, according to one embodiment.
  • FIG. 3B includes a content type quality table 320 illustrating results of applying the Latent Dirichlet allocation algorithm to a data set, and the content type quality table 320 shows that topic 32 receives nearly twice the percentage of up votes (82.5%) as topic 6 (46.8%).
  • quality refers to customer satisfaction, as reflected in more up votes.
  • the content type quality table 320 includes columns of terms and frequency of terms for each of the topics 32 and 6. The terms for topic 32 indicate that topic 32 predominantly includes tax -related question content and that the terms for topic 6 indicate that topic 6 predominately includes product-related question content.
  • FIG. 3C includes a question content quality graph 330, which indicates that tax- related question content receives a higher percentage of up votes than product-related question content.
  • the question content quality graph 330 relies on the same 62,000 question content data set that is used to generate the graphs of FIGs.
  • the question content quality graph 330 includes an x-axis that provides a scale of predominantly tax-related question content on the left side of the graph (at "0.0") that extends to predominantly product-related question content on the right side of the graph (at "1.0").
  • the y-axis indicates a percentage of up votes. The closer a topic is to a tax- related question content, the higher (on average) the up vote percentages were.
  • the capability of the probabilistic model to distinguish tax-related question content from product-related question content enables the question and answer customer support system 111 to route questions to the appropriate responding users so that a question can be adequately and efficiently addressed, according to one embodiment.
  • the analytics module 115 uses the customer support content generator 130 to generate new customer support content 119 to provide guidance to a responding user while the user is creating/generating the response content 121 that is associated with the question content 120, according to one embodiment.
  • the guidance can include one or more user interface elements 133 and/or phrases 135 that include
  • recommendations and/or suggestions for improving the likelihood for user satisfaction of an answer to a question may include question quality indicators and/or question popularity indicators to motivate the question answer to compile high-quality and popular response content
  • the analytics module 115 uses the customer support content generator 130 to generate new customer support content 119 to customize the navigation of and search results displayed in the user experience display for the user interface 117, to assist the searching user in efficiently finding existing submission content 123 that is relevant to the search query content
  • the analytics module 115 receives the search query content 122 and applies the probabilistic topic model 125 to determine the dominant topics, e.g., highest ranked of the submission content topics 128, of the search query content 122.
  • the analytics module 115 applies the probabilistic topic model 125 to the existing submission content 123 of the customer support content database 116 to determine the topics and frequencies of topics of the existing submission content 123, according to one embodiment.
  • the analytics module 115 is configured to sort the model output 127 for the existing submission content 123 in an order that is based on the dominant topics of the search query content 122, according to one embodiment.
  • the new customer support content 119 and the submission content topics 128 are ranked by relevance to the search query content 122 in a navigation bar of a webpage or web interface and/or as search results in a webpage or web interface, in response to receipt of the search query content 122 from a searching user, according to one embodiment.
  • FIGs. 4A and 4B illustrate example user experience displays 400 and 420 that have topics and search results that are sorted based on the application of the Latent Dirichlet allocation algorithm to a data set of at least part of a question and answer database (i.e., an example embodiment of the existing submission content 123) and to a searching user's search query (i.e., an example embodiment of the search query content 122), according to one embodiment.
  • the user experience display 400 includes a navigation bar 401 that provides a sorted list of topics that are related to a search query received from a searching user, according to one embodiment.
  • the sorted list of the navigation bar 401 can be sorted/ranked in order based on the popularity of the topics, based on the quality of the products, and/or based on the relevance of the topics to the search query submitted by the searching user.
  • the user experience display 420 includes search results 421 that are generated and/or sorted based on the application of the Latent Dirichlet allocation algorithm to at least part of the question and answer database, and to search query content submitted by the searching user.
  • the search results 421 are sorted based on popularity, quality, and/or relevance to the search query submitted by the searching user, according to various embodiments.
  • the question and answer customer support system 111 can use the analytics module 115 to apply the probabilistic topic model 125 to the customer support content database 116 to provide quality control analyses of the customer support content database 116.
  • the analytics module 115 can use the probabilistic topic model 125 to search the existing submission content 123 for redundant entries, for entries that are likely to be low quality, for entries that are likely to the unpopular, and can be configured to remove low- quality, unpopular, and/or redundant entries from the existing submission content 123.
  • the question and answer customer support system 111 increases the likelihood that searching users will find content that is high-quality, popular, and/or relevant to the search query content 122 submitted by the searching user, according to one embodiment.
  • Applying probabilistic topic models to submission content in the question and answer customer support system 111 can provide a number of benefits to the question and answer customer support system 111, according to one embodiment.
  • the probabilistic topic model 125 is scalable and can be applied to customer support systems that are associated with the one or more additional service provider systems 113, i.e., software systems other than the tax return preparation system 112. The probabilistic topic model 125 can be operated
  • the probabilistic topic model 125 can be used without being trained by confirmed or known data sets. This benefit enables providing question and answer customer support services using less processing power, fewer computing cycles, and less computing system bandwidth than traditional techniques for providing question and answer customer support services, according to one embodiment.
  • the probabilistic topic model 125 can be used to determine whether the content type of submission content is predominantly tax- related or predominantly product-related so that the submission content can be routed to the appropriate responding users or responding user group, and to enable the submission content to be correctly identified in the customer support content database 116, to facilitate the delivery of efficient and accurate search results to searching users, according to one embodiment.
  • the probabilistic topic model 125 can be used to improve user interactions with the question and answer customer support system 111 by, assisting a user in improving question content 120, assisting a user to improve response content 121, and customizing search results to the search query content 122, among other benefits, according to one embodiment.
  • FIG. 5 is a flow diagram of a process 500 for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, in accordance with one embodiment.
  • the process receives submission content from a user, according to one embodiment.
  • the user can be an asking user, a responding user, or a searching user, according to various embodiments.
  • the submission content can include question content, response content, or search query content, according to various embodiments.
  • the process proceeds to operation 502, according to one embodiment.
  • the process determines if the submission content is question content, according to one embodiment. If the submission content is question content, the process proceeds to operation 503, and if the submission content is not question content, the process proceeds to operation 505, according to one embodiment.
  • the process applies a probabilistic topic model to the question content to identify the question content as tax-related or product-related, and applies the probabilistic topic model to the question content to assist a user in crafting quality and/or popular question content, according to one embodiment. Based on the output of the
  • the process provides recommendations, question quality indicators, and/or question popularity indicators, to assist the user in improving the question content, according to one embodiment.
  • the process proceeds to operation 504, according to one embodiment.
  • the process updates a customer support content database, according to one embodiment.
  • the process returns to operation 501, according to one embodiment.
  • the process determines if the submission content is response content, according to one embodiment. If the submission content is response content, the process proceeds to operation 506, and if the submission content is not response content, the process proceeds to operation 507, wherein to one embodiment.
  • the process applies probabilistic topic model to response content to identify the response content as tax -related or product-related, and applies a probabilistic topic model to response content to assist the user in crafting quality and/or popular response content, according to one embodiment.
  • the process Based on the output of the probabilistic topic model, the process provides recommendations, response quality indicators, and/or response popularity indicators, to assist the user in improving the response content, according to one embodiment.
  • the process proceeds to operation 504, according to one embodiment.
  • the process determines if the submission content is search query content, according to one embodiment. If the submission content is not search query content, the process proceeds to operation 508, and if the submission content is search query content, the process proceeds to operation 509, according to one embodiment.
  • the process request clarification from a user, according to one embodiment.
  • the process returns to operation 501, according to one embodiment.
  • the process applies a probabilistic topic model to search query content and to existing submission content to determine which existing submission content is relevant to the search query content, according to one embodiment.
  • the process proceeds to operation 510, according to one embodiment.
  • the process sorts and displays relevant existing submission content, in response to receipt of the search query content, according to one embodiment.
  • the process proceeds to operation 501, according to one embodiment.
  • FIG. 6 is a flow diagram of a method for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, in accordance with one embodiment.
  • the process includes receiving, with a computing system, submission content from a user through a user interface for the question and answer customer support system, according to one embodiment.
  • the process includes applying a probabilistic topic model to the submission content to determine submission content topics and submission content statistics, according to one embodiment.
  • the process includes generating customer support content at least partially based on the submission content topics and at least partially based on the submission content statistics, to facilitate use of the question and answer customer support system by the user, according to one embodiment.
  • the process includes providing the customer support content to the user through the user interface, in response to receiving the submission content from the user through the user interface, according to one embodiment.
  • a computer-implemented method applies probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system.
  • the method includes receiving, with a computing system, submission content from a user through a user interface for the question and answer customer support system, according to one embodiment.
  • the method includes applying a probabilistic topic model to the submission content to determine submission content topics and submission content statistics, according to one embodiment.
  • the method includes generating customer support content at least partially based on the submission content topics and at least partially based on the submission content statistics, to facilitate use of the question and answer customer support system by the user, according to one embodiment.
  • the method includes providing the customer support content to the user through the user interface, in response to receiving the submission content from the user through the user interface, according to one embodiment.
  • a non-transitory computer-readable medium has instructions which, when executed by one or more computer processors, perform a method for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system.
  • the instructions include a customer support content database configured to maintain existing submission content to support operations for a question and answer customer support system, according to one embodiment.
  • the instructions include an analytics module configured to apply a probabilistic topic model to new submission content to generate new customer support content, according to one
  • the new customer support content is at least partially based on submission content topics and submission content topics statistics that are generated by the probabilistic topic model from the new submission content, according to one embodiment.
  • the instructions include a customer support engine configured to receive new submission content from a user, to update the existing submission content in the customer support content database with the new submission content, to provide the new submission content to the analytics module, and to receive customer support content from the analytics module that is at least partially based on the new submission content provided to the analytics module, according to one embodiment.
  • a system applies probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, according to one embodiment.
  • the system includes at least one processor, and at least one memory coupled to the at least one processor, according to one embodiment.
  • the at least one memory stores instructions which, when executed by any set of the one or more processors, perform a process for applying probabilistic topic models to content in a tax environment, according to one embodiment.
  • the process includes receiving, with a computing system, submission content from a user through a user interface for the question and answer customer support system, according to one embodiment.
  • the process includes applying a probabilistic topic model to the submission content to determine submission content topics and submission content statistics, according to one embodiment.
  • the process includes generating customer support content at least partially based on the submission content topics and at least partially based on the submission content statistics, to facilitate use of the question and answer customer support system by the user, according to one embodiment.
  • the process includes providing the customer support content to the user through the user interface, in response to receiving the submission content from the user through the user interface.
  • Certain aspects of the present invention include process steps or operations and instructions described herein in an algorithmic and/or algorithmic-like form. It should be noted that the process steps and/or operations and instructions of the present invention can be embodied in software, firmware, and/or hardware, and when embodied in software, can be downloaded to reside on and be operated from different platforms used by real time network operating systems.
  • the present invention also relates to an apparatus or system for performing the operations described herein.
  • This apparatus or system may be specifically constructed for the required purposes by a computer program stored via a computer program product as defined herein that can be accessed by a computing system or other device to transform the computing system or other device into a specifically and specially programmed computing system or other device.
  • the present invention is well suited to a wide variety of computer network systems operating over numerous topologies.
  • the configuration and management of large networks comprise storage devices and computers that are
  • a private network a LAN, a WAN, a private network, or a public network, such as the Internet.

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Accounting & Taxation (AREA)
  • Finance (AREA)
  • Development Economics (AREA)
  • Strategic Management (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • General Engineering & Computer Science (AREA)
  • General Business, Economics & Management (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Technology Law (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Computational Mathematics (AREA)
  • Pure & Applied Mathematics (AREA)
  • Algebra (AREA)
  • Probability & Statistics with Applications (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

A method and system applies a probabilistic topic model to content in a tax environment to improve user satisfaction with a question and answer customer support system, according to one embodiment. The probabilistic topic model may be the Latent Dirichlet allocation algorithm or other implementations of probabilistic topic models, according to various embodiments. The method and system include receiving submission content from a user, according to one embodiment. The method and system include applying a probabilistic topic model to the submission content to determine submission content topics and submission content statistics, according to one embodiment. The method and system include generating and providing customer support content at least partially based on the submission content topics and at least partially based on the submission content statistics, to facilitate use of the question and answer customer support system by the user, according to one embodiment.

Description

METHOD AND SYSTEM FOR APPLYING PROBABILISTIC TOPIC MODELS TO CONTENT IN A TAX ENVIRONMENT TO IMPROVE USER SATISFACTION WITH A QUESTION AND ANSWER CUSTOMER SUPPORT SYSTEM
Igor A. Podgorny
Benjamin John Koonse
BACKGROUND
[ 0001 ] Software applications and systems have become indispensable tools for helping consumers, i.e., users, perform a wide variety of tasks in their daily professional and personal lives. Currently, numerous types of desktop, web-based, and cloud-based software systems are available to help users perform a plethora of tasks ranging from basic computing system operations and word processing, to financial management, small business management, tax preparation, health tracking and healthcare management, as well as other personal and business endeavors, operations, and functions far too numerous to individually delineate here.
[ 0002 ] One major, if not determinative, factor in the utility, and ultimate commercial success, of a given software system of any type is the ability to implement and provide a customer support system through which a given user can obtain assistance and, in particular, get answers to questions that arise during the installation and operation of the software system. However, providing potentially millions of software system users with specialized advice and answers to their specific questions is a huge undertaking that can easily, and rapidly, become economically infeasible.
[ 0003 ] To address this problem, many providers of software systems implement or sponsor one or more question and answer based customer support systems. Typically, a question and answer based customer support system includes a hosted forum through which a user can direct their specific questions, typically in a text format, to a support community that often includes other users and/or professional support personal.
[ 0004 ] In many cases, once a user's specific question is answered by one or more members of the support community through the question and answer based customer support system, the user's specific question, and the answer to the specific question provided by the support community, is categorized and added to a customer support question and answer database associated with the question and answer based customer support system. In this way, subsequent users of the software system can access the user's specific question or topic, and find the answer to the user's question, via a search of the customer support question and answer database. As a result, a dynamic customer support question and answer database of
categorized/indexed user questions and answers is made available to users of the software system through the question and answer based customer support system.
[ 0005 ] The development of customer support question and answer databases has numerous advantages including a self-help element whereby a searching user, i.e., a user accessing the resulting question and answer pair, can find an answer to their particular question by simply searching the customer support question and answer database for topics, questions, and answers related to their issue. In addition, if the answer to the user's specific question is not in the customer support question and answer database, the user can then become an asking user by submitting their question to the question and answer based customer support system, typically through the same web-site and/or user interface. Consequently, by using a question and answer based customer support system that includes a customer support question and answer database, potentially millions of user questions can be answered in an efficient and effective manner, and with minimal duplicative effort.
[ 0006 ] The content that users are exposed to, within the question and answer based customer support system, may affect the reputation of the service provider of the question and answer based customer support system. If users' queries consistently are directed to answers and/or questions that are unsatisfactory, the users will communicate dissatisfaction by, for example, using competitor question and answer systems, providing negative reviews in forums, and avoiding other products that are available from the service provider. Furthermore, the more dissatisfied users are with answers found in the question and answer based customer support system, the more likely the users are to request live customer support. This is financially undesirable for a service provider because providing live customer support, such as telephone call and web chats, is significantly more expensive than maintaining a database of answers provided by a support community (e.g., provided at least partially by volunteers). [ 0007 ] What is needed is a method and system for an automated content categorization system in a tax environment to improve user satisfaction with a question and answer customer support system.
SUMMARY
[ 0008 ] Embodiments of the present disclosure address some of the shortcomings associated with traditional question and answer based customer support systems for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, according to one embodiment. By improving user satisfaction with the question and answer customer support system, the production environment assists the service provider in achieving business objectives such as, but not limited to, converting potential customers into paying customers of other services; reducing costs associated with user requests for live customer assistance; and attracting/directing/introducing new potential customers to products offered by the service provider, according to one embodiment.
[ 0009] The question and answer customer support system uses the probabilistic topic models to analyze various types of submission content, which can originate from different types of users, in order to generate various types of customer support content to facilitate and/or improve the user experience in the question and answer customer support system, according to one embodiment. Probabilistic topic models extract hidden topics or summaries from content objects (e.g., database entries, webpages, in documents), without requiring the training of the model with known (e.g., manually verified) data sets, according to one embodiment. The submission content can include question content (e.g., a question), response content (e.g., a comment or an answer to a question), and search query content (e.g., from a searching user), according to one embodiment. The question and answer customer support system applies a probabilistic topic model to the submission content to generate customer support content such as, but not limited to, recommendations for improving question content, recommendations for improving response content, question quality indicators, question popularity indicators, answer quality indicators, answer popularity indicators, categorization of question content as product- related or tax -related, topically categorized navigation interfaces, and topically categorized search results, according to various embodiments. By applying probabilistic topic models to content in the tax environment, the question and answer customer support system facilitates/enables: asking users to submit high-quality questions that result in more satisfying responses; asking users to submit popular questions, which can increase the likelihood that searching users will be directed to high-quality content; responding users to receive question content that is related to the responding users' particular areas of expertise (e.g., product-related topics or tax-related topics); and searching users to receive topically categorized and/or relevance-sorted search results in response to submission of search query content, according to various embodiments. In one embodiment, the probabilistic topic model that is applied to content in the tax environment is a Latent Dirichlet allocation ("LDA") algorithm or another version of a probabilistic topic model.
[0010 ] These and other embodiments are disclosed in more detail below.
[0011 ] The disclosed method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, provides for significant improvements to the technical fields of customer support, information dissemination, software implementation, and user experience. In addition, using the disclosed method and system for applying probabilistic topic models to content in a tax environment results in more efficient use of human and non-human resources, fewer processor cycles being utilized, reduced memory utilization, and less communications bandwidth being utilized to relay data to and from backend systems because users are less likely to request live customer support and because improving question quality and popularity results in users being directed to questions that are likely to result in user satisfaction and result in fewer additional question submissions. As a result, computing systems are transformed into faster, more efficient, and more effective computing systems by implementing the method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system.
BRIEF DESCRIPTION OF THE DRAWINGS
[0012 ] FIG. 1 is a block diagram representing one example of a hardware system and production environment for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system in accordance with one embodiment; [0013] FIGs. 2A, 2B, and 2C are illustrative example graphs showing relationships between question quality and popularity for various topics that are determined by applying a probabilistic topic model to a sample data set in accordance with one embodiment;
[0014] FIGs. 3A, 3B, and 3C are illustrative example tables and a graph showing relationships between topic content, question popularity, and question quality, determined by applying a probabilistic topic model to a sample data set in accordance with one embodiment;
[0015] FIGs. 4A and 4B are illustrate example user experience displays having topics and search results that are selected and sorted based on the application of a probabilistic topic model to a sample data set in accordance with one embodiment;
[0016] FIG. 5 is a flow diagram of a method for applying probabilistic topic models to content in a tax environment in accordance with one embodiment; and
[0017 ] FIG. 6 is a flow diagram of a method for applying probabilistic topic models to content in a tax environment in accordance with one embodiment.
[0018 ] Common reference numerals are used throughout the FIG.s and the detailed description to indicate like elements. One skilled in the art will readily recognize that the above FIG.s are examples and that other architectures, modes of operation, orders of operation, and elements/functions can be provided and implemented without departing from the characteristics and features of the invention, as set forth in the claims.
TERM DEFINITIONS
[0019] Herein, a software system can be, but is not limited to, any data management system implemented on a computing system, accessed through one or more servers, accessed through a network, accessed through a cloud, and/or provided through any system or by any means, as discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing, that gathers/obtains data, from one or more sources and/or has the capability to analyze at least part of the data.
[0020] As used herein, the term software system includes, but is not limited to the following: computing system implemented, and/or online, and/or web-based, personal and/or business tax preparation systems; computing system implemented, and/or online, and/or web- based, personal and/or business financial management systems, services, packages, programs, modules, or applications; computing system implemented, and/or online, and/or web-based, personal and/or business management systems, services, packages, programs, modules, or applications; computing system implemented, and/or online, and/or web-based, personal and/or business accounting and/or invoicing systems, services, packages, programs, modules, or applications; and various other personal and/or business electronic data management systems, services, packages, programs, modules, or applications, whether known at the time of filling or as developed later.
[ 0021 ] Specific examples of software systems include, but are not limited to the following: TurboTax™ available from Intuit, Inc. of Mountain View, California; TurboTax Online™ available from Intuit, Inc. of Mountain View, California; Quicken™, available from Intuit, Inc. of Mountain View, California; Quicken Online™, available from Intuit, Inc. of Mountain View, California; QuickBooks™, available from Intuit, Inc. of Mountain View, California; QuickBooks Online™, available from Intuit, Inc. of Mountain View, California; Mint™, available from Intuit, Inc. of Mountain View, California; Mint Online™, available from Intuit, Inc. of Mountain View, California; and/or various other software systems discussed herein, and/or known to those of skill in the art at the time of filing, and/or as developed after the time of filing.
[ 0022 ] As used herein, the terms "computing system," "computing device," and
"computing entity," include, but are not limited to, the following: a server computing system; a workstation; a desktop computing system; a mobile computing system, including, but not limited to, smart phones, portable devices, and/or devices worn or carried by a user; a database system or storage cluster; a virtual asset; a switching system; a router; any hardware system; any communications system; any form of proxy system; a gateway system; a firewall system; a load balancing system; or any device, subsystem, or mechanism that includes components that can execute all, or part, of any one of the processes and/or operations as described herein.
[ 0023 ] In addition, as used herein, the terms "computing system" and "computing entity," can denote, but are not limited to the following: systems made up of multiple virtual assets, server computing systems, workstations, desktop computing systems, mobile computing systems, database systems or storage clusters, switching systems, routers, hardware systems, communications systems, proxy systems, gateway systems, firewall systems, load balancing systems, or any devices that can be used to perform the processes and/or operations as described herein.
[ 0024 ] Herein, the term "production environment" includes the various components, or assets, used to deploy, implement, access, and use, a given software system as that software system is intended to be used. In various embodiments, production environments include multiple computing systems and/or assets that are combined, communicatively coupled, virtually and/or physically connected, and/or associated with one another, to provide the production environment implementing the application.
[ 0025 ] As specific illustrative examples, the assets making up a given production environment can include, but are not limited to, the following: one or more computing environments used to implement at least part of the software system in the production environment such as a data center, a cloud computing environment, a dedicated hosting environment, and/or one or more other computing environments in which one or more assets used by the application in the production environment are implemented; one or more computing systems or computing entities used to implement at least part of the software system in the production environment; one or more virtual assets used to implement at least part of the software system in the production environment; one or more supervisory or control systems, such as hypervisors, or other monitoring and management systems used to monitor and control assets and/or components of the production environment; one or more communications channels for sending and receiving data used to implement at least part of the software system in the production environment; one or more access control systems for limiting access to various components of the production environment, such as firewalls and gateways; one or more traffic and/or routing systems used to direct, control, and/or buffer data traffic to components of the production environment, such as routers and switches; one or more communications endpoint proxy systems used to buffer, process, and/or direct data traffic, such as load balancers or buffers; one or more secure communication protocols and/or endpoints used to encrypt/decrypt data, such as Secure Sockets Layer (SSL) protocols, used to implement at least part of the software system in the production environment; one or more databases used to store data in the production environment; one or more internal or external services used to implement at least part of the software system in the production environment; one or more backend systems, such as backend servers or other hardware used to process data and implement at least part of the software system in the production environment; one or more software modules/functions used to implement at least part of the software system in the production environment; and/or any other assets/components making up an actual production environment in which at least part of the software system is deployed, implemented, accessed, and run, e.g., operated, as discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing.
[ 0026 ] As used herein, the term "computing environment" includes, but is not limited to, a logical or physical grouping of connected or networked computing systems and/or virtual assets using the same infrastructure and systems such as, but not limited to, hardware systems, software systems, and networking/communications systems. Typically, computing environments are either known, "trusted" environments or unknown, "untrusted" environments. Typically, trusted computing environments are those where the assets, infrastructure, communication and networking systems, and security systems associated with the computing systems and/or virtual assets making up the trusted computing environment, are either under the control of, or known to, a party.
[ 0027 ] In various embodiments, each computing environment includes allocated assets and virtual assets associated with, and controlled or used to create, and/or deploy, and/or operate at least part of the software system.
[ 0028 ] In various embodiments, one or more cloud computing environments are used to create, and/or deploy, and/or operate at least part of the software system that can be any form of cloud computing environment, such as, but not limited to, a public cloud; a private cloud; a virtual private network (VPN); a subnet; a Virtual Private Cloud (VPC); a sub-net or any security/communications grouping; or any other cloud-based infrastructure, sub-structure, or architecture, as discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing.
[ 0029 ] In many cases, a given software system or service may utilize, and interface with, multiple cloud computing environments, such as multiple VPCs, in the course of being created, and/or deployed, and/or operated.
[ 0030 ] As used herein, the term "virtual asset" includes any virtualized entity or resource, and/or virtualized part of an actual, or "bare metal" entity. In various embodiments, the virtual assets can be, but are not limited to, the following: virtual machines, virtual servers, and instances implemented in a cloud computing environment; databases associated with a cloud computing environment, and/or implemented in a cloud computing environment; services associated with, and/or delivered through, a cloud computing environment; communications systems used with, part of, or provided through a cloud computing environment; and/or any other virtualized assets and/or sub-systems of "bare metal" physical devices such as mobile devices, remote sensors, laptops, desktops, point-of-sale devices, etc., located within a data center, within a cloud computing environment, and/or any other physical or logical location, as discussed herein, and/or as known/available in the art at the time of filing, and/or as
developed/made available after the time of filing.
[ 0031 ] In various embodiments, any, or all, of the assets making up a given production environment discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing can be implemented as one or more virtual assets.
[ 0032 ] In one embodiment, two or more assets, such as computing systems and/or virtual assets, and/or two or more computing environments are connected by one or more
communications channels including but not limited to, Secure Sockets Layer (SSL)
communications channels and various other secure communications channels, and/or distributed computing system networks, such as, but not limited to the following: a public cloud; a private cloud; a virtual private network (VPN); a subnet; any general network, communications network, or general network/communications network system; a combination of different network types; a public network; a private network; a satellite network; a cable network; or any other network capable of allowing communication between two or more assets, computing systems, and/or virtual assets, as discussed herein, and/or available or known at the time of filing, and/or as developed after the time of filing.
[ 0033 ] As used herein, the term "network" includes, but is not limited to, any network or network system such as, but not limited to, the following: a peer-to-peer network; a hybrid peer- to-peer network; a Local Area Network (LAN); a Wide Area Network (WAN); a public network, such as the Internet; a private network; a cellular network; any general network, communications network, or general network/communications network system; a wireless network; a wired network; a wireless and wired combination network; a satellite network; a cable network; any combination of different network types; or any other system capable of allowing communication between two or more assets, virtual assets, and/or computing systems, whether available or known at the time of filing or as later developed.
[ 0034 ] As used herein, the term "user experience display" includes not only data entry and question submission user interfaces, but also other user experience features provided or displayed to the user such as, but not limited to the following: data entry fields; question quality indicators; images; backgrounds; avatars; highlighting mechanisms; icons; and any other features that individually, or in combination, create a user experience, as discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing.
[ 0035 ] As used herein, the term "question quality indicator" includes any mechanism, means, or feature/function provided to indicate to a user a determined quality of a question being provided by the user. Specific examples of question quality indicators include, but are not limited to, meter displays; line displays; score displays; audio content; visual content; images; backgrounds; avatars; highlighting mechanisms; icons; and any other features that individually, or in combination, relay to a user a determined quality of a question being submitted.
[ 0036 ] As used herein, the term "question popularity indicator" includes any mechanism, means, or feature/function provided to indicate to a user a determined/estimated popularity of a question being provided by the user. Specific examples of question popularity indicators include, but are not limited to, meter displays; line displays; score displays; audio content; visual content; images; backgrounds; avatars; highlighting mechanisms; icons; and any other features that individually, or in combination, relay to a user a determined, estimated, or predicted quality of a question being submitted.
[ 0037 ] Herein, the term "party," "user," "user consumer," and "customer" are used interchangeably to denote any party and/or entity that interfaces with, and/or to whom
information is provided by, the method and system for determining a level of popularity of submission content, prior to publicizing the submission content with a question and answer support system as described herein, and/or a person and/or entity that interfaces with, and/or to whom information is provided by, the method and system for determining a level of popularity of submission content, prior to publicizing the submission content with a question and answer support system as described herein, and/or a legal guardian of person and/or entity that interfaces with, and/or to whom information is provided by, the method and system for determining a level of popularity of submission content, prior to publicizing the submission content with a question and answer support system as described herein, and/or an authorized agent of any party and/or person and/or entity that interfaces with, and/or to whom information is provided by, the method and system for determining a level of popularity of submission content, prior to publicizing the submission content with a question and answer support system as described herein. For instance, in various embodiments, a user can be, but is not limited to, a person, a commercial entity, an application, a service, and/or a computing system. [0038 ] As used herein, the term "asking user" includes a user of a software system submitting submission content (e.g., a question) to a question and answer based customer support system.
[0039] As used herein, the term "searching user" includes a user of a software system submitting a search query to a customer support question and answer database associated with a question and answer customer support system. An "answer recipient" includes one or both of the "asking user" and the "searching user", according to one embodiment.
[0040] As used herein, the term "responding user" includes a user of a software system who submits a response to submission content for the question and answer based customer support system. In one embodiment, the term "response" is interchangeably used with the term "reply", and the term "responding user" is interchangeably used with the term(s) "replying user" and/or "answering user". In one embodiment, a "responding user", "replying user", and/or "answering user" is a user who submits an answer to submission content (e.g., a question) and/or one who submits a comment to submission content in the question and answer based customer support system.
[0041] As used herein, submission content includes a question content (inclusive of question summary and question details), response content, and search query content and the term "submission content" is used interchangeably with the term "question". A question summary is a character limited summary or articulation of the question, and the question details are additional information about the user or the circumstances surrounding the question summary. As used herein, a "post" is used to refer to a publicized or published version of the submission content, and may include comments and/or answers submitted by users in response to publicizing, publishing, hosting, and/or posting the submission content. Although submission content and a post may include similar information, one references content that has not been made publically available by a question and answer customer support system and the other has been made available for review, response, and comment by the public.
[0042 ] As used herein, the term "probabilistic topic model" or "probabilistic model" denotes one or more individual or combined algorithms or sets of equations that describe, determine, or predict characteristics of or the performance of a datum, a data set, multiple data sets, data objects, a computing system, and/or multiple computing system. The probabilistic topic model include algorithms configured to discover the hidden thematic (or semantic) structure in large data objects, text, and archives of documents. DETAILED DISCLOSURE
[ 0043 ] Embodiments will now be discussed with reference to the accompanying FIG.s, which depict one or more exemplary embodiments. Embodiments may be implemented in many different forms and should not be construed as limited to the embodiments set forth herein, shown in the FIG.s, and/or described below. Rather, these exemplary embodiments are provided to allow a complete disclosure that conveys the principles of the invention, as set forth in the claims, to those of skill in the art.
[ 0044 ] FIG. 1 illustrates a block diagram of a production environment 100 for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, according to one embodiment. By improving user satisfaction with the question and answer customer support system, the production environment assists the service provider in achieving business objectives such as, but not limited to, converting potential customers into paying customers of other services; reducing user requests for live customer assistance; and attracting/directing/introducing new potential customers to products offered by the service provider, according to one embodiment.
Probabilistic topic models extract hidden topics or extract content summaries from content objects (e.g., database entries, webpages, in documents), without requiring the training of the model with known (e.g., manually verified) data sets, according to one embodiment. The hidden topics or content summaries can then be labeled by system administrators or other users, based on the terms returned from the model for the hidden topics or content summaries, according to one embodiment. The production environment 100 uses the probabilistic topic models to analyze various types of submission content, which can originate from different types of users, in order to generate various types of customer support content to facilitate and/or improve the user experience in the question and answer customer support system, according to one embodiment. The submission content can include question content (e.g., a question), response content (e.g., a comment or an answer to a question), and search query content, according to one embodiment. The production environment 100 applies a probabilistic topic model to the submission content to generate a customer support content such as, but not limited to, recommendations for improving question content, recommendations for improving response content, question quality indicators, question popularity indicators, answer quality indicators, answer popularity indicators, categorization of question content as product-related or tax-related, topically categorized navigation interfaces, topically categorized search results, and recommendations for improving a response, according to various embodiments. By applying probabilistic topic models to content in the tax environment, the production environment 100 facilitates/enables: asking users to submit high-quality questions that result in more satisfying responses; asking users to submit popular questions, which can increase the likelihood that searching users will be directed to high-quality content; responding users to receive question content that is related to the responding users' particular areas of expertise (e.g., product-related topics or tax-related topics); and searching users to receive topically categorized and/or relevance-sorted search results in response to submission of search query content, according to various embodiments. In one embodiment, the probabilistic topic model that is applied to content in the tax environment is a Latent Dirichlet allocation ("LDA") algorithm or another version of a probabilistic topic model.
[ 0045 ] The production environment 100 includes a service provider computing environment 110, an asking user computing environment 140, a responding user computing environment 145, and a searching user computing environment 150, according to one embodiment. The service provider computing environment 110 includes a question and answer customer support system 111 that is associated with and/or configured to support a tax return preparation system 112 and/or one or more additional service provider systems 113, according to one embodiment. The question and answer customer support system 111, the tax return preparation system 112, and the one or more additional service provider systems 113 are software systems, according to one embodiment. As noted above, herein, the term software system includes, but is not limited to the following: computing system implemented, and/or online, and/or web-based, personal and/or business tax preparation systems; computing system implemented, and/or online, and/or web-based, personal and/or business financial management systems, services, packages, programs, modules, or applications; computing system
implemented, and/or online, and/or web-based, personal and/or business management systems, services, packages, programs, modules, or applications; computing system implemented, and/or online, and/or web-based, personal and/or business accounting and/or invoicing systems, services, packages, programs, modules, or applications; and various other personal and/or business electronic data management systems, services, packages, programs, modules, or applications, whether known at the time of filling or as developed later. [ 0046 ] Specific examples of software systems of the question and answer customer support system 111, the tax return preparation system 112, and the one or more additional service provider systems 113 include, but are not limited to the following: TurboTax
AnswerXchange™ available from Intuit, Inc. of Mountain View, California; TurboTax™ available from Intuit, Inc. of Mountain View, California; TurboTax Online™ available from Intuit, Inc. of Mountain View, California; Quicken™, available from Intuit, Inc. of Mountain View, California; Quicken Online™, available from Intuit, Inc. of Mountain View, California; QuickBooks™, available from Intuit, Inc. of Mountain View, California; QuickBooks Online™, available from Intuit, Inc. of Mountain View, California; Mint™, available from Intuit, Inc. of Mountain View, California; Mint Online™, available from Intuit, Inc. of Mountain View, California; and/or various other software systems discussed herein, and/or known to those of skill in the art at the time of filing, and/or as developed after the time of filing.
[ 0047 ] In one embodiment the question and answer customer support system 111, e.g., a social question and answer (Q&A) system, is provided to support users of the software system (e.g., the tax return preparation system 112 and/or one or more additional service provider systems 113).
[ 0048 ] The question and answer customer support system 111 includes a customer support engine 114, an analytics module 115, and a customer support content database 116 for applying probabilistic topic models to content in a tax environment to improve/maintain customer satisfaction with the question and answer customer support system 111, according to one embodiment.
[ 0049 ] The customer support engine 114 includes a user interface 117 for providing a user interface display that receives new submission content 118 from a user and that delivers new customer support content 119 to the user, according to one embodiment. The user interface 117 includes, but is not limited to one or more data entry fields; question quality indicators; images; backgrounds; avatars; highlighting mechanisms; icons; boxes; slides; buttons; and any other user interface elements or features that individually, or in combination, create a user experience, as discussed herein, and/or as known in the art at the time of filing, and/or as developed after the time of filing.
[ 0050 ] The customer support engine 114 uses the user interface 117 to receive different types of new submission content 118 from different types of users, according to one
embodiment. The new submission content 118 includes question content 120, response content 121, and search query content 122, according to one embodiment. The question content 120 is received from an asking user through the asking user computing environment 140, the response content 121 is received from a responding user through the responding user computing environment 145, and the search query content 122 is received from a searching user through the searching user computing environment 150, according to one embodiment. In one embodiment, question and answer customer support system 111 is provided to support the tax return preparation system 112 and therefore the question content 120, the response content 121, and the search query content 122 are related to tax -related questions (e.g., federal and state taxation and tax preparation) and/or product-related questions (e.g., the installation and/or operations of the tax return preparation system 112), according to one embodiment. The question and answer customer support system 111 uses the customer support engine 114 and/or the user interface 117 to enable: asking users to submit question content 120 (e.g., questions); responding users to submit response content 121 (e.g., answers to questions); and searching users to submit search query content 122 (e.g., one or more search terms) to find answers to their questions that are already stored/maintained by the question and answer customer support system 111, according to one embodiment.
[ 0051 ] The customer support engine 114 uses the user interface 117 to provide different types of new customer support content 119 to the relevant type of user, e.g., asking user, responding user, searching user, etc., according to one embodiment. The new customer support content 119 is generated by the analytics module 115 and includes, but is not limited to, recommendations, indicators, user interface ("UI") elements, topics, and/or search results that facilitate/improve users' experience/interactions with the question and answer customer support system 111, according to one embodiment. Additional example embodiments of the new customer support content 119 will be described below after a discussion of embodiments of applying a probabilistic topic model to the new submission content 118 with the analytics module 115.
[ 0052 ] The customer support engine 114 updates the customer support content database 116 in response to receiving the new submission content 118 and/or the new customer support content 119, according to one embodiment. The customer support content database 116 includes existing submission content 123 (e.g., question content, response content, and search query content) and existing customer support content 124 (e.g., a history of recommendations and customer service provided to users), according to one embodiment. In other words, the customer support content database 116 stores and maintains one or more tables or other data structures of previously received questions, responses, comments, and search queries received from users of the question and answer customer support system 111, according to one embodiment. The customer support engine 114 updates the existing submission content 123 to reflect the new submission content 118, and the customer support engine 114 updates the existing customer support content 124 to reflect the new customer support content 119, according to one embodiment. The question and answer customer support system 111 publicizes the contents of the customer support content database 116 to enable users to submit question content, submit/review response content associated with the question content, submit search query content to find response content that is relevant to the user's current needs, and otherwise view and/or interact with content hosted by the question and answer customer support system 111, according to one embodiment.
[ 0053 ] The question content 120 submitted to the question and answer customer support system 111 can be related to very different broad categories, be of various question types, have varying predicted answer lengths, and be formatted in various different ways, according to one embodiment. The question content 120 includes a question summary (which provide a character-limited overview or description of the question), question details (which provide a non-character-limited description of the circumstances and/or background and/or context for the question summary, and user click-stream data (e.g., user IP address, web browsing history, geographical location, click speeds, hover durations, hardware identifier(s), and the like), according to one embodiment.
[ 0054 ] The response content 121 is provided by responding users who include paid support personnel in the employ of the service provider and volunteer experts, according to one embodiment. The response content 121 can include comments and answers to questions in the question content 120, according to one embodiment.
[ 0055 ] The search query content 122 includes one or more terms, phrases, or sentences used by searching users to search the question and answer customer support system for answers to product-related and/or substantive questions about the tax return preparation system 112 and/or the one or more additional service provider systems 113, according to one embodiment. The search query content 122 also includes other miscellaneous information about the searching user such as, but not limited to, the browsing history of the user, how the user arrived at the interface for the question and answer customer support system 111 (e.g., the landing page), the relationship of the user with the tax return preparation system 112 (e.g., system authenticated user, guest user, redirected user from search engine) and/or the one or more additional service provider systems 113, the IP address and/or geographic location of the user, and the like.
[ 0056 ] The customer support engine 114 receives the new customer support content 119, in response to providing the new submission content 118 to the analytics module 115 for processing with a probabilistic topic model 125, according to one embodiment. Probabilistic topic models are algorithms that processes text from a database elements, documents, files, and other content objects to determine topics, themes, and/or subjects of text-based content. The topic is a term/phrase for which the text-based content includes words having a high probability of relevance to the term/phrase, according to one embodiment. The probabilistic topic models populate a list, array, or other data structure with the topics of the text-based content, along with statistical information associated with the topics, according to one embodiment. The statistical information associated with the topics include, but are not limited to, quantity of occurrences, distribution of topics in the content, distribution of words in the topic, probability of a word occurring in a topic, probability of a topic occurring in the text-based content, and the like, according to various embodiments. In one embodiment, the probabilistic topic model 125 is the Latent Dirichlet allocation algorithm or another version of a probabilistic topic model.
[ 0057 ] The probabilistic topic model 125 receives the new submission content 118 and model parameters 126, to generate model output 127, according to one embodiment. The model parameters 126 include, but are not limited to, the number of topics for the probabilistic topic model 125 to generate and the number of iterations for the probabilistic topic model 125 to execute while processing the new submission content 118, according to one embodiment. The probabilistic topic model 125 can be configured to process hundreds, thousands, or tens of thousands of data samples, e.g., question and answer pairs with corresponding view and vote data, in a matter of hours, when a similar manual processing of the data samples might take weeks or months of manual human processing, according to one embodiment.
[ 0058 ] The model output 127 includes submission content topics 128 and submission content topic statistics 129, according to one embodiment. The submission content topics 128 are the terms/phrases for which the new submission content 118 includes words having a high probability of relevance to the terms/phrases, according to one embodiment. In one
embodiment, submission content topics 128 are discrete portions of the new submission content 118 that provide quantifiable summaries of the submission content. The submission content topics 128 include the identity of a topic of a word in the new submission content 118, the identity of a topic of all words in the new submission content 118, the identity of a word in the new submission content 118, and/or the identity of all words and all more than one instance of submission content, according to one embodiment. The submission content topic statistics 129 include, but are not limited to, the probability of a word occurring in a topic, the distribution of words in a topic, the probability of a topic occurring in the new submission content 118 or in another text-based content, and/or the distribution of topics in the new submission content 118, according to one embodiment.
[ 0059 ] The analytics module 115 uses the customer support content generator 130 to generate different types of new customer support content 119, based on the model output 127 for the new submission content 118 and based on the content generator database 131, according to one embodiment. The customer support content generator 130 is configured to index, search, and/or rank the submission content topics 128 based at least partially on the submission content topic statistics 129 to determine which content from the content generator database 131 to use to populate the new customer support content 119, according to one embodiment. For example, the customer support content generator 130 may apply one or more thresholds 132 to the model output 127 to determine whether the question content 120 is estimated or predicted to receive enough votes to be deemed "popular". The customer support content generator 130 may display one or more user interface elements 133, such as meters, slides, digital displays, and the like to indicate a level of popularity and/or a level of quality of question content 120, according to one embodiment. The customer support content generator 130 may use topics 134 and/or phrases 135 to populate templates for recommendations, suggestions, and/or encouragement for a user to alter the question summary and/or the question details used in question content, according to one embodiment. Similarly, the customer support content generator 130 may use the thresholds 132, the user interface elements 133, the topics 134, and/or the phrases 135, to encourage or recommend that a responding user alter response content 121, according to one embodiment. The customer support content generator 130 may also use the contents of the content generator database 131 to display the submission content topics 128, based on the relevance of the search query content 122 to facilitate navigation of the user experience display for the user interface 117, according to one embodiment.
[ 0060 ] The analytics module 115 uses the customer support content generator 130 to generate new customer support content 119 that provides guidance to an asking user while the user is creating/generating the question content 120, according to one embodiment. In particular, the analytics module 115 can be configured to populate the new customer support content 119 with real-time recommendations for improving the quality and/or the popularity of the question content 120 that the asking user is creating. Because certain topics, words, question lengths, and question types of a question can determine the likelihood of user satisfaction with the answer to the question and can determine the likelihood of popularity (e.g., quantity of views) of a question, the customer support content generator 130 is configured to analyze model output 127 for the question content 120 and populate the new customer support content with one or more of a question quality indicator, a question popularity indicator, and/or recommendations for improving the quality and/or the popularity of the question content 120, according to one embodiment.
[ 0061 ] Briefly turning to FIGs. 2A, 2B, and 2C, FIG. 2A illustrates a question quality graph 200, FIG. 2B illustrates a question popularity graph 220, and FIG. 2C illustrates a question popularity and quality correlation graph 230, which are each generated by applying Latent Dirichlet allocation algorithm (an embodiment of the probabilistic topic model 125) to approximately 62,000 question content samples from 2013 (an embodiment of existing submission content 123). In particular, the Latent Dirichlet allocation algorithm was applied to question subjects and question details of the 62,000 question content to rank 50 topics within the question content samples based on the percentages of up votes received by each of the 50 topics and based on the quantity of posts made for each of the 50 topics. The question quality graph 200 includes an x-axis that identifies each one of 50 topics and a y-axis that identifies the percentage of up vote received for each of the 50 topics. The plot 201 includes the dark circles in the graph. The question popularity graph 220 includes an x-axis that identifies each one of 50 topics and a y-axis that identifies the number of posts, i.e., question submissions, made for each of the 50 topics. The plot 221 illustrates the relationship between the 50 topics and the number of posts made to a question and answer customer support system for each of the 50 topics. Each of the 50 topics (derived from the 62,000 questions) were manually evaluated, and the Latent Dirichlet allocation algorithm proved to be consistent with the findings that were manually determined by people. The question popularity and quality correlation graph 230 includes an x- axis that identifies the percentage of up vote for each of the 50 topics and a y-axis that identifies the views per post (i.e., an indication of popularity) for the 50 topics. As illustrated, there is very little correlation between the topics that users tend to up vote and the topics that receive above average views per post. Accordingly, the probabilistic topic model 125 can be used to predict question quality and question popularity separately, and can be used by the analytics module 115 to provide indicators and recommendations to the asking user to assist the asking user in strengthening the quality and/or the popularity of the question content formulated by the asking user, according to one embodiment. Because there is very little, if any, correlation between topics that receive up votes and topics that receive above average views per post, it may be unlikely for asking users to create both quality question content and popular question content without the assistance of the question and answer customer support system 111. In other words, it is just as likely for a user to create high-quality and very unpopular question content as it is for the user to create high-quality and very popular content. However, using the probabilistic topic model 125, the question and answer customer support system 111 can be configured to guide users to modify their question content so that it is likely to be of high-quality (e.g., receive customer satisfaction) and be popular (e.g., receive average or above average quantity of views), according to one embodiment.
[ 0062 ] Returning to FIG. 1, the analytics module 115 also uses the model output 127 to determine whether the question content 120 contains a predominantly product-related question or a predominantly tax -related question, to properly route the question content 120 to the type of responding user who can effectively generate satisfactory response content associated with the question content 120, according to one embodiment. For example, some questions submitted to the question and answer based customer support system are product-related questions, e.g., questions related to pricing, installation, version choice, etc. for the software systems that often have little or no relation to the subject matter/job of the software system, i.e., the endeavor supported by the software system. On the other hand, some questions submitted to the question and answer based customer support system are subject matter related, or substantive, questions directly related to the subject matter/endeavor of the software system.
[ 0063 ] As an illustrative example, in the case of a tax preparation software system, the questions "What version of the tax preparation software system should I use?" or "How do I install the tax preparation software system?" would be product-related questions while the questions "Can I deduct my computer?" or "What is my adjusted gross income?" would be subject matter related questions.
[ 0064 ] In general, product-related questions are best answered by paid support personnel in the employ of the provider of the software system while subject matter related questions are often best answered by volunteer expert users of the software system. Consequently, in one embodiment, the analytics module 115 uses the probabilistic topic model 125 to identify broad category/subject matter of the questions, e.g., product-related questions and subject matter related questions, to facilitate appropriately routing the questions to support personnel or volunteer expert users of the software system.
[ 0065 ] FIGs. 3A, 3B, and 3C illustrate example tables and a graph that demonstrate the effectiveness of applying the Latent Dirichlet allocation algorithm (i.e., an embodiment of the probabilistic topic model 125) to characterize the popularity and quality of topics based on whether the topic is predominantly a product-related question or predominantly a tax -related question. FIG. 3A includes a content type popularity table 300 illustrating results of applying the Latent Dirichlet allocation algorithm to a data set, and shows that topic 18 received more views than topic 27. The content type popularity table 300 includes columns of terms and frequency of the terms for each of the topics 18 and 27. The terms for topic 18 indicate that topic 18 predominantly includes product-related question content and is almost 7 times more popular than topic 27, which has terms that indicate the topic 27 permanently includes tax- related question content. One conclusion that can be drawn from the content type popularity table 300 is that procedure or product-related question content may be more popular than tax- related question content, according to one embodiment.
[ 0066 ] FIG. 3B includes a content type quality table 320 illustrating results of applying the Latent Dirichlet allocation algorithm to a data set, and the content type quality table 320 shows that topic 32 receives nearly twice the percentage of up votes (82.5%) as topic 6 (46.8%). As used herein, quality refers to customer satisfaction, as reflected in more up votes. The content type quality table 320 includes columns of terms and frequency of terms for each of the topics 32 and 6. The terms for topic 32 indicate that topic 32 predominantly includes tax -related question content and that the terms for topic 6 indicate that topic 6 predominately includes product-related question content. Although the tax -related question content of topic 32 received a significantly higher percentage of up votes than the product-related question content of topic 6, it may be difficult to conclude that the topic 32 was of higher quality than topic 6 because users tend to up vote tax-related content more than product-related content. Users therefore may be confusing quality of content (deserving of an up vote) with having to do their taxes or with the tax return preparation system (which may not have met the user's expectations). [ 0067 ] FIG. 3C includes a question content quality graph 330, which indicates that tax- related question content receives a higher percentage of up votes than product-related question content. The question content quality graph 330 relies on the same 62,000 question content data set that is used to generate the graphs of FIGs. 2A, 2B, and 2C, using output generated by the Latent Dirichlet allocation algorithm. The question content quality graph 330 includes an x-axis that provides a scale of predominantly tax-related question content on the left side of the graph (at "0.0") that extends to predominantly product-related question content on the right side of the graph (at "1.0"). The y-axis indicates a percentage of up votes. The closer a topic is to a tax- related question content, the higher (on average) the up vote percentages were. The capability of the probabilistic model to distinguish tax-related question content from product-related question content enables the question and answer customer support system 111 to route questions to the appropriate responding users so that a question can be adequately and efficiently addressed, according to one embodiment.
[ 0068 ] Returning to FIG. 1, the analytics module 115 uses the customer support content generator 130 to generate new customer support content 119 to provide guidance to a responding user while the user is creating/generating the response content 121 that is associated with the question content 120, according to one embodiment. As described above, the guidance can include one or more user interface elements 133 and/or phrases 135 that include
recommendations and/or suggestions for improving the likelihood for user satisfaction of an answer to a question, and may include question quality indicators and/or question popularity indicators to motivate the question answer to compile high-quality and popular response content
121, according to one embodiment.
[ 0069 ] The analytics module 115 uses the customer support content generator 130 to generate new customer support content 119 to customize the navigation of and search results displayed in the user experience display for the user interface 117, to assist the searching user in efficiently finding existing submission content 123 that is relevant to the search query content
122, according to one embodiment. The analytics module 115 receives the search query content 122 and applies the probabilistic topic model 125 to determine the dominant topics, e.g., highest ranked of the submission content topics 128, of the search query content 122. The analytics module 115 applies the probabilistic topic model 125 to the existing submission content 123 of the customer support content database 116 to determine the topics and frequencies of topics of the existing submission content 123, according to one embodiment. The analytics module 115 is configured to sort the model output 127 for the existing submission content 123 in an order that is based on the dominant topics of the search query content 122, according to one embodiment. In one embodiment, the new customer support content 119 and the submission content topics 128 are ranked by relevance to the search query content 122 in a navigation bar of a webpage or web interface and/or as search results in a webpage or web interface, in response to receipt of the search query content 122 from a searching user, according to one embodiment.
[ 0070 ] FIGs. 4A and 4B illustrate example user experience displays 400 and 420 that have topics and search results that are sorted based on the application of the Latent Dirichlet allocation algorithm to a data set of at least part of a question and answer database (i.e., an example embodiment of the existing submission content 123) and to a searching user's search query (i.e., an example embodiment of the search query content 122), according to one embodiment. The user experience display 400 includes a navigation bar 401 that provides a sorted list of topics that are related to a search query received from a searching user, according to one embodiment. The sorted list of the navigation bar 401 can be sorted/ranked in order based on the popularity of the topics, based on the quality of the products, and/or based on the relevance of the topics to the search query submitted by the searching user. The user experience display 420 includes search results 421 that are generated and/or sorted based on the application of the Latent Dirichlet allocation algorithm to at least part of the question and answer database, and to search query content submitted by the searching user. The search results 421 are sorted based on popularity, quality, and/or relevance to the search query submitted by the searching user, according to various embodiments.
[ 0071 ] Returning to FIG. 1, the question and answer customer support system 111 can use the analytics module 115 to apply the probabilistic topic model 125 to the customer support content database 116 to provide quality control analyses of the customer support content database 116. For example, the analytics module 115 can use the probabilistic topic model 125 to search the existing submission content 123 for redundant entries, for entries that are likely to be low quality, for entries that are likely to the unpopular, and can be configured to remove low- quality, unpopular, and/or redundant entries from the existing submission content 123. By performing quality control analyses of the customer support content database 116, the question and answer customer support system 111 increases the likelihood that searching users will find content that is high-quality, popular, and/or relevant to the search query content 122 submitted by the searching user, according to one embodiment. [ 0072 ] Applying probabilistic topic models to submission content in the question and answer customer support system 111 can provide a number of benefits to the question and answer customer support system 111, according to one embodiment. The probabilistic topic model 125 is scalable and can be applied to customer support systems that are associated with the one or more additional service provider systems 113, i.e., software systems other than the tax return preparation system 112. The probabilistic topic model 125 can be operated
"unsupervised". In other words, the probabilistic topic model 125 can be used without being trained by confirmed or known data sets. This benefit enables providing question and answer customer support services using less processing power, fewer computing cycles, and less computing system bandwidth than traditional techniques for providing question and answer customer support services, according to one embodiment. The probabilistic topic model 125 can be used to determine whether the content type of submission content is predominantly tax- related or predominantly product-related so that the submission content can be routed to the appropriate responding users or responding user group, and to enable the submission content to be correctly identified in the customer support content database 116, to facilitate the delivery of efficient and accurate search results to searching users, according to one embodiment. The probabilistic topic model 125 can be used to improve user interactions with the question and answer customer support system 111 by, assisting a user in improving question content 120, assisting a user to improve response content 121, and customizing search results to the search query content 122, among other benefits, according to one embodiment.
PROCESS
[ 0073 ] FIG. 5 is a flow diagram of a process 500 for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, in accordance with one embodiment.
[ 0074 ] At operation 501, the process receives submission content from a user, according to one embodiment. The user can be an asking user, a responding user, or a searching user, according to various embodiments. The submission content can include question content, response content, or search query content, according to various embodiments. The process proceeds to operation 502, according to one embodiment.
[ 0075 ] At operation 502, the process determines if the submission content is question content, according to one embodiment. If the submission content is question content, the process proceeds to operation 503, and if the submission content is not question content, the process proceeds to operation 505, according to one embodiment.
[0076] At operation 503, the process applies a probabilistic topic model to the question content to identify the question content as tax-related or product-related, and applies the probabilistic topic model to the question content to assist a user in crafting quality and/or popular question content, according to one embodiment. Based on the output of the
probabilistic topic model, the process provides recommendations, question quality indicators, and/or question popularity indicators, to assist the user in improving the question content, according to one embodiment. The process proceeds to operation 504, according to one embodiment.
[0077] At operation 504, the process updates a customer support content database, according to one embodiment. The process returns to operation 501, according to one embodiment.
[0078] At operation 505, the process determines if the submission content is response content, according to one embodiment. If the submission content is response content, the process proceeds to operation 506, and if the submission content is not response content, the process proceeds to operation 507, wherein to one embodiment.
[0079] At operation 506, the process applies probabilistic topic model to response content to identify the response content as tax -related or product-related, and applies a probabilistic topic model to response content to assist the user in crafting quality and/or popular response content, according to one embodiment. Based on the output of the probabilistic topic model, the process provides recommendations, response quality indicators, and/or response popularity indicators, to assist the user in improving the response content, according to one embodiment. The process proceeds to operation 504, according to one embodiment.
[0080 ] At operation 507, the process determines if the submission content is search query content, according to one embodiment. If the submission content is not search query content, the process proceeds to operation 508, and if the submission content is search query content, the process proceeds to operation 509, according to one embodiment.
[0081 ] At operation 508, the process request clarification from a user, according to one embodiment. The process returns to operation 501, according to one embodiment.
[0082 ] At operation 509, the process applies a probabilistic topic model to search query content and to existing submission content to determine which existing submission content is relevant to the search query content, according to one embodiment. The process proceeds to operation 510, according to one embodiment.
[0083] At operation 510, the process sorts and displays relevant existing submission content, in response to receipt of the search query content, according to one embodiment. The process proceeds to operation 501, according to one embodiment.
[0084 ] FIG. 6 is a flow diagram of a method for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, in accordance with one embodiment.
[0085 ] At operation 602, the process begins.
[0086] At operation 604, the process includes receiving, with a computing system, submission content from a user through a user interface for the question and answer customer support system, according to one embodiment.
[0087 ] At operation 606, the process includes applying a probabilistic topic model to the submission content to determine submission content topics and submission content statistics, according to one embodiment.
[0088 ] At operation 608, the process includes generating customer support content at least partially based on the submission content topics and at least partially based on the submission content statistics, to facilitate use of the question and answer customer support system by the user, according to one embodiment.
[0089] At operation 610, the process includes providing the customer support content to the user through the user interface, in response to receiving the submission content from the user through the user interface, according to one embodiment.
[0090] At operation 612, the process ends.
[0091] In accordance with an embodiment, a computer-implemented method applies probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system. The method includes receiving, with a computing system, submission content from a user through a user interface for the question and answer customer support system, according to one embodiment. The method includes applying a probabilistic topic model to the submission content to determine submission content topics and submission content statistics, according to one embodiment. The method includes generating customer support content at least partially based on the submission content topics and at least partially based on the submission content statistics, to facilitate use of the question and answer customer support system by the user, according to one embodiment. The method includes providing the customer support content to the user through the user interface, in response to receiving the submission content from the user through the user interface, according to one embodiment.
[ 0092 ] In accordance with an embodiment, a non-transitory computer-readable medium has instructions which, when executed by one or more computer processors, perform a method for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system. The instructions include a customer support content database configured to maintain existing submission content to support operations for a question and answer customer support system, according to one embodiment. The instructions include an analytics module configured to apply a probabilistic topic model to new submission content to generate new customer support content, according to one
embodiment. The new customer support content is at least partially based on submission content topics and submission content topics statistics that are generated by the probabilistic topic model from the new submission content, according to one embodiment. The instructions include a customer support engine configured to receive new submission content from a user, to update the existing submission content in the customer support content database with the new submission content, to provide the new submission content to the analytics module, and to receive customer support content from the analytics module that is at least partially based on the new submission content provided to the analytics module, according to one embodiment.
[ 0093 ] In accordance with an embodiment, a system applies probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, according to one embodiment. The system includes at least one processor, and at least one memory coupled to the at least one processor, according to one embodiment. The at least one memory stores instructions which, when executed by any set of the one or more processors, perform a process for applying probabilistic topic models to content in a tax environment, according to one embodiment. The process includes receiving, with a computing system, submission content from a user through a user interface for the question and answer customer support system, according to one embodiment. The process includes applying a probabilistic topic model to the submission content to determine submission content topics and submission content statistics, according to one embodiment. The process includes generating customer support content at least partially based on the submission content topics and at least partially based on the submission content statistics, to facilitate use of the question and answer customer support system by the user, according to one embodiment. The process includes providing the customer support content to the user through the user interface, in response to receiving the submission content from the user through the user interface.
[ 0094 ] The present invention has been described in particular detail with respect to specific possible embodiments. Those of skill in the art will appreciate that the invention may be practiced in other embodiments. For example, the nomenclature used for components, capitalization of component designations and terms, the attributes, data structures, or any other programming or structural aspect is not significant, mandatory, or limiting, and the mechanisms that implement the invention or its features can have various different names, formats, and/or protocols. Further, the system and/or functionality of the invention may be implemented via various combinations of software and hardware, as described, or entirely in hardware elements. Also, particular divisions of functionality between the various components described herein, are merely exemplary, and not mandatory or significant. Consequently, functions performed by a single component may, in other embodiments, be performed by multiple components, and functions performed by multiple components may, in other embodiments, be performed by a single component.
[ 0095 ] Some portions of the above description present the features of the present invention in terms of algorithms and symbolic representations of operations, or algorithm-like representations, of operations on information/data. These algorithmic and/or algorithm-like descriptions and representations are the means used by those of skill in the art to most effectively and efficiently convey the substance of their work to others of skill in the art. These operations, while described functionally or logically, are understood to be implemented by computer programs and/or computing systems. Furthermore, it has also proven convenient at times to refer to these arrangements of operations as steps or modules or by functional names, without loss of generality.
[ 0096 ] Unless specifically stated otherwise, as would be apparent from the above discussion, it is appreciated that throughout the above description, discussions utilizing terms such as "accessing," "analyzing," "obtaining," "identifying," "associating," "aggregating," "initiating," "collecting," "creating," "transferring," "storing," "searching," "comparing," "providing," "processing" etc., refer to the action and processes of a computing system or similar electronic device that manipulates and operates on data represented as physical (electronic) quantities within the computing system memories, resisters, caches or other information storage, transmission or display devices.
[ 0097 ] Certain aspects of the present invention include process steps or operations and instructions described herein in an algorithmic and/or algorithmic-like form. It should be noted that the process steps and/or operations and instructions of the present invention can be embodied in software, firmware, and/or hardware, and when embodied in software, can be downloaded to reside on and be operated from different platforms used by real time network operating systems.
[ 0098 ] The present invention also relates to an apparatus or system for performing the operations described herein. This apparatus or system may be specifically constructed for the required purposes by a computer program stored via a computer program product as defined herein that can be accessed by a computing system or other device to transform the computing system or other device into a specifically and specially programmed computing system or other device.
[ 0099] Those of skill in the art will readily recognize that the algorithms and operations presented herein are not inherently related to any particular computing system, computer architecture, computer or industry standard, or any other specific apparatus. It may prove convenient/efficient to construct or transform one or more specialized apparatuses to perform the required operations described herein. The required structure for a variety of these systems will be apparent to those of skill in the art, along with equivalent variations. In addition, the present invention is not described with reference to any particular programming language and it is appreciated that a variety of programming languages may be used to implement the teachings of the present invention as described herein, and any references to a specific language or languages are provided for illustrative purposes only and for enablement of the contemplated best mode of the invention at the time of filing.
[ 0100 ] The present invention is well suited to a wide variety of computer network systems operating over numerous topologies. Within this field, the configuration and management of large networks comprise storage devices and computers that are
communicatively coupled to similar and/or dissimilar computers and storage devices over a private network, a LAN, a WAN, a private network, or a public network, such as the Internet.
[ 0101 ] It should also be noted that the language used in the specification has been principally selected for readability, clarity, and instructional purposes, and may not have been selected to delineate or circumscribe the inventive subject matter. Accordingly, the disclosure of the present invention is intended to be illustrative, but not limiting, of the scope of the invention, which is set forth in the claims below.
[0102 ] In addition, the operations shown in the FIG.s are identified using a particular nomenclature for ease of description and understanding, but other nomenclature is often used in the art to identify equivalent operations.
[0103] In the discussion above, certain aspects of one embodiment include process steps and/or operations and/or instructions described herein for illustrative purposes in a particular order and/or grouping. However, the particular order and/or grouping shown and discussed herein is illustrative only and not limiting. Those of skill in the art will recognize that other orders and/or grouping of the process steps and/or operations and/or instructions are possible and, in some embodiments, one or more of the process steps and/or operations and/or instructions discussed above can be combined and/or deleted. In addition, portions of one or more of the process steps and/or operations and/or instructions can be re-grouped as portions of one or more other of the process steps and/or operations and/or instructions discussed herein. Consequently, the particular order and/or grouping of the process steps and/or operations and/or instructions discussed herein does not limit the scope of the invention as claimed below.
[0104 ] Therefore, numerous variations, whether explicitly provided for by the specification or implied by the specification or not, may be implemented by one of skill in the art in view of this disclosure.

Claims

CLAIMS What is claimed is:
1. A computer-implemented method for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, the method comprising:
receiving, with a computing system, submission content from a user through a user interface for the question and answer customer support system;
applying a probabilistic topic model to the submission content to determine submission content topics and submission content statistics;
generating customer support content at least partially based on the submission content topics and at least partially based on the submission content statistics, to facilitate use of the question and answer customer support system by the user; and providing the customer support content to the user through the user interface, in response to receiving the submission content from the user through the user interface.
2. The method of claim 1, wherein the probabilistic topic model includes a Latent Dirichlet allocation algorithm.
3. The method of claim 1, wherein the submission content includes content selected from a group consisting of:
question content from an asking user,
response content from a responding user, and
query content from a searching user.
4. The method of claim 1, wherein the submission content topics are discrete portions of the submission content that provide quantifiable summaries of the submission content.
5. The method of claim 1, wherein the customer support content includes content selected from a group consisting of: recommendations for modifying question content that is received in the submission content;
recommendations for modifying answer content that is received in the submission
content;
indicators for strength of question content that is received in the submission content; indicators for strength of response content that is received in the submission content; and search results for query content that is received in the submission content.
6. The method of claim 1, wherein applying the probabilistic topic model includes applying the probabilistic topic model to the submission content to determine submission content topics and submission content statistics, without training the probabilistic topic model with existing content from the question and answer customer support system.
7. The method of claim 1, wherein the submission content topics include individual terms that are discrete portions of the submission content, wherein each of the terms are unique to each of other ones of the terms.
8. The method of claim 7, wherein each term is an individual word.
9. The method of claim 1, further comprising:
categorizing the submission content as product-related content or tax-related content, at least partially based on one or more of the submission content topics and the submission content statistics; and
identifying the submission content as product-related content or tax-related content.
10. The method of claim 9, further comprising:
updating a question and answer customer support database to include the submission content that has been identified as product-related content or tax-related content, wherein updating the question and answer customer support database includes associating a product-related content identifier or a tax -related content identifier with the submission content, in the question and answer customer support database.
11. The method of claim 9, further comprising:
if the submission content is identified as product-related content, routing the submission content to a first responding user to generate response content for the submission content; and
if the submission content is identified as tax-related content, routing the submission content to a second responding user to generate the response content for the submission content.
12. The method of claim 11, wherein the first responding user is a customer service representative for the question and answer customer support system.
13. The method of claim 1, wherein the submission content includes query content from a searching user, the method further comprising:
determining, with the probabilistic topic model, whether the query content is more
relevant to product-related content or tax -related content; and
searching product-related content or tax -related content in a question and answer
customer support database for response content that satisfies search criteria of the query content, to increase a likelihood of returning relevant response criteria in response to the query content.
14. The method of claim 1, wherein the submission content includes query content from a searching user, the method further comprising:
determining existing submission content topics of existing submission content, at least partially based on the probabilistic topic model;
determining relevant ones of the existing submission content topics; and
providing weblinks to the relevant ones of the existing submission content topics, to improve relevant navigation of search results for the user.
15. The method of claim 14, wherein the weblinks of the relevant ones of the existing submission content topics are sorted by at least one of: popularity, quality, and relevance to the query content.
16. The method of claim 1, further comprising:
applying the probabilistic topic model to content of a question and answer customer support database to determine submission content topics for existing submission content in the question and answer customer support database; and correcting mis-categorized submission content topics for the existing submission content, at least partially based on model output from the probabilistic topic model, to improve an accuracy of queries to the question and answer customer support database.
17. The method of claim 1, further comprising:
applying the probabilistic topic model to content of a question and answer customer support database to determine submission content topics for existing submission content in the question and answer customer support database; and removing portions of the existing submission content, at least partially based on model output from the probabilistic topic model, to improve user satisfaction with queries to the question and answer customer support database,
wherein the portions of the existing submission content include one or more of low-quality content, low-popularity content, and redundant content.
18. A non-transitory computer-readable medium having instructions which, when executed by one or more computer processors, perform a method for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, the instructions including:
a customer support content database configured to maintain existing submission content to support operations for a question and answer customer support system;
an analytics module configured to apply a probabilistic topic model to new submission content to generate new customer support content,
wherein the new customer support content is at least partially based on
submission content topics and submission content topics statistics that are generated by the probabilistic topic model from the new submission content; and a customer support engine configured to receive new submission content from a user, to update the existing submission content in the customer support content database with the new submission content, to provide the new submission content to the analytics module, and to receive customer support content from the analytics module that is at least partially based on the new submission content provided to the analytics module.
19. The computer-readable medium of claim 18, wherein the probabilistic topic model includes a Latent Dirichlet allocation algorithm.
20. The computer-readable medium of claim 18, wherein the new submission content includes content selected from a group consisting of:
question content from an asking user,
response content from a responding user, and
query content from a searching user.
21. The computer-readable medium of claim 18, wherein the submission content topics are discrete portions of the new submission content that provide quantifiable summaries of the submission content.
22. The computer-readable medium of claim 18, wherein the new customer support content includes content selected from a group consisting of:
recommendations for modifying question content that is received in the submission
content;
recommendations for modifying answer content that is received in the submission
content;
indicators for strength of question content that is received in the submission content; indicators for strength of response content that is received in the submission content; and search results for query content that is received in the submission content.
23. The computer-readable medium of claim 18, wherein the analytics module is configured to categorized the new submission content as product-related content or tax-related content, at least partially based on one or more of the submission content topics and the submission content statistics, and to tag the new submission content as product-related content or tax-related content for identification in the customer support content database.
24. A system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system, the system comprising:
at least one processor; and
at least one memory coupled to the at least one processor, the at least one memory
having stored therein instructions which, when executed by any set of the one or more processors, perform a process for applying probabilistic topic models to content in a tax environment, the process including:
receiving, with a computing system, submission content from a user through a user
interface for the question and answer customer support system;
applying a probabilistic topic model to the submission content to determine submission content topics and submission content statistics;
generating customer support content at least partially based on the submission content topics and at least partially based on the submission content statistics, to facilitate use of the question and answer customer support system by the user; and providing the customer support content to the user through the user interface, in response to receiving the submission content from the user through the user interface.
25. The system of claim 24, wherein the probabilistic topic model includes a Latent Dirichlet allocation algorithm.
26. The system of claim 24, wherein the submission content includes content selected from a group consisting of:
question content from an asking user,
response content from a responding user, and
query content from a searching user.
27. The system of claim 24, wherein the submission content topics are discrete portions of the submission content that provide quantifiable summaries of the submission content.
28. The system of claim 24, wherein the customer support content includes content selected from a group consisting of:
recommendations for modifying question content that is received in the submission
content;
recommendations for modifying answer content that is received in the submission
content;
indicators for strength of question content that is received in the submission content; indicators for strength of response content that is received in the submission content; and search results for query content that is received in the submission content.
29. The system of claim 24, wherein applying the probabilistic topic model includes applying the probabilistic topic model to the submission content to determine submission content topics and submission content statistics, without training the probabilistic topic model with existing content from the question and answer customer support system.
30. The system of claim 24, wherein the submission content topics include individual terms that are discrete portions of the submission content, wherein each of the terms are unique to each of other ones of the terms.
31. The system of claim 30, wherein each term is an individual word.
32. The system of claim 24, wherein the process further comprises:
categorizing the submission content as product-related content or tax-related content, at least partially based on one or more of the submission content topics and the submission content statistics; and
identifying the submission content as product-related content or tax-related content.
33. The system of claim 32, wherein the process further comprises: updating a question and answer customer support database to include the submission content that has been identified as product-related content or tax-related content, wherein updating the question and answer customer support database includes associating a product-related content identifier or a tax -related content identifier with the submission content, in the question and answer customer support database.
34. The system of claim 32, wherein the process further comprises:
if the submission content is identified as product-related content, routing the submission content to a first responding user to generate response content for the submission content; and
if the submission content is identified as tax-related content, routing the submission content to a second responding user to generate the response content for the submission content.
35. The system of claim 34, wherein the first responding user is a customer service representative for the question and answer customer support system.
36. The system of claim 24, wherein the submission content includes query content from a searching user, wherein the process further comprises:
determining, with the probabilistic topic model, whether the query content is more
relevant to product-related content or tax -related content; and
searching product-related content or tax -related content in a question and answer
customer support database for response content that satisfies search criteria of the query content, to increase a likelihood of returning relevant response criteria in response to the query content.
37. The system of claim 24, wherein the submission content includes query content from a searching user, wherein the process further comprises:
determining existing submission content topics of existing submission content, at least partially based on the probabilistic topic model;
determining relevant ones of the existing submission content topics; and providing weblinks to the relevant ones of the existing submission content topics, to improve relevant navigation of search results for the user.
38. The system of claim 37, wherein the weblinks of the relevant ones of the existing submission content topics are sorted by at least one of: popularity, quality, and relevance to the query content.
39. The system of claim 24, wherein the process further comprises:
applying the probabilistic topic model to content of a question and answer customer support database to determine submission content topics for existing submission content in the question and answer customer support database; and correcting mis-categorized submission content topics for the existing submission content, at least partially based on model output from the probabilistic topic model, to improve an accuracy of queries to the question and answer customer support database.
40. The system of claim 24, wherein the process further comprises:
applying the probabilistic topic model to content of a question and answer customer support database to determine submission content topics for existing submission content in the question and answer customer support database; and removing portions of the existing submission content, at least partially based on model output from the probabilistic topic model, to improve user satisfaction with queries to the question and answer customer support database,
wherein the portions of the existing submission content include one or more of low-quality content, low-popularity content, and redundant content.
PCT/US2016/044687 2015-07-31 2016-07-29 Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system WO2017023742A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CA2992563A CA2992563C (en) 2015-07-31 2016-07-29 Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system
AU2016303436A AU2016303436B2 (en) 2015-07-31 2016-07-29 Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system
EP16833601.4A EP3329401A4 (en) 2015-07-31 2016-07-29 Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US14/814,765 2015-07-31
US14/814,765 US10268956B2 (en) 2015-07-31 2015-07-31 Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system

Publications (1)

Publication Number Publication Date
WO2017023742A1 true WO2017023742A1 (en) 2017-02-09

Family

ID=57883560

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2016/044687 WO2017023742A1 (en) 2015-07-31 2016-07-29 Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system

Country Status (5)

Country Link
US (1) US10268956B2 (en)
EP (1) EP3329401A4 (en)
AU (1) AU2016303436B2 (en)
CA (1) CA2992563C (en)
WO (1) WO2017023742A1 (en)

Families Citing this family (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10475043B2 (en) 2015-01-28 2019-11-12 Intuit Inc. Method and system for pro-active detection and correction of low quality questions in a question and answer based customer support system
US10083213B1 (en) 2015-04-27 2018-09-25 Intuit Inc. Method and system for routing a question based on analysis of the question content and predicted user satisfaction with answer content before the answer content is generated
US10755294B1 (en) 2015-04-28 2020-08-25 Intuit Inc. Method and system for increasing use of mobile devices to provide answer content in a question and answer based customer support system
US10134050B1 (en) 2015-04-29 2018-11-20 Intuit Inc. Method and system for facilitating the production of answer content from a mobile device for a question and answer based customer support system
US10447777B1 (en) 2015-06-30 2019-10-15 Intuit Inc. Method and system for providing a dynamically updated expertise and context based peer-to-peer customer support system within a software application
US10147037B1 (en) 2015-07-28 2018-12-04 Intuit Inc. Method and system for determining a level of popularity of submission content, prior to publicizing the submission content with a question and answer support system
US10475044B1 (en) 2015-07-29 2019-11-12 Intuit Inc. Method and system for question prioritization based on analysis of the question content and predicted asker engagement before answer content is generated
US10268956B2 (en) 2015-07-31 2019-04-23 Intuit Inc. Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system
US10803399B1 (en) * 2015-09-10 2020-10-13 EMC IP Holding Company LLC Topic model based clustering of text data with machine learning utilizing interface feedback
US10289740B2 (en) * 2015-09-24 2019-05-14 Searchmetrics Gmbh Computer systems to outline search content and related methods therefor
US10394804B1 (en) 2015-10-08 2019-08-27 Intuit Inc. Method and system for increasing internet traffic to a question and answer customer support system
US10242093B2 (en) 2015-10-29 2019-03-26 Intuit Inc. Method and system for performing a probabilistic topic analysis of search queries for a customer support system
US10902340B2 (en) * 2016-01-25 2021-01-26 International Business Machines Corporation Dynamic templates for incremental addition of specialized content into a cognitive question answer system
US10599699B1 (en) 2016-04-08 2020-03-24 Intuit, Inc. Processing unstructured voice of customer feedback for improving content rankings in customer support systems
US11334800B2 (en) 2016-05-12 2022-05-17 International Business Machines Corporation Altering input search terms
US10162734B1 (en) 2016-07-20 2018-12-25 Intuit Inc. Method and system for crowdsourcing software quality testing and error detection in a tax return preparation system
US10467541B2 (en) 2016-07-27 2019-11-05 Intuit Inc. Method and system for improving content searching in a question and answer customer support system by using a crowd-machine learning hybrid predictive model
US10460398B1 (en) 2016-07-27 2019-10-29 Intuit Inc. Method and system for crowdsourcing the detection of usability issues in a tax return preparation system
US10242002B2 (en) * 2016-08-01 2019-03-26 International Business Machines Corporation Phenomenological semantic distance from latent dirichlet allocations (LDA) classification
US10229184B2 (en) * 2016-08-01 2019-03-12 International Business Machines Corporation Phenomenological semantic distance from latent dirichlet allocations (LDA) classification
US10445332B2 (en) 2016-09-28 2019-10-15 Intuit Inc. Method and system for providing domain-specific incremental search results with a customer self-service system for a financial management system
US10572954B2 (en) 2016-10-14 2020-02-25 Intuit Inc. Method and system for searching for and navigating to user content and other user experience pages in a financial management system with a customer self-service system for the financial management system
US10733677B2 (en) 2016-10-18 2020-08-04 Intuit Inc. Method and system for providing domain-specific and dynamic type ahead suggestions for search query terms with a customer self-service system for a tax return preparation system
US10552843B1 (en) 2016-12-05 2020-02-04 Intuit Inc. Method and system for improving search results by recency boosting customer support content for a customer self-help system associated with one or more financial management systems
US11100171B1 (en) * 2016-12-30 2021-08-24 X Development Llc Personalized decision engine
US10748157B1 (en) 2017-01-12 2020-08-18 Intuit Inc. Method and system for determining levels of search sophistication for users of a customer self-help system to personalize a content search user experience provided to the users and to increase a likelihood of user satisfaction with the search experience
US11087334B1 (en) 2017-04-04 2021-08-10 Intuit Inc. Method and system for identifying potential fraud activity in a tax return preparation system, at least partially based on data entry characteristics of tax return content
CN107403381A (en) * 2017-05-27 2017-11-28 阿里巴巴集团控股有限公司 Collecting method and device and electronic equipment for risk test and appraisal
US20190018692A1 (en) * 2017-07-14 2019-01-17 Intuit Inc. System and method for identifying and providing personalized self-help content with artificial intelligence in a customer self-help system
US10922367B2 (en) 2017-07-14 2021-02-16 Intuit Inc. Method and system for providing real time search preview personalization in data management systems
US20190066248A1 (en) * 2017-08-25 2019-02-28 Intuit Inc. Method and system for identifying potential fraud activity in a tax return preparation system to trigger an identity verification challenge through the tax return preparation system
US11093951B1 (en) 2017-09-25 2021-08-17 Intuit Inc. System and method for responding to search queries using customer self-help systems associated with a plurality of data management systems
CN107808132A (en) * 2017-10-23 2018-03-16 重庆邮电大学 A kind of scene image classification method for merging topic model
US11829866B1 (en) 2017-12-27 2023-11-28 Intuit Inc. System and method for hierarchical deep semi-supervised embeddings for dynamic targeted anomaly detection
US11436642B1 (en) 2018-01-29 2022-09-06 Intuit Inc. Method and system for generating real-time personalized advertisements in data management self-help systems
US11269665B1 (en) 2018-03-28 2022-03-08 Intuit Inc. Method and system for user experience personalization in data management systems using machine learning
US11036774B2 (en) * 2018-10-04 2021-06-15 Robert Bosch Gmbh Knowledge-based question answering system for the DIY domain
US11675787B2 (en) * 2019-03-15 2023-06-13 International Business Machines Corporation Multiple search collections based on relevancy value
CN112215230B (en) * 2020-09-18 2024-06-28 支付宝实验室(新加坡)有限公司 Information auditing method and device and electronic equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090248659A1 (en) * 2008-03-27 2009-10-01 Yahoo! Inc. System and method for maintenance of questions and answers through collaborative and community editing
US20090253112A1 (en) * 2008-04-07 2009-10-08 Microsoft Corporation Recommending questions to users of community qiestion answering
US20120166438A1 (en) * 2010-12-23 2012-06-28 Yahoo! Inc. System and method for recommending queries related to trending topics based on a received query
US20120233191A1 (en) * 2010-11-22 2012-09-13 Salesforce.Com, Inc. Method and system for making content-based recommendations
JP2014112316A (en) 2012-12-05 2014-06-19 Kddi Corp Question answering program using large amounts of comment sentences, server and method
US8892539B2 (en) 2012-11-28 2014-11-18 International Business Machines Corporation Building, reusing and managing authored content for incident management

Family Cites Families (122)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5999908A (en) 1992-08-06 1999-12-07 Abelow; Daniel H. Customer-based product design module
US5519608A (en) 1993-06-24 1996-05-21 Xerox Corporation Method for extracting from a text corpus answers to questions stated in natural language by using linguistic analysis and hypothesis generation
US5471382A (en) 1994-01-10 1995-11-28 Informed Access Systems, Inc. Medical network management system and process
US6601055B1 (en) 1996-12-27 2003-07-29 Linda M. Roberts Explanation generation system for a diagnosis support tool employing an inference system
US6147975A (en) 1999-06-02 2000-11-14 Ac Properties B.V. System, method and article of manufacture of a proactive threhold manager in a hybrid communication system architecture
US6711585B1 (en) 1999-06-15 2004-03-23 Kanisa Inc. System and method for implementing a knowledge management system
US7385716B1 (en) 1999-09-02 2008-06-10 Hewlett-Packard Development Company, L.P. Authoring tool for bayesian network troubleshooters
US6981040B1 (en) 1999-12-28 2005-12-27 Utopy, Inc. Automatic, personalized online information and product services
WO2002021370A1 (en) 2000-09-01 2002-03-14 Intuit, Inc. Automated tax return with universal data import
US20020123983A1 (en) 2000-10-20 2002-09-05 Riley Karen E. Method for implementing service desk capability
DE10107928A1 (en) 2001-02-15 2002-08-29 Unite Software Gmbh Ebmatingen expert system
US7269545B2 (en) 2001-03-30 2007-09-11 Nec Laboratories America, Inc. Method for retrieving answers from an information retrieval system
US7594176B1 (en) 2001-09-05 2009-09-22 Intuit Inc. Automated retrieval, evaluation, and presentation of context-sensitive user support
US7013263B1 (en) 2001-10-25 2006-03-14 Mindfabric, Inc. Online interaction processing
US20030099924A1 (en) 2001-11-28 2003-05-29 Sougo Tsuboi Learning support method and learning support system
US7574427B2 (en) 2002-05-17 2009-08-11 Delefevre Patrick Y Neutral sales consultant
JP3981734B2 (en) 2003-11-21 2007-09-26 独立行政法人情報通信研究機構 Question answering system and question answering processing method
US7899698B2 (en) 2004-03-19 2011-03-01 Accenture Global Services Limited Real-time sales support and learning tool
US20060265232A1 (en) 2005-05-20 2006-11-23 Microsoft Corporation Adaptive customer assistance system for software products
WO2007052285A2 (en) 2005-07-22 2007-05-10 Yogesh Chunilal Rathod Universal knowledge management and desktop search system
CN1794233A (en) 2005-12-28 2006-06-28 刘文印 Network user interactive asking answering method and its system
US7685144B1 (en) 2005-12-29 2010-03-23 Google Inc. Dynamically autocompleting a data entry
WO2007099812A1 (en) 2006-03-01 2007-09-07 Nec Corporation Question answering device, question answering method, and question answering program
US8930282B2 (en) 2006-03-20 2015-01-06 Amazon Technologies, Inc. Content generation revenue sharing
US20070259325A1 (en) 2006-04-22 2007-11-08 Clapper Rock L Internet-based method and system for human-human question and answer sessions in delayed and real time
NO330128B1 (en) 2006-10-25 2011-02-21 Reliant Exams As System and procedures for improving the quality of computer-generated exam papers
US7818222B2 (en) 2006-11-30 2010-10-19 Hrb Innovations, Inc. Method and system for organizing tax information and providing tax advice
US20080215541A1 (en) 2007-03-01 2008-09-04 Microsoft Corporation Techniques for searching web forums
US8200527B1 (en) 2007-04-25 2012-06-12 Convergys Cmg Utah, Inc. Method for prioritizing and presenting recommendations regarding organizaion's customer care capabilities
US8494978B2 (en) 2007-11-02 2013-07-23 Ebay Inc. Inferring user preferences from an internet based social interactive construct
US8713029B2 (en) 2007-11-05 2014-04-29 Verizon Patent And Licensing Inc. Method and apparatus for providing auto-completion of information
US7809664B2 (en) 2007-12-21 2010-10-05 Yahoo! Inc. Automated learning from a question and answering network of humans
US8385812B2 (en) 2008-03-18 2013-02-26 Jones International, Ltd. Assessment-driven cognition system
EP2277105A4 (en) 2008-04-07 2012-09-19 Telecomm Systems Inc Proximity search for point-of-interest names combining inexact string match with an expanding radius search
US7974860B1 (en) 2008-05-09 2011-07-05 ExperienceLab, Inc. Consumer directed health plan (CDHP) and high deductible health plan (HDHP) counseling and information distribution
US8010545B2 (en) 2008-08-28 2011-08-30 Palo Alto Research Center Incorporated System and method for providing a topic-directed search
US8019753B2 (en) 2008-09-11 2011-09-13 Intuit Inc. Method and system for generating a dynamic help document
US8751559B2 (en) 2008-09-16 2014-06-10 Microsoft Corporation Balanced routing of questions to experts
US20100088262A1 (en) 2008-09-29 2010-04-08 Neuric Technologies, Llc Emulated brain
US20100185630A1 (en) 2008-12-30 2010-07-22 Microsoft Corporation Morphing social networks based on user context
US8346701B2 (en) 2009-01-23 2013-01-01 Microsoft Corporation Answer ranking in community question-answering sites
US8341167B1 (en) 2009-01-30 2012-12-25 Intuit Inc. Context based interactive search
EP2394244A4 (en) 2009-02-05 2013-02-20 Bazaarvoice Method and system for providing content generation capabilities
US20100235361A1 (en) 2009-03-12 2010-09-16 International Business Machines Corporation Optimizing Questionnaires
CN101520802A (en) 2009-04-13 2009-09-02 腾讯科技(深圳)有限公司 Question-answer pair quality evaluation method and system
US9460092B2 (en) 2009-06-16 2016-10-04 Rovi Technologies Corporation Media asset recommendation service
US20110005560A1 (en) * 2009-07-07 2011-01-13 Mark Nair Portable solar canopy with modular connections
EP2462517A4 (en) 2009-08-04 2014-08-06 Google Inc Query suggestions from documents
US8291319B2 (en) * 2009-08-28 2012-10-16 International Business Machines Corporation Intelligent self-enabled solution discovery
US20110106895A1 (en) 2009-10-30 2011-05-05 Google Inc. Social search engine
US20110125734A1 (en) 2009-11-23 2011-05-26 International Business Machines Corporation Questions and answers generation
US8943145B1 (en) 2010-02-08 2015-01-27 Intuit Inc. Customer support via social network
US9009134B2 (en) 2010-03-16 2015-04-14 Microsoft Technology Licensing, Llc Named entity recognition in query
US8706580B2 (en) 2010-04-26 2014-04-22 Hrb Tax Group, Inc. Method, system, and computer program for predicting tax liabilities and benefits
US9563656B2 (en) 2010-05-17 2017-02-07 Xerox Corporation Method and system to guide formulations of questions for digital investigation activities
US8732222B2 (en) 2010-06-30 2014-05-20 Microsoft Corporation Integrating specialized knowledge sources into a general search service
US8606739B2 (en) 2010-06-30 2013-12-10 Microsoft Corporation Using computational engines to improve search relevance
US20120022983A1 (en) 2010-07-22 2012-01-26 Hughes Jr Terrell D Web Based Tax Preparation
US20130325992A1 (en) 2010-08-05 2013-12-05 Solariat, Inc. Methods and apparatus for determining outcomes of on-line conversations and similar discourses through analysis of expressions of sentiment during the conversations
US9002773B2 (en) 2010-09-24 2015-04-07 International Business Machines Corporation Decision-support application and system for problem solving using a question-answering system
US8943051B2 (en) 2010-09-24 2015-01-27 International Business Machines Corporation Lexical answer type confidence estimation and application
WO2012044894A2 (en) 2010-10-01 2012-04-05 Hrb Tax Group, Inc. System, computer program, and method for online, real-time delivery of consumer tax services
US8645298B2 (en) 2010-10-26 2014-02-04 Microsoft Corporation Topic models
US20120130910A1 (en) 2010-11-19 2012-05-24 Salesforce.Com, Inc. Customer support flow
US20120136764A1 (en) 2010-11-29 2012-05-31 Intuit Inc. On-line tax preparation application screen navigation using url bookmarks
US8630975B1 (en) 2010-12-06 2014-01-14 The Research Foundation For The State University Of New York Knowledge discovery from citation networks
US20130304730A1 (en) 2011-01-18 2013-11-14 Google Inc. Automated answers to online questions
US20120219142A1 (en) 2011-02-27 2012-08-30 General Motors Llc Call transfer process and system
US8484228B2 (en) * 2011-03-17 2013-07-09 Indian Institute Of Science Extraction and grouping of feature words
US9146917B2 (en) 2011-07-15 2015-09-29 International Business Machines Corporation Validating that a user is human
US9015031B2 (en) 2011-08-04 2015-04-21 International Business Machines Corporation Predicting lexical answer types in open domain question and answering (QA) systems
US8583568B2 (en) 2011-08-26 2013-11-12 SurveyMonkey.com, LLC Systems and methods for detection of satisficing in surveys
US20130066693A1 (en) 2011-09-14 2013-03-14 Microsoft Corporation Crowd-sourced question and answering
US10776431B2 (en) * 2011-10-26 2020-09-15 Oath Inc. System and method for recommending content based on search history and trending topics
US9310879B2 (en) 2011-11-09 2016-04-12 Xerox Corporation Methods and systems for displaying web pages based on a user-specific browser history analysis
US20130297553A1 (en) 2012-05-04 2013-11-07 Pearl.com LLC Method and apparatus for predicting question answerability in an online consultation system
US9679568B1 (en) 2012-06-01 2017-06-13 Google Inc. Training a dialog system using user feedback
US9063983B1 (en) 2012-06-01 2015-06-23 Google Inc. Detecting name-triggering queries
US9128579B2 (en) 2012-06-14 2015-09-08 Intuit Inc. Software localization through user contribution
US20140006012A1 (en) 2012-07-02 2014-01-02 Microsoft Corporation Learning-Based Processing of Natural Language Questions
US9887887B2 (en) 2012-07-09 2018-02-06 Eturi Corp. Information throttle based on compliance with electronic communication rules
US9852239B2 (en) 2012-09-24 2017-12-26 Adobe Systems Incorporated Method and apparatus for prediction of community reaction to a post
US9042540B2 (en) 2012-10-30 2015-05-26 Teletech Holdings, Inc. Method for providing support using answer engine and dialog rules
US9158772B2 (en) 2012-12-17 2015-10-13 International Business Machines Corporation Partial and parallel pipeline processing in a deep question answering system
US9043887B2 (en) 2012-12-31 2015-05-26 Apple Inc. Adaptive secondary authentication criteria based on account data
US20140201045A1 (en) 2013-01-17 2014-07-17 Yogish Pai Determining local tax structures in an accounting application through user contribution
US10387858B2 (en) 2013-02-07 2019-08-20 Jpmorgan Chase Bank, N.A. Integrated electronic cash flow management system and method
EP2965271A4 (en) 2013-03-08 2016-10-26 Inmoment Inc Method and system for conducting a deductive survey
US9886500B2 (en) 2013-03-13 2018-02-06 Open Text Holdings, Inc. System and method for providing technology assisted data review with optimizing features
US9336211B1 (en) 2013-03-13 2016-05-10 Google Inc. Associating an entity with a search query
US20170270159A1 (en) 2013-03-14 2017-09-21 Google Inc. Determining query results in response to natural language queries
US20140308648A1 (en) 2013-04-15 2014-10-16 Rishabh Jain System for question answering
US20140324856A1 (en) 2013-04-27 2014-10-30 Microsoft Corporation Application discoverability
US9471883B2 (en) 2013-05-09 2016-10-18 Moodwire, Inc. Hybrid human machine learning system and method
US9430227B2 (en) 2013-06-13 2016-08-30 Intuit Inc. Automatic customization of a software application
US9342608B2 (en) 2013-08-01 2016-05-17 International Business Machines Corporation Clarification of submitted questions in a question and answer system
US9256826B2 (en) 2013-08-14 2016-02-09 Adobe Systems Incorporated Predicting reactions to short-text posts
US20150095267A1 (en) 2013-09-30 2015-04-02 Avaya Inc. Techniques to dynamically generate real time frequently asked questions from forum data
US9292597B2 (en) 2013-10-24 2016-03-22 International Business Machines Corporation Smart question routing and answerer growing for online community question-answer services
US9635175B2 (en) 2013-11-19 2017-04-25 Avaya Inc. Aggregated multi-topic agent desktop
US10915973B2 (en) 2014-03-05 2021-02-09 Transform Sr Brands Llc System and method providing expert audience targeting
US10475042B2 (en) 2014-05-08 2019-11-12 Avaya Inc. Public non-company controlled social forum response method
US9633309B2 (en) 2014-06-19 2017-04-25 International Business Machines Corporation Displaying quality of question being asked a question answering system
US20160022537A1 (en) * 2014-07-24 2016-01-28 Seung Woo Lee Handling Useful Sauna Equipment
US9582757B1 (en) 2014-07-25 2017-02-28 Amazon Technologies, Inc. Scalable curation system
US10366332B2 (en) 2014-08-14 2019-07-30 International Business Machines Corporation Tailoring question answering system output based on user expertise
US9727637B2 (en) 2014-08-19 2017-08-08 International Business Machines Corporation Retrieving text from a corpus of documents in an information handling system
US20160148321A1 (en) 2014-11-20 2016-05-26 Hrb Innovations, Inc. Simplified screening for predicting errors in tax returns
US20160180470A1 (en) 2014-12-23 2016-06-23 Intuit Inc. Method and system for evaluating interchangeable analytics modules used to provide customized tax return preparation interviews
US9373086B1 (en) 2015-01-07 2016-06-21 International Business Machines Corporation Crowdsource reasoning process to facilitate question answering
US10475043B2 (en) 2015-01-28 2019-11-12 Intuit Inc. Method and system for pro-active detection and correction of low quality questions in a question and answer based customer support system
US9875296B2 (en) 2015-03-25 2018-01-23 Google Llc Information extraction from question and answer websites
US9910886B2 (en) 2015-04-17 2018-03-06 International Business Machines Corporation Visual representation of question quality
US10025799B2 (en) 2015-07-22 2018-07-17 International Business Machines Corporation Access and presentation of files based on semantic proximity to current interests
US20170024424A1 (en) 2015-07-26 2017-01-26 Saad Almohizea Suggestive search engine
US10402913B2 (en) 2015-07-30 2019-09-03 Intuit Inc. Generation of personalized and hybrid responses to queries submitted from within tax return preparation system during preparation of electronic tax return
US10268956B2 (en) 2015-07-31 2019-04-23 Intuit Inc. Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system
US10242093B2 (en) 2015-10-29 2019-03-26 Intuit Inc. Method and system for performing a probabilistic topic analysis of search queries for a customer support system
US10467541B2 (en) 2016-07-27 2019-11-05 Intuit Inc. Method and system for improving content searching in a question and answer customer support system by using a crowd-machine learning hybrid predictive model
US10445332B2 (en) 2016-09-28 2019-10-15 Intuit Inc. Method and system for providing domain-specific incremental search results with a customer self-service system for a financial management system
US10572954B2 (en) 2016-10-14 2020-02-25 Intuit Inc. Method and system for searching for and navigating to user content and other user experience pages in a financial management system with a customer self-service system for the financial management system
US10733677B2 (en) 2016-10-18 2020-08-04 Intuit Inc. Method and system for providing domain-specific and dynamic type ahead suggestions for search query terms with a customer self-service system for a tax return preparation system

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090248659A1 (en) * 2008-03-27 2009-10-01 Yahoo! Inc. System and method for maintenance of questions and answers through collaborative and community editing
US20090253112A1 (en) * 2008-04-07 2009-10-08 Microsoft Corporation Recommending questions to users of community qiestion answering
US20120233191A1 (en) * 2010-11-22 2012-09-13 Salesforce.Com, Inc. Method and system for making content-based recommendations
US20120166438A1 (en) * 2010-12-23 2012-06-28 Yahoo! Inc. System and method for recommending queries related to trending topics based on a received query
US8892539B2 (en) 2012-11-28 2014-11-18 International Business Machines Corporation Building, reusing and managing authored content for incident management
JP2014112316A (en) 2012-12-05 2014-06-19 Kddi Corp Question answering program using large amounts of comment sentences, server and method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
See also references of EP3329401A4
ZEINAB ZOLAKTAF ZADEH: "PROBABILISTIC MODELING IN COMMUNITY-BASED QUESTION ANSWERING SERVICES", MASTER DALHOUSIE UNIVERSITY, February 2012 (2012-02-01), XP055362041, Retrieved from the Internet <URL:http://www.cs.ubc.ca/-zolaktaf/ZolaktafZadeh,Zeinab,MCS, CS ,APRIL2012.pdf> *

Also Published As

Publication number Publication date
CA2992563A1 (en) 2017-02-09
US10268956B2 (en) 2019-04-23
EP3329401A1 (en) 2018-06-06
AU2016303436A1 (en) 2018-02-08
CA2992563C (en) 2022-07-19
US20170032251A1 (en) 2017-02-02
EP3329401A4 (en) 2018-12-12
AU2016303436B2 (en) 2020-07-23

Similar Documents

Publication Publication Date Title
AU2016303436B2 (en) Method and system for applying probabilistic topic models to content in a tax environment to improve user satisfaction with a question and answer customer support system
US10467541B2 (en) Method and system for improving content searching in a question and answer customer support system by using a crowd-machine learning hybrid predictive model
US10242093B2 (en) Method and system for performing a probabilistic topic analysis of search queries for a customer support system
US10162734B1 (en) Method and system for crowdsourcing software quality testing and error detection in a tax return preparation system
US10134050B1 (en) Method and system for facilitating the production of answer content from a mobile device for a question and answer based customer support system
US10147037B1 (en) Method and system for determining a level of popularity of submission content, prior to publicizing the submission content with a question and answer support system
US10861023B2 (en) Method and system for question prioritization based on analysis of the question content and predicted asker engagement before answer content is generated
US10475043B2 (en) Method and system for pro-active detection and correction of low quality questions in a question and answer based customer support system
US10552843B1 (en) Method and system for improving search results by recency boosting customer support content for a customer self-help system associated with one or more financial management systems
US11429988B2 (en) Method and system for increasing use of mobile devices to provide answer content in a question and answer based customer support system
US20190018692A1 (en) System and method for identifying and providing personalized self-help content with artificial intelligence in a customer self-help system
US10748157B1 (en) Method and system for determining levels of search sophistication for users of a customer self-help system to personalize a content search user experience provided to the users and to increase a likelihood of user satisfaction with the search experience
US10460398B1 (en) Method and system for crowdsourcing the detection of usability issues in a tax return preparation system
US20170178199A1 (en) Method and system for adaptively providing personalized marketing experiences to potential customers and users of a tax return preparation system
AU2017249007A1 (en) Method and system for updating analytics models that are used to dynamically and adaptively provide personalized user experiences in a software system
US20130117204A1 (en) Inferring procedural knowledge from data sources
US11436642B1 (en) Method and system for generating real-time personalized advertisements in data management self-help systems
US20190057404A1 (en) Jobs forecasting
US10387526B2 (en) Content analysis and scoring system and method
US11410203B1 (en) Optimized management of online advertising auctions
CN109075987B (en) Optimizing digital component analysis systems
US10394804B1 (en) Method and system for increasing internet traffic to a question and answer customer support system
US20140289217A1 (en) Aggregate Crowdsourcing Systems and Methods
US11580322B2 (en) Scalable attributed graph embedding for large-scale graph analytics

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16833601

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2992563

Country of ref document: CA

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2016303436

Country of ref document: AU

Date of ref document: 20160729

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 2016833601

Country of ref document: EP