US20200081907A1 - Updating natural language interfaces by processing usage data - Google Patents

Updating natural language interfaces by processing usage data Download PDF

Info

Publication number
US20200081907A1
US20200081907A1 US16/680,906 US201916680906A US2020081907A1 US 20200081907 A1 US20200081907 A1 US 20200081907A1 US 201916680906 A US201916680906 A US 201916680906A US 2020081907 A1 US2020081907 A1 US 2020081907A1
Authority
US
United States
Prior art keywords
natural language
intent
language interface
messages
intents
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/680,906
Inventor
Satchuthananthavale Rasiah Kuhan Branavan
Joseph Ellsworth Hackman
Frederick William Poe Heckel
Aaron Isaksen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
ASAPP Inc
Original Assignee
ASAPP Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ASAPP Inc filed Critical ASAPP Inc
Priority to US16/680,906 priority Critical patent/US20200081907A1/en
Assigned to ASAPP, INC. reassignment ASAPP, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BRANAVAN, SATCHUTHANANTHAVALE RASIAH KUHAN, HACKMAN, JOSEPH ELLSWORTH, HECKEL, FREDERICK WILLIAM POE, ISAKSEN, AARON
Publication of US20200081907A1 publication Critical patent/US20200081907A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F17/2785
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/10Machine learning using kernel methods, e.g. support vector machines [SVM]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • G06N3/0445
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • G06N3/0454
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/02Knowledge representation; Symbolic representation
    • G06N5/022Knowledge engineering; Knowledge acquisition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models
    • G06N5/041Abduction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/01Customer relationship services
    • G06Q30/015Providing customer assistance, e.g. assisting a customer within a business location or via helpdesk
    • G06Q30/016After-sales

Definitions

  • the present invention relates to using semantic processing to improve customer support.
  • Companies need to efficiently interact with customers to provide services to their customers. For example, customers may need to obtain information about services of the company, may have a question about billing, or may need technical support from the company. Companies interact with customers in a variety of different ways. Companies may have a website and the customer may navigate the website to perform various actions. Companies may have an application (“app”) that runs on a user device, such as a smart phone or a tablet, that provides similar services as a website. Companies may have a phone number that customers can call to obtain information via interactive voice response or to speak with a customer service representative. Companies may also respond to customers using various social media services, such as Facebook or Twitter.
  • Some existing techniques for allowing customers to interact with companies may be a nuisance to the customer. Navigating to the right page on a website or an app or navigating a voice menu on a phone call may be time consuming. Some existing techniques for allowing customers to interact with companies may be expensive for a company to implement. Hiring customer service representatives to manually respond to requests and answer phone calls may be a significant expense.
  • FIG. 1A is an example user interface for a natural language interface
  • FIG. 1B is a flowchart of an example implementation of the natural language interface.
  • FIG. 2A is an example user interface for a natural language interface
  • FIG. 2B is a flowchart of an example implementation of the natural language interface.
  • FIG. 3 is an example system for a third-party company to provide services to multiple companies.
  • FIG. 4 is a flowchart of an example implementation of configuring and improving a natural language interface.
  • FIG. 5 is a flowchart of an example implementation of configuring a natural language interface.
  • FIG. 6 is an example list of intents and hierarchical information relating to the list of intents.
  • FIG. 7 is a flowchart of an example implementation of a natural language interface with a customer.
  • FIGS. 8A and 8B are examples of graphs and reduced graphs that may be used with a natural language interface.
  • FIG. 9A is an example of usage data corresponding to an intent
  • FIG. 9B is an example of clustered usage data for an intent.
  • FIG. 10 is an example clustering of usage data for an intent.
  • FIG. 11 is a flowchart of an example implementation of improving a natural language interface by processing usage data.
  • FIG. 12 is an example system that a third-party company may use to assist companies in configuring and improving a natural language interface.
  • FIG. 13 is an exemplary computing device that may be used to configure and improve a natural language interface.
  • Described herein are techniques for implementing a natural language interface to respond to a request of a user.
  • the techniques described herein may be used for a wide variety of users and requests, for clarity of presentation, an example of a company providing a response to a request of a customer will be used.
  • the techniques described herein are not limited to customers and companies, responses may be provided to requests from users who are not customers, and responses may be from any entity or person.
  • the techniques described herein may use or be combined with any of the techniques described in U.S. Pat. No. 9,715,496 or U.S. patent application Ser. No. 15/789,241, each of which is incorporated herein by reference in their entireties for all purposes (“incorporated patents”).
  • a natural language interface may be used to provide a fully automated experience for a customer.
  • a natural language interface may use semantic processing to allow a customer to obtain needed information or needed support by making a request using natural language.
  • the use of natural language for making requests may allow the customer to receive a resolution to his request more quickly than with other methods of making requests.
  • the customer's request may be resolved with a single interaction. For example, where the customer asks, “When is my package arriving?,” the response may be, “Your package is scheduled to arrive on Wednesday and has tracking number 1234 .”
  • Implementing, diagnosing, and improving a natural language interface may be challenging for a company.
  • Customers may express a concept using a wide variety of different expressions. For example, a customer wanting to change his address may state “I'd like to change my address” or “I've moved.” Additionally, a company may want to handle many different types of requests, such as address changes, bill requests, refunds, requests for new services, technical support, and any other issue that may normally be handled by a human customer support representative. Between a large number of different types of requests, and a large number of ways of expressing each request, it may be challenging to implement a natural language interface that works properly for most ways of expressing each type of request.
  • a company may desire an accurate natural language interface because it may lower costs for the company and also provide a better experience for customers of the company.
  • the techniques described herein improve the ability of a company to initially create a natural language interface and to later improve the natural language interface based on usage data of customers using the natural language interface.
  • the techniques described herein allow a company to identify portions of the natural language interface that are operating well (e.g. with low error rates) and portions of the natural language interface that could be improved.
  • the usage data of customers using the natural language interface may include a large amount of data that may be challenging for a computer or a human to review to understand why portions of the natural language interface are working well or not.
  • the techniques described herein include clustering the usage data into clusters of similar usage data (e.g., different ways of expressing the same request) and selecting prototype messages for each cluster. By using information about clusters of usage data and their prototype messages, the usage data may be processed more efficiently to improve the operation of the natural language interface, as described in greater detail below.
  • FIG. 1A illustrates an example automated customer support session on a customer device
  • FIG. 1B is a flowchart of an example implementation of providing automated customer support.
  • the ordering of the steps is exemplary and other orders are possible, not all steps are required, steps may be combined (in whole or part) or sub-divided and, in some implementations, some steps may be omitted or other steps may be added.
  • the methods described by any flowcharts described herein may be implemented, for example, by any of the computers or systems described herein.
  • the customer has used customer device 100 to access a user interface for obtaining support from a company.
  • Customer device 100 may be any appropriate device, such as a smart phone, tablet, wearable device, or Internet of things device.
  • the customer may request support using any appropriate techniques, such as typing or speaking a message to an app running on the customer device (e.g., an app of the company or a third-party app created for processing customer requests), typing or speaking a message on a web page, sending a text message, or sending an email.
  • a text message includes any message sent as text including but not limited to a message sent using SMS (short message service) or a special-purpose application (e.g., Facebook Messenger, Apple iMessage, Google Hangouts, or WhatsApp).
  • SMS short message service
  • a special-purpose application e.g., Facebook Messenger, Apple iMessage, Google Hangouts, or WhatsApp.
  • My address is incorrect.”
  • User interfaces described throughout the present disclosure may be of any type, including at least a mobile application, a web-based interface, a texting interface (e.g., receipt, responses, and/or prompts via text), a dedicated program operating on a computer, and/or any hardware device, computer executable media, and/or display accessible to customers, customer service representatives, and/or company agents or representatives.
  • User interfaces include any display devices configured to depict elements of the user interface, and/or input/output devices configured to receive responses from users of any aspect of embodiments described herein.
  • user interfaces include, in certain embodiments, one or more of network communications, graphical displays, mobile devices, keyboards, mouses, and/or touch screens.
  • a first user interface may be provided to a user for training a natural language interface
  • a second user interface may be provided to a user (e.g., at a first company) for data collection or display about a natural language interface (e.g., displaying cluster information)
  • a third user interface may be provided to a user (e.g., at a second company) for training and/or data collection or display about a second natural language interface
  • a fourth user interface may be provided to a customer for obtaining support from a company (e.g., the first company, the second company, or another company).
  • Any given interface infrastructure e.g., computer executable media operating on a server, and/or network communications
  • Any one or more of the example user interfaces may be present or not present in given embodiments.
  • a company receives the message of the customer.
  • the company processes the message to determine an intent of the message. For example, the company may process the message with an intent classifier that provides a score for each intent of a list of available intents, and an intent having a highest score may be selected as the intent of message 110 .
  • the company obtains an action corresponding to the determined intent of message 110 .
  • An action may correspond to any sequence of interactions defined by the company.
  • an action may correspond to launching an application with the user (e.g., to change an address), providing a message to the user with requested information, directing the user to a resource (e.g., a web page), or connecting the user with a customer service representative.
  • the selected action corresponds to launching an application with the user to change the user's address.
  • the action is implemented with the user.
  • an intent may include a user intent to acquire assistance with a problem (e.g., assistance with a malfunctioning product or reduction in service), to make a request (e.g., request a status of an account, replacement of a product, disconnection of service), input information (e.g., change of address, change of purchase order information, change a status of a previous request), or reporting an emergency condition (e.g., electrical wires down, smell of gas, weather condition).
  • a person of skill in the art may use available information regarding a company (such as its products, services, and customers) to determine an appropriate list of intents.
  • the list of intents may be modified, such as by changing an intent or adding a new intent.
  • a message may have more than one intent (e.g., a longer message with a compound sentence or multiple sentences) or multiple messages may be processed to determine an intent (e.g., where a customer partially enters a message and completes it in a subsequent message).
  • the implemented action is shown with box 120 to inform the user the address change application is being started, and box 130 that is the next step in the address change application. The customer may then continue with the action to complete the address update (not shown in FIG. 1A ).
  • FIG. 2A illustrates another example automated customer support session on a customer device and FIG. 2B is a flowchart of this example implementation of providing automated customer support.
  • the customer has used customer device 200 to access a user interface for obtaining support from a company, and the customer has entered message 210 “My address is incorrect.”
  • a company receives the message of the customer.
  • the company processes the message to determine one or more possible intents of the message. Multiple intents may be selected, for example, because there is some uncertainty as to the correct intent (e.g., if the scores of the two highest-scoring intents are close to each other).
  • the company may present information about the possible intents to the customer as shown in box 220 of FIG. 2A .
  • Clarification information may then be received from the customer to determine a final intent.
  • the intents may be presented to the customer, and the customer may then select the intent that best matches his request.
  • clarification information refers to any information received from a customer to determine a final intent from one or more possible intents.
  • the customer has selected the “Change Address” intent.
  • the company receives the clarification information (e.g., the selection of the intent) from the customer and determines that the final intent is the change address intent.
  • the company obtains an action corresponding to the selected intent, such as the address change application described above.
  • the company implements the address change application, such as by presenting box 230 and box 240 to the customer.
  • a natural language interface is any communications interface where a person communicating using the interface is able to transmit natural language using the interface.
  • a natural language interface may include a person entering a typed message or speaking to a microphone.
  • FIG. 3 illustrates an example architecture that may be used by a company to obtain assistance from a third-party company in providing customer support to its customers.
  • FIG. 3 illustrates a system 301 that allows a third-party company 310 to provide customer support services to multiple companies.
  • third-party company 310 is providing customer support services to company A 330 , company B 331 , and company C 332 .
  • Third-party company 310 may provide customer support services to any number of companies.
  • Each company may seek customer support from a company where the support process uses the services of third-party company 310 .
  • customer A 320 may be seeking support from company A 330
  • customer B 321 may be seeking support from company B 331
  • customer C 322 may be seeking support from company C 332 . It may or may not be apparent to the customers whether they are seeking customer support directly from the company or using services of third-party company 310 .
  • Third-party company 310 may assist a company in providing customer support in a variety of ways.
  • third-party company 310 may assist in connecting a customer with a customer service representative working on behalf of the company.
  • third-party company 310 may select a customer service representative, may provide a user interface to a customer to make it easier for a customer to request support, and may provide a user interface to a customer service representative to assist the customer service representative in responding to a request of a customer.
  • a customer service representative may have any appropriate relationship with the company on behalf of which it is providing customer support.
  • a customer service representative may be an employee or contractor of a company and providing customer support to only customers of that company, or a customer service representative may be providing services to multiple companies and providing support to customers of the multiple companies at the same time.
  • third-party company 310 The communications between third-party company 310 , customers, and companies may be architected in a variety of ways. In some implementations, all communications between a customer and a company may be via third-party company 310 and there may not be any direct connection between the customer and the company. In some implementations, third-party company 310 may communicate with the company but may not communicate directly with the customer. In some implementations, a customer may communicate directly with the company and also third-party company 310 .
  • each of the two connections may be used for different kinds of requests.
  • the customer may use the network connection with the company.
  • the customer may use the network connection with third-party company. It may not be apparent to the customer whether the customer is using a network connection with the company or with third-party company 310 .
  • FIG. 4 is a flowchart of an example implementation of creating a natural language interface and improving the natural language interface over time by processing usage data of the natural language interface.
  • FIG. 4 may be implemented by a company on its own or by a third-party company that provides services to companies for at least some aspects of implementing a natural language interface.
  • an initial configuration for a natural language interface is determined.
  • the configuration of the natural language may include specifying a list of intents that are to be understood by the natural language interface.
  • the configuration may also include an initial set of seed messages for each intent, such as example messages illustrating how customers may express the intent.
  • the natural language interface is deployed so that it may be used by customers.
  • Deployment may include training models, such as training an intent classifier that may classify a message as corresponding to one or more intents, and then implementing them on a server that may be accessed by customers.
  • the natural language interface is used by customers, and usage data is collected.
  • the usage data may include interactions of customers with the natural language interface and each interaction may include a message received from a customer, one or more intents corresponding to the message as determined by an intent classifier, and an indication as to whether the interaction was successful (e.g., whether the message was classified correctly or the customer was able to obtain the desired assistance).
  • the natural language interface is improved by processing the usage data. For example, it may be determined that a first intent is being confused with a second intent and that the configuration should be updated to more clearly distinguish those intents. For another example, it may be determined that an existing intent should be split into two intents.
  • a company may need to configure a natural language interface. For example, the first time a natural language interface is implemented, the company may need to specify the intents to be handled by the natural language interface (e.g., address change, cancel service, etc.).
  • a company may configure a natural language interface on its own or use the services of a third-party company to do so. In the following, the use of a third-party company is presented for clarity of presentation, but the same techniques may be implemented by a company for itself.
  • FIG. 5 is a flowchart of an example implementation of configuring a natural language interface.
  • a user interface is presented by third-party company to a company for providing configuration information relating to the natural language interface.
  • the third-party company may use any appropriate techniques in providing a user interface to companies, such as creating a web page that allows the companies to provide information to configure the natural language interface.
  • third-party company receives information about a list of intents from the company.
  • the information about the list of intents may be specified using any appropriate techniques. For example, for each intent provided by the company, the company may specify a name for the intent (e.g., a change address intent), a longer description of the intent (e.g., “This intent allows a customer to change his address.”), and an action to perform when the intent is selected (e.g., an address of a web application to be launched with the customer).
  • a name for the intent e.g., a change address intent
  • a longer description of the intent e.g., “This intent allows a customer to change his address.”
  • an action to perform when the intent is selected e.g., an address of a web application to be launched with the customer.
  • An intent may be associated with any appropriate action, such as launching a program with the customer (e.g., web application or smart phone application), providing a message to the customer (“Your package is arriving on Thursday.”), providing a resource to the customer (e.g., directing the user to a web page that provides the customer with information), or connecting the customer with a customer service representative (e.g., via a messaging interface or a phone call).
  • launching a program with the customer e.g., web application or smart phone application
  • providing a message to the customer (“Your package is arriving on Thursday.”
  • providing a resource to the customer e.g., directing the user to a web page that provides the customer with information
  • connecting the customer with a customer service representative e.g., via a messaging interface or a phone call.
  • third-party company receives hierarchical information about the list of intents, such as a graph or a tree indicating relationships between intents.
  • FIG. 6 illustrates example hierarchical information for a list of intents in the form of a graph.
  • the intents are shown along the right side of the figure (e.g., “Change Address”).
  • the top or root node of the graph is on the left side of the figure.
  • the graph may group or categorize the intents based on a similarity of the subject matter of the intents.
  • FIG. 6 illustrates nodes for three categories: Billing, Services, and Tech Support. Each of these categories has multiple intents associated with it.
  • the Billing category includes nodes for the intents Change Address, Make Payment, and Cancel Service.
  • An intent may belong to more than one category.
  • the Cancel Service intent is included in the Billing Category and the Services category.
  • the hierarchical information may include additional levels, such as categories, sub-categories, and then intents.
  • the hierarchical information may facilitate a customer or a customer service representative in navigating to the intent needed by the customer.
  • the customer may provide a message that does not clearly specify the needs of the customer. For example, the customer may transmit the message “Need Internet,” and it may be unclear from the message if the customer would like to add a service relating to the Internet or may need tech support relating to the Internet.
  • the hierarchical information may be used to disambiguate the customer's request.
  • a company may not provide hierarchical information about the list of intents, and the intents may be used without any hierarchical relationship.
  • the list of intents may include an “Other” intent to capture messages from a customer that do not match any of the other intents provided by the customer.
  • the Other intent may operate as a universal background model.
  • third-party company receives example messages or seed messages for each of the provided intents from the company.
  • the example messages may be created by a person or may be extracted from existing logs of customer support sessions.
  • the messages may be provided using any appropriate techniques, such as by providing a text file for each intent where each line of the text file is an example message for the intent. Where the list of intents includes an Other intent, the Other intent may be seeded with messages that do not match any of the intents provided by the customer.
  • third-party trains an intent classifier using the seed messages.
  • Any appropriate classifier may be used for the intent classifier, such as a neural network, a recurrent neural network, a support vector machine, or a logistic regression classifier.
  • the intent classifier may process a message (e.g., the text of the message or word embeddings corresponding to the message) and output a score for each of the available intents, where a score indicates a match between the message and the intent.
  • the performance of the intent classifier may be evaluated. Any appropriate evaluation procedures may be used. For example, each of the seed messages may be processed by the intent classifier to make sure that they are classified as belonging to the intent for which the message was seeded. During evaluation, a person doing the evaluation may be able to enter a message for evaluation by the classifier to see the highest scoring intent for the entered message or the scores for each intent. The intent classifier may also be tested against a test set of messages where each message of the test set has been labelled with an intent, and an error rate for the test set may be computed and presented to a user. During the evaluation process, the configuration for the natural language interface may be changed. For example, the hierarchy may be altered, intents may be added, intents may be removed or merged, and seed messages may be added, removed, or moved to a different intent.
  • the natural language interface is deployed.
  • the software implementing the natural language interface may be deployed to one or more servers using the trained intent classifier.
  • the natural language interface After the natural language interface is deployed, it may be used by customers to obtain assistance from the company.
  • Third-party company may deploy natural language interfaces for multiple companies simultaneously. For example, customers of a first company may use a first natural language interface, customers of a second company may use a second natural language interface, and so forth.
  • FIG. 7 is a flowchart of an example implementation of a natural language interface by a company on behalf of a company by third-party company.
  • a message is received from a customer, such as by using any of the techniques described above.
  • one or more intents are selected by processing the message with an intent classifier. Any appropriate techniques may be used for selecting the one or more intents using the output of the classifier. For example, all intents with a score above a threshold may be selected, or the top N scoring intents may be selected. In some instances, a single intent may be selected, and, in such instances, steps 730 and 740 may be skipped and processing may proceed to step 750 .
  • the intent classifier may have been trained using unsupervised, semi-supervised, or supervised techniques.
  • An intent classifier may process feature vectors computed from the words (or possibly n-grams of words) of the message. For example, for each word of the message, a 1-hot vector or a word embedding may be obtained to represent the word.
  • the message may be represented as a bag of words.
  • the recurrent neural network may sequentially process the feature vectors of the words of the message, and compute one or both of an internal state vector or an output vector at each iteration.
  • a final vector e.g., an internal state vector or output vector
  • a transformation of the final vector may be computed (e.g., a linear transform, affine transform, or a multi-layer perceptron) where the transformed final vector has a length corresponding to the number of possible intents.
  • the largest element of the transformed final vector may be selected as corresponding to the intent of the message (or one or more largest elements may be selected as corresponding to one or more intents).
  • the convolutional neural network may process a matrix created from the feature vectors of the words of the message.
  • the convolutional neural network may perform one or more operations of filtering and max pooling to generate an output vector.
  • a transformation of the output vector may be computed (such as the transforms described above) where the transformed output vector has a length corresponding to the number of possible intents.
  • One or more intents may be selected using the transformed output vector as described above.
  • the intent classifier may be a multiclass support vector machine (SVM). Any appropriate SVM may be used, such as an SVM that makes classification decisions between individual intents and all other intents, that makes classification decisions between all pairs of possible intents, or that makes classification decisions using a directed acyclic graph. After performing one or more classification decisions comparing intents, the SVM may select a one or more intents as corresponding to the message.
  • SVM multiclass support vector machine
  • a reduced intent graph is created using the one or more selected intents.
  • FIGS. 8A and 8B illustrate examples of how an intent hierarchy graph (e.g., the graph of FIG. 6 ) may be reduced using the selected intents.
  • a complete intent graph 810 is shown, where a complete intent graph corresponds to the entire hierarchy of intents provided by the company.
  • A, B, and C represent categories of intents and A 1 , A 2 , A 3 , B 1 , B 2 , C 1 , C 2 , and C 3 represent individual intents.
  • intents A 1 , A 3 , and C 2 are highlighted because these were the intents selected at step 720 .
  • a score or probability of each of the selected intents are also shown.
  • a pruned intent graph 820 is shown, where intents that were not selected are pruned from the graph, and categories with no remaining intents are also pruned. If FIG.
  • a reduced intent graph 830 is shown, where categories with only a single intent are removed and the single intent remains on its own. Reduced intent graph 830 may be used to disambiguate among the selected intents to select a final intent for the message as described in greater detail below.
  • FIG. 8B illustrates another example of generating a reduced intent graph.
  • complete intent graph 850 is the same as complete intent graph 810 , except that a different set of intents was selected at step 730 .
  • intents A 1 , A 2 , and C 2 are selected with the scores or probabilities as shown.
  • a pruned intent graph 860 is shown, where intents that were not selected are pruned, and categories with no remaining intents are also pruned.
  • a reduced intent graph 870 is shown, where the graph is converted to a tree. Reduced intent graph 870 may be used to disambiguate among the selected intents to select a final intent for the message as described in greater detail below.
  • a final intent is selected using the reduced intent graph.
  • Any appropriate techniques may be used to select the final intent using the reduced intent graph, and the techniques may depend on the type of device used by the customer (e.g., the size of the display of the device).
  • the complete intent graph (e.g., complete intent graph 810 ) may be presented to the customer along with other information such as the selected intents from step 720 , probabilities of the selected intents, and any other information about each of the intents.
  • the customer may be presented with a portion of the complete intent graph and may be able to navigate to other portions of the complete intent graph. The customer may then select a final intent that best matches the customer's needs.
  • the reduced intent graph (e.g., reduced intent graph 830 or reduced intent graph 870 ) may be presented to the customer along with any of the information about the intents or categories described above.
  • the reduced intent graph may be sorted so that intents or categories with the highest probability are shown at the top, and intents or categories with the lowest probability are shown at the bottom.
  • the customer may be able to navigate to categories or intents that are not part of the reduced intent graph. The customer may then select a final intent that best matches the customer's needs.
  • the customer may not be presented with the intent graph, but instead the customer may be presented with questions to select a final intent.
  • the customer may be asked to select between category A and intent C 2 . Where the customer selects category A, the customer may then be asked to select between intent A 1 and intent A 3 . Alternatively, the customer may be asked to select among intents A 1 , A 3 , and C 2 .
  • the customer may similarly be asked to select category C or category A, and then asked to select an intent from the selected category. Alternatively, the customer may be asked to select among intents C 2 , A 1 , and A 2 .
  • the number of options in a question may depend on the size of the display of the customer's device.
  • a customer service representative may assist in the selection of a final intent. For a customer service representative, it may be preferred to show the complete intent graph since a customer service representative may be more likely to have a larger display and may be more accustomed to viewing and navigating an intent graph. The customer service representative may also be able to view the message entered by the customer. The customer service representative may select the final intent for the customer after viewing the message from the customer.
  • a customer service representative may select the final intent on behalf of the customer without the customer being aware of the involvement of the customer service representative.
  • the customer service representative may select an intent to cause an action to be implemented with the customer (e.g., an address change action), and the action may be immediately implemented without the customer and customer service representative exchanging messages.
  • the customer and customer service representative may exchange messages while the customer service representative is selecting a final intent on behalf of the customer.
  • hierarchical information may not be available or may not be used with a list of intents, and accordingly a reduced intent graph may not be created.
  • a final intent may be selected from the one or more intents selected at step 720 and steps 730 and 740 may be omitted.
  • step 750 an action is obtained corresponding to the final intent, and at step 760 , the action is implemented with the customer.
  • the natural language interface may be similarly used with a large number of customers of the company. Usage data of the natural language interface may be collected and used to improve the natural language interface as described below.
  • Usage data of the natural language interface may be collected and used to improve the natural language interface.
  • Each session with a customer may be referred to as an interaction, and data may be collected for each interaction. Any appropriate data may be collected for an interaction, such as the message sent by the customer, the scores computed by the intent classifier, one or more intents selected by the intent classifier, a final intent such as a final intent selected as described above, and an indication of success or failure for the interaction.
  • the usage data may include data collected from some or all of the interactions during a time period. The usage data may then be analyzed to determine the performance of the natural language interface and to determine changes to the natural language interface to improve performance.
  • the indication may be a boolean value indicating success or not.
  • the indicate may be a value, such as a value in a range of 0 to 100 with values above a threshold (e.g., a threshold of 50) indicating success.
  • an indication of success may be determined for an interaction: the highest scoring intent of the classifier was the final intent used for the interaction, the customer completing the action (e.g., the customer completing the address change process), a survey of the customer after the action has been completed (e.g., asking the customer if he was happy with the outcome), comparing the output of the intent classifier with an intent selected by the customer, comparing the output of the intent classifier with the action implemented with the customer, or where the customer has interacted with a customer service representative, asking the customer service representative to indicate the success of the interaction.
  • the customer completing the action e.g., the customer completing the address change process
  • a survey of the customer after the action has been completed e.g., asking the customer if he was happy with the outcome
  • comparing the output of the intent classifier with an intent selected by the customer comparing the output of the intent classifier with the action implemented with the customer, or where the customer has interacted with a customer service representative, asking the customer service representative to indicate the success of
  • FIG. 11 is a flowchart of an example implementation of processing usage data to improve a natural language interface.
  • usage data of the natural language interface is obtained, such as any of the usage data described above.
  • usage data is compiled for each intent of the list of intents.
  • Each interaction may be assigned to an intent using any appropriate criteria. For example, an interaction may be assigned to an intent corresponding to the highest score of the intent classifier or to a final intent selected as described above.
  • FIG. 9A illustrates example usage data that has been compiled for the change address intent.
  • a table shows messages that were received from a customer corresponding to the change address intent, the number of times that message was received, and a success rate for that message (using any of the indications of success described above.).
  • the messages assigned to an intent are clustered.
  • any intent there may be a large number of ways that a person could express that intent.
  • Some of the ways for expressing an intent may be similar to other ways and some ways may be quite different from other ways.
  • FIG. 9A for the change address intent the first three messages are similar to one another: “I want to change my address”; “I need to change my address”; and “I would like to change my address.”
  • Other ways of expressing the intent may be quite different, such as the message “I moved.”
  • the table in FIG. 9A could have hundreds, thousands, or even more unique messages.
  • the messages corresponding to the intent may be clustered so that messages that are similar to one another are likely in the same cluster and messages that are very different from each are likely in different clusters. Any appropriate techniques may be used to cluster messages.
  • a message embedding may be computed for each message, where a message embedding is a vector in an N-dimensional vector space that represents the message but does so in a manner that preserves useful information about the content of the message. Any appropriate techniques may be used to compute message embeddings from a training corpus (e.g., stored messages between customers and CSRs).
  • message embeddings may be computed using latent semantic indexing.
  • a feature vector e.g., word counts or term frequency-inverse document frequency values
  • a matrix may be created by combining the feature vectors
  • the message embeddings may be obtained by performing a singular value decomposition of the matrix. See Deerwester, “Improving Information Retrieval with Latent Semantic Indexing,” Proceedings of the 51st ASIS Annual Meeting (ASIS '88), Vol. 25 (October 1988).
  • message embeddings may be computed using latent Dirichlet allocation and computing a message embedding as a mixture of topics from a topic model. See Blei et al., “Latent Dirichlet Allocation,” Journal of Machine Learning Research, vol. 3, pp. 993-1022, (2003).
  • message embeddings may be computed as paragraph vectors where the paragraph vectors are trained to predict the next word of the paragraph given a context of the paragraph. See Le et al., “Distributed Representations of Sentences and Documents,” Proceedings of the 31st International Conference on Machine Learning, W&CP vol. 32, Beijing, China, 2014.
  • message embeddings may be computed as the final hidden state of a recurrent neural network using an encoder-decoder language model. See Kiros et al., “Skip-Thought Vectors,” Proceedings of the 28th International Conference on Neural Information Processing Systems, vol. 2., pp. 3294-3302, 2015.
  • FIG. 10 illustrates a conceptual representation of message embeddings in a vector space.
  • the vector space is two dimensions, but in a real application, the number of dimensions would be much higher.
  • Each message is represented by an “x” or an “o” (the difference will be explained below).
  • the message embeddings may be clustered using any appropriate clustering algorithm, such as k-means or agglomerative clustering.
  • possible clusters of messages are indicated by the dashed ovals.
  • Each cluster is likely to contain similar manners of expressing the intent. For example, the messages “I want to change my address”; “I need to change my address”; and “I would like to change my address” are likely to be in the same cluster.
  • a message may be generated or selected from the cluster and referred to as a prototype message. Any appropriate techniques may be used to obtain a prototype message for a cluster.
  • the prototype message may be a most frequent message, a message with a largest count, a message closest to the center of the cluster (e.g., geometrically or by center of mass), a message having a selected language type (e.g., simplest language, most commonly used words, or the inclusion or exclusion of jargon or selected words), or a randomly selected message.
  • the prototype message may be generated instead of being selected and the prototype message may be different from the messages in the cluster.
  • a prototype message may be generated by processing the messages of the cluster with an encoder neural network and then generating the prototype message with a decoder neural network.
  • a prototype message may be generated by creating parse trees for the messages in the cluster, transforming the parse trees into a single parse tree, and then generating the prototype message using the transformed parse tree.
  • the prototype message is represented as an “o” instead of an “x.”
  • FIG. 9B is an example list of clusters for the change address intent that may be determined from the usage data of FIG. 9A .
  • the first cluster has “I need to change my address” as a prototype message and this cluster also includes the messages “I want to change my address” and “I would like to change my address.”
  • Each cluster may include additional information about the cluster, such as the total count of messages in the cluster and a success rate of the messages in the cluster.
  • information about the intents and clusters is presented to a user to allow the user to improve the natural language interface.
  • third-party company may present a user interface to one of its own employees, an employee of the company for whom it is providing the natural language interface, or another person. Any appropriate information may be presented about the natural language interface, such as any of the following: the list of intents, hierarchical information corresponding to the intents, statistics relating to the intents, and information about clusters (e.g., a prototype message and statistics for the clusters).
  • third-party company may present a user interface that shows a success rate for each of the intents.
  • a first intent may have a low success rate and a person may select the first intent to see additional details relating to that intent.
  • the user interface may present information about the clusters in the first intent, such as the prototype messages and statistics.
  • the usage data for a natural language interface may include a very large number of interactions, and accordingly there may be a very large number of unique messages. Because of the large number of messages, it may be difficult for a user to view individual messages and understand the performance of the natural language interface. By combining similar messages into a cluster and providing a prototype message for the messages in the cluster, it may be much easier for a person to understand which aspects of the natural language interface are working well and which aspects need improvement. For example, a single intent may have tens of thousands of unique messages, but these messages may be clustered into five clusters that are presented to a user. The user can then view information about each of the five clusters to better understand why some clusters are working well and some clusters are not.
  • the user interface may include controls to allow the user to modify the natural language interface. Any appropriate controls for any appropriate modifications may be included.
  • the user interface may include a control to create a new intent, a control to move a cluster from a first intent to a second intent, or a control to retrain the natural language interface using the usage data. From step 1140 , processing may proceed to step 1150 , step 1160 , or 1160 according to the actions of the user.
  • the user may be able to make other changes at step 1140 .
  • the user may be able to change the hierarchical information relating to the list of intents to improve the ability of a customer or other person to navigate the intent hierarchy.
  • suggestions may be presented to the user at step 1140 . Any appropriate suggestions may be presented, such as a changes to the intents or the hierarchical information of the intents.
  • a cluster of an intent has a success rate below a threshold
  • that cluster may be highlighted or a suggestion may be presented to the user to create a new intent with the cluster or move the cluster (or a portion thereof) to a new or different intent.
  • a cluster with the low success rate may be compared with clusters of other intents using any appropriate distance metric (e.g., a Mahalanobis distance).
  • the suggestion may be to move the cluster to the second intent.
  • the suggestion may be to create a new intent with the cluster.
  • suggestions may be presented to improve the hierarchical information.
  • the usage data may include, for each interaction, navigation information relating to a user (e.g., the customer or customer service representative) navigating the intent graph.
  • a suggestion may be made to modify the hierarchical information to include a connection between the first node and the second node.
  • third-party company has received an input from the user to create a new intent using a cluster.
  • the user may determine that customers are frequently requesting an operation that is not covered by any of the existing intents. For example, from the example clusters of FIG. 9B , a user may determine that the cluster with prototype “I'm on vacation” is sufficiently different from the prototypes for the other clusters that a new intent should be created for providing services to customers when they go on vacation (e.g., temporarily halting services or placing a hold on sending packages to the customer).
  • the user may specify any appropriate information for the new cluster, such as a name, a description, or seed messages.
  • the user may also specify that some or all of the messages from an existing cluster (e.g., cluster 5 of FIG. 9B ) may be assigned to the new intent.
  • third-party company may update the configuration for the natural language interface, such as by updating a configuration stored in a data store of third-party company.
  • third-party may retrain the intent classifier using the configuration and the usage data.
  • the intent classifier was trained using seed messages, which may be relatively small in number when a natural language interface is first created because of a lack of available data.
  • the intent classifier may be trained using some or all of the messages provided by customers.
  • the messages used to train each intent may be the messages as compiled at step 1120 , except that a new intent was created at step 1155 using the cluster of messages specified by the user.
  • the user may instead decide that the natural language interface would likely perform better if a cluster of messages currently at a first intent are moved to a second intent. For example, there may be an existing intent for temporary suspension of services, and the user may decide that the cluster relating to the customer going on vacation should be moved to the intent for temporary suspension of services.
  • third-party company receives an input from the user to move a cluster of messages from a first intent to a second intent.
  • third-party company modifies the configuration of the natural language interface so that this cluster of messages is now assigned to the second intent.
  • the intent classifier is retrained using the updated configuration and the usage data as described above.
  • step 1140 the user may decide to not make any changes to the configuration and instead to simply retrain the intent classifier using the usage data. Accordingly, processing may proceed to step 1170 where the intent classifier is retrained without any changes to the configuration.
  • the retrained intent classifier may be evaluated or tested to make sure that the retrained intent classifier is performing well. Any appropriate evaluations or tests may be performed.
  • the retrained classifier may be backtested with the usage data or with other data. For example, the retrained classifier may be used to obtain classification results for each message from the usage data that had been marked with a positive indication of success. It may be desired that the retrained intent classifier produce that same intent that was produced by the previous classifier. Where the retrained intent classifier produces a different intent for more than a threshold number of messages from the usage data, processing may proceed back to step 1140 to allow further actions to be taken. Where the retrained classifier passes the evaluations or tests, then processing may proceed to step 1180 .
  • the natural language interface is deployed with the updated intent classifier that was trained at step 1170 .
  • Customers may now access the natural language interface with hopefully improved performance.
  • the user interface may allow user to access information about individual messages in a cluster in addition to the prototype. For example, the user interface may allow the user to view any of the messages for the intent as shown in FIG. 9A . In some implementations, the user interface may allow the user to provide an input regarding one or more messages, such as to create a new intent with the one or more messages or to move the one or more messages from a first intent to a second intent.
  • the processing of FIG. 11 may be repeated any number of times to improve the natural language interface.
  • the processing of FIG. 11 may be performed at regular intervals, such as once a week or once a month.
  • the processing may be performed more frequently when the natural language interface is first launched and less frequently after the natural language interface has reached a desired performance level.
  • FIG. 12 is an example system that may be used by third-party company to provide these services to other companies.
  • the system of FIG. 12 may be implemented in a variety of ways, and the system of FIG. 12 is exemplary and not a limiting example. For example, other components may be added, components may be removed or merged, and connections between components may be added or removed.
  • Third-party company may provide services to a company and company user 1260 may be employee of the company or other person assisting the company in configuring a natural language interface.
  • User 1210 , user 1211 , and user 1212 may be customers of the company who are using the natural language interface to obtain support from the company.
  • Third-party company may provide services to any number of companies, but for clarity of presentation, only one company is shown.
  • Third-party company may have an intent console component 1250 that presents a user interface to company user 1260 to provide an initial configuration for a natural language interface and to later update the configuration.
  • Intent console component 1250 may receive the configuration information and store it in configuration data store 1290 so that it may be accessed by other components.
  • Intent console may also receive analysis results (e.g., analysis of usage data) from analysis component 1230 and present the analysis results to company user 1260 .
  • Third-party company may have a training component 1240 that may train an intent classifier for classifying messages and may also train a clustering model for clustering messages of an intent. Training component 1240 may also retrain or update classifiers and models after usage data has been collected. Training component 1240 may access any of configuration data store 1290 , usage data store 1280 , and models data store 1270 .
  • Third-party company may have an analysis component 1230 that analyzes usage data of the natural language, such as any of the analysis described above. Analysis component may access usage data store 1280 .
  • Third-party company may have a production component 1220 that allows users to interact with the natural language interface.
  • production component 1220 may receive messages from customers, determine an intent of a message, cause an action to be performed with a customer, or connect a customer with a customer service representative.
  • Production component 1220 may access models data store 1270 and usage data store 1280 .
  • Models data store 1270 may store any appropriate models or classifiers, such as an intent classifier or a clustering model.
  • Usage data store 1280 may store any appropriate usage data, such as the usage data described herein.
  • Configuration data store 1290 may store any appropriate configuration information for a natural language interface, such as any of the configuration information described herein.
  • FIG. 13 illustrates components of one implementation of a computing device 1300 for implementing any of the techniques described above.
  • the components are shown as being on a single computing device, but the components may be distributed among multiple computing devices, such as a system of computing devices, including, for example, an end-user computing device (e.g., a smart phone or a tablet) and/or a server computing device (e.g., cloud computing).
  • an end-user computing device e.g., a smart phone or a tablet
  • server computing device e.g., cloud computing
  • Computing device 1300 may include any components typical of a computing device, such as volatile or nonvolatile memory 1310 , one or more processors 1311 , and one or more network interfaces 1312 .
  • Computing device 1300 may also include any input and output components, such as displays, keyboards, and touch screens.
  • Computing device 1300 may also include a variety of components or modules providing specific functionality, and these components or modules may be implemented in software, hardware, or a combination thereof.
  • computing device 1300 may include any of production component 1220 , analysis component 1230 , training component 1240 , or intent console component 1250 , as described above.
  • Other implementations may include additional components or exclude some of the components.
  • Computing device 1300 may include or have access to various data stores, such as models data store 1270 , usage data store 1280 , and configuration data store 1290 .
  • Data stores may use any known storage technology such as files, relational or non-relational databases, or any non-transitory computer-readable media.
  • processor as used herein is meant to include at least one processor and unless context clearly indicates otherwise, the plural and the singular should be understood to be interchangeable. Any aspects of the present disclosure may be implemented as a method on the machine, as a system or apparatus as part of or in relation to the machine, or as a computer program product embodied in a computer readable medium executing on one or more of the machines.
  • the processor may be part of a server, client, network infrastructure, mobile computing platform, stationary computing platform, or other computing platform.
  • a processor may be any kind of computational or processing device capable of executing program instructions, codes, binary instructions and the like.
  • the processor may be or include a signal processor, digital processor, embedded processor, microprocessor or any variant such as a co-processor (math co-processor, graphic co-processor, communication co-processor and the like) and the like that may directly or indirectly facilitate execution of program code or program instructions stored thereon.
  • the processor may enable execution of multiple programs, threads, and codes. The threads may be executed simultaneously to enhance the performance of the processor and to facilitate simultaneous operations of the application.
  • methods, program codes, program instructions and the like described herein may be implemented in one or more thread.
  • the thread may spawn other threads that may have assigned priorities associated with them; the processor may execute these threads based on priority or any other order based on instructions provided in the program code.
  • the processor may include memory that stores methods, codes, instructions and programs as described herein and elsewhere.
  • the processor may access a storage medium through an interface that may store methods, codes, and instructions as described herein and elsewhere.
  • the storage medium associated with the processor for storing methods, programs, codes, program instructions or other type of instructions capable of being executed by the computing or processing device may include but may not be limited to one or more of a CD-ROM, DVD, memory, hard disk, flash drive, RAM, ROM, cache and the like.
  • a processor may include one or more cores that may enhance speed and performance of a multiprocessor.
  • the process may be a dual core processor, quad core processors, other chip-level multiprocessor and the like that combine two or more independent cores (called a die).
  • the methods and systems described herein may be deployed in part or in whole through a machine that executes computer software on a server, client, firewall, gateway, hub, router, or other such computer and/or networking hardware.
  • the software program may be associated with a server that may include a file server, print server, domain server, internet server, intranet server and other variants such as secondary server, host server, distributed server and the like.
  • the server may include one or more of memories, processors, computer readable media, storage media, ports (physical and virtual), communication devices, and interfaces capable of accessing other servers, clients, machines, and devices through a wired or a wireless medium, and the like.
  • the methods, programs, or codes as described herein and elsewhere may be executed by the server.
  • other devices required for execution of methods as described in this application may be considered as a part of the infrastructure associated with the server.
  • the server may provide an interface to other devices including, without limitation, clients, other servers, printers, database servers, print servers, file servers, communication servers, distributed servers and the like. Additionally, this coupling and/or connection may facilitate remote execution of program across the network. The networking of some or all of these devices may facilitate parallel processing of a program or method at one or more locations without deviating from the scope of the disclosure.
  • any of the devices attached to the server through an interface may include at least one storage medium capable of storing methods, programs, code and/or instructions.
  • a central repository may provide program instructions to be executed on different devices.
  • the remote repository may act as a storage medium for program code, instructions, and programs.
  • the software program may be associated with a client that may include a file client, print client, domain client, internet client, intranet client and other variants such as secondary client, host client, distributed client and the like.
  • the client may include one or more of memories, processors, computer readable media, storage media, ports (physical and virtual), communication devices, and interfaces capable of accessing other clients, servers, machines, and devices through a wired or a wireless medium, and the like.
  • the methods, programs, or codes as described herein and elsewhere may be executed by the client.
  • other devices required for execution of methods as described in this application may be considered as a part of the infrastructure associated with the client.
  • the client may provide an interface to other devices including, without limitation, servers, other clients, printers, database servers, print servers, file servers, communication servers, distributed servers and the like. Additionally, this coupling and/or connection may facilitate remote execution of program across the network. The networking of some or all of these devices may facilitate parallel processing of a program or method at one or more locations without deviating from the scope of the disclosure.
  • any of the devices attached to the client through an interface may include at least one storage medium capable of storing methods, programs, applications, code and/or instructions.
  • a central repository may provide program instructions to be executed on different devices.
  • the remote repository may act as a storage medium for program code, instructions, and programs.
  • the methods and systems described herein may be deployed in part or in whole through network infrastructures.
  • the network infrastructure may include elements such as computing devices, servers, routers, hubs, firewalls, clients, personal computers, communication devices, routing devices and other active and passive devices, modules and/or components as known in the art.
  • the computing and/or non-computing device(s) associated with the network infrastructure may include, apart from other components, a storage medium such as flash memory, buffer, stack, RAM, ROM and the like.
  • the processes, methods, program codes, instructions described herein and elsewhere may be executed by one or more of the network infrastructural elements.
  • the methods, program codes, and instructions described herein and elsewhere may be implemented on a cellular network having multiple cells.
  • the cellular network may either be frequency division multiple access (FDMA) network or code division multiple access (CDMA) network.
  • FDMA frequency division multiple access
  • CDMA code division multiple access
  • the cellular network may include mobile devices, cell sites, base stations, repeaters, antennas, towers, and the like.
  • the cell network may be a GSM, GPRS, 3G, EVDO, mesh, or other networks types.
  • the mobile devices may include navigation devices, cell phones, mobile phones, mobile personal digital assistants, laptops, palmtops, netbooks, pagers, electronic books readers, music players and the like. These devices may include, apart from other components, a storage medium such as a flash memory, buffer, RAM, ROM and one or more computing devices.
  • the computing devices associated with mobile devices may be enabled to execute program codes, methods, and instructions stored thereon. Alternatively, the mobile devices may be configured to execute instructions in collaboration with other devices.
  • the mobile devices may communicate with base stations interfaced with servers and configured to execute program codes.
  • the mobile devices may communicate on a peer-to-peer network, mesh network, or other communications network.
  • the program code may be stored on the storage medium associated with the server and executed by a computing device embedded within the server.
  • the base station may include a computing device and a storage medium.
  • the storage device may store program codes and instructions executed by the computing devices associated with the base station.
  • the computer software, program codes, and/or instructions may be stored and/or accessed on machine readable media that may include: computer components, devices, and recording media that retain digital data used for computing for some interval of time; semiconductor storage known as random access memory (RAM); mass storage typically for more permanent storage, such as optical discs, forms of magnetic storage like hard disks, tapes, drums, cards and other types; processor registers, cache memory, volatile memory, non-volatile memory; optical storage such as CD, DVD; removable media such as flash memory (e.g.
  • RAM random access memory
  • mass storage typically for more permanent storage, such as optical discs, forms of magnetic storage like hard disks, tapes, drums, cards and other types
  • processor registers cache memory, volatile memory, non-volatile memory
  • optical storage such as CD, DVD
  • removable media such as flash memory (e.g.
  • USB sticks or keys floppy disks, magnetic tape, paper tape, punch cards, standalone RAM disks, Zip drives, removable mass storage, off-line, and the like; other computer memory such as dynamic memory, static memory, read/write storage, mutable storage, read only, random access, sequential access, location addressable, file addressable, content addressable, network attached storage, storage area network, bar codes, magnetic ink, and the like.
  • the methods and systems described herein may transform physical and/or or intangible items from one state to another.
  • the methods and systems described herein may also transform data representing physical and/or intangible items from one state to another.
  • machines may include, but may not be limited to, personal digital assistants, laptops, personal computers, mobile phones, other handheld computing devices, medical equipment, wired or wireless communication devices, transducers, chips, calculators, satellites, tablet PCs, electronic books, gadgets, electronic devices, devices having artificial intelligence, computing devices, networking equipment, servers, routers and the like.
  • the elements depicted in the flow chart and block diagrams or any other logical component may be implemented on a machine capable of executing program instructions.
  • the methods and/or processes described above, and steps thereof, may be realized in hardware, software or any combination of hardware and software suitable for a particular application.
  • the hardware may include a general-purpose computer and/or dedicated computing device or specific computing device or particular aspect or component of a specific computing device.
  • the processes may be realized in one or more microprocessors, microcontrollers, embedded microcontrollers, programmable digital signal processors or other programmable device, along with internal and/or external memory.
  • the processes may also, or instead, be embodied in an application specific integrated circuit, a programmable gate array, programmable array logic, or any other device or combination of devices that may be configured to process electronic signals. It will further be appreciated that one or more of the processes may be realized as a computer executable code capable of being executed on a machine-readable medium.
  • the computer executable code may be created using a structured programming language such as C, an object oriented programming language such as C++, or any other high-level or low-level programming language (including assembly languages, hardware description languages, and database programming languages and technologies) that may be stored, compiled or interpreted to run on one of the above devices, as well as heterogeneous combinations of processors, processor architectures, or combinations of different hardware and software, or any other machine capable of executing program instructions.
  • a structured programming language such as C
  • an object oriented programming language such as C++
  • any other high-level or low-level programming language including assembly languages, hardware description languages, and database programming languages and technologies
  • each method described above and combinations thereof may be embodied in computer executable code that, when executing on one or more computing devices, performs the steps thereof.
  • the methods may be embodied in systems that perform the steps thereof, and may be distributed across devices in a number of ways, or all of the functionality may be integrated into a dedicated, standalone device or other hardware.
  • the means for performing the steps associated with the processes described above may include any of the hardware and/or software described above. All such permutations and combinations are intended to fall within the scope of the present disclosure.

Abstract

A third-party company may assist companies in providing natural language interfaces for their customers. To implement a natural language interface for a company, a configuration may be received that includes information, such as a list intents, seed messages for the intents, and hierarchical information of the intents. An intent classifier may be trained using the configuration, and the natural language interface may be deployed for use with customers. Usage data of the natural language classifier may be collected and used to improve the natural language interface. Messages corresponding to an intent may be clustered into clusters of similar messages, and a prototype message may be obtained for each cluster to provide a human understandable description of the cluster. The information about the clusters may be used to improve the natural language interface, such as by creating a new intent with a cluster or moving a cluster to a different intent.

Description

    CLAIM OF PRIORITY
  • The present application is a continuation of U.S. patent application Ser. No. 16/240,958, filed on 7 Jan. 2019, and entitled “UPDATING NATURAL LANGUAGE INTERFACES BY PROCESSING USAGE DATA” (ASAP-0010-U01-C01).
  • U.S. patent application Ser. No. 16/240,958 is a continuation of U.S. patent application Ser. No. 15/894,504, filed on 12 Feb. 2018, entitled “UPDATING NATURAL LANGUAGE INTERFACES BY PROCESSING USAGE DATA”, and now issued as U.S. Pat. No. 10,210,244 on 19 Feb. 2019 (ASAP-0010-U01).
  • Each of the foregoing applications/patents is incorporated herein by reference in its entirety for all purposes.
  • FIELD OF THE INVENTION
  • The present invention relates to using semantic processing to improve customer support.
  • BACKGROUND
  • Companies need to efficiently interact with customers to provide services to their customers. For example, customers may need to obtain information about services of the company, may have a question about billing, or may need technical support from the company. Companies interact with customers in a variety of different ways. Companies may have a website and the customer may navigate the website to perform various actions. Companies may have an application (“app”) that runs on a user device, such as a smart phone or a tablet, that provides similar services as a website. Companies may have a phone number that customers can call to obtain information via interactive voice response or to speak with a customer service representative. Companies may also respond to customers using various social media services, such as Facebook or Twitter.
  • Some existing techniques for allowing customers to interact with companies may be a nuisance to the customer. Navigating to the right page on a website or an app or navigating a voice menu on a phone call may be time consuming. Some existing techniques for allowing customers to interact with companies may be expensive for a company to implement. Hiring customer service representatives to manually respond to requests and answer phone calls may be a significant expense.
  • BRIEF DESCRIPTION OF THE FIGURES
  • The invention and the following detailed description of certain embodiments thereof may be understood by reference to the following figures:
  • FIG. 1A is an example user interface for a natural language interface, and FIG. 1B is a flowchart of an example implementation of the natural language interface.
  • FIG. 2A is an example user interface for a natural language interface, and FIG. 2B is a flowchart of an example implementation of the natural language interface.
  • FIG. 3 is an example system for a third-party company to provide services to multiple companies.
  • FIG. 4 is a flowchart of an example implementation of configuring and improving a natural language interface.
  • FIG. 5 is a flowchart of an example implementation of configuring a natural language interface.
  • FIG. 6 is an example list of intents and hierarchical information relating to the list of intents.
  • FIG. 7 is a flowchart of an example implementation of a natural language interface with a customer.
  • FIGS. 8A and 8B are examples of graphs and reduced graphs that may be used with a natural language interface.
  • FIG. 9A is an example of usage data corresponding to an intent, and FIG. 9B is an example of clustered usage data for an intent.
  • FIG. 10 is an example clustering of usage data for an intent.
  • FIG. 11 is a flowchart of an example implementation of improving a natural language interface by processing usage data.
  • FIG. 12 is an example system that a third-party company may use to assist companies in configuring and improving a natural language interface.
  • FIG. 13 is an exemplary computing device that may be used to configure and improve a natural language interface.
  • DETAILED DESCRIPTION
  • Described herein are techniques for implementing a natural language interface to respond to a request of a user. Although the techniques described herein may be used for a wide variety of users and requests, for clarity of presentation, an example of a company providing a response to a request of a customer will be used. The techniques described herein, however, are not limited to customers and companies, responses may be provided to requests from users who are not customers, and responses may be from any entity or person. The techniques described herein may use or be combined with any of the techniques described in U.S. Pat. No. 9,715,496 or U.S. patent application Ser. No. 15/789,241, each of which is incorporated herein by reference in their entireties for all purposes (“incorporated patents”).
  • A natural language interface may be used to provide a fully automated experience for a customer. A natural language interface may use semantic processing to allow a customer to obtain needed information or needed support by making a request using natural language. The use of natural language for making requests may allow the customer to receive a resolution to his request more quickly than with other methods of making requests. In some instances, the customer's request may be resolved with a single interaction. For example, where the customer asks, “When is my package arriving?,” the response may be, “Your package is scheduled to arrive on Wednesday and has tracking number 1234.”
  • Implementing, diagnosing, and improving a natural language interface may be challenging for a company. Customers may express a concept using a wide variety of different expressions. For example, a customer wanting to change his address may state “I'd like to change my address” or “I've moved.” Additionally, a company may want to handle many different types of requests, such as address changes, bill requests, refunds, requests for new services, technical support, and any other issue that may normally be handled by a human customer support representative. Between a large number of different types of requests, and a large number of ways of expressing each request, it may be challenging to implement a natural language interface that works properly for most ways of expressing each type of request. A company may desire an accurate natural language interface because it may lower costs for the company and also provide a better experience for customers of the company.
  • The techniques described herein improve the ability of a company to initially create a natural language interface and to later improve the natural language interface based on usage data of customers using the natural language interface. The techniques described herein allow a company to identify portions of the natural language interface that are operating well (e.g. with low error rates) and portions of the natural language interface that could be improved.
  • The usage data of customers using the natural language interface may include a large amount of data that may be challenging for a computer or a human to review to understand why portions of the natural language interface are working well or not. The techniques described herein include clustering the usage data into clusters of similar usage data (e.g., different ways of expressing the same request) and selecting prototype messages for each cluster. By using information about clusters of usage data and their prototype messages, the usage data may be processed more efficiently to improve the operation of the natural language interface, as described in greater detail below.
  • FIG. 1A illustrates an example automated customer support session on a customer device and FIG. 1B is a flowchart of an example implementation of providing automated customer support. In FIG. 1B and other flowcharts herein, the ordering of the steps is exemplary and other orders are possible, not all steps are required, steps may be combined (in whole or part) or sub-divided and, in some implementations, some steps may be omitted or other steps may be added. The methods described by any flowcharts described herein may be implemented, for example, by any of the computers or systems described herein.
  • In FIG. 1A, the customer has used customer device 100 to access a user interface for obtaining support from a company. Customer device 100 may be any appropriate device, such as a smart phone, tablet, wearable device, or Internet of things device. The customer may request support using any appropriate techniques, such as typing or speaking a message to an app running on the customer device (e.g., an app of the company or a third-party app created for processing customer requests), typing or speaking a message on a web page, sending a text message, or sending an email. As used herein, a text message includes any message sent as text including but not limited to a message sent using SMS (short message service) or a special-purpose application (e.g., Facebook Messenger, Apple iMessage, Google Hangouts, or WhatsApp). In FIG. 1A, the customer has entered message 110 “My address is incorrect.”
  • User interfaces described throughout the present disclosure may be of any type, including at least a mobile application, a web-based interface, a texting interface (e.g., receipt, responses, and/or prompts via text), a dedicated program operating on a computer, and/or any hardware device, computer executable media, and/or display accessible to customers, customer service representatives, and/or company agents or representatives. User interfaces include any display devices configured to depict elements of the user interface, and/or input/output devices configured to receive responses from users of any aspect of embodiments described herein. Without limitation, user interfaces include, in certain embodiments, one or more of network communications, graphical displays, mobile devices, keyboards, mouses, and/or touch screens. In certain embodiments, the same or distinct, or re-configured, user interfaces may be provided to various users in an embodiment of the disclosure. For example, a first user interface may be provided to a user for training a natural language interface, a second user interface may be provided to a user (e.g., at a first company) for data collection or display about a natural language interface (e.g., displaying cluster information), a third user interface may be provided to a user (e.g., at a second company) for training and/or data collection or display about a second natural language interface, and a fourth user interface may be provided to a customer for obtaining support from a company (e.g., the first company, the second company, or another company). The number of user interface examples is provided to distinguish user interfaces, and does not imply any order or relationship between the interfaces. Any given interface infrastructure (e.g., computer executable media operating on a server, and/or network communications) may be utilized in one or more interfaces, and any one or more of the example user interfaces may be present or not present in given embodiments.
  • At step 150 of FIG. 1B, a company receives the message of the customer. At step 160, the company processes the message to determine an intent of the message. For example, the company may process the message with an intent classifier that provides a score for each intent of a list of available intents, and an intent having a highest score may be selected as the intent of message 110. At step 170, the company obtains an action corresponding to the determined intent of message 110. An action may correspond to any sequence of interactions defined by the company. For example, an action may correspond to launching an application with the user (e.g., to change an address), providing a message to the user with requested information, directing the user to a resource (e.g., a web page), or connecting the user with a customer service representative. In this example, the selected action corresponds to launching an application with the user to change the user's address. At step 180, the action is implemented with the user.
  • The term intent, as used herein should be understood broadly and any appropriate intents may be used. For example, an intent may include a user intent to acquire assistance with a problem (e.g., assistance with a malfunctioning product or reduction in service), to make a request (e.g., request a status of an account, replacement of a product, disconnection of service), input information (e.g., change of address, change of purchase order information, change a status of a previous request), or reporting an emergency condition (e.g., electrical wires down, smell of gas, weather condition). For any particular application, a person of skill in the art may use available information regarding a company (such as its products, services, and customers) to determine an appropriate list of intents. In some implementations, described in greater detail below, the list of intents may be modified, such as by changing an intent or adding a new intent. In some implementations, a message may have more than one intent (e.g., a longer message with a compound sentence or multiple sentences) or multiple messages may be processed to determine an intent (e.g., where a customer partially enters a message and completes it in a subsequent message).
  • In FIG. 1A, the implemented action is shown with box 120 to inform the user the address change application is being started, and box 130 that is the next step in the address change application. The customer may then continue with the action to complete the address update (not shown in FIG. 1A).
  • FIG. 2A illustrates another example automated customer support session on a customer device and FIG. 2B is a flowchart of this example implementation of providing automated customer support. In FIG. 2A, the customer has used customer device 200 to access a user interface for obtaining support from a company, and the customer has entered message 210 “My address is incorrect.”
  • At step 250 of FIG. 2B, a company receives the message of the customer. At step 260, the company processes the message to determine one or more possible intents of the message. Multiple intents may be selected, for example, because there is some uncertainty as to the correct intent (e.g., if the scores of the two highest-scoring intents are close to each other). At step 260, the company may present information about the possible intents to the customer as shown in box 220 of FIG. 2A.
  • Clarification information may then be received from the customer to determine a final intent. For example, the intents may be presented to the customer, and the customer may then select the intent that best matches his request. As used herein, clarification information refers to any information received from a customer to determine a final intent from one or more possible intents.
  • In box 220, the customer has selected the “Change Address” intent. At step 265, the company receives the clarification information (e.g., the selection of the intent) from the customer and determines that the final intent is the change address intent. At step 270 the company obtains an action corresponding to the selected intent, such as the address change application described above. At step 275, the company implements the address change application, such as by presenting box 230 and box 240 to the customer.
  • The interactions between the customer and the company may be referred to as a natural language interface because the customer is able to request assistance from the company by entering a natural language message. As used herein, a natural language interface is any communications interface where a person communicating using the interface is able to transmit natural language using the interface. For example, a natural language interface may include a person entering a typed message or speaking to a microphone.
  • A company providing customer support to its customers may desire to use the services of a third-party company to improve the customer support experience and to reduce overall customer support costs. For example, a company may find it more cost effective to use semantic processing services of a third-party company than to implement its own semantic processing services. FIG. 3 illustrates an example architecture that may be used by a company to obtain assistance from a third-party company in providing customer support to its customers.
  • FIG. 3 illustrates a system 301 that allows a third-party company 310 to provide customer support services to multiple companies. In FIG. 3, third-party company 310 is providing customer support services to company A 330, company B 331, and company C 332. Third-party company 310 may provide customer support services to any number of companies.
  • Customers of each company may seek customer support from a company where the support process uses the services of third-party company 310. For example, customer A 320 may be seeking support from company A 330, customer B 321 may be seeking support from company B 331, and customer C 322 may be seeking support from company C 332. It may or may not be apparent to the customers whether they are seeking customer support directly from the company or using services of third-party company 310.
  • Third-party company 310 may assist a company in providing customer support in a variety of ways. In some implementations, third-party company 310 may assist in connecting a customer with a customer service representative working on behalf of the company. For example, third-party company 310 may select a customer service representative, may provide a user interface to a customer to make it easier for a customer to request support, and may provide a user interface to a customer service representative to assist the customer service representative in responding to a request of a customer. A customer service representative may have any appropriate relationship with the company on behalf of which it is providing customer support. For example, a customer service representative may be an employee or contractor of a company and providing customer support to only customers of that company, or a customer service representative may be providing services to multiple companies and providing support to customers of the multiple companies at the same time.
  • The communications between third-party company 310, customers, and companies may be architected in a variety of ways. In some implementations, all communications between a customer and a company may be via third-party company 310 and there may not be any direct connection between the customer and the company. In some implementations, third-party company 310 may communicate with the company but may not communicate directly with the customer. In some implementations, a customer may communicate directly with the company and also third-party company 310.
  • Where a customer is connected to both a company and third-party company 310, each of the two connections may be used for different kinds of requests. For example, where the customer is interacting with the company in a way that does not require the services of third-party company 310 (e.g., navigating a web site of the company), the customer may use the network connection with the company. Where the customer is interacting with the company in a way that uses the services of third-party company 310, the customer may use the network connection with third-party company. It may not be apparent to the customer whether the customer is using a network connection with the company or with third-party company 310.
  • FIG. 4 is a flowchart of an example implementation of creating a natural language interface and improving the natural language interface over time by processing usage data of the natural language interface. FIG. 4 may be implemented by a company on its own or by a third-party company that provides services to companies for at least some aspects of implementing a natural language interface.
  • At step 410 an initial configuration for a natural language interface is determined. For example, the configuration of the natural language may include specifying a list of intents that are to be understood by the natural language interface. The configuration may also include an initial set of seed messages for each intent, such as example messages illustrating how customers may express the intent.
  • At step 420, the natural language interface is deployed so that it may be used by customers. Deployment may include training models, such as training an intent classifier that may classify a message as corresponding to one or more intents, and then implementing them on a server that may be accessed by customers.
  • At step 430, the natural language interface is used by customers, and usage data is collected. The usage data may include interactions of customers with the natural language interface and each interaction may include a message received from a customer, one or more intents corresponding to the message as determined by an intent classifier, and an indication as to whether the interaction was successful (e.g., whether the message was classified correctly or the customer was able to obtain the desired assistance).
  • At step 440, the natural language interface is improved by processing the usage data. For example, it may be determined that a first intent is being confused with a second intent and that the configuration should be updated to more clearly distinguish those intents. For another example, it may be determined that an existing intent should be split into two intents.
  • Each of these steps are described in greater detail below.
  • Configuration of Natural Language Interface
  • A company may need to configure a natural language interface. For example, the first time a natural language interface is implemented, the company may need to specify the intents to be handled by the natural language interface (e.g., address change, cancel service, etc.). A company may configure a natural language interface on its own or use the services of a third-party company to do so. In the following, the use of a third-party company is presented for clarity of presentation, but the same techniques may be implemented by a company for itself.
  • FIG. 5 is a flowchart of an example implementation of configuring a natural language interface. At step 510, a user interface is presented by third-party company to a company for providing configuration information relating to the natural language interface. The third-party company may use any appropriate techniques in providing a user interface to companies, such as creating a web page that allows the companies to provide information to configure the natural language interface.
  • At step 520, third-party company receives information about a list of intents from the company. The information about the list of intents may be specified using any appropriate techniques. For example, for each intent provided by the company, the company may specify a name for the intent (e.g., a change address intent), a longer description of the intent (e.g., “This intent allows a customer to change his address.”), and an action to perform when the intent is selected (e.g., an address of a web application to be launched with the customer). An intent may be associated with any appropriate action, such as launching a program with the customer (e.g., web application or smart phone application), providing a message to the customer (“Your package is arriving on Thursday.”), providing a resource to the customer (e.g., directing the user to a web page that provides the customer with information), or connecting the customer with a customer service representative (e.g., via a messaging interface or a phone call).
  • At step 530, third-party company receives hierarchical information about the list of intents, such as a graph or a tree indicating relationships between intents. FIG. 6 illustrates example hierarchical information for a list of intents in the form of a graph. In FIG. 6, the intents are shown along the right side of the figure (e.g., “Change Address”). The top or root node of the graph is on the left side of the figure. The graph may group or categorize the intents based on a similarity of the subject matter of the intents. FIG. 6 illustrates nodes for three categories: Billing, Services, and Tech Support. Each of these categories has multiple intents associated with it. For example, the Billing category includes nodes for the intents Change Address, Make Payment, and Cancel Service. An intent may belong to more than one category. For example, the Cancel Service intent is included in the Billing Category and the Services category. The hierarchical information may include additional levels, such as categories, sub-categories, and then intents.
  • The hierarchical information may facilitate a customer or a customer service representative in navigating to the intent needed by the customer. In some situations, the customer may provide a message that does not clearly specify the needs of the customer. For example, the customer may transmit the message “Need Internet,” and it may be unclear from the message if the customer would like to add a service relating to the Internet or may need tech support relating to the Internet. As described in greater detail below, the hierarchical information may be used to disambiguate the customer's request.
  • In some implementations, a company may not provide hierarchical information about the list of intents, and the intents may be used without any hierarchical relationship.
  • In some implementations, the list of intents may include an “Other” intent to capture messages from a customer that do not match any of the other intents provided by the customer. For example, the Other intent may operate as a universal background model.
  • At step 540, third-party company receives example messages or seed messages for each of the provided intents from the company. The example messages may be created by a person or may be extracted from existing logs of customer support sessions. The messages may be provided using any appropriate techniques, such as by providing a text file for each intent where each line of the text file is an example message for the intent. Where the list of intents includes an Other intent, the Other intent may be seeded with messages that do not match any of the intents provided by the customer.
  • At step 550, third-party trains an intent classifier using the seed messages. Any appropriate classifier may be used for the intent classifier, such as a neural network, a recurrent neural network, a support vector machine, or a logistic regression classifier. The intent classifier may process a message (e.g., the text of the message or word embeddings corresponding to the message) and output a score for each of the available intents, where a score indicates a match between the message and the intent.
  • At step 560, the performance of the intent classifier may be evaluated. Any appropriate evaluation procedures may be used. For example, each of the seed messages may be processed by the intent classifier to make sure that they are classified as belonging to the intent for which the message was seeded. During evaluation, a person doing the evaluation may be able to enter a message for evaluation by the classifier to see the highest scoring intent for the entered message or the scores for each intent. The intent classifier may also be tested against a test set of messages where each message of the test set has been labelled with an intent, and an error rate for the test set may be computed and presented to a user. During the evaluation process, the configuration for the natural language interface may be changed. For example, the hierarchy may be altered, intents may be added, intents may be removed or merged, and seed messages may be added, removed, or moved to a different intent.
  • At step 570, the natural language interface is deployed. For example, the software implementing the natural language interface may be deployed to one or more servers using the trained intent classifier.
  • Deployment of Natural Language Interface
  • After the natural language interface is deployed, it may be used by customers to obtain assistance from the company. Third-party company may deploy natural language interfaces for multiple companies simultaneously. For example, customers of a first company may use a first natural language interface, customers of a second company may use a second natural language interface, and so forth.
  • FIG. 7 is a flowchart of an example implementation of a natural language interface by a company on behalf of a company by third-party company. At step 710, a message is received from a customer, such as by using any of the techniques described above.
  • At step 720, one or more intents are selected by processing the message with an intent classifier. Any appropriate techniques may be used for selecting the one or more intents using the output of the classifier. For example, all intents with a score above a threshold may be selected, or the top N scoring intents may be selected. In some instances, a single intent may be selected, and, in such instances, steps 730 and 740 may be skipped and processing may proceed to step 750.
  • Any appropriate intent classifier may be used. For example, the intent classifier may have been trained using unsupervised, semi-supervised, or supervised techniques. An intent classifier may process feature vectors computed from the words (or possibly n-grams of words) of the message. For example, for each word of the message, a 1-hot vector or a word embedding may be obtained to represent the word. In another example, the message may be represented as a bag of words.
  • For example, where the intent classifier is a recurrent neural network, the recurrent neural network may sequentially process the feature vectors of the words of the message, and compute one or both of an internal state vector or an output vector at each iteration. A final vector (e.g., an internal state vector or output vector) from a final iteration may then be used to classify the intent of the message. A transformation of the final vector may be computed (e.g., a linear transform, affine transform, or a multi-layer perceptron) where the transformed final vector has a length corresponding to the number of possible intents. The largest element of the transformed final vector may be selected as corresponding to the intent of the message (or one or more largest elements may be selected as corresponding to one or more intents).
  • For another example, where the intent classifier is a convolutional neural network, the convolutional neural network may process a matrix created from the feature vectors of the words of the message. The convolutional neural network may perform one or more operations of filtering and max pooling to generate an output vector. A transformation of the output vector may be computed (such as the transforms described above) where the transformed output vector has a length corresponding to the number of possible intents. One or more intents may be selected using the transformed output vector as described above.
  • For another example, the intent classifier may be a multiclass support vector machine (SVM). Any appropriate SVM may be used, such as an SVM that makes classification decisions between individual intents and all other intents, that makes classification decisions between all pairs of possible intents, or that makes classification decisions using a directed acyclic graph. After performing one or more classification decisions comparing intents, the SVM may select a one or more intents as corresponding to the message.
  • At step 730, a reduced intent graph is created using the one or more selected intents. FIGS. 8A and 8B illustrate examples of how an intent hierarchy graph (e.g., the graph of FIG. 6) may be reduced using the selected intents.
  • In FIG. 8A, a complete intent graph 810 is shown, where a complete intent graph corresponds to the entire hierarchy of intents provided by the company. In this example, A, B, and C represent categories of intents and A1, A2, A3, B1, B2, C1, C2, and C3 represent individual intents. In the complete intent graph 810, intents A1, A3, and C2 are highlighted because these were the intents selected at step 720. A score or probability of each of the selected intents are also shown. In FIG. 8A, a pruned intent graph 820 is shown, where intents that were not selected are pruned from the graph, and categories with no remaining intents are also pruned. If FIG. 8A, a reduced intent graph 830 is shown, where categories with only a single intent are removed and the single intent remains on its own. Reduced intent graph 830 may be used to disambiguate among the selected intents to select a final intent for the message as described in greater detail below.
  • FIG. 8B illustrates another example of generating a reduced intent graph. In FIG. 8B, complete intent graph 850 is the same as complete intent graph 810, except that a different set of intents was selected at step 730. In complete intent graph 850, intents A1, A2, and C2 are selected with the scores or probabilities as shown. In FIG. 8B, a pruned intent graph 860 is shown, where intents that were not selected are pruned, and categories with no remaining intents are also pruned. If FIG. 8B, a reduced intent graph 870 is shown, where the graph is converted to a tree. Reduced intent graph 870 may be used to disambiguate among the selected intents to select a final intent for the message as described in greater detail below.
  • At step 740, a final intent is selected using the reduced intent graph. Any appropriate techniques may be used to select the final intent using the reduced intent graph, and the techniques may depend on the type of device used by the customer (e.g., the size of the display of the device). Where the customer is using a personal computer with a large display, the complete intent graph (e.g., complete intent graph 810) may be presented to the customer along with other information such as the selected intents from step 720, probabilities of the selected intents, and any other information about each of the intents. Depending on the size of the graph and the display, the customer may be presented with a portion of the complete intent graph and may be able to navigate to other portions of the complete intent graph. The customer may then select a final intent that best matches the customer's needs.
  • In some implementations, the reduced intent graph (e.g., reduced intent graph 830 or reduced intent graph 870) may be presented to the customer along with any of the information about the intents or categories described above. The reduced intent graph may be sorted so that intents or categories with the highest probability are shown at the top, and intents or categories with the lowest probability are shown at the bottom. In some implementations, the customer may be able to navigate to categories or intents that are not part of the reduced intent graph. The customer may then select a final intent that best matches the customer's needs.
  • In some implementations, the customer may not be presented with the intent graph, but instead the customer may be presented with questions to select a final intent. For example, for the reduced intent graph 830, the customer may be asked to select between category A and intent C2. Where the customer selects category A, the customer may then be asked to select between intent A1 and intent A3. Alternatively, the customer may be asked to select among intents A1, A3, and C2. For reduced intent graph 870, the customer may similarly be asked to select category C or category A, and then asked to select an intent from the selected category. Alternatively, the customer may be asked to select among intents C2, A1, and A2. The number of options in a question may depend on the size of the display of the customer's device.
  • In some implementations, a customer service representative may assist in the selection of a final intent. For a customer service representative, it may be preferred to show the complete intent graph since a customer service representative may be more likely to have a larger display and may be more accustomed to viewing and navigating an intent graph. The customer service representative may also be able to view the message entered by the customer. The customer service representative may select the final intent for the customer after viewing the message from the customer.
  • In some implementations or instances, a customer service representative may select the final intent on behalf of the customer without the customer being aware of the involvement of the customer service representative. The customer service representative may select an intent to cause an action to be implemented with the customer (e.g., an address change action), and the action may be immediately implemented without the customer and customer service representative exchanging messages. In some implementations or instances, the customer and customer service representative may exchange messages while the customer service representative is selecting a final intent on behalf of the customer.
  • In some implementations, hierarchical information may not be available or may not be used with a list of intents, and accordingly a reduced intent graph may not be created. A final intent may be selected from the one or more intents selected at step 720 and steps 730 and 740 may be omitted.
  • At step 750 an action is obtained corresponding to the final intent, and at step 760, the action is implemented with the customer. These steps may use any of the techniques described above.
  • The natural language interface may be similarly used with a large number of customers of the company. Usage data of the natural language interface may be collected and used to improve the natural language interface as described below.
  • Improvement of Natural Language Interface
  • Usage data of the natural language interface may be collected and used to improve the natural language interface. Each session with a customer may be referred to as an interaction, and data may be collected for each interaction. Any appropriate data may be collected for an interaction, such as the message sent by the customer, the scores computed by the intent classifier, one or more intents selected by the intent classifier, a final intent such as a final intent selected as described above, and an indication of success or failure for the interaction. The usage data may include data collected from some or all of the interactions during a time period. The usage data may then be analyzed to determine the performance of the natural language interface and to determine changes to the natural language interface to improve performance.
  • Any appropriate techniques may be used to determine an indication of success for failure for an interaction. In some implementations, the indication may be a boolean value indicating success or not. In some implementations, the indicate may be a value, such as a value in a range of 0 to 100 with values above a threshold (e.g., a threshold of 50) indicating success. The following are some non-limiting examples of how an indication of success may be determined for an interaction: the highest scoring intent of the classifier was the final intent used for the interaction, the customer completing the action (e.g., the customer completing the address change process), a survey of the customer after the action has been completed (e.g., asking the customer if he was happy with the outcome), comparing the output of the intent classifier with an intent selected by the customer, comparing the output of the intent classifier with the action implemented with the customer, or where the customer has interacted with a customer service representative, asking the customer service representative to indicate the success of the interaction.
  • FIG. 11 is a flowchart of an example implementation of processing usage data to improve a natural language interface. At step 1110, usage data of the natural language interface is obtained, such as any of the usage data described above.
  • At step 1120, usage data is compiled for each intent of the list of intents. Each interaction may be assigned to an intent using any appropriate criteria. For example, an interaction may be assigned to an intent corresponding to the highest score of the intent classifier or to a final intent selected as described above. FIG. 9A illustrates example usage data that has been compiled for the change address intent. In FIG. 9A, a table shows messages that were received from a customer corresponding to the change address intent, the number of times that message was received, and a success rate for that message (using any of the indications of success described above.).
  • At step 1130, the messages assigned to an intent are clustered. For any intent, there may be a large number of ways that a person could express that intent. Some of the ways for expressing an intent may be similar to other ways and some ways may be quite different from other ways. For example, in FIG. 9A for the change address intent, the first three messages are similar to one another: “I want to change my address”; “I need to change my address”; and “I would like to change my address.” Other ways of expressing the intent may be quite different, such as the message “I moved.” When collecting messages corresponding to an intent, there may be a large number of ways of expressing that intent and thus the table in FIG. 9A could have hundreds, thousands, or even more unique messages.
  • To facilitate analysis of the intent, the messages corresponding to the intent may be clustered so that messages that are similar to one another are likely in the same cluster and messages that are very different from each are likely in different clusters. Any appropriate techniques may be used to cluster messages.
  • In some implementations, a message embedding may be computed for each message, where a message embedding is a vector in an N-dimensional vector space that represents the message but does so in a manner that preserves useful information about the content of the message. Any appropriate techniques may be used to compute message embeddings from a training corpus (e.g., stored messages between customers and CSRs).
  • In some implementations, message embeddings may be computed using latent semantic indexing. A feature vector (e.g., word counts or term frequency-inverse document frequency values) may be computed for each message of a training corpus of messages, a matrix may be created by combining the feature vectors, and the message embeddings may be obtained by performing a singular value decomposition of the matrix. See Deerwester, “Improving Information Retrieval with Latent Semantic Indexing,” Proceedings of the 51st ASIS Annual Meeting (ASIS '88), Vol. 25 (October 1988).
  • In some implementations, message embeddings may be computed using latent Dirichlet allocation and computing a message embedding as a mixture of topics from a topic model. See Blei et al., “Latent Dirichlet Allocation,” Journal of Machine Learning Research, vol. 3, pp. 993-1022, (2003).
  • In some implementations, message embeddings may be computed as paragraph vectors where the paragraph vectors are trained to predict the next word of the paragraph given a context of the paragraph. See Le et al., “Distributed Representations of Sentences and Documents,” Proceedings of the 31st International Conference on Machine Learning, W&CP vol. 32, Beijing, China, 2014.
  • In some implementations, message embeddings may be computed as the final hidden state of a recurrent neural network using an encoder-decoder language model. See Kiros et al., “Skip-Thought Vectors,” Proceedings of the 28th International Conference on Neural Information Processing Systems, vol. 2., pp. 3294-3302, 2015.
  • FIG. 10 illustrates a conceptual representation of message embeddings in a vector space. In the example of FIG. 10 the vector space is two dimensions, but in a real application, the number of dimensions would be much higher. Each message is represented by an “x” or an “o” (the difference will be explained below). After a message embedding has been computed for each message, the message embeddings may be clustered using any appropriate clustering algorithm, such as k-means or agglomerative clustering. In FIG. 10, possible clusters of messages are indicated by the dashed ovals. Each cluster is likely to contain similar manners of expressing the intent. For example, the messages “I want to change my address”; “I need to change my address”; and “I would like to change my address” are likely to be in the same cluster.
  • To describe the cluster, a message may be generated or selected from the cluster and referred to as a prototype message. Any appropriate techniques may be used to obtain a prototype message for a cluster. In some implementations, the prototype message may be a most frequent message, a message with a largest count, a message closest to the center of the cluster (e.g., geometrically or by center of mass), a message having a selected language type (e.g., simplest language, most commonly used words, or the inclusion or exclusion of jargon or selected words), or a randomly selected message. In some implementations, the prototype message may be generated instead of being selected and the prototype message may be different from the messages in the cluster. For example, a prototype message may be generated by processing the messages of the cluster with an encoder neural network and then generating the prototype message with a decoder neural network. For another example, a prototype message may be generated by creating parse trees for the messages in the cluster, transforming the parse trees into a single parse tree, and then generating the prototype message using the transformed parse tree. In the example of FIG. 10, the prototype message is represented as an “o” instead of an “x.”
  • FIG. 9B is an example list of clusters for the change address intent that may be determined from the usage data of FIG. 9A. In FIG. 9B, the first cluster has “I need to change my address” as a prototype message and this cluster also includes the messages “I want to change my address” and “I would like to change my address.” Each cluster may include additional information about the cluster, such as the total count of messages in the cluster and a success rate of the messages in the cluster.
  • At step 1140, information about the intents and clusters is presented to a user to allow the user to improve the natural language interface. In some implementations, third-party company may present a user interface to one of its own employees, an employee of the company for whom it is providing the natural language interface, or another person. Any appropriate information may be presented about the natural language interface, such as any of the following: the list of intents, hierarchical information corresponding to the intents, statistics relating to the intents, and information about clusters (e.g., a prototype message and statistics for the clusters).
  • For example, third-party company may present a user interface that shows a success rate for each of the intents. A first intent may have a low success rate and a person may select the first intent to see additional details relating to that intent. In response, the user interface may present information about the clusters in the first intent, such as the prototype messages and statistics.
  • The usage data for a natural language interface may include a very large number of interactions, and accordingly there may be a very large number of unique messages. Because of the large number of messages, it may be difficult for a user to view individual messages and understand the performance of the natural language interface. By combining similar messages into a cluster and providing a prototype message for the messages in the cluster, it may be much easier for a person to understand which aspects of the natural language interface are working well and which aspects need improvement. For example, a single intent may have tens of thousands of unique messages, but these messages may be clustered into five clusters that are presented to a user. The user can then view information about each of the five clusters to better understand why some clusters are working well and some clusters are not.
  • The user interface may include controls to allow the user to modify the natural language interface. Any appropriate controls for any appropriate modifications may be included. For example, the user interface may include a control to create a new intent, a control to move a cluster from a first intent to a second intent, or a control to retrain the natural language interface using the usage data. From step 1140, processing may proceed to step 1150, step 1160, or 1160 according to the actions of the user.
  • In some implementations, the user may be able to make other changes at step 1140. For example, the user may be able to change the hierarchical information relating to the list of intents to improve the ability of a customer or other person to navigate the intent hierarchy.
  • In some implementations, suggestions may be presented to the user at step 1140. Any appropriate suggestions may be presented, such as a changes to the intents or the hierarchical information of the intents. In some implementations, where a cluster of an intent has a success rate below a threshold, that cluster may be highlighted or a suggestion may be presented to the user to create a new intent with the cluster or move the cluster (or a portion thereof) to a new or different intent. For example, a cluster with the low success rate may be compared with clusters of other intents using any appropriate distance metric (e.g., a Mahalanobis distance). Where the cluster with the low success rate is close to a cluster of a second intent, the suggestion may be to move the cluster to the second intent. Where the cluster with the low success rate is not close to any existing cluster, the suggestion may be to create a new intent with the cluster.
  • In some implementations, suggestions may be presented to improve the hierarchical information. The usage data may include, for each interaction, navigation information relating to a user (e.g., the customer or customer service representative) navigating the intent graph. Where a user commonly navigates from a first node of the graph to a second node, and there is not a direct connection between the first node and the second node, a suggestion may be made to modify the hierarchical information to include a connection between the first node and the second node.
  • At step 1150, third-party company has received an input from the user to create a new intent using a cluster. In reviewing the information about the clusters of an intent, the user may determine that customers are frequently requesting an operation that is not covered by any of the existing intents. For example, from the example clusters of FIG. 9B, a user may determine that the cluster with prototype “I'm on vacation” is sufficiently different from the prototypes for the other clusters that a new intent should be created for providing services to customers when they go on vacation (e.g., temporarily halting services or placing a hold on sending packages to the customer). The user may specify any appropriate information for the new cluster, such as a name, a description, or seed messages. The user may also specify that some or all of the messages from an existing cluster (e.g., cluster 5 of FIG. 9B) may be assigned to the new intent.
  • At step 1155, third-party company may update the configuration for the natural language interface, such as by updating a configuration stored in a data store of third-party company.
  • At step 1170, third-party may retrain the intent classifier using the configuration and the usage data. Previously, the intent classifier was trained using seed messages, which may be relatively small in number when a natural language interface is first created because of a lack of available data. Once the natural language interface has been used by customers, the intent classifier may be trained using some or all of the messages provided by customers. In training the intent classifier the messages used to train each intent may be the messages as compiled at step 1120, except that a new intent was created at step 1155 using the cluster of messages specified by the user.
  • Returning back to step 1140, the user may instead decide that the natural language interface would likely perform better if a cluster of messages currently at a first intent are moved to a second intent. For example, there may be an existing intent for temporary suspension of services, and the user may decide that the cluster relating to the customer going on vacation should be moved to the intent for temporary suspension of services.
  • At step 1160, third-party company receives an input from the user to move a cluster of messages from a first intent to a second intent. At step 1165, third-party company modifies the configuration of the natural language interface so that this cluster of messages is now assigned to the second intent.
  • At step 1170, the intent classifier is retrained using the updated configuration and the usage data as described above.
  • Returning back to step 1140, the user may decide to not make any changes to the configuration and instead to simply retrain the intent classifier using the usage data. Accordingly, processing may proceed to step 1170 where the intent classifier is retrained without any changes to the configuration.
  • At step 1175, the retrained intent classifier may be evaluated or tested to make sure that the retrained intent classifier is performing well. Any appropriate evaluations or tests may be performed. In some implementations, the retrained classifier may be backtested with the usage data or with other data. For example, the retrained classifier may be used to obtain classification results for each message from the usage data that had been marked with a positive indication of success. It may be desired that the retrained intent classifier produce that same intent that was produced by the previous classifier. Where the retrained intent classifier produces a different intent for more than a threshold number of messages from the usage data, processing may proceed back to step 1140 to allow further actions to be taken. Where the retrained classifier passes the evaluations or tests, then processing may proceed to step 1180.
  • At step 1180, the natural language interface is deployed with the updated intent classifier that was trained at step 1170. Customers may now access the natural language interface with hopefully improved performance.
  • In some implementations, the user interface may allow user to access information about individual messages in a cluster in addition to the prototype. For example, the user interface may allow the user to view any of the messages for the intent as shown in FIG. 9A. In some implementations, the user interface may allow the user to provide an input regarding one or more messages, such as to create a new intent with the one or more messages or to move the one or more messages from a first intent to a second intent.
  • The processing of FIG. 11 may be repeated any number of times to improve the natural language interface. For example, the processing of FIG. 11 may be performed at regular intervals, such as once a week or once a month. In some implementations, the processing may be performed more frequently when the natural language interface is first launched and less frequently after the natural language interface has reached a desired performance level.
  • Implementation
  • As described above, a third-party company may provide services to companies to assist those companies in implementing a natural language interface for their customers. FIG. 12 is an example system that may be used by third-party company to provide these services to other companies. The system of FIG. 12 may be implemented in a variety of ways, and the system of FIG. 12 is exemplary and not a limiting example. For example, other components may be added, components may be removed or merged, and connections between components may be added or removed.
  • In FIG. 12, the portions of the service likely provided by third-party company are enclosed in box 1200. Third-party company may provide services to a company and company user 1260 may be employee of the company or other person assisting the company in configuring a natural language interface. User 1210, user 1211, and user 1212 may be customers of the company who are using the natural language interface to obtain support from the company. Third-party company may provide services to any number of companies, but for clarity of presentation, only one company is shown.
  • Third-party company may have an intent console component 1250 that presents a user interface to company user 1260 to provide an initial configuration for a natural language interface and to later update the configuration. Intent console component 1250 may receive the configuration information and store it in configuration data store 1290 so that it may be accessed by other components. Intent console may also receive analysis results (e.g., analysis of usage data) from analysis component 1230 and present the analysis results to company user 1260.
  • Third-party company may have a training component 1240 that may train an intent classifier for classifying messages and may also train a clustering model for clustering messages of an intent. Training component 1240 may also retrain or update classifiers and models after usage data has been collected. Training component 1240 may access any of configuration data store 1290, usage data store 1280, and models data store 1270.
  • Third-party company may have an analysis component 1230 that analyzes usage data of the natural language, such as any of the analysis described above. Analysis component may access usage data store 1280.
  • Third-party company may have a production component 1220 that allows users to interact with the natural language interface. For example, production component 1220 may receive messages from customers, determine an intent of a message, cause an action to be performed with a customer, or connect a customer with a customer service representative. Production component 1220 may access models data store 1270 and usage data store 1280.
  • Models data store 1270 may store any appropriate models or classifiers, such as an intent classifier or a clustering model. Usage data store 1280 may store any appropriate usage data, such as the usage data described herein. Configuration data store 1290 may store any appropriate configuration information for a natural language interface, such as any of the configuration information described herein.
  • FIG. 13 illustrates components of one implementation of a computing device 1300 for implementing any of the techniques described above. In FIG. 13, the components are shown as being on a single computing device, but the components may be distributed among multiple computing devices, such as a system of computing devices, including, for example, an end-user computing device (e.g., a smart phone or a tablet) and/or a server computing device (e.g., cloud computing).
  • Computing device 1300 may include any components typical of a computing device, such as volatile or nonvolatile memory 1310, one or more processors 1311, and one or more network interfaces 1312. Computing device 1300 may also include any input and output components, such as displays, keyboards, and touch screens. Computing device 1300 may also include a variety of components or modules providing specific functionality, and these components or modules may be implemented in software, hardware, or a combination thereof. For example, computing device 1300 may include any of production component 1220, analysis component 1230, training component 1240, or intent console component 1250, as described above. Other implementations may include additional components or exclude some of the components.
  • Computing device 1300 may include or have access to various data stores, such as models data store 1270, usage data store 1280, and configuration data store 1290. Data stores may use any known storage technology such as files, relational or non-relational databases, or any non-transitory computer-readable media.
  • The methods and systems described herein may be deployed in part or in whole through a machine that executes computer software, program codes, and/or instructions on a processor. “Processor” as used herein is meant to include at least one processor and unless context clearly indicates otherwise, the plural and the singular should be understood to be interchangeable. Any aspects of the present disclosure may be implemented as a method on the machine, as a system or apparatus as part of or in relation to the machine, or as a computer program product embodied in a computer readable medium executing on one or more of the machines. The processor may be part of a server, client, network infrastructure, mobile computing platform, stationary computing platform, or other computing platform. A processor may be any kind of computational or processing device capable of executing program instructions, codes, binary instructions and the like. The processor may be or include a signal processor, digital processor, embedded processor, microprocessor or any variant such as a co-processor (math co-processor, graphic co-processor, communication co-processor and the like) and the like that may directly or indirectly facilitate execution of program code or program instructions stored thereon. In addition, the processor may enable execution of multiple programs, threads, and codes. The threads may be executed simultaneously to enhance the performance of the processor and to facilitate simultaneous operations of the application. By way of implementation, methods, program codes, program instructions and the like described herein may be implemented in one or more thread. The thread may spawn other threads that may have assigned priorities associated with them; the processor may execute these threads based on priority or any other order based on instructions provided in the program code. The processor may include memory that stores methods, codes, instructions and programs as described herein and elsewhere. The processor may access a storage medium through an interface that may store methods, codes, and instructions as described herein and elsewhere. The storage medium associated with the processor for storing methods, programs, codes, program instructions or other type of instructions capable of being executed by the computing or processing device may include but may not be limited to one or more of a CD-ROM, DVD, memory, hard disk, flash drive, RAM, ROM, cache and the like.
  • A processor may include one or more cores that may enhance speed and performance of a multiprocessor. In embodiments, the process may be a dual core processor, quad core processors, other chip-level multiprocessor and the like that combine two or more independent cores (called a die).
  • The methods and systems described herein may be deployed in part or in whole through a machine that executes computer software on a server, client, firewall, gateway, hub, router, or other such computer and/or networking hardware. The software program may be associated with a server that may include a file server, print server, domain server, internet server, intranet server and other variants such as secondary server, host server, distributed server and the like. The server may include one or more of memories, processors, computer readable media, storage media, ports (physical and virtual), communication devices, and interfaces capable of accessing other servers, clients, machines, and devices through a wired or a wireless medium, and the like. The methods, programs, or codes as described herein and elsewhere may be executed by the server. In addition, other devices required for execution of methods as described in this application may be considered as a part of the infrastructure associated with the server.
  • The server may provide an interface to other devices including, without limitation, clients, other servers, printers, database servers, print servers, file servers, communication servers, distributed servers and the like. Additionally, this coupling and/or connection may facilitate remote execution of program across the network. The networking of some or all of these devices may facilitate parallel processing of a program or method at one or more locations without deviating from the scope of the disclosure. In addition, any of the devices attached to the server through an interface may include at least one storage medium capable of storing methods, programs, code and/or instructions. A central repository may provide program instructions to be executed on different devices. In this implementation, the remote repository may act as a storage medium for program code, instructions, and programs.
  • The software program may be associated with a client that may include a file client, print client, domain client, internet client, intranet client and other variants such as secondary client, host client, distributed client and the like. The client may include one or more of memories, processors, computer readable media, storage media, ports (physical and virtual), communication devices, and interfaces capable of accessing other clients, servers, machines, and devices through a wired or a wireless medium, and the like. The methods, programs, or codes as described herein and elsewhere may be executed by the client. In addition, other devices required for execution of methods as described in this application may be considered as a part of the infrastructure associated with the client.
  • The client may provide an interface to other devices including, without limitation, servers, other clients, printers, database servers, print servers, file servers, communication servers, distributed servers and the like. Additionally, this coupling and/or connection may facilitate remote execution of program across the network. The networking of some or all of these devices may facilitate parallel processing of a program or method at one or more locations without deviating from the scope of the disclosure. In addition, any of the devices attached to the client through an interface may include at least one storage medium capable of storing methods, programs, applications, code and/or instructions. A central repository may provide program instructions to be executed on different devices. In this implementation, the remote repository may act as a storage medium for program code, instructions, and programs.
  • The methods and systems described herein may be deployed in part or in whole through network infrastructures. The network infrastructure may include elements such as computing devices, servers, routers, hubs, firewalls, clients, personal computers, communication devices, routing devices and other active and passive devices, modules and/or components as known in the art. The computing and/or non-computing device(s) associated with the network infrastructure may include, apart from other components, a storage medium such as flash memory, buffer, stack, RAM, ROM and the like. The processes, methods, program codes, instructions described herein and elsewhere may be executed by one or more of the network infrastructural elements.
  • The methods, program codes, and instructions described herein and elsewhere may be implemented on a cellular network having multiple cells. The cellular network may either be frequency division multiple access (FDMA) network or code division multiple access (CDMA) network. The cellular network may include mobile devices, cell sites, base stations, repeaters, antennas, towers, and the like. The cell network may be a GSM, GPRS, 3G, EVDO, mesh, or other networks types.
  • The methods, programs codes, and instructions described herein and elsewhere may be implemented on or through mobile devices. The mobile devices may include navigation devices, cell phones, mobile phones, mobile personal digital assistants, laptops, palmtops, netbooks, pagers, electronic books readers, music players and the like. These devices may include, apart from other components, a storage medium such as a flash memory, buffer, RAM, ROM and one or more computing devices. The computing devices associated with mobile devices may be enabled to execute program codes, methods, and instructions stored thereon. Alternatively, the mobile devices may be configured to execute instructions in collaboration with other devices. The mobile devices may communicate with base stations interfaced with servers and configured to execute program codes. The mobile devices may communicate on a peer-to-peer network, mesh network, or other communications network. The program code may be stored on the storage medium associated with the server and executed by a computing device embedded within the server. The base station may include a computing device and a storage medium. The storage device may store program codes and instructions executed by the computing devices associated with the base station.
  • The computer software, program codes, and/or instructions may be stored and/or accessed on machine readable media that may include: computer components, devices, and recording media that retain digital data used for computing for some interval of time; semiconductor storage known as random access memory (RAM); mass storage typically for more permanent storage, such as optical discs, forms of magnetic storage like hard disks, tapes, drums, cards and other types; processor registers, cache memory, volatile memory, non-volatile memory; optical storage such as CD, DVD; removable media such as flash memory (e.g. USB sticks or keys), floppy disks, magnetic tape, paper tape, punch cards, standalone RAM disks, Zip drives, removable mass storage, off-line, and the like; other computer memory such as dynamic memory, static memory, read/write storage, mutable storage, read only, random access, sequential access, location addressable, file addressable, content addressable, network attached storage, storage area network, bar codes, magnetic ink, and the like.
  • The methods and systems described herein may transform physical and/or or intangible items from one state to another. The methods and systems described herein may also transform data representing physical and/or intangible items from one state to another.
  • The elements described and depicted herein, including in flow charts and block diagrams throughout the figures, imply logical boundaries between the elements. However, according to software or hardware engineering practices, the depicted elements and the functions thereof may be implemented on machines through computer executable media having a processor capable of executing program instructions stored thereon as a monolithic software structure, as standalone software modules, or as modules that employ external routines, code, services, and so forth, or any combination of these, and all such implementations may be within the scope of the present disclosure. Examples of such machines may include, but may not be limited to, personal digital assistants, laptops, personal computers, mobile phones, other handheld computing devices, medical equipment, wired or wireless communication devices, transducers, chips, calculators, satellites, tablet PCs, electronic books, gadgets, electronic devices, devices having artificial intelligence, computing devices, networking equipment, servers, routers and the like. Furthermore, the elements depicted in the flow chart and block diagrams or any other logical component may be implemented on a machine capable of executing program instructions. Thus, while the foregoing drawings and descriptions set forth functional aspects of the disclosed systems, no particular arrangement of software for implementing these functional aspects should be inferred from these descriptions unless explicitly stated or otherwise clear from the context. Similarly, it will be appreciated that the various steps identified and described above may be varied, and that the order of steps may be adapted to particular applications of the techniques disclosed herein. All such variations and modifications are intended to fall within the scope of this disclosure. As such, the depiction and/or description of an order for various steps should not be understood to require a particular order of execution for those steps, unless required by a particular application, or explicitly stated or otherwise clear from the context.
  • The methods and/or processes described above, and steps thereof, may be realized in hardware, software or any combination of hardware and software suitable for a particular application. The hardware may include a general-purpose computer and/or dedicated computing device or specific computing device or particular aspect or component of a specific computing device. The processes may be realized in one or more microprocessors, microcontrollers, embedded microcontrollers, programmable digital signal processors or other programmable device, along with internal and/or external memory. The processes may also, or instead, be embodied in an application specific integrated circuit, a programmable gate array, programmable array logic, or any other device or combination of devices that may be configured to process electronic signals. It will further be appreciated that one or more of the processes may be realized as a computer executable code capable of being executed on a machine-readable medium.
  • The computer executable code may be created using a structured programming language such as C, an object oriented programming language such as C++, or any other high-level or low-level programming language (including assembly languages, hardware description languages, and database programming languages and technologies) that may be stored, compiled or interpreted to run on one of the above devices, as well as heterogeneous combinations of processors, processor architectures, or combinations of different hardware and software, or any other machine capable of executing program instructions.
  • Thus, in one aspect, each method described above and combinations thereof may be embodied in computer executable code that, when executing on one or more computing devices, performs the steps thereof. In another aspect, the methods may be embodied in systems that perform the steps thereof, and may be distributed across devices in a number of ways, or all of the functionality may be integrated into a dedicated, standalone device or other hardware. In another aspect, the means for performing the steps associated with the processes described above may include any of the hardware and/or software described above. All such permutations and combinations are intended to fall within the scope of the present disclosure.
  • While the invention has been disclosed in connection with the preferred embodiments shown and described in detail, various modifications and improvements thereon will become readily apparent to those skilled in the art. Accordingly, the spirit and scope of the present invention is not to be limited by the foregoing examples, but is to be understood in the broadest sense allowable by law.
  • All documents referenced herein are hereby incorporated by reference.

Claims (20)

What is claimed is:
1. A system for implementing a natural language interface, the system comprising:
at least one server computer comprising at least one processor and at least one memory, the at least one server computer configured to:
deploy the natural language interface, wherein the natural language interface is configured to recognize a list of intents;
collect usage data for the natural language interface, wherein the usage data comprises messages received from users of the natural language interface and intents selected by the natural language interface as corresponding to the messages;
obtain a first set of messages from the usage data, wherein the first set of messages corresponds to a first intent of the list of intents;
cluster the first set of messages into a plurality of clusters;
obtain a first prototype message for a first cluster of the plurality of clusters, wherein the first prototype message represents messages of the first cluster;
present the first prototype message;
receive an input to (i) create a new intent using the first cluster or (ii) move the first cluster to a second intent of the list of intents;
update the natural language interface using the received input;
deploy the updated natural language interface.
2. The system of claim 1, wherein the system is implemented by a third-party company providing services to a company and the at least one server computer is configured to:
provide a user interface to the company to allow the company to configure the natural language interface; and
receive a configuration for the natural language interface from the company.
3. The system of claim 2, wherein the at least one server computer is configured to:
receive a second configuration for a second natural language interface for a second company; and
deploy the second natural language interface, wherein the second natural language interface is used by customers of the second company.
4. The system of claim 1, wherein the usage data includes an indication of success for one or more of the messages.
5. The system of claim 4, wherein the at least one server computer is configured to compute a success rate for the first cluster.
6. The system of claim 4, wherein the at least one server computer is configured to compute a success rate for the first intent.
7. The system of claim 4, wherein a first indication of success for a first message of a first user is determined by one or more of:
asking the first user if a first interaction corresponding to the first message was a success;
asking a customer service representative assisting the first user if the first interaction was a success; or
determining that the first user completed an action.
8. The system of claim 1, wherein the at least one server computer is configured to:
determine to present a suggestion for improving the natural language interface;
present a suggestion (i) to create the new intent using the first cluster or (ii) to move the first cluster to the second intent.
9. The system of claim 1, wherein the natural language interface processes a message received from a user, determines an intent of the message, selects an action using the intent, and causes the action to be implemented with the user.
10. A computer-implemented method for implementing a natural language interface, the method comprising:
deploying the natural language interface, wherein the natural language interface is configured to recognize a list of intents;
collecting usage data for the natural language interface, wherein the usage data comprises wherein the usage data comprises messages received from users of the natural language interface and intents selected by the natural language interface as corresponding to the messages;
obtaining a first set of messages from the usage data, wherein the first set of messages corresponds to a first intent of the list of intents,
clustering the first set of messages into a plurality of clusters, and
obtaining a first prototype message for a first cluster of the plurality of clusters, wherein the first prototype message represents messages of the first cluster;
presenting the first prototype message;
receiving an input to (i) create a new intent using a first cluster or (ii) move the first cluster to a second intent of the list of intents;
updating the natural language interface using the received input; and
deploying the updated natural language interface.
11. The method of claim 10, wherein the natural language interface comprises a recurrent neural network.
12. The method of claim 10, wherein the natural language interface comprises a specification of an action to be performed with a user for each intent of the list of intents.
13. The method of claim 10, wherein clustering the first set of messages comprises:
computing message embeddings for the first set of messages; and
clustering the message embeddings.
14. The method of claim 10, wherein obtaining the first prototype message comprises selecting a message with a largest count or selecting a message closest to a center of the first cluster.
15. The method of claim 10, wherein obtaining the first prototype message comprises generating the first prototype message with a neural network.
16. The method of claim 10, comprising computing a success rate for the first intent.
17. One or more non-transitory computer-readable media comprising computer executable instructions that, when executed, cause at least one processor to perform actions comprising:
deploying a natural language interface, wherein the natural language interface is configured to recognize a list of intents;
collecting usage data for the natural language interface, wherein the usage data comprises wherein the usage data comprises messages received from users of the natural language interface and intents selected by the natural language interface as corresponding to the messages;
obtaining a first set of messages from the usage data, wherein the first set of messages corresponds to a first intent of the list of intents,
clustering the first set of messages into a plurality of clusters, and
obtaining a first prototype message for a first cluster of the plurality of clusters, wherein the first prototype message represents messages of the first cluster;
presenting the first prototype message;
receiving an input to (i) create a new intent using a first cluster or (ii) move the first cluster to a second intent of the list of intents;
updating the natural language interface using the received input; and
deploying the updated natural language interface.
18. The one or more non-transitory computer-readable media of claim 17, wherein the natural language interface was trained using a plurality of seed messages for each intent of the list of intents.
19. The one or more non-transitory computer-readable media of claim 17, wherein the actions comprise:
training an intent classifier using a plurality of seed messages for each intent of the list of intents; and
wherein the natural language interface comprises the intent classifier.
20. The one or more non-transitory computer-readable media of claim 17, wherein the actions comprise presenting a suggestion (i) to create the new intent using the first cluster or (ii) to move the first cluster to the second intent.
US16/680,906 2018-02-12 2019-11-12 Updating natural language interfaces by processing usage data Abandoned US20200081907A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/680,906 US20200081907A1 (en) 2018-02-12 2019-11-12 Updating natural language interfaces by processing usage data

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US15/894,504 US10210244B1 (en) 2018-02-12 2018-02-12 Updating natural language interfaces by processing usage data
US16/240,958 US10515104B2 (en) 2018-02-12 2019-01-07 Updating natural language interfaces by processing usage data
US16/680,906 US20200081907A1 (en) 2018-02-12 2019-11-12 Updating natural language interfaces by processing usage data

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/240,958 Continuation US10515104B2 (en) 2018-02-12 2019-01-07 Updating natural language interfaces by processing usage data

Publications (1)

Publication Number Publication Date
US20200081907A1 true US20200081907A1 (en) 2020-03-12

Family

ID=65322256

Family Applications (3)

Application Number Title Priority Date Filing Date
US15/894,504 Active US10210244B1 (en) 2018-02-12 2018-02-12 Updating natural language interfaces by processing usage data
US16/240,958 Active US10515104B2 (en) 2018-02-12 2019-01-07 Updating natural language interfaces by processing usage data
US16/680,906 Abandoned US20200081907A1 (en) 2018-02-12 2019-11-12 Updating natural language interfaces by processing usage data

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US15/894,504 Active US10210244B1 (en) 2018-02-12 2018-02-12 Updating natural language interfaces by processing usage data
US16/240,958 Active US10515104B2 (en) 2018-02-12 2019-01-07 Updating natural language interfaces by processing usage data

Country Status (4)

Country Link
US (3) US10210244B1 (en)
EP (1) EP3724792A4 (en)
CN (1) CN111801674A (en)
WO (1) WO2019156803A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022189925A1 (en) * 2021-03-11 2022-09-15 International Business Machines Corporation Generating natural languages interface from graphic user interfaces
US11615422B2 (en) 2016-07-08 2023-03-28 Asapp, Inc. Automatically suggesting completions of text
US11790376B2 (en) 2016-07-08 2023-10-17 Asapp, Inc. Predicting customer support requests

Families Citing this family (384)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070084897A1 (en) 2003-05-20 2007-04-19 Shelton Frederick E Iv Articulating surgical stapling instrument incorporating a two-piece e-beam firing mechanism
US9060770B2 (en) 2003-05-20 2015-06-23 Ethicon Endo-Surgery, Inc. Robotically-driven surgical instrument with E-beam driver
US11896225B2 (en) 2004-07-28 2024-02-13 Cilag Gmbh International Staple cartridge comprising a pan
US8215531B2 (en) 2004-07-28 2012-07-10 Ethicon Endo-Surgery, Inc. Surgical stapling instrument having a medical substance dispenser
US11246590B2 (en) 2005-08-31 2022-02-15 Cilag Gmbh International Staple cartridge including staple drivers having different unfired heights
US11484312B2 (en) 2005-08-31 2022-11-01 Cilag Gmbh International Staple cartridge comprising a staple driver arrangement
US7934630B2 (en) 2005-08-31 2011-05-03 Ethicon Endo-Surgery, Inc. Staple cartridges for forming staples having differing formed staple heights
US9237891B2 (en) 2005-08-31 2016-01-19 Ethicon Endo-Surgery, Inc. Robotically-controlled surgical stapling devices that produce formed staples having different lengths
US10159482B2 (en) 2005-08-31 2018-12-25 Ethicon Llc Fastener cartridge assembly comprising a fixed anvil and different staple heights
US7669746B2 (en) 2005-08-31 2010-03-02 Ethicon Endo-Surgery, Inc. Staple cartridges for forming staples having differing formed staple heights
US20070106317A1 (en) 2005-11-09 2007-05-10 Shelton Frederick E Iv Hydraulically and electrically actuated articulation joints for surgical instruments
US7753904B2 (en) 2006-01-31 2010-07-13 Ethicon Endo-Surgery, Inc. Endoscopic surgical instrument with a handle that can articulate with respect to the shaft
US20110024477A1 (en) 2009-02-06 2011-02-03 Hall Steven G Driven Surgical Stapler Improvements
US11224427B2 (en) 2006-01-31 2022-01-18 Cilag Gmbh International Surgical stapling system including a console and retraction assembly
US11278279B2 (en) 2006-01-31 2022-03-22 Cilag Gmbh International Surgical instrument assembly
US20120292367A1 (en) 2006-01-31 2012-11-22 Ethicon Endo-Surgery, Inc. Robotically-controlled end effector
US8186555B2 (en) 2006-01-31 2012-05-29 Ethicon Endo-Surgery, Inc. Motor-driven surgical cutting and fastening instrument with mechanical closure system
US20110295295A1 (en) 2006-01-31 2011-12-01 Ethicon Endo-Surgery, Inc. Robotically-controlled surgical instrument having recording capabilities
US8820603B2 (en) 2006-01-31 2014-09-02 Ethicon Endo-Surgery, Inc. Accessing data stored in a memory of a surgical instrument
US8708213B2 (en) 2006-01-31 2014-04-29 Ethicon Endo-Surgery, Inc. Surgical instrument having a feedback system
US7845537B2 (en) 2006-01-31 2010-12-07 Ethicon Endo-Surgery, Inc. Surgical instrument having recording capabilities
US11793518B2 (en) 2006-01-31 2023-10-24 Cilag Gmbh International Powered surgical instruments with firing system lockout arrangements
US8992422B2 (en) 2006-03-23 2015-03-31 Ethicon Endo-Surgery, Inc. Robotically-controlled endoscopic accessory channel
US8322455B2 (en) 2006-06-27 2012-12-04 Ethicon Endo-Surgery, Inc. Manually driven surgical cutting and fastening instrument
US10568652B2 (en) 2006-09-29 2020-02-25 Ethicon Llc Surgical staples having attached drivers of different heights and stapling instruments for deploying the same
US8652120B2 (en) 2007-01-10 2014-02-18 Ethicon Endo-Surgery, Inc. Surgical instrument with wireless communication between control unit and sensor transponders
US8684253B2 (en) 2007-01-10 2014-04-01 Ethicon Endo-Surgery, Inc. Surgical instrument with wireless communication between a control unit of a robotic system and remote sensor
US11291441B2 (en) 2007-01-10 2022-04-05 Cilag Gmbh International Surgical instrument with wireless communication between control unit and remote sensor
US8701958B2 (en) 2007-01-11 2014-04-22 Ethicon Endo-Surgery, Inc. Curved end effector for a surgical stapling device
US11039836B2 (en) 2007-01-11 2021-06-22 Cilag Gmbh International Staple cartridge for use with a surgical stapling instrument
US7669747B2 (en) 2007-03-15 2010-03-02 Ethicon Endo-Surgery, Inc. Washer for use with a surgical stapling instrument
US8931682B2 (en) 2007-06-04 2015-01-13 Ethicon Endo-Surgery, Inc. Robotically-controlled shaft based rotary drive systems for surgical instruments
US11564682B2 (en) 2007-06-04 2023-01-31 Cilag Gmbh International Surgical stapler device
US7753245B2 (en) 2007-06-22 2010-07-13 Ethicon Endo-Surgery, Inc. Surgical stapling instruments
US11849941B2 (en) 2007-06-29 2023-12-26 Cilag Gmbh International Staple cartridge having staple cavities extending at a transverse angle relative to a longitudinal cartridge axis
US8573465B2 (en) 2008-02-14 2013-11-05 Ethicon Endo-Surgery, Inc. Robotically-controlled surgical end effector system with rotary actuated closure systems
US7866527B2 (en) 2008-02-14 2011-01-11 Ethicon Endo-Surgery, Inc. Surgical stapling apparatus with interlockable firing system
US8758391B2 (en) 2008-02-14 2014-06-24 Ethicon Endo-Surgery, Inc. Interchangeable tools for surgical instruments
JP5410110B2 (en) 2008-02-14 2014-02-05 エシコン・エンド−サージェリィ・インコーポレイテッド Surgical cutting / fixing instrument with RF electrode
US9179912B2 (en) 2008-02-14 2015-11-10 Ethicon Endo-Surgery, Inc. Robotically-controlled motorized surgical cutting and fastening instrument
US7819298B2 (en) 2008-02-14 2010-10-26 Ethicon Endo-Surgery, Inc. Surgical stapling apparatus with control features operable with one hand
US8636736B2 (en) 2008-02-14 2014-01-28 Ethicon Endo-Surgery, Inc. Motorized surgical cutting and fastening instrument
US20130153641A1 (en) 2008-02-15 2013-06-20 Ethicon Endo-Surgery, Inc. Releasable layer of material and surgical end effector having the same
US9386983B2 (en) 2008-09-23 2016-07-12 Ethicon Endo-Surgery, Llc Robotically-controlled motorized surgical instrument
US8210411B2 (en) 2008-09-23 2012-07-03 Ethicon Endo-Surgery, Inc. Motor-driven surgical cutting instrument
US11648005B2 (en) 2008-09-23 2023-05-16 Cilag Gmbh International Robotically-controlled motorized surgical instrument with an end effector
US9005230B2 (en) 2008-09-23 2015-04-14 Ethicon Endo-Surgery, Inc. Motorized surgical instrument
US8608045B2 (en) 2008-10-10 2013-12-17 Ethicon Endo-Sugery, Inc. Powered surgical cutting and stapling apparatus with manually retractable firing system
US8517239B2 (en) 2009-02-05 2013-08-27 Ethicon Endo-Surgery, Inc. Surgical stapling instrument comprising a magnetic element driver
EP2393430A1 (en) 2009-02-06 2011-12-14 Ethicon Endo-Surgery, Inc. Driven surgical stapler improvements
US8220688B2 (en) 2009-12-24 2012-07-17 Ethicon Endo-Surgery, Inc. Motor-driven surgical cutting instrument with electric actuator directional control assembly
US8851354B2 (en) 2009-12-24 2014-10-07 Ethicon Endo-Surgery, Inc. Surgical cutting instrument that analyzes tissue thickness
US8783543B2 (en) 2010-07-30 2014-07-22 Ethicon Endo-Surgery, Inc. Tissue acquisition arrangements and methods for surgical stapling devices
US9386988B2 (en) 2010-09-30 2016-07-12 Ethicon End-Surgery, LLC Retainer assembly including a tissue thickness compensator
US11812965B2 (en) 2010-09-30 2023-11-14 Cilag Gmbh International Layer of material for a surgical end effector
US11298125B2 (en) 2010-09-30 2022-04-12 Cilag Gmbh International Tissue stapler having a thickness compensator
US10945731B2 (en) 2010-09-30 2021-03-16 Ethicon Llc Tissue thickness compensator comprising controlled release and expansion
US9566061B2 (en) 2010-09-30 2017-02-14 Ethicon Endo-Surgery, Llc Fastener cartridge comprising a releasably attached tissue thickness compensator
US9629814B2 (en) 2010-09-30 2017-04-25 Ethicon Endo-Surgery, Llc Tissue thickness compensator configured to redistribute compressive forces
US9839420B2 (en) 2010-09-30 2017-12-12 Ethicon Llc Tissue thickness compensator comprising at least one medicament
US8746535B2 (en) 2010-09-30 2014-06-10 Ethicon Endo-Surgery, Inc. Tissue thickness compensator comprising detachable portions
US11849952B2 (en) 2010-09-30 2023-12-26 Cilag Gmbh International Staple cartridge comprising staples positioned within a compressible portion thereof
US8695866B2 (en) 2010-10-01 2014-04-15 Ethicon Endo-Surgery, Inc. Surgical instrument having a power control circuit
AU2012250197B2 (en) 2011-04-29 2017-08-10 Ethicon Endo-Surgery, Inc. Staple cartridge comprising staples positioned within a compressible portion thereof
US11207064B2 (en) 2011-05-27 2021-12-28 Cilag Gmbh International Automated end effector component reloading system for use with a robotic system
US9072535B2 (en) 2011-05-27 2015-07-07 Ethicon Endo-Surgery, Inc. Surgical stapling instruments with rotatable staple deployment arrangements
RU2644272C2 (en) 2012-03-28 2018-02-08 Этикон Эндо-Серджери, Инк. Limitation node with tissue thickness compensator
MX358135B (en) 2012-03-28 2018-08-06 Ethicon Endo Surgery Inc Tissue thickness compensator comprising a plurality of layers.
RU2639857C2 (en) 2012-03-28 2017-12-22 Этикон Эндо-Серджери, Инк. Tissue thickness compensator containing capsule for medium with low pressure
US9101358B2 (en) 2012-06-15 2015-08-11 Ethicon Endo-Surgery, Inc. Articulatable surgical instrument comprising a firing drive
US9649111B2 (en) 2012-06-28 2017-05-16 Ethicon Endo-Surgery, Llc Replaceable clip cartridge for a clip applier
US9226751B2 (en) 2012-06-28 2016-01-05 Ethicon Endo-Surgery, Inc. Surgical instrument system including replaceable end effectors
BR112014032776B1 (en) 2012-06-28 2021-09-08 Ethicon Endo-Surgery, Inc SURGICAL INSTRUMENT SYSTEM AND SURGICAL KIT FOR USE WITH A SURGICAL INSTRUMENT SYSTEM
US11202631B2 (en) 2012-06-28 2021-12-21 Cilag Gmbh International Stapling assembly comprising a firing lockout
US20140001231A1 (en) 2012-06-28 2014-01-02 Ethicon Endo-Surgery, Inc. Firing system lockout arrangements for surgical instruments
US9289256B2 (en) 2012-06-28 2016-03-22 Ethicon Endo-Surgery, Llc Surgical end effectors having angled tissue-contacting surfaces
EP2866686A1 (en) 2012-06-28 2015-05-06 Ethicon Endo-Surgery, Inc. Empty clip cartridge lockout
RU2669463C2 (en) 2013-03-01 2018-10-11 Этикон Эндо-Серджери, Инк. Surgical instrument with soft stop
RU2672520C2 (en) 2013-03-01 2018-11-15 Этикон Эндо-Серджери, Инк. Hingedly turnable surgical instruments with conducting ways for signal transfer
US9629629B2 (en) 2013-03-14 2017-04-25 Ethicon Endo-Surgey, LLC Control systems for surgical instruments
US9332987B2 (en) 2013-03-14 2016-05-10 Ethicon Endo-Surgery, Llc Control arrangements for a drive member of a surgical instrument
BR112015026109B1 (en) 2013-04-16 2022-02-22 Ethicon Endo-Surgery, Inc surgical instrument
US9867612B2 (en) 2013-04-16 2018-01-16 Ethicon Llc Powered surgical stapler
MX369362B (en) 2013-08-23 2019-11-06 Ethicon Endo Surgery Llc Firing member retraction devices for powered surgical instruments.
US9775609B2 (en) 2013-08-23 2017-10-03 Ethicon Llc Tamper proof circuit for surgical instrument battery pack
US9962161B2 (en) 2014-02-12 2018-05-08 Ethicon Llc Deliverable surgical instrument
BR112016021943B1 (en) 2014-03-26 2022-06-14 Ethicon Endo-Surgery, Llc SURGICAL INSTRUMENT FOR USE BY AN OPERATOR IN A SURGICAL PROCEDURE
US20150272557A1 (en) 2014-03-26 2015-10-01 Ethicon Endo-Surgery, Inc. Modular surgical instrument system
US9826977B2 (en) 2014-03-26 2017-11-28 Ethicon Llc Sterilization verification circuit
CN106456176B (en) 2014-04-16 2019-06-28 伊西康内外科有限责任公司 Fastener cartridge including the extension with various configuration
US20150297223A1 (en) 2014-04-16 2015-10-22 Ethicon Endo-Surgery, Inc. Fastener cartridges including extensions having different configurations
US9844369B2 (en) 2014-04-16 2017-12-19 Ethicon Llc Surgical end effectors with firing element monitoring arrangements
JP6612256B2 (en) 2014-04-16 2019-11-27 エシコン エルエルシー Fastener cartridge with non-uniform fastener
JP6532889B2 (en) 2014-04-16 2019-06-19 エシコン エルエルシーEthicon LLC Fastener cartridge assembly and staple holder cover arrangement
US11311294B2 (en) 2014-09-05 2022-04-26 Cilag Gmbh International Powered medical device including measurement of closure state of jaws
US20160066913A1 (en) 2014-09-05 2016-03-10 Ethicon Endo-Surgery, Inc. Local display of tissue parameter stabilization
BR112017004361B1 (en) 2014-09-05 2023-04-11 Ethicon Llc ELECTRONIC SYSTEM FOR A SURGICAL INSTRUMENT
US10105142B2 (en) 2014-09-18 2018-10-23 Ethicon Llc Surgical stapler with plurality of cutting elements
JP6648119B2 (en) 2014-09-26 2020-02-14 エシコン エルエルシーEthicon LLC Surgical stapling buttress and accessory materials
US11523821B2 (en) 2014-09-26 2022-12-13 Cilag Gmbh International Method for creating a flexible staple line
US10076325B2 (en) 2014-10-13 2018-09-18 Ethicon Llc Surgical stapling apparatus comprising a tissue stop
US9924944B2 (en) 2014-10-16 2018-03-27 Ethicon Llc Staple cartridge comprising an adjunct material
US11141153B2 (en) 2014-10-29 2021-10-12 Cilag Gmbh International Staple cartridges comprising driver arrangements
US10517594B2 (en) 2014-10-29 2019-12-31 Ethicon Llc Cartridge assemblies for surgical staplers
US9844376B2 (en) 2014-11-06 2017-12-19 Ethicon Llc Staple cartridge comprising a releasable adjunct material
US10736636B2 (en) 2014-12-10 2020-08-11 Ethicon Llc Articulatable surgical instrument system
US10085748B2 (en) 2014-12-18 2018-10-02 Ethicon Llc Locking arrangements for detachable shaft assemblies with articulatable surgical end effectors
US10004501B2 (en) 2014-12-18 2018-06-26 Ethicon Llc Surgical instruments with improved closure arrangements
RU2703684C2 (en) 2014-12-18 2019-10-21 ЭТИКОН ЭНДО-СЕРДЖЕРИ, ЭлЭлСи Surgical instrument with anvil which is selectively movable relative to staple cartridge around discrete fixed axis
US9987000B2 (en) 2014-12-18 2018-06-05 Ethicon Llc Surgical instrument assembly comprising a flexible articulation system
US9844375B2 (en) 2014-12-18 2017-12-19 Ethicon Llc Drive arrangements for articulatable surgical instruments
US9844374B2 (en) 2014-12-18 2017-12-19 Ethicon Llc Surgical instrument systems comprising an articulatable end effector and means for adjusting the firing stroke of a firing member
US11154301B2 (en) 2015-02-27 2021-10-26 Cilag Gmbh International Modular stapling assembly
US10245033B2 (en) 2015-03-06 2019-04-02 Ethicon Llc Surgical instrument comprising a lockable battery housing
US9901342B2 (en) 2015-03-06 2018-02-27 Ethicon Endo-Surgery, Llc Signal and power communication system positioned on a rotatable shaft
US10052044B2 (en) 2015-03-06 2018-08-21 Ethicon Llc Time dependent evaluation of sensor data to determine stability, creep, and viscoelastic elements of measures
US9993248B2 (en) 2015-03-06 2018-06-12 Ethicon Endo-Surgery, Llc Smart sensors with local signal processing
US10617412B2 (en) 2015-03-06 2020-04-14 Ethicon Llc System for detecting the mis-insertion of a staple cartridge into a surgical stapler
JP2020121162A (en) 2015-03-06 2020-08-13 エシコン エルエルシーEthicon LLC Time dependent evaluation of sensor data to determine stability element, creep element and viscoelastic element of measurement
US10687806B2 (en) 2015-03-06 2020-06-23 Ethicon Llc Adaptive tissue compression techniques to adjust closure rates for multiple tissue types
US10441279B2 (en) 2015-03-06 2019-10-15 Ethicon Llc Multiple level thresholds to modify operation of powered surgical instruments
US10433844B2 (en) 2015-03-31 2019-10-08 Ethicon Llc Surgical instrument with selectively disengageable threaded drive systems
US10755294B1 (en) 2015-04-28 2020-08-25 Intuit Inc. Method and system for increasing use of mobile devices to provide answer content in a question and answer based customer support system
US10475044B1 (en) 2015-07-29 2019-11-12 Intuit Inc. Method and system for question prioritization based on analysis of the question content and predicted asker engagement before answer content is generated
US10835249B2 (en) 2015-08-17 2020-11-17 Ethicon Llc Implantable layers for a surgical instrument
US10238386B2 (en) 2015-09-23 2019-03-26 Ethicon Llc Surgical stapler having motor control based on an electrical parameter related to a motor current
US10105139B2 (en) 2015-09-23 2018-10-23 Ethicon Llc Surgical stapler having downstream current-based motor control
US10299878B2 (en) 2015-09-25 2019-05-28 Ethicon Llc Implantable adjunct systems for determining adjunct skew
US10271849B2 (en) 2015-09-30 2019-04-30 Ethicon Llc Woven constructs with interlocked standing fibers
US10603039B2 (en) 2015-09-30 2020-03-31 Ethicon Llc Progressively releasable implantable adjunct for use with a surgical stapling instrument
US10980539B2 (en) 2015-09-30 2021-04-20 Ethicon Llc Implantable adjunct comprising bonded layers
US11890015B2 (en) 2015-09-30 2024-02-06 Cilag Gmbh International Compressible adjunct with crossing spacer fibers
US10368865B2 (en) 2015-12-30 2019-08-06 Ethicon Llc Mechanisms for compensating for drivetrain failure in powered surgical instruments
US10292704B2 (en) 2015-12-30 2019-05-21 Ethicon Llc Mechanisms for compensating for battery pack failure in powered surgical instruments
US10265068B2 (en) 2015-12-30 2019-04-23 Ethicon Llc Surgical instruments with separable motors and motor control circuits
US11213293B2 (en) 2016-02-09 2022-01-04 Cilag Gmbh International Articulatable surgical instruments with single articulation link arrangements
BR112018016098B1 (en) 2016-02-09 2023-02-23 Ethicon Llc SURGICAL INSTRUMENT
US11224426B2 (en) 2016-02-12 2022-01-18 Cilag Gmbh International Mechanisms for compensating for drivetrain failure in powered surgical instruments
US10448948B2 (en) 2016-02-12 2019-10-22 Ethicon Llc Mechanisms for compensating for drivetrain failure in powered surgical instruments
US11179150B2 (en) 2016-04-15 2021-11-23 Cilag Gmbh International Systems and methods for controlling a surgical stapling and cutting instrument
US10492783B2 (en) 2016-04-15 2019-12-03 Ethicon, Llc Surgical instrument with improved stop/start control during a firing motion
US11607239B2 (en) 2016-04-15 2023-03-21 Cilag Gmbh International Systems and methods for controlling a surgical stapling and cutting instrument
US10335145B2 (en) 2016-04-15 2019-07-02 Ethicon Llc Modular surgical instrument with configurable operating mode
US10828028B2 (en) 2016-04-15 2020-11-10 Ethicon Llc Surgical instrument with multiple program responses during a firing motion
US10426467B2 (en) 2016-04-15 2019-10-01 Ethicon Llc Surgical instrument with detection sensors
US10456137B2 (en) 2016-04-15 2019-10-29 Ethicon Llc Staple formation detection mechanisms
US10357247B2 (en) 2016-04-15 2019-07-23 Ethicon Llc Surgical instrument with multiple program responses during a firing motion
US11317917B2 (en) 2016-04-18 2022-05-03 Cilag Gmbh International Surgical stapling system comprising a lockable firing assembly
US10426469B2 (en) 2016-04-18 2019-10-01 Ethicon Llc Surgical instrument comprising a primary firing lockout and a secondary firing lockout
US20170296173A1 (en) 2016-04-18 2017-10-19 Ethicon Endo-Surgery, Llc Method for operating a surgical instrument
US10733677B2 (en) 2016-10-18 2020-08-04 Intuit Inc. Method and system for providing domain-specific and dynamic type ahead suggestions for search query terms with a customer self-service system for a tax return preparation system
US10552843B1 (en) 2016-12-05 2020-02-04 Intuit Inc. Method and system for improving search results by recency boosting customer support content for a customer self-help system associated with one or more financial management systems
US10109275B2 (en) 2016-12-19 2018-10-23 Asapp, Inc. Word hash language model
US10568625B2 (en) 2016-12-21 2020-02-25 Ethicon Llc Staple cartridges and arrangements of staples and staple cavities therein
JP6983893B2 (en) 2016-12-21 2021-12-17 エシコン エルエルシーEthicon LLC Lockout configuration for surgical end effectors and replaceable tool assemblies
US11090048B2 (en) 2016-12-21 2021-08-17 Cilag Gmbh International Method for resetting a fuse of a surgical instrument shaft
US10610224B2 (en) 2016-12-21 2020-04-07 Ethicon Llc Lockout arrangements for surgical end effectors and replaceable tool assemblies
MX2019007311A (en) 2016-12-21 2019-11-18 Ethicon Llc Surgical stapling systems.
US10695055B2 (en) 2016-12-21 2020-06-30 Ethicon Llc Firing assembly comprising a lockout
US10888322B2 (en) 2016-12-21 2021-01-12 Ethicon Llc Surgical instrument comprising a cutting member
US10675025B2 (en) 2016-12-21 2020-06-09 Ethicon Llc Shaft assembly comprising separately actuatable and retractable systems
US11419606B2 (en) 2016-12-21 2022-08-23 Cilag Gmbh International Shaft assembly comprising a clutch configured to adapt the output of a rotary firing member to two different systems
JP7010956B2 (en) 2016-12-21 2022-01-26 エシコン エルエルシー How to staple tissue
US11191540B2 (en) 2016-12-21 2021-12-07 Cilag Gmbh International Protective cover arrangements for a joint interface between a movable jaw and actuator shaft of a surgical instrument
US20180168598A1 (en) 2016-12-21 2018-06-21 Ethicon Endo-Surgery, Llc Staple forming pocket arrangements comprising zoned forming surface grooves
US10588630B2 (en) 2016-12-21 2020-03-17 Ethicon Llc Surgical tool assemblies with closure stroke reduction features
US20180168625A1 (en) 2016-12-21 2018-06-21 Ethicon Endo-Surgery, Llc Surgical stapling instruments with smart staple cartridges
US10568624B2 (en) 2016-12-21 2020-02-25 Ethicon Llc Surgical instruments with jaws that are pivotable about a fixed axis and include separate and distinct closure and firing systems
US20180168615A1 (en) 2016-12-21 2018-06-21 Ethicon Endo-Surgery, Llc Method of deforming staples from two different types of staple cartridges with the same surgical stapling instrument
US10856868B2 (en) 2016-12-21 2020-12-08 Ethicon Llc Firing member pin configurations
US11134942B2 (en) 2016-12-21 2021-10-05 Cilag Gmbh International Surgical stapling instruments and staple-forming anvils
US10748157B1 (en) 2017-01-12 2020-08-18 Intuit Inc. Method and system for determining levels of search sophistication for users of a customer self-help system to personalize a content search user experience provided to the users and to increase a likelihood of user satisfaction with the search experience
US10307170B2 (en) 2017-06-20 2019-06-04 Ethicon Llc Method for closed loop control of motor velocity of a surgical stapling and cutting instrument
US11090046B2 (en) 2017-06-20 2021-08-17 Cilag Gmbh International Systems and methods for controlling displacement member motion of a surgical stapling and cutting instrument
US10980537B2 (en) 2017-06-20 2021-04-20 Ethicon Llc Closed loop feedback control of motor velocity of a surgical stapling and cutting instrument based on measured time over a specified number of shaft rotations
USD890784S1 (en) 2017-06-20 2020-07-21 Ethicon Llc Display panel with changeable graphical user interface
US11517325B2 (en) 2017-06-20 2022-12-06 Cilag Gmbh International Closed loop feedback control of motor velocity of a surgical stapling and cutting instrument based on measured displacement distance traveled over a specified time interval
US10779820B2 (en) 2017-06-20 2020-09-22 Ethicon Llc Systems and methods for controlling motor speed according to user input for a surgical instrument
US11653914B2 (en) 2017-06-20 2023-05-23 Cilag Gmbh International Systems and methods for controlling motor velocity of a surgical stapling and cutting instrument according to articulation angle of end effector
US11382638B2 (en) 2017-06-20 2022-07-12 Cilag Gmbh International Closed loop feedback control of motor velocity of a surgical stapling and cutting instrument based on measured time over a specified displacement distance
US10881399B2 (en) 2017-06-20 2021-01-05 Ethicon Llc Techniques for adaptive control of motor velocity of a surgical stapling and cutting instrument
US10888321B2 (en) 2017-06-20 2021-01-12 Ethicon Llc Systems and methods for controlling velocity of a displacement member of a surgical stapling and cutting instrument
US11071554B2 (en) 2017-06-20 2021-07-27 Cilag Gmbh International Closed loop feedback control of motor velocity of a surgical stapling and cutting instrument based on magnitude of velocity error measurements
US11324503B2 (en) 2017-06-27 2022-05-10 Cilag Gmbh International Surgical firing member arrangements
US11266405B2 (en) 2017-06-27 2022-03-08 Cilag Gmbh International Surgical anvil manufacturing methods
US10631859B2 (en) 2017-06-27 2020-04-28 Ethicon Llc Articulation systems for surgical instruments
US10993716B2 (en) 2017-06-27 2021-05-04 Ethicon Llc Surgical anvil arrangements
US10856869B2 (en) 2017-06-27 2020-12-08 Ethicon Llc Surgical anvil arrangements
US10765427B2 (en) 2017-06-28 2020-09-08 Ethicon Llc Method for articulating a surgical instrument
USD906355S1 (en) 2017-06-28 2020-12-29 Ethicon Llc Display screen or portion thereof with a graphical user interface for a surgical instrument
US20190000461A1 (en) 2017-06-28 2019-01-03 Ethicon Llc Surgical cutting and fastening devices with pivotable anvil with a tissue locating arrangement in close proximity to an anvil pivot axis
US11564686B2 (en) 2017-06-28 2023-01-31 Cilag Gmbh International Surgical shaft assemblies with flexible interfaces
US11678880B2 (en) 2017-06-28 2023-06-20 Cilag Gmbh International Surgical instrument comprising a shaft including a housing arrangement
US11246592B2 (en) 2017-06-28 2022-02-15 Cilag Gmbh International Surgical instrument comprising an articulation system lockable to a frame
US10903685B2 (en) 2017-06-28 2021-01-26 Ethicon Llc Surgical shaft assemblies with slip ring assemblies forming capacitive channels
EP4070740A1 (en) 2017-06-28 2022-10-12 Cilag GmbH International Surgical instrument comprising selectively actuatable rotatable couplers
US10716614B2 (en) 2017-06-28 2020-07-21 Ethicon Llc Surgical shaft assemblies with slip ring assemblies with increased contact pressure
US11259805B2 (en) 2017-06-28 2022-03-01 Cilag Gmbh International Surgical instrument comprising firing member supports
US10932772B2 (en) 2017-06-29 2021-03-02 Ethicon Llc Methods for closed loop velocity control for robotic surgical instrument
US10898183B2 (en) 2017-06-29 2021-01-26 Ethicon Llc Robotic surgical instrument with closed loop feedback techniques for advancement of closure member during firing
US11007022B2 (en) 2017-06-29 2021-05-18 Ethicon Llc Closed loop velocity control techniques based on sensed tissue parameters for robotic surgical instrument
US10922367B2 (en) 2017-07-14 2021-02-16 Intuit Inc. Method and system for providing real time search preview personalization in data management systems
US11304695B2 (en) 2017-08-03 2022-04-19 Cilag Gmbh International Surgical system shaft interconnection
US11944300B2 (en) 2017-08-03 2024-04-02 Cilag Gmbh International Method for operating a surgical system bailout
US11471155B2 (en) 2017-08-03 2022-10-18 Cilag Gmbh International Surgical system bailout
WO2019035765A1 (en) * 2017-08-14 2019-02-21 Dathena Science Pte. Ltd. Methods, machine learning engines and file management platform systems for content and context aware data classification and security anomaly detection
US11093951B1 (en) 2017-09-25 2021-08-17 Intuit Inc. System and method for responding to search queries using customer self-help systems associated with a plurality of data management systems
US10729501B2 (en) * 2017-09-29 2020-08-04 Ethicon Llc Systems and methods for language selection of a surgical instrument
USD907647S1 (en) 2017-09-29 2021-01-12 Ethicon Llc Display screen or portion thereof with animated graphical user interface
USD917500S1 (en) 2017-09-29 2021-04-27 Ethicon Llc Display screen or portion thereof with graphical user interface
US10743872B2 (en) 2017-09-29 2020-08-18 Ethicon Llc System and methods for controlling a display of a surgical instrument
USD907648S1 (en) 2017-09-29 2021-01-12 Ethicon Llc Display screen or portion thereof with animated graphical user interface
US11399829B2 (en) 2017-09-29 2022-08-02 Cilag Gmbh International Systems and methods of initiating a power shutdown mode for a surgical instrument
US10765429B2 (en) 2017-09-29 2020-09-08 Ethicon Llc Systems and methods for providing alerts according to the operational state of a surgical instrument
US11134944B2 (en) 2017-10-30 2021-10-05 Cilag Gmbh International Surgical stapler knife motion controls
US11090075B2 (en) 2017-10-30 2021-08-17 Cilag Gmbh International Articulation features for surgical end effector
US10779903B2 (en) 2017-10-31 2020-09-22 Ethicon Llc Positive shaft rotation lock activated by jaw closure
US10842490B2 (en) 2017-10-31 2020-11-24 Ethicon Llc Cartridge body design with force reduction based on firing completion
US10497004B2 (en) 2017-12-08 2019-12-03 Asapp, Inc. Automating communications using an intent classifier
US10828033B2 (en) 2017-12-15 2020-11-10 Ethicon Llc Handheld electromechanical surgical instruments with improved motor control arrangements for positioning components of an adapter coupled thereto
US11071543B2 (en) 2017-12-15 2021-07-27 Cilag Gmbh International Surgical end effectors with clamping assemblies configured to increase jaw aperture ranges
US10687813B2 (en) 2017-12-15 2020-06-23 Ethicon Llc Adapters with firing stroke sensing arrangements for use in connection with electromechanical surgical instruments
US10869666B2 (en) 2017-12-15 2020-12-22 Ethicon Llc Adapters with control systems for controlling multiple motors of an electromechanical surgical instrument
US10779826B2 (en) 2017-12-15 2020-09-22 Ethicon Llc Methods of operating surgical end effectors
US11033267B2 (en) 2017-12-15 2021-06-15 Ethicon Llc Systems and methods of controlling a clamping member firing rate of a surgical instrument
US10966718B2 (en) 2017-12-15 2021-04-06 Ethicon Llc Dynamic clamping assemblies with improved wear characteristics for use in connection with electromechanical surgical instruments
US11197670B2 (en) 2017-12-15 2021-12-14 Cilag Gmbh International Surgical end effectors with pivotal jaws configured to touch at their respective distal ends when fully closed
US10779825B2 (en) 2017-12-15 2020-09-22 Ethicon Llc Adapters with end effector position sensing and control arrangements for use in connection with electromechanical surgical instruments
US10743875B2 (en) 2017-12-15 2020-08-18 Ethicon Llc Surgical end effectors with jaw stiffener arrangements configured to permit monitoring of firing member
US10743874B2 (en) 2017-12-15 2020-08-18 Ethicon Llc Sealed adapters for use with electromechanical surgical instruments
US10716565B2 (en) 2017-12-19 2020-07-21 Ethicon Llc Surgical instruments with dual articulation drivers
US11045270B2 (en) 2017-12-19 2021-06-29 Cilag Gmbh International Robotic attachment comprising exterior drive actuator
US11020112B2 (en) 2017-12-19 2021-06-01 Ethicon Llc Surgical tools configured for interchangeable use with different controller interfaces
US10729509B2 (en) 2017-12-19 2020-08-04 Ethicon Llc Surgical instrument comprising closure and firing locking mechanism
USD910847S1 (en) 2017-12-19 2021-02-16 Ethicon Llc Surgical instrument assembly
US10835330B2 (en) 2017-12-19 2020-11-17 Ethicon Llc Method for determining the position of a rotatable jaw of a surgical instrument attachment assembly
US11311290B2 (en) 2017-12-21 2022-04-26 Cilag Gmbh International Surgical instrument comprising an end effector dampener
US11076853B2 (en) 2017-12-21 2021-08-03 Cilag Gmbh International Systems and methods of displaying a knife position during transection for a surgical instrument
US11337691B2 (en) 2017-12-21 2022-05-24 Cilag Gmbh International Surgical instrument configured to determine firing path
US11129680B2 (en) 2017-12-21 2021-09-28 Cilag Gmbh International Surgical instrument comprising a projector
US10489792B2 (en) 2018-01-05 2019-11-26 Asapp, Inc. Maintaining quality of customer support messages
US11436642B1 (en) 2018-01-29 2022-09-06 Intuit Inc. Method and system for generating real-time personalized advertisements in data management self-help systems
US10210244B1 (en) 2018-02-12 2019-02-19 Asapp, Inc. Updating natural language interfaces by processing usage data
US11269665B1 (en) * 2018-03-28 2022-03-08 Intuit Inc. Method and system for user experience personalization in data management systems using machine learning
US11301777B1 (en) * 2018-04-19 2022-04-12 Meta Platforms, Inc. Determining stages of intent using text processing
US10169315B1 (en) 2018-04-27 2019-01-01 Asapp, Inc. Removing personal information from text using a neural network
US11194968B2 (en) * 2018-05-31 2021-12-07 Siemens Aktiengesellschaft Automatized text analysis
US10721190B2 (en) * 2018-07-31 2020-07-21 Microsoft Technology Licensing, Llc Sequence to sequence to classification model for generating recommended messages
US11216510B2 (en) 2018-08-03 2022-01-04 Asapp, Inc. Processing an incomplete message with a neural network to generate suggested messages
US11324501B2 (en) 2018-08-20 2022-05-10 Cilag Gmbh International Surgical stapling devices with improved closure members
US11253256B2 (en) 2018-08-20 2022-02-22 Cilag Gmbh International Articulatable motor powered surgical instruments with dedicated articulation motor arrangements
US10779821B2 (en) 2018-08-20 2020-09-22 Ethicon Llc Surgical stapler anvils with tissue stop features configured to avoid tissue pinch
US11207065B2 (en) 2018-08-20 2021-12-28 Cilag Gmbh International Method for fabricating surgical stapler anvils
US11083458B2 (en) 2018-08-20 2021-08-10 Cilag Gmbh International Powered surgical instruments with clutching arrangements to convert linear drive motions to rotary drive motions
USD914878S1 (en) 2018-08-20 2021-03-30 Ethicon Llc Surgical instrument anvil
US11045192B2 (en) 2018-08-20 2021-06-29 Cilag Gmbh International Fabricating techniques for surgical stapler anvils
US11039834B2 (en) 2018-08-20 2021-06-22 Cilag Gmbh International Surgical stapler anvils with staple directing protrusions and tissue stability features
US10912559B2 (en) 2018-08-20 2021-02-09 Ethicon Llc Reinforced deformable anvil tip for surgical stapler anvil
US10856870B2 (en) 2018-08-20 2020-12-08 Ethicon Llc Switching arrangements for motor powered articulatable surgical instruments
US11291440B2 (en) 2018-08-20 2022-04-05 Cilag Gmbh International Method for operating a powered articulatable surgical instrument
US10842492B2 (en) 2018-08-20 2020-11-24 Ethicon Llc Powered articulatable surgical instruments with clutching and locking arrangements for linking an articulation drive system to a firing drive system
US11061955B2 (en) * 2018-09-21 2021-07-13 Salesforce.Com, Inc. Intent classification system
US11551004B2 (en) 2018-11-13 2023-01-10 Asapp, Inc. Intent discovery with a prototype classifier
US10747957B2 (en) 2018-11-13 2020-08-18 Asapp, Inc. Processing communications using a prototype classifier
US10984781B2 (en) 2019-01-18 2021-04-20 Asapp, Inc. Identifying representative conversations using a state model
US10956474B2 (en) 2019-03-14 2021-03-23 Microsoft Technology Licensing, Llc Determination of best set of suggested responses
US11147551B2 (en) 2019-03-25 2021-10-19 Cilag Gmbh International Firing drive arrangements for surgical systems
US11696761B2 (en) 2019-03-25 2023-07-11 Cilag Gmbh International Firing drive arrangements for surgical systems
US11172929B2 (en) 2019-03-25 2021-11-16 Cilag Gmbh International Articulation drive arrangements for surgical systems
US11147553B2 (en) 2019-03-25 2021-10-19 Cilag Gmbh International Firing drive arrangements for surgical systems
US11373044B2 (en) 2019-04-12 2022-06-28 Asapp, Inc. Maintaining machine language model state across communications channels
US11903581B2 (en) 2019-04-30 2024-02-20 Cilag Gmbh International Methods for stapling tissue using a surgical instrument
US11253254B2 (en) 2019-04-30 2022-02-22 Cilag Gmbh International Shaft rotation actuator on a surgical instrument
US11452528B2 (en) 2019-04-30 2022-09-27 Cilag Gmbh International Articulation actuators for a surgical instrument
US11426251B2 (en) 2019-04-30 2022-08-30 Cilag Gmbh International Articulation directional lights on a surgical instrument
US11432816B2 (en) 2019-04-30 2022-09-06 Cilag Gmbh International Articulation pin for a surgical instrument
US11471157B2 (en) 2019-04-30 2022-10-18 Cilag Gmbh International Articulation control mapping for a surgical instrument
US11648009B2 (en) 2019-04-30 2023-05-16 Cilag Gmbh International Rotatable jaw tip for a surgical instrument
US11593589B2 (en) * 2019-05-17 2023-02-28 Robert Bosch Gmbh System and method for interpretable sequence and time-series data modeling
US11219455B2 (en) 2019-06-28 2022-01-11 Cilag Gmbh International Surgical instrument including a lockout key
US11464601B2 (en) 2019-06-28 2022-10-11 Cilag Gmbh International Surgical instrument comprising an RFID system for tracking a movable component
US11426167B2 (en) 2019-06-28 2022-08-30 Cilag Gmbh International Mechanisms for proper anvil attachment surgical stapling head assembly
US11051807B2 (en) 2019-06-28 2021-07-06 Cilag Gmbh International Packaging assembly including a particulate trap
US11627959B2 (en) 2019-06-28 2023-04-18 Cilag Gmbh International Surgical instruments including manual and powered system lockouts
US11399837B2 (en) 2019-06-28 2022-08-02 Cilag Gmbh International Mechanisms for motor control adjustments of a motorized surgical instrument
US11638587B2 (en) 2019-06-28 2023-05-02 Cilag Gmbh International RFID identification systems for surgical instruments
US11523822B2 (en) 2019-06-28 2022-12-13 Cilag Gmbh International Battery pack including a circuit interrupter
US11684434B2 (en) 2019-06-28 2023-06-27 Cilag Gmbh International Surgical RFID assemblies for instrument operational setting control
US11246678B2 (en) 2019-06-28 2022-02-15 Cilag Gmbh International Surgical stapling system having a frangible RFID tag
US11241235B2 (en) 2019-06-28 2022-02-08 Cilag Gmbh International Method of using multiple RFID chips with a surgical assembly
US11497492B2 (en) 2019-06-28 2022-11-15 Cilag Gmbh International Surgical instrument including an articulation lock
US11259803B2 (en) 2019-06-28 2022-03-01 Cilag Gmbh International Surgical stapling system having an information encryption protocol
US11478241B2 (en) 2019-06-28 2022-10-25 Cilag Gmbh International Staple cartridge including projections
US11298127B2 (en) 2019-06-28 2022-04-12 Cilag GmbH Interational Surgical stapling system having a lockout mechanism for an incompatible cartridge
US11224497B2 (en) 2019-06-28 2022-01-18 Cilag Gmbh International Surgical systems with multiple RFID tags
US11553971B2 (en) 2019-06-28 2023-01-17 Cilag Gmbh International Surgical RFID assemblies for display and communication
US11291451B2 (en) 2019-06-28 2022-04-05 Cilag Gmbh International Surgical instrument with battery compatibility verification functionality
US11298132B2 (en) 2019-06-28 2022-04-12 Cilag GmbH Inlernational Staple cartridge including a honeycomb extension
US11771419B2 (en) 2019-06-28 2023-10-03 Cilag Gmbh International Packaging for a replaceable component of a surgical stapling system
US11660163B2 (en) 2019-06-28 2023-05-30 Cilag Gmbh International Surgical system with RFID tags for updating motor assembly parameters
US11376098B2 (en) 2019-06-28 2022-07-05 Cilag Gmbh International Surgical instrument system comprising an RFID system
US11425064B2 (en) 2019-10-25 2022-08-23 Asapp, Inc. Customized message suggestion with user embedding vectors
US11593608B2 (en) 2019-10-28 2023-02-28 Paypal, Inc. Systems and methods for predicting and providing automated online chat assistance
US11304696B2 (en) 2019-12-19 2022-04-19 Cilag Gmbh International Surgical instrument comprising a powered articulation system
US11234698B2 (en) 2019-12-19 2022-02-01 Cilag Gmbh International Stapling system comprising a clamp lockout and a firing lockout
US11576672B2 (en) 2019-12-19 2023-02-14 Cilag Gmbh International Surgical instrument comprising a closure system including a closure member and an opening member driven by a drive screw
US11607219B2 (en) 2019-12-19 2023-03-21 Cilag Gmbh International Staple cartridge comprising a detachable tissue cutting knife
US11701111B2 (en) 2019-12-19 2023-07-18 Cilag Gmbh International Method for operating a surgical stapling instrument
US11504122B2 (en) 2019-12-19 2022-11-22 Cilag Gmbh International Surgical instrument comprising a nested firing member
US11911032B2 (en) 2019-12-19 2024-02-27 Cilag Gmbh International Staple cartridge comprising a seating cam
US11931033B2 (en) 2019-12-19 2024-03-19 Cilag Gmbh International Staple cartridge comprising a latch lockout
US11464512B2 (en) 2019-12-19 2022-10-11 Cilag Gmbh International Staple cartridge comprising a curved deck surface
US11529137B2 (en) 2019-12-19 2022-12-20 Cilag Gmbh International Staple cartridge comprising driver retention members
US11844520B2 (en) 2019-12-19 2023-12-19 Cilag Gmbh International Staple cartridge comprising driver retention members
US11446029B2 (en) 2019-12-19 2022-09-20 Cilag Gmbh International Staple cartridge comprising projections extending from a curved deck surface
US11529139B2 (en) 2019-12-19 2022-12-20 Cilag Gmbh International Motor driven surgical instrument
US11559304B2 (en) 2019-12-19 2023-01-24 Cilag Gmbh International Surgical instrument comprising a rapid closure mechanism
US11291447B2 (en) 2019-12-19 2022-04-05 Cilag Gmbh International Stapling instrument comprising independent jaw closing and staple firing systems
WO2021178901A1 (en) * 2020-03-05 2021-09-10 Brain Technologies, Inc. Collaboration user interface for computing device
USD967421S1 (en) 2020-06-02 2022-10-18 Cilag Gmbh International Staple cartridge
USD975850S1 (en) 2020-06-02 2023-01-17 Cilag Gmbh International Staple cartridge
USD966512S1 (en) 2020-06-02 2022-10-11 Cilag Gmbh International Staple cartridge
USD975851S1 (en) 2020-06-02 2023-01-17 Cilag Gmbh International Staple cartridge
USD974560S1 (en) 2020-06-02 2023-01-03 Cilag Gmbh International Staple cartridge
USD975278S1 (en) 2020-06-02 2023-01-10 Cilag Gmbh International Staple cartridge
USD976401S1 (en) 2020-06-02 2023-01-24 Cilag Gmbh International Staple cartridge
US11883024B2 (en) 2020-07-28 2024-01-30 Cilag Gmbh International Method of operating a surgical instrument
US11452526B2 (en) 2020-10-29 2022-09-27 Cilag Gmbh International Surgical instrument comprising a staged voltage regulation start-up system
US11534259B2 (en) 2020-10-29 2022-12-27 Cilag Gmbh International Surgical instrument comprising an articulation indicator
US11844518B2 (en) 2020-10-29 2023-12-19 Cilag Gmbh International Method for operating a surgical instrument
US11517390B2 (en) 2020-10-29 2022-12-06 Cilag Gmbh International Surgical instrument comprising a limited travel switch
USD1013170S1 (en) 2020-10-29 2024-01-30 Cilag Gmbh International Surgical instrument assembly
USD980425S1 (en) 2020-10-29 2023-03-07 Cilag Gmbh International Surgical instrument assembly
US11617577B2 (en) 2020-10-29 2023-04-04 Cilag Gmbh International Surgical instrument comprising a sensor configured to sense whether an articulation drive of the surgical instrument is actuatable
US11779330B2 (en) 2020-10-29 2023-10-10 Cilag Gmbh International Surgical instrument comprising a jaw alignment system
US11931025B2 (en) 2020-10-29 2024-03-19 Cilag Gmbh International Surgical instrument comprising a releasable closure drive lock
US11896217B2 (en) 2020-10-29 2024-02-13 Cilag Gmbh International Surgical instrument comprising an articulation lock
US11717289B2 (en) 2020-10-29 2023-08-08 Cilag Gmbh International Surgical instrument comprising an indicator which indicates that an articulation drive is actuatable
US11737751B2 (en) 2020-12-02 2023-08-29 Cilag Gmbh International Devices and methods of managing energy dissipated within sterile barriers of surgical instrument housings
US11653915B2 (en) 2020-12-02 2023-05-23 Cilag Gmbh International Surgical instruments with sled location detection and adjustment features
US11849943B2 (en) 2020-12-02 2023-12-26 Cilag Gmbh International Surgical instrument with cartridge release mechanisms
US11890010B2 (en) 2020-12-02 2024-02-06 Cllag GmbH International Dual-sided reinforced reload for surgical instruments
US11944296B2 (en) 2020-12-02 2024-04-02 Cilag Gmbh International Powered surgical instruments with external connectors
US11678882B2 (en) 2020-12-02 2023-06-20 Cilag Gmbh International Surgical instruments with interactive features to remedy incidental sled movements
US11744581B2 (en) 2020-12-02 2023-09-05 Cilag Gmbh International Powered surgical instruments with multi-phase tissue treatment
US11653920B2 (en) 2020-12-02 2023-05-23 Cilag Gmbh International Powered surgical instruments with communication interfaces through sterile barrier
US11627960B2 (en) 2020-12-02 2023-04-18 Cilag Gmbh International Powered surgical instruments with smart reload with separately attachable exteriorly mounted wiring connections
US11049023B1 (en) * 2020-12-08 2021-06-29 Moveworks, Inc. Methods and systems for evaluating and improving the content of a knowledge datastore
US11393475B1 (en) * 2021-01-13 2022-07-19 Artificial Solutions Iberia S.L Conversational system for recognizing, understanding, and acting on multiple intents and hypotheses
US11744583B2 (en) 2021-02-26 2023-09-05 Cilag Gmbh International Distal communication array to tune frequency of RF systems
US11812964B2 (en) 2021-02-26 2023-11-14 Cilag Gmbh International Staple cartridge comprising a power management circuit
US11730473B2 (en) 2021-02-26 2023-08-22 Cilag Gmbh International Monitoring of manufacturing life-cycle
US11751869B2 (en) 2021-02-26 2023-09-12 Cilag Gmbh International Monitoring of multiple sensors over time to detect moving characteristics of tissue
US11723657B2 (en) 2021-02-26 2023-08-15 Cilag Gmbh International Adjustable communication based on available bandwidth and power capacity
US11950779B2 (en) 2021-02-26 2024-04-09 Cilag Gmbh International Method of powering and communicating with a staple cartridge
US11701113B2 (en) 2021-02-26 2023-07-18 Cilag Gmbh International Stapling instrument comprising a separate power antenna and a data transfer antenna
US11696757B2 (en) 2021-02-26 2023-07-11 Cilag Gmbh International Monitoring of internal systems to detect and track cartridge motion status
US11950777B2 (en) 2021-02-26 2024-04-09 Cilag Gmbh International Staple cartridge comprising an information access control system
US11925349B2 (en) 2021-02-26 2024-03-12 Cilag Gmbh International Adjustment to transfer parameters to improve available power
US11793514B2 (en) 2021-02-26 2023-10-24 Cilag Gmbh International Staple cartridge comprising sensor array which may be embedded in cartridge body
US11749877B2 (en) 2021-02-26 2023-09-05 Cilag Gmbh International Stapling instrument comprising a signal antenna
US11759202B2 (en) 2021-03-22 2023-09-19 Cilag Gmbh International Staple cartridge comprising an implantable layer
US11806011B2 (en) 2021-03-22 2023-11-07 Cilag Gmbh International Stapling instrument comprising tissue compression systems
US11826012B2 (en) 2021-03-22 2023-11-28 Cilag Gmbh International Stapling instrument comprising a pulsed motor-driven firing rack
US11737749B2 (en) 2021-03-22 2023-08-29 Cilag Gmbh International Surgical stapling instrument comprising a retraction system
US11826042B2 (en) 2021-03-22 2023-11-28 Cilag Gmbh International Surgical instrument comprising a firing drive including a selectable leverage mechanism
US11717291B2 (en) 2021-03-22 2023-08-08 Cilag Gmbh International Staple cartridge comprising staples configured to apply different tissue compression
US11723658B2 (en) 2021-03-22 2023-08-15 Cilag Gmbh International Staple cartridge comprising a firing lockout
US11786239B2 (en) 2021-03-24 2023-10-17 Cilag Gmbh International Surgical instrument articulation joint arrangements comprising multiple moving linkage features
US11832816B2 (en) 2021-03-24 2023-12-05 Cilag Gmbh International Surgical stapling assembly comprising nonplanar staples and planar staples
US11896219B2 (en) 2021-03-24 2024-02-13 Cilag Gmbh International Mating features between drivers and underside of a cartridge deck
US11903582B2 (en) 2021-03-24 2024-02-20 Cilag Gmbh International Leveraging surfaces for cartridge installation
US11944336B2 (en) 2021-03-24 2024-04-02 Cilag Gmbh International Joint arrangements for multi-planar alignment and support of operational drive shafts in articulatable surgical instruments
US11793516B2 (en) 2021-03-24 2023-10-24 Cilag Gmbh International Surgical staple cartridge comprising longitudinal support beam
US11786243B2 (en) 2021-03-24 2023-10-17 Cilag Gmbh International Firing members having flexible portions for adapting to a load during a surgical firing stroke
US11857183B2 (en) 2021-03-24 2024-01-02 Cilag Gmbh International Stapling assembly components having metal substrates and plastic bodies
US11896218B2 (en) 2021-03-24 2024-02-13 Cilag Gmbh International Method of using a powered stapling device
US11849944B2 (en) 2021-03-24 2023-12-26 Cilag Gmbh International Drivers for fastener cartridge assemblies having rotary drive screws
US11744603B2 (en) 2021-03-24 2023-09-05 Cilag Gmbh International Multi-axis pivot joints for surgical instruments and methods for manufacturing same
US11849945B2 (en) 2021-03-24 2023-12-26 Cilag Gmbh International Rotary-driven surgical stapling assembly comprising eccentrically driven firing member
US11826047B2 (en) 2021-05-28 2023-11-28 Cilag Gmbh International Stapling instrument comprising jaw mounts
US11877745B2 (en) 2021-10-18 2024-01-23 Cilag Gmbh International Surgical stapling assembly having longitudinally-repeating staple leg clusters
US11957337B2 (en) 2021-10-18 2024-04-16 Cilag Gmbh International Surgical stapling assembly with offset ramped drive surfaces
US11937816B2 (en) 2021-10-28 2024-03-26 Cilag Gmbh International Electrical lead arrangements for surgical instruments

Citations (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080052071A1 (en) * 2006-02-10 2008-02-28 Spinvox Limited Mass-Scale, User-Independent, Device-Independent Voice Messaging System
US20110066634A1 (en) * 2007-03-07 2011-03-17 Phillips Michael S Sending a communications header with voice recording to send metadata for use in speech recognition, formatting, and search in mobile search application
US8024196B1 (en) * 2005-09-19 2011-09-20 Sap Ag Techniques for creating and translating voice applications
US20130144605A1 (en) * 2011-12-06 2013-06-06 Mehrman Law Office, PC Text Mining Analysis and Output System
US20130143587A1 (en) * 2011-12-02 2013-06-06 Yellowpages.Com, Llc System and Method for Coordinating Meetings Between Users of a Mobile Communication Network
US20150073798A1 (en) * 2013-09-08 2015-03-12 Yael Karov Automatic generation of domain models for virtual personal assistants
US20150149177A1 (en) * 2013-11-27 2015-05-28 Sri International Sharing Intents to Provide Virtual Assistance in a Multi-Person Dialog
US20150242385A1 (en) * 2013-04-23 2015-08-27 Facebook, Inc. Methods and systems for generation of flexible sentences in a social networking system
US20160063067A1 (en) * 2014-08-26 2016-03-03 Accenture Global Services Limited Automatic assistance for resource reuse based on context extracted from a user workspace
US20160105386A1 (en) * 2014-10-09 2016-04-14 International Business Machines Corporation System for handling abbreviation related text
US20170116982A1 (en) * 2015-10-21 2017-04-27 Google Inc. Parameter collection and automatic dialog generation in dialog systems
US20170148073A1 (en) * 2015-11-20 2017-05-25 Jagadeshwar Nomula Systems and methods for virtual agents to help customers and businesses
US9723151B2 (en) * 2015-10-19 2017-08-01 Genesys Telecommunications Laboratories, Inc. Optimized routing of interactions to contact center agents based on forecast agent availability and customer patience
US9762733B1 (en) * 2016-09-21 2017-09-12 Genesys Telecommunications Laboratories, Inc. System and method for recommending communication mediums based on predictive analytics
US20180052664A1 (en) * 2016-08-16 2018-02-22 Rulai, Inc. Method and system for developing, training, and deploying effective intelligent virtual agent
US20180181807A1 (en) * 2016-12-22 2018-06-28 Microsoft Technology Licensing, Llc Conversion of static images into interactive maps
US10049663B2 (en) * 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US20180232434A1 (en) * 2017-02-16 2018-08-16 Microsoft Technology Licensing, Llc Proactive and retrospective joint weight attribution in a streaming environment
US20180239830A1 (en) * 2017-02-17 2018-08-23 Microsoft Technology Licensing, Llc Using log data to train for automated sourcing
US10067938B2 (en) * 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10083690B2 (en) * 2014-05-30 2018-09-25 Apple Inc. Better resolution when referencing to concepts
US10089072B2 (en) * 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US20180329882A1 (en) * 2017-05-12 2018-11-15 International Business Machines Corporation Automatic, Unsupervised Paraphrase Detection
US20190034780A1 (en) * 2017-07-31 2019-01-31 Microsoft Technology Licensing, Llc Knowledge Graph For Conversational Semantic Search

Family Cites Families (65)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH01314373A (en) 1988-06-15 1989-12-19 Hitachi Ltd Translated word selecting system in machine translating system
US5077669A (en) 1989-12-27 1991-12-31 International Business Machines Corporation Method for quasi-key search within a national language support (nls) data processing system
US5369575A (en) 1992-05-15 1994-11-29 International Business Machines Corporation Constrained natural language interface for a computer system
US6381645B1 (en) 1997-12-08 2002-04-30 Siemens Information And Communication Networks, Inc. Method of implementing push techniques in conventional web browsers
US6173279B1 (en) 1998-04-09 2001-01-09 At&T Corp. Method of using a natural language interface to retrieve information from one or more data resources
US6915254B1 (en) 1998-07-30 2005-07-05 A-Life Medical, Inc. Automatically assigning medical codes using natural language processing
US6453292B2 (en) 1998-10-28 2002-09-17 International Business Machines Corporation Command boundary identifier for conversational natural language
US6981040B1 (en) 1999-12-28 2005-12-27 Utopy, Inc. Automatic, personalized online information and product services
US7644057B2 (en) 2001-01-03 2010-01-05 International Business Machines Corporation System and method for electronic communication management
US6820075B2 (en) 2001-08-13 2004-11-16 Xerox Corporation Document-centric system with auto-completion
US7606714B2 (en) 2003-02-11 2009-10-20 Microsoft Corporation Natural language classification within an automated response system
US20050228790A1 (en) 2004-04-12 2005-10-13 Christopher Ronnewinkel Coherent categorization scheme
US20060112127A1 (en) 2004-11-23 2006-05-25 Microsoft Corporation Method for efficiently mapping error messages to unique identifiers
US20060173776A1 (en) 2005-01-28 2006-08-03 Barry Shalley A Method of Authentication
US7512580B2 (en) 2005-08-04 2009-03-31 Sap Ag Confidence indicators for automated suggestions
KR100755677B1 (en) 2005-11-02 2007-09-05 삼성전자주식회사 Apparatus and method for dialogue speech recognition using topic detection
EP1963959A2 (en) 2005-12-09 2008-09-03 Fraunhofer-Gesellschaft zur Förderung der Angewandten Forschung e.V. A method and apparatus for automatic comparison of data sequences
US7668922B2 (en) 2006-01-19 2010-02-23 International Business Machines Corporation Identifying and displaying relevant shared entities in an instant messaging system
US7603330B2 (en) 2006-02-01 2009-10-13 Honda Motor Co., Ltd. Meta learning for question classification
US7664644B1 (en) * 2006-06-09 2010-02-16 At&T Intellectual Property Ii, L.P. Multitask learning for spoken language understanding
US20080091423A1 (en) 2006-10-13 2008-04-17 Shourya Roy Generation of domain models from noisy transcriptions
US7890326B2 (en) 2006-10-13 2011-02-15 Google Inc. Business listing search
US7707162B2 (en) 2007-01-08 2010-04-27 International Business Machines Corporation Method and apparatus for classifying multimedia artifacts using ontology selection and semantic classification
US8156138B2 (en) 2007-06-26 2012-04-10 Richrelevance, Inc. System and method for providing targeted content
WO2010086855A2 (en) 2009-01-29 2010-08-05 Fortress Applications Ltd. System and methods for encryption with authentication integrity
WO2010093750A2 (en) 2009-02-12 2010-08-19 President And Fellows Of Harvard College Adaptive event-guided system and method for avoiding voltage emergencies
US8543532B2 (en) 2009-10-05 2013-09-24 Nokia Corporation Method and apparatus for providing a co-creation platform
US8612432B2 (en) 2010-06-16 2013-12-17 Microsoft Corporation Determining query intent
WO2012030838A1 (en) 2010-08-30 2012-03-08 Honda Motor Co., Ltd. Belief tracking and action selection in spoken dialog systems
US8788443B2 (en) 2011-12-23 2014-07-22 Sap Ag Automated observational decision tree classifier
US20130317808A1 (en) 2012-05-24 2013-11-28 About, Inc. System for and method of analyzing and responding to user generated content
US8577671B1 (en) 2012-07-20 2013-11-05 Veveo, Inc. Method of and system for using conversation state information in a conversational interaction system
US9461876B2 (en) 2012-08-29 2016-10-04 Loci System and method for fuzzy concept mapping, voting ontology crowd sourcing, and technology prediction
US9372674B2 (en) 2013-03-12 2016-06-21 Microsoft Technology Licensing, Llc Dynamic generation of user interface placeholder templates
US9336269B1 (en) 2013-03-14 2016-05-10 Google Inc. Determining question and answer alternatives
US9558743B2 (en) 2013-03-15 2017-01-31 Google Inc. Integration of semantic context information
JP6221301B2 (en) 2013-03-28 2017-11-01 富士通株式会社 Audio processing apparatus, audio processing system, and audio processing method
US9177318B2 (en) 2013-04-22 2015-11-03 Palo Alto Research Center Incorporated Method and apparatus for customizing conversation agents based on user characteristics using a relevance score for automatic statements, and a response prediction function
US9367880B2 (en) 2013-05-03 2016-06-14 Facebook, Inc. Search intent for queries on online social networks
US10599765B2 (en) 2013-06-27 2020-03-24 Avaya Inc. Semantic translation model training
US10162884B2 (en) 2013-07-23 2018-12-25 Conduent Business Services, Llc System and method for auto-suggesting responses based on social conversational contents in customer care services
US9672556B2 (en) 2013-08-15 2017-06-06 Nook Digital, Llc Systems and methods for programatically classifying text using topic classification
CN104598445B (en) 2013-11-01 2019-05-10 腾讯科技(深圳)有限公司 Automatically request-answering system and method
US9189742B2 (en) 2013-11-20 2015-11-17 Justin London Adaptive virtual intelligent agent
DE112014005354T5 (en) 2013-11-25 2016-08-04 Mitsubishi Electric Corporation DIALOG MANAGEMENT SYSTEM AND DIALOG MANAGEMENT PROCESS
US9986242B2 (en) 2014-01-28 2018-05-29 Vmware, Inc. Enhanced image encoding in a virtual desktop infrastructure environment
JP6588449B2 (en) 2014-01-31 2019-10-09 グーグル エルエルシー Generating a vector representation of a document
US9837075B2 (en) 2014-02-10 2017-12-05 Mitsubishi Electric Research Laboratories, Inc. Statistical voice dialog system and method
US20150310377A1 (en) 2014-04-24 2015-10-29 Videodesk Sa Methods, devices and systems for providing online customer service
US10108608B2 (en) 2014-06-12 2018-10-23 Microsoft Technology Licensing, Llc Dialog state tracking using web-style ranking and multiple language understanding engines
US11038864B2 (en) 2014-06-12 2021-06-15 Mastercard International Incorporated Systems and methods for customer service access to a consumer interface system
US9658854B2 (en) 2014-09-26 2017-05-23 Intel Corporation Instructions and logic to provide SIMD SM3 cryptographic hashing functionality
US9466297B2 (en) 2014-12-09 2016-10-11 Microsoft Technology Licensing, Llc Communication system
EP3796235A1 (en) 2014-12-17 2021-03-24 Google LLC Generating numeric embeddings of images
WO2016106155A1 (en) 2014-12-22 2016-06-30 Zenedge, Inc. Dynamic content caching system
US9508340B2 (en) 2014-12-22 2016-11-29 Google Inc. User specified keyword spotting using long short term memory neural network feature extractor
US20160239846A1 (en) 2015-02-12 2016-08-18 Mastercard International Incorporated Payment Networks and Methods for Processing Support Messages Associated With Features of Payment Networks
US9672814B2 (en) 2015-05-08 2017-06-06 International Business Machines Corporation Semi-supervised learning of word embeddings
US10185803B2 (en) 2015-06-15 2019-01-22 Deep Genomics Incorporated Systems and methods for classifying, prioritizing and interpreting genetic variants and therapies using a deep neural network
US10635949B2 (en) 2015-07-07 2020-04-28 Xerox Corporation Latent embeddings for word images and their semantics
WO2018009231A1 (en) 2016-07-08 2018-01-11 Asapp, Inc. Automatically responding to a request of a user
US10083451B2 (en) 2016-07-08 2018-09-25 Asapp, Inc. Using semantic processing for customer support
US10109275B2 (en) 2016-12-19 2018-10-23 Asapp, Inc. Word hash language model
US10650311B2 (en) 2016-12-19 2020-05-12 Asaap, Inc. Suggesting resources using context hashing
US10210244B1 (en) 2018-02-12 2019-02-19 Asapp, Inc. Updating natural language interfaces by processing usage data

Patent Citations (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8024196B1 (en) * 2005-09-19 2011-09-20 Sap Ag Techniques for creating and translating voice applications
US20080052071A1 (en) * 2006-02-10 2008-02-28 Spinvox Limited Mass-Scale, User-Independent, Device-Independent Voice Messaging System
US20110066634A1 (en) * 2007-03-07 2011-03-17 Phillips Michael S Sending a communications header with voice recording to send metadata for use in speech recognition, formatting, and search in mobile search application
US20130143587A1 (en) * 2011-12-02 2013-06-06 Yellowpages.Com, Llc System and Method for Coordinating Meetings Between Users of a Mobile Communication Network
US20130144605A1 (en) * 2011-12-06 2013-06-06 Mehrman Law Office, PC Text Mining Analysis and Output System
US20150242385A1 (en) * 2013-04-23 2015-08-27 Facebook, Inc. Methods and systems for generation of flexible sentences in a social networking system
US20150073798A1 (en) * 2013-09-08 2015-03-12 Yael Karov Automatic generation of domain models for virtual personal assistants
US20150149177A1 (en) * 2013-11-27 2015-05-28 Sri International Sharing Intents to Provide Virtual Assistance in a Multi-Person Dialog
US10083690B2 (en) * 2014-05-30 2018-09-25 Apple Inc. Better resolution when referencing to concepts
US20160063067A1 (en) * 2014-08-26 2016-03-03 Accenture Global Services Limited Automatic assistance for resource reuse based on context extracted from a user workspace
US20160105386A1 (en) * 2014-10-09 2016-04-14 International Business Machines Corporation System for handling abbreviation related text
US9723151B2 (en) * 2015-10-19 2017-08-01 Genesys Telecommunications Laboratories, Inc. Optimized routing of interactions to contact center agents based on forecast agent availability and customer patience
US20170116982A1 (en) * 2015-10-21 2017-04-27 Google Inc. Parameter collection and automatic dialog generation in dialog systems
US20170148073A1 (en) * 2015-11-20 2017-05-25 Jagadeshwar Nomula Systems and methods for virtual agents to help customers and businesses
US10049663B2 (en) * 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10067938B2 (en) * 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10089072B2 (en) * 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US20180052664A1 (en) * 2016-08-16 2018-02-22 Rulai, Inc. Method and system for developing, training, and deploying effective intelligent virtual agent
US9762733B1 (en) * 2016-09-21 2017-09-12 Genesys Telecommunications Laboratories, Inc. System and method for recommending communication mediums based on predictive analytics
US20180181807A1 (en) * 2016-12-22 2018-06-28 Microsoft Technology Licensing, Llc Conversion of static images into interactive maps
US20180232434A1 (en) * 2017-02-16 2018-08-16 Microsoft Technology Licensing, Llc Proactive and retrospective joint weight attribution in a streaming environment
US20180239830A1 (en) * 2017-02-17 2018-08-23 Microsoft Technology Licensing, Llc Using log data to train for automated sourcing
US20180329882A1 (en) * 2017-05-12 2018-11-15 International Business Machines Corporation Automatic, Unsupervised Paraphrase Detection
US20190034780A1 (en) * 2017-07-31 2019-01-31 Microsoft Technology Licensing, Llc Knowledge Graph For Conversational Semantic Search

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11615422B2 (en) 2016-07-08 2023-03-28 Asapp, Inc. Automatically suggesting completions of text
US11790376B2 (en) 2016-07-08 2023-10-17 Asapp, Inc. Predicting customer support requests
WO2022189925A1 (en) * 2021-03-11 2022-09-15 International Business Machines Corporation Generating natural languages interface from graphic user interfaces
US11966562B2 (en) 2021-03-11 2024-04-23 International Business Machines Corporation Generating natural languages interface from graphic user interfaces

Also Published As

Publication number Publication date
EP3724792A4 (en) 2021-10-06
WO2019156803A1 (en) 2019-08-15
US10210244B1 (en) 2019-02-19
CN111801674A (en) 2020-10-20
EP3724792A1 (en) 2020-10-21
US20190251104A1 (en) 2019-08-15
US10515104B2 (en) 2019-12-24

Similar Documents

Publication Publication Date Title
US10515104B2 (en) Updating natural language interfaces by processing usage data
US11651043B2 (en) Leveraging analytics across disparate computing devices
US11816436B2 (en) Automated summarization of extracted insight data
US11663409B2 (en) Systems and methods for training machine learning models using active learning
US10489792B2 (en) Maintaining quality of customer support messages
US11282020B2 (en) Dynamic playback of synchronized narrated analytics playlists
US11853107B2 (en) Dynamic phase generation and resource load reduction for a query
US20220284013A1 (en) Processing a logical query
US11551004B2 (en) Intent discovery with a prototype classifier
US11841854B2 (en) Differentiation of search results for accurate query output
US20200019609A1 (en) Suggesting a response to a message by selecting a template using a neural network
US20190180196A1 (en) Systems and methods for generating and updating machine hybrid deep learning models
US11663405B2 (en) Machine learning applications for temporally-related events
US10339470B1 (en) Techniques for generating machine learning training data
US8903711B2 (en) System and methods for semiautomatic generation and tuning of natural language interaction applications
CN114503115A (en) Generating rich action items
GB2568363A (en) Rule determination for black-box machine-learning models
US11610064B2 (en) Clarification of natural language requests using neural networks
US20190179903A1 (en) Systems and methods for multi language automated action response
WO2019113122A1 (en) Systems and methods for improved machine learning for conversations
US11797591B2 (en) Abstractive multi-document summarization through self-supervision and control
US10984781B2 (en) Identifying representative conversations using a state model
US11880666B2 (en) Generating conversation descriptions using neural networks
Muñoz et al. A cognitive agent for mining bugs reports, feature suggestions and sentiment in a mobile application store
US20230237276A1 (en) System and Method for Incremental Estimation of Interlocutor Intents and Goals in Turn-Based Electronic Conversational Flow

Legal Events

Date Code Title Description
AS Assignment

Owner name: ASAPP, INC., NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BRANAVAN, SATCHUTHANANTHAVALE RASIAH KUHAN;HACKMAN, JOSEPH ELLSWORTH;HECKEL, FREDERICK WILLIAM POE;AND OTHERS;SIGNING DATES FROM 20180221 TO 20180223;REEL/FRAME:050995/0156

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION