US20180226072A1 - Real-time conflict mitigator - Google Patents

Real-time conflict mitigator Download PDF

Info

Publication number
US20180226072A1
US20180226072A1 US15/700,250 US201715700250A US2018226072A1 US 20180226072 A1 US20180226072 A1 US 20180226072A1 US 201715700250 A US201715700250 A US 201715700250A US 2018226072 A1 US2018226072 A1 US 2018226072A1
Authority
US
United States
Prior art keywords
communicator
recipient
sentiment
program
mitigator
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/700,250
Inventor
Richard Gorzela
Asima Silva
Jaime M. Stockton
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US15/700,250 priority Critical patent/US20180226072A1/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GORZELA, RICHARD, SILVA, ASIMA, STOCKTON, JAIME M.
Publication of US20180226072A1 publication Critical patent/US20180226072A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1815Semantic context, e.g. disambiguation of the recognition hypotheses based on word meaning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/432Query formulation
    • G06F16/433Query formulation using audio data
    • G06F17/30026
    • G06K9/00302
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1822Parsing for meaning understanding
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/63Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/90Pitch determination of speech signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/04Real-time or near real-time messaging, e.g. instant messaging [IM]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/52User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail for supporting social networking services
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/183Speech classification or search using natural language modelling using context dependencies, e.g. language models
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue

Definitions

  • the present invention relates generally to the field of computing, and more particularly to conflict mitigation.
  • Embodiments of the present invention disclose a method, computer system, and a computer program product for mitigating conflicts.
  • the present invention may include receiving communication by a recipient, wherein the received communication has a communicator sentiment associated with a communicator and a recipient sentiment associated with a recipient.
  • the present invention may also include determining the communicator sentiment associated with the communicator message, wherein the communicator sentiment is measured by a plurality of tone identifiers, a plurality of biometric identifiers and a plurality of facial expression identifiers, wherein the tone identifiers are captured by a microphone on a device, wherein the biometric identifiers are selected from a group consisting of body heat, pupil dilation and heart rate, and wherein the facial expression identifiers are captured by a camera on a device or a sensor on a device.
  • the present invention may then include determining the recipient sentiment associated with a received communicator message, wherein the recipient sentiment is measured by a plurality of tone identifiers, a plurality of biometric identifiers and a plurality of facial expression identifiers, wherein the tone identifiers are captured by a microphone on a device, wherein the biometric identifiers are selected from a group consisting of body heat, pupil dilation and heart rate, and wherein the facial expression identifiers are captured by a camera on a device or a sensor on a device.
  • the present invention may further include determining that the determined recipient sentiment does not match the determined communicator sentiment based on the received communication.
  • the present invention may also include determining an appropriate response based on determining that the communicator sentiment does not match the recipient sentiment.
  • the present invention may then include presenting the determined appropriate response, wherein presenting the appropriate response comprises delivering a message selected from the group consisting of an audio message, a type-written message, and a combination audio message and type-written message, wherein the determined appropriate response is searched for in a positive response database, wherein the positive response database accumulates a plurality of logged conversation responses, wherein presenting the determined appropriate response comprises presenting the determined appropriate response to the communicator and the recipient.
  • the present invention may further include logging a conversation result in a database.
  • FIG. 1 illustrates a networked computer environment according to at least one embodiment
  • FIG. 2 is a block diagram of the conflict mitigator program components according to at least one embodiment
  • FIG. 3 is an operational flowchart illustrating a process for conflict mitigating according to at least one embodiment
  • FIG. 4 is a block diagram of internal and external components of computers and servers depicted in FIG. 1 according to at least one embodiment
  • FIG. 5 is a block diagram of an illustrative cloud computing environment including the computer system depicted in FIG. 1 , in accordance with an embodiment of the present disclosure.
  • FIG. 6 is a block diagram of functional layers of the illustrative cloud computing environment of FIG. 5 , in accordance with an embodiment of the present disclosure.
  • the present invention may be a system, a method, and/or a computer program product at any possible technical detail level of integration
  • the computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention
  • the computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device.
  • the computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
  • a non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read-only memory
  • EPROM or Flash memory erasable programmable read-only memory
  • SRAM static random access memory
  • CD-ROM compact disc read-only memory
  • DVD digital versatile disk
  • memory stick a floppy disk
  • a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon
  • a computer readable storage medium is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
  • Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network.
  • the network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
  • a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
  • Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, configuration data for integrated circuitry, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++, or the like, and procedural programming languages, such as the “C” programming language or similar programming languages.
  • the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
  • These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
  • the computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s).
  • the functions noted in the blocks may occur out of the order noted in the Figures.
  • two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
  • the following described exemplary embodiments provide a system, method and program product for conflict mitigation.
  • the present embodiment has the capacity to improve the technical field of conflict mitigation by providing real-time response feedback during a conversation that may lead to more conversations with positive results.
  • the conflict mitigator program may analyze each person's facial expressions, tone of speech and biometric responses during a real-time conversation in real-time in order to provide positive real-time responses to negative reactions.
  • the real-time responses may be provided to an individual through a message which may be typed on a computing device screen or verbally communicated via a computing device speaker or headphones.
  • the positive responses, provided by the conflict mitigator program, to negative reactions during communications between individuals may foster better and more trusting relationships between parties.
  • a conflict mitigator program may provide real-time feedback for communicators and recipients in the event that there may be a misunderstanding or a mismatch between the recipients internal response and the intent of the communicator.
  • Examples of communicators and recipients may include a doctor and patient relationship, a bully and victim situation, an assisted living support worker and an assisted living resident relationship, a parent and child relationship, or between spouses.
  • a misunderstanding and a proper real-time response may help the individuals in any of the previously mentioned relationships, or any other kind of communicating situation, understand one another and get a more positive result out of the interaction. A better, more positive response and result may lead to stronger relationships and a better understanding between the communicator and the recipient.
  • Assessing sentiment and emotions through voice identifiers, facial expression identifiers and biometric identifiers may assist in understanding how a message is communicated and how a message is received. Therefore, it may be advantageous to, among other things, provide a way to offer real-time feedback to both the communicator and the recipient while providing each with a more positive response in order to foster better relationships.
  • the conflict mitigator program may identify personalities of communicators and recipients and during a conversation may compare the real-time interaction with previous interactions between the same parties. The comparison may identify differences in previous reactions or consistencies in previous reactions whether the reactions are positive or negative.
  • the conflict mitigator program may have a response component that provides the parties with real-time feedback on how a message could have been worded better based on similar conversations that had positive emotive results, personalities of the parties and a history of interactions between the parties.
  • a personality example that the conflict mitigator program may take into consideration is someone who has difficulty with presenting constructive criticism and may need specific advice specific to that person's task. Alternatively, someone who does not handle criticism well might need specific advice on how to respond to criticism.
  • An example of a history of interactions between parties may include a manager who consistently gets negative reactions from specific team members when giving direction. The manager may need feedback on how to solicit the team members advice before giving direction.
  • the conflict mitigator program may provide an individual with specific recommendations for better and more positively phrased real-time communication responses based on real-time communication with another person.
  • the conflict mitigator program may provide real-time responses to individuals by using a cognitive analyzer to analyze both the communicator's and the recipient's tone, biometrics, facial expressions and sentiment.
  • the tone identifier and the facial expression identifier may take data from the communicator while capturing the recipient's facial expression and biometric response.
  • the conflict mitigator program may use the currently received data along with the personalities involved and a history of reactions between the parties to identify exemplar conversations.
  • the cognitive analyzer component of the conflict mitigator program may compare data in order to provide a more positive conversational response by using a database that provides similar and more successful conversational responses.
  • the cognitive analyzer may provide comparisons of conversations that have happened in the past with the current conversation.
  • the response feature may provide the communicator and the receiver of the communication with a better real-time response (e.g., feedback) during the conversation.
  • the conflict mitigator program may provide the person that delivered a negatively received message with an alternative message based on similar but more positive resulting conversations, specific personalities involved and the history of interactions between the parties. Similarly, the conflict mitigator program may also provide the person who received a message that they reacted to in a negative way with a positive response based on similar criteria. When providing a response, the conflict mitigator program may remove identifying data (e.g., individual's names, locations, etc.) so that the better worded responses may provide positive responses without presenting information to the parties (e.g., personality types, history of interactions, etc.). While the conflict mitigator program may analyze personality types and a history of interactions, the information may not be provided in specific responses to communicators and recipients. The conflict mitigator program may be used and be beneficial for first instances of conversations between more than one person and with already developed relationships where the individuals have communicated in multiple instances.
  • identifying data e.g., individual's names, locations, etc.
  • One scenario of how the conflict mitigator program may foster a better relationship may be amongst co-workers, for example, Party A often tells jokes about a specific professional group to co-worker Party B. Party A thinks his jokes are funny and help to put people in a good mood in the morning, however, Party A does not know that Party B's spouse is in the specific professional group that Party A jokes about.
  • the conflict mitigator program may detect Party B's increasing annoyance through facial expressions and tone and the conflict mitigator program may then recommend that Party A try a different small talk subject such as sports.
  • Sports may be a topic chosen by the conflict mitigator program if there is a history of positive results amongst co-workers having a positive response to sports talk.
  • Another scenario of how the conflict mitigator program may foster a more trusting relationship may be between a doctor and a patient, for example when a doctor reviews test data with a patient after only one previous meeting between the two.
  • the doctor had ordered blood tests and an ultrasound since the initial physical examination proved inconclusive.
  • the patient is 18 years old and this is the first medical encounter with a doctor other than the patient's pediatrician.
  • the patient is apprehensive to receive the results of the blood tests and the ultrasound since the patient has not previously had any serious conditions.
  • Another apprehension the patient has is that the patient's pediatrician was friendly and was only visited for an occasional cold or sports related abrasions.
  • the doctor explains to the young patient that the tests were inconclusive so a biopsy or surgery may be necessary. Although the doctor also explains that the patient should wait a month for a re-examination before getting a biopsy or surgery, the words surgery and inconclusive have already made the patient concerned. The 18-year-old patient does not understand what a biopsy is and surgery sounds scary so all the patient can muster up in a response is to say “ok.”
  • the conflict mitigator program may identify the patient's heightened tension through facial expression analysis and biometrics while the doctor's tone is analyzed as confident and even tempered.
  • the conflict mitigator program identifies a misunderstanding (i.e., a mismatch) based on the patient's reaction to the doctor's words.
  • the cognitive analyzer component of the program may compare the mismatch and find similar conversations between a doctor and a younger, more inexperienced, patient from a database where the doctor had to deliver news regarding a continuing diagnosis.
  • the conflict mitigator program may then provide a real-time recommendation to the doctor to slow down and explain the terms to the new patient.
  • Another provided real-time recommendation for the doctor's response may be an adjustment in the phrasing of the information to “While we may determine the need for some further tests, we are confident that the best next step is for you to first come back in a month for a re-examination to see how things many have changed.”
  • the conflict mitigator program may also provide the patient with recommendations, such as to ask for clarification of the terms that were not understood and recommend the patient say “Doctor, please explain what a ‘biopsy’ is and how it is used?”
  • Another scenario of how the conflict mitigator program may foster a more trusting relationship may be between an employee and a resident of an assisted living facility, for example in a situation where an assisted living resident has resided at a facility for a year.
  • the resident does not like physical contact and a newer employee of the facility likes to give friendly hugs when greeting people.
  • the quick affection from the newer employee to the resident may cause the resident to be apprehensive and react negatively to the employee even though the employee believes they are being friendly.
  • the conflict mitigator program may identify the resident's apprehension when interacting with the friendly employee through biometrics and facial expression analysis.
  • the conflict mitigator program may also identify the employee's joyful sentiment and openness which would create a misunderstanding (i.e., a mismatch) with the resident's reaction.
  • the conflict mitigator program may then search a database to find similar scenarios and provide a recommendation of alternative real-time responses.
  • An alternative real-time response for the employee may be to ask the resident permission before doing anything physical such as giving the resident a hug.
  • the conflict mitigator program may recommend that the resident state more directly to the employee about what kind of interaction is preferred.
  • the conflict mitigator program in the above stated examples may generate a real-time response by analyzing past behavior and by analyzing other similarly situated conversation environments.
  • the conflict mitigator program may recognize that even if the communicators are prepared for a conversation, that face-to-face communication may still lead to misunderstandings that are specific to the parties involved and may provide guidance to the communicators in real-time. If there is a mismatch in sentiment during the conversation, the conflict mitigator program may notice that one person may be upset while the other is demonstrating positive characteristics such as joy or empathy.
  • a real-time response may provide the communicators with similar conversations that had more positive results where the participants did not have a negative mood, were not in a heightened response and showed positive sentiments such as empathy, confidence, joy and/or openness. Previous similar conversations that the conflict mitigator program uses may be between the two parties conversing which had positive emotive results or between two different parties which had positive emotive results.
  • the networked computer environment 100 may include a computer 102 with a processor 104 and a data storage device 106 that is enabled to run a software program 108 and a conflict mitigator program 110 a .
  • the networked computer environment 100 may also include a server 112 that is enabled to run a conflict mitigator program 110 b that may interact with a database 114 and a communication network 116 .
  • the networked computer environment 100 may include a plurality of computers 102 and servers 112 , only one of which is shown.
  • the communication network 116 may include various types of communication networks, such as a wide area network (WAN), local area network (LAN), a telecommunication network, a wireless network, a public switched network and/or a satellite network.
  • WAN wide area network
  • LAN local area network
  • the client computer 102 may communicate with the server computer 112 via the communications network 116 .
  • the communications network 116 may include connections, such as wire, wireless communication links, or fiber optic cables.
  • server computer 112 may include internal components 902 a and external components 904 a , respectively, and client computer 102 may include internal components 902 b and external components 904 b , respectively.
  • Server computer 112 may also operate in a cloud computing service model, such as Software as a Service (SaaS), Platform as a Service (PaaS), or Infrastructure as a Service (IaaS).
  • Server 112 may also be located in a cloud computing deployment model, such as a private cloud, community cloud, public cloud, or hybrid cloud.
  • Client computer 102 may be, for example, a mobile device, a telephone, a personal digital assistant, a netbook, a laptop computer, a tablet computer, a desktop computer, or any type of computing devices capable of running a program, accessing a network, and accessing a database 114 .
  • the conflict mitigator program 110 a , 110 b may interact with a database 114 that may be embedded in various storage devices, such as, but not limited to a computer/mobile device 102 , a networked server 112 , or a cloud storage service.
  • a user using a client computer 102 or a server computer 112 may use the conflict mitigator program 110 a , 110 b (respectively) to obtain real-time responses to communication with another person where a misunderstanding may be present.
  • the conflict mitigator method is explained in more detail below with respect to FIGS. 2 and 3 .
  • the program components 200 may include identifier components such as a tone identifier 202 , a biometric identifier 204 and a facial expression identifier 206 .
  • the program components 200 may also include a cognitive analyzer 208 , a response feature 210 and a logging feature 212 .
  • the tone identifier 202 may analyze the sentiment in the communicator's tone of voice (e.g., volume, inflection, sharpness, softness, etc.). The tone identifier 202 may also identify the tone of a digital message by analyzing the words written with other symbols and the use of capital letters.
  • the biometric identifier 204 may measure heightened responses (e.g., changes in blood pressure, heart rate, breathing, pupil dilation, body temperature, etc.).
  • the facial expression identifier 206 may recognize emotions (e.g., joy, fear, duress, anger, etc.).
  • the cognitive analyzer 208 may use computing technology (e.g., Watson AnalyticsTM, Watson Analytics and all Watson Analytics-based trademarks and logos are trademarks or registered trademarks of International Business Machines Corporation and/or its affiliates) to analyze the tone identifiers 202 , biometric identifiers 204 , and facial expression identifiers 206 that have been inputted into the conflict mitigator program 110 a , 110 b .
  • the cognitive analyzer 208 may analyze the personalities of the communicator and the recipient, the history of reactions between the personalities (e.g., positive or negative reactions) and may identify exemplary conversations that relate to the same subject that is being conversed.
  • the cognitive analyzer 208 may detect patterns of conversations and reactions made by an individual. For example, whether a doctor has had consistent negative or positive responses from the nurse staff. This data may be useful to both the doctor to know how best to talk to co-workers and for the hospital to be able to track the doctor's effectiveness in communicating.
  • the response feature 210 may provide real-time feedback (e.g., through headphones, speakers, text, type written message, etc.) for both the communicator and the recipient based on the cognitive analyzer's identification of exemplar conversations.
  • the response feature 210 may provide both parties with real-time feedback that may be more positive in nature and that might defuse any situation (e.g., alternative reactions and message phrasing). Further, the response feature 210 may also allow the communicators to build trust by communicating through any misunderstandings in real-time.
  • the logging feature 212 may log the tone identifiers 202 , biometric identifiers 204 , facial expression identifiers 206 , cognitive analysis and real-time response to capture the history of conversations and reactions between communicators for use in the present conversation or use in a future conversation.
  • the conflict mitigator program 110 a , 110 b may use a database 114 to store the communications between communicators, the reactions and resolutions.
  • the database 114 may provide better responses (i.e., feedback) as the database obtains more data to pull from. For example, the more conversations and reactions that are added to the database 114 (e.g., Watson) may create more data for the conflict mitigator program 110 a , 110 b to use when providing real-time feedback to communicators.
  • FIG. 3 an operational flowchart illustrating the exemplary real-time conflict mitigator process 300 used by the conflict mitigator program 110 a , 110 b according to at least one embodiment is depicted.
  • the conflict mitigator program 110 a , 110 b receives real-time communication.
  • the communication may be received either verbally or type-written.
  • both the recipient and the conflict mitigator program 110 a , 110 b receive the verbal or type-written message.
  • the conflict mitigator program 110 a , 110 b may receive communication from any individual during a conversation.
  • Verbal communication may be captured using a microphone on a device in an audio file format (e.g., .mp3 and .wav formats).
  • the audio file may be filed in a data repository as an audio message.
  • the type-written communication (i.e., type-written message) may be captured from an email or a text message and may be filed in a data repository.
  • the conflict mitigator program 110 a , 110 b determines if there is a mismatch in communication between the sender and the recipient.
  • a mismatch may be a misunderstanding between the people conversing.
  • a mismatch may occur when the intent of the communicator does not match the resulting emotional response of the recipient. The mismatch may even occur in the form of body contact such as a hug, as presented in the previous example between the employee and the assisted living facility resident.
  • An example of a communication misunderstanding may occur if the communicator makes a statement or asks a question with good intentions and the statement is met with a response from the recipient that is upset, sad or angry.
  • Good intentions may be measured by the communicator's tone of voice, facial expression features and biometric features.
  • the recipient's opposite response of sadness or anger may be measured by tone of voice, facial expression features and biometric features.
  • An example of good intentions may show an even-level heart rate while anger may show an elevated heart rate.
  • the tone of voice feature may be captured by a microphone input into a digital audio file format and the conflict mitigator program 110 a , 110 b may identify certain patterns and compare the patterns with known patterns of tone of voice associated with positive and negative reactions. For example, a softer and even tone may convey a calm and nice tone of voice while a sharp and loud tone may convey an angry tone of voice.
  • Facial expression features may be captured through a camera sensor and the conflict mitigator program 110 a , 110 b may conduct an image analysis or pattern matching, or both image analysis and pattern matching, to identify facial features associated with positive and negative reactions. For example, a smile may show a positive facial feature while furrowed eyebrows may convey a negative or confused facial feature.
  • the biometric feature may use known technology to capture body reactions (e.g., body heat, blood flow, heart rate/pulse, dilated pupils, etc.) to comments or physical contact.
  • body reactions e.g., body heat, blood flow, heart rate/pulse, dilated pupils, etc.
  • a LED (light-emitting diode) light may be shined from a device worn on the wrist onto the wrist to detect changes in blood flow. For example, a sharp rise in blood flow during a conversation may convey a more negative reaction.
  • the mismatch may be recognized by the conflict mitigator program 110 a , 110 b by noticing, for example, that the communicator's intent was good, however, the recipient had a negative response to the comment made by the communicator.
  • the tone identifiers 202 , biometric identifiers 204 and facial feature identifiers 206 may identify the communicator is positively communicating with good intent through a nice tone, regular heartrate and joyful facial features.
  • the recipient may provide tone identifiers 202 , biometric identifiers 204 and facial expression identifiers 206 to show a negative reaction by body temperature rising, facial features that represent a stressful emotion and a distressed tone of voice in a response.
  • This mismatch (e.g., misunderstanding) will cause the conflict mitigator program 110 a , 110 b to search a positive response database (e.g., database 114 ) containing positive responses for better responses to provide for each party to use.
  • the communicator may be presented with real-time feedback to reword the comment while the recipient may be presented with real-time feedback to ask the communicator to clarify what they meant.
  • a doctor communicates to a young patient that the results to the tests that were run were inconclusive, a biopsy may be needed next month and surgery may follow.
  • the young patient does not fully understand what is being communicated and gets scared of the word biopsy and the word surgery.
  • An older patient in this scenario might not react negatively to this communication from the doctor because an older patient would most likely understand what the doctor is saying.
  • the younger patient does not know what a biopsy is and is frightened at the thought of surgery because the young patient has not been to a doctor for more than a common cold. Therefore, the doctor's good intent with empathetic emotions while communicating to the patient is a mismatch with the patients resulting fear.
  • the conflict mitigator program 110 a , 110 b determines that there is not a mismatch in communication between the communicator and the recipient at 304 . If the conflict mitigator program 110 a , 110 b determined that there is not a mismatch in communication between the communicator and the recipient at 304 , then the conflict mitigator program 110 a , 110 b will log the results in a database (e.g., database 114 ) at 306 . If no mismatch in the tone identifiers 202 , biometric identifiers 204 , and facial feature identifier 206 between the communicator's intent and the recipient's reaction is present, then the conflict mitigator program 110 a , 110 b may not need to provide any real-time response to parties in the conversation.
  • a database e.g., database 114
  • the conversation may be saved in a database 114 to be marked as successful and the data from the conversation may be used to make the cognitive analyzer 208 component more robust with potential positive responses. For example, if a doctor provides test results to a young patient who understood what the doctor was saying, and the young patient provided positive tone identifiers 202 , biometric identifiers 204 and facial expression identifiers 206 , then the details of the conversation may be used for a doctor in a similar situation but in an instance where the young patient reacted negatively to the conversation.
  • the conflict mitigator program 110 a , 110 b determines that there is a mismatch in communication between the communicator and the recipient at 304 . If the conflict mitigator program 110 a , 110 b determined that there is a mismatch in communication between the communicator and the recipient at 304 , then the conflict mitigator program 110 a , 110 b will provide real-time feedback to both the communicator and the recipient at 308 .
  • the conflict mitigator program 110 a , 110 b provides real-time feedback for communicators and recipients in the event that there may be a misunderstanding or a mismatch between the recipients internal response and the intent of the communicator. Responses may be in the form of a message on a screen, a voice played through speakers, or through headphones. Responses may also be a combination of an audio message and a type-written message.
  • Real-time feedback may turn negative conversations into positive conversations that yield stronger and more trusting relationships.
  • the real-time component of the conflict mitigator program 110 a , 110 b allows the conversation to be corrected instantly while the conversation is happening. A person could go from having bad feelings associated with a person or conversation in the beginning of conversing to later walking away with good feelings associated with the same person or conversation.
  • the real-time component of the conflict mitigator program 110 a , 110 b may create a more positive environment where the people communicating may get more value and feel better after having the conversation.
  • the conflict mitigator program 110 a , 110 b may provide real-time responses to both the doctor and the patient.
  • Real-time feedback for the doctor may tell the doctor to explain what a biopsy is to the patient and that nothing further needs to be done between now and the next checkup in a month when more information will be available after more tests are run.
  • the conflict mitigator program 110 a , 110 b may provide real-time feedback for the doctor by searching for similar conversations between doctors and young patients in the positive response database (e.g., database 114 ), which may be more explanatory in nature by providing definitions to words that an older patient may already know.
  • the conflict mitigator program 110 a , 110 b may provide real-time feedback for the patient to ask the doctor to explain what a biopsy is. Through the dialog that may come from the conflict mitigator program's 110 a , 110 b real-time feedback component, the doctor may earn the trust of the patient and the patient may walk away from the doctor appointment without fear. This more positive interaction may foster better trust and relationships.
  • the conflict mitigator program 110 a , 110 b will log the results in a database 114 .
  • the results are logged into a database 114 as stated above at 306 .
  • the log may continually collect data and continually grow, providing more responses to more types of situations thereby increasing the effectiveness of the conflict mitigator program 110 a , 110 b over time.
  • FIGS. 2 and 3 provide only an illustration of one embodiment and do not imply any limitations with regard to how different embodiments may be implemented. Many modifications to the depicted embodiment(s) may be made based on design and implementation requirements.
  • An alternate embodiment may be where one party receives recommended (i.e., appropriate) responses from the conflict mitigator program 110 a , 110 b . For example, if the communicator has a device running the conflict mitigator program 110 a , 110 b , and was speaking to a recipient who did not have a device running the conflict mitigator program 110 a , 110 b , then the communicator's device may provide the communicator with proper responses based on sentiment analysis of both the communicator and the recipient.
  • FIG. 4 is a block diagram 900 of internal and external components of computers depicted in FIG. 1 in accordance with an illustrative embodiment of the present invention. It should be appreciated that FIG. 4 provides only an illustration of one implementation and does not imply any limitations with regard to the environments in which different embodiments may be implemented. Many modifications to the depicted environments may be made based on design and implementation requirements.
  • Data processing system 902 , 904 is representative of any electronic device capable of executing machine-readable program instructions.
  • Data processing system 902 , 904 may be representative of a smart phone, a computer system, PDA, or other electronic devices.
  • Examples of computing systems, environments, and/or configurations that may represented by data processing system 902 , 904 include, but are not limited to, personal computer systems, server computer systems, thin clients, thick clients, hand-held or laptop devices, multiprocessor systems, microprocessor-based systems, network PCs, minicomputer systems, and distributed cloud computing environments that include any of the above systems or devices.
  • User client computer 102 and network server 112 may include respective sets of internal components 902 a, b and external components 904 a, b illustrated in FIG. 4 .
  • Each of the sets of internal components 902 a, b includes one or more processors 906 , one or more computer-readable RAMs 908 , and one or more computer-readable ROMs 910 on one or more buses 912 , and one or more operating systems 914 and one or more computer-readable tangible storage devices 916 .
  • the one or more operating systems 914 , the software program 108 and the conflict mitigator program 110 a in client computer 102 , and the conflict mitigator program 110 b in network server 112 may be stored on one or more computer-readable tangible storage devices 916 for execution by one or more processors 906 via one or more RAMs 908 (which typically include cache memory).
  • each of the computer-readable tangible storage devices 916 is a magnetic disk storage device of an internal hard drive.
  • each of the computer-readable tangible storage devices 916 is a semiconductor storage device such as ROM 910 , EPROM, flash memory or any other computer-readable tangible storage device that can store a computer program and digital information.
  • Each set of internal components 902 a, b also includes a R/W drive or interface 918 to read from and write to one or more portable computer-readable tangible storage devices 920 such as a CD-ROM, DVD, memory stick, magnetic tape, magnetic disk, optical disk or semiconductor storage device.
  • a software program such as the software program 108 and the conflict mitigator program 110 a , 110 b can be stored on one or more of the respective portable computer-readable tangible storage devices 920 , read via the respective R/W drive or interface 918 , and loaded into the respective hard drive 916 .
  • Each set of internal components 902 a, b may also include network adapters (or switch port cards) or interfaces 922 such as a TCP/IP adapter cards, wireless wi-fi interface cards, or 3G or 4G wireless interface cards or other wired or wireless communication links.
  • the software program 108 and the conflict mitigator program 110 a in client computer 102 and the conflict mitigator program 110 b in network server computer 112 can be downloaded from an external computer (e.g., server) via a network (for example, the Internet, a local area network or other, wide area network) and respective network adapters or interfaces 922 .
  • the network may comprise copper wires, optical fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
  • Each of the sets of external components 904 a, b can include a computer display monitor 924 , a keyboard 926 , and a computer mouse 928 .
  • External components 904 a, b can also include touch screens, virtual keyboards, touch pads, pointing devices, and other human interface devices.
  • Each of the sets of internal components 902 a, b also includes device drivers 930 to interface to computer display monitor 924 , keyboard 926 , and computer mouse 928 .
  • the device drivers 930 , R/W drive or interface 918 , and network adapter or interface 922 comprise hardware and software (stored in storage device 916 and/or ROM 910 ).
  • Cloud computing is a model of service delivery for enabling convenient, on-demand network access to a shared pool of configurable computing resources (e.g. networks, network bandwidth, servers, processing, memory, storage, applications, virtual machines, and services) that can be rapidly provisioned and released with minimal management effort or interaction with a provider of the service.
  • This cloud model may include at least five characteristics, at least three service models, and at least four deployment models.
  • a cloud computing environment is service oriented with a focus on statelessness, low coupling, modularity, and semantic interoperability.
  • An infrastructure comprising a network of interconnected nodes.
  • cloud computing environment 1000 comprises one or more cloud computing nodes 100 with which local computing devices used by cloud consumers, such as, for example, personal digital assistant (PDA) or cellular telephone 1000 A, desktop computer 1000 B, laptop computer 1000 C, and/or automobile computer system 1000 N may communicate.
  • Nodes 100 may communicate with one another. They may be grouped (not shown) physically or virtually, in one or more networks, such as Private, Community, Public, or Hybrid clouds as described hereinabove, or a combination thereof.
  • This allows cloud computing environment 1000 to offer infrastructure, platforms and/or software as services for which a cloud consumer does not need to maintain resources on a local computing device.
  • computing devices 1000 A-N shown in FIG. 5 are intended to be illustrative only and that computing nodes 100 and cloud computing environment 1000 can communicate with any type of computerized device over any type of network and/or network addressable connection (e.g., using a web browser).
  • FIG. 6 a set of functional abstraction layers 1100 provided by cloud computing environment 1000 is shown. It should be understood in advance that the components, layers, and functions shown in FIG. 6 are intended to be illustrative only and embodiments of the invention are not limited thereto. As depicted, the following layers and corresponding functions are provided:
  • Hardware and software layer 1102 includes hardware and software components.
  • hardware components include: mainframes 1104 ; RISC (Reduced Instruction Set Computer) architecture based servers 1106 ; servers 1108 ; blade servers 1110 ; storage devices 1112 ; and networks and networking components 1114 .
  • software components include network application server software 1116 and database software 1118 .
  • Virtualization layer 1120 provides an abstraction layer from which the following examples of virtual entities may be provided: virtual servers 1122 ; virtual storage 1124 ; virtual networks 1126 , including virtual private networks; virtual applications and operating systems 1128 ; and virtual clients 1130 .
  • management layer 1132 may provide the functions described below.
  • Resource provisioning 1134 provides dynamic procurement of computing resources and other resources that are utilized to perform tasks within the cloud computing environment.
  • Metering and Pricing 1136 provide cost tracking as resources are utilized within the cloud computing environment, and billing or invoicing for consumption of these resources. In one example, these resources may comprise application software licenses.
  • Security provides identity verification for cloud consumers and tasks, as well as protection for data and other resources.
  • User portal 1138 provides access to the cloud computing environment for consumers and system administrators.
  • Service level management 1140 provides cloud computing resource allocation and management such that required service levels are met.
  • Service Level Agreement (SLA) planning and fulfillment 1142 provide pre-arrangement for, and procurement of, cloud computing resources for which a future requirement is anticipated in accordance with an SLA.
  • SLA Service Level Agreement
  • Workloads layer 1144 provides examples of functionality for which the cloud computing environment may be utilized. Examples of workloads and functions which may be provided from this layer include: mapping and navigation 1146 ; software development and lifecycle management 1148 ; virtual classroom education delivery 1150 ; data analytics processing 1152 ; transaction processing 1154 ; and conflict mitigation 1156 .
  • a conflict mitigator program 110 a , 110 b provides real-time feedback for communicators and recipients in the event that there may be a misunderstanding or a mismatch between the recipients internal response and the intent of the communicator.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Computational Linguistics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Psychiatry (AREA)
  • Hospice & Palliative Care (AREA)
  • General Physics & Mathematics (AREA)
  • Child & Adolescent Psychology (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Information Transfer Between Computers (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Embodiments of the present invention disclose a method, computer system, and a computer program product for mitigating conflicts. The present invention may include receiving communication by a recipient. The present invention may include determining the communicator sentiment and recipient sentiment measured by tone identifiers, biometric identifiers and facial expression identifiers. The present invention may include determining that the recipient sentiment does not match the communicator sentiment. The present invention may include determining and presenting an appropriate response that comprises delivering a message selected from the group consisting of an audio message, a type-written message, and a combination audio message and type-written message, wherein the determined appropriate response is searched for in a positive response database, wherein the positive response database accumulates a plurality of logged conversation responses. The present invention may further include logging a conversation result in a database.

Description

    BACKGROUND
  • The present invention relates generally to the field of computing, and more particularly to conflict mitigation.
  • Sometimes communication between people can break down and leave participants confused as to why they are receiving a reaction they did not anticipate. Even well worded expressions may carry a tone that can negatively impact the recipient. The causes may be that the communicator could have difficulty reading others or have a reduced focus due to stress factors. Another level of communication breakdown can happen even if the communicator does understand the negative impact of their message. A breakdown may occur if the communicator does not know how to respond to the recipient who received the communication negatively. Further, the recipient may also not know how to respond to the negatively received message, which can lead to a degraded conversation and an untrusting relationship.
  • SUMMARY
  • Embodiments of the present invention disclose a method, computer system, and a computer program product for mitigating conflicts. The present invention may include receiving communication by a recipient, wherein the received communication has a communicator sentiment associated with a communicator and a recipient sentiment associated with a recipient. The present invention may also include determining the communicator sentiment associated with the communicator message, wherein the communicator sentiment is measured by a plurality of tone identifiers, a plurality of biometric identifiers and a plurality of facial expression identifiers, wherein the tone identifiers are captured by a microphone on a device, wherein the biometric identifiers are selected from a group consisting of body heat, pupil dilation and heart rate, and wherein the facial expression identifiers are captured by a camera on a device or a sensor on a device. The present invention may then include determining the recipient sentiment associated with a received communicator message, wherein the recipient sentiment is measured by a plurality of tone identifiers, a plurality of biometric identifiers and a plurality of facial expression identifiers, wherein the tone identifiers are captured by a microphone on a device, wherein the biometric identifiers are selected from a group consisting of body heat, pupil dilation and heart rate, and wherein the facial expression identifiers are captured by a camera on a device or a sensor on a device. The present invention may further include determining that the determined recipient sentiment does not match the determined communicator sentiment based on the received communication. The present invention may also include determining an appropriate response based on determining that the communicator sentiment does not match the recipient sentiment. The present invention may then include presenting the determined appropriate response, wherein presenting the appropriate response comprises delivering a message selected from the group consisting of an audio message, a type-written message, and a combination audio message and type-written message, wherein the determined appropriate response is searched for in a positive response database, wherein the positive response database accumulates a plurality of logged conversation responses, wherein presenting the determined appropriate response comprises presenting the determined appropriate response to the communicator and the recipient. The present invention may further include logging a conversation result in a database.
  • BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
  • These and other objects, features and advantages of the present invention will become apparent from the following detailed description of illustrative embodiments thereof, which is to be read in connection with the accompanying drawings. The various features of the drawings are not to scale as the illustrations are for clarity in facilitating one skilled in the art in understanding the invention in conjunction with the detailed description. In the drawings:
  • FIG. 1 illustrates a networked computer environment according to at least one embodiment;
  • FIG. 2 is a block diagram of the conflict mitigator program components according to at least one embodiment;
  • FIG. 3 is an operational flowchart illustrating a process for conflict mitigating according to at least one embodiment;
  • FIG. 4 is a block diagram of internal and external components of computers and servers depicted in FIG. 1 according to at least one embodiment;
  • FIG. 5 is a block diagram of an illustrative cloud computing environment including the computer system depicted in FIG. 1, in accordance with an embodiment of the present disclosure; and
  • FIG. 6 is a block diagram of functional layers of the illustrative cloud computing environment of FIG. 5, in accordance with an embodiment of the present disclosure.
  • DETAILED DESCRIPTION
  • Detailed embodiments of the claimed structures and methods are disclosed herein; however, it can be understood that the disclosed embodiments are merely illustrative of the claimed structures and methods that may be embodied in various forms. This invention may, however, be embodied in many different forms and should not be construed as limited to the exemplary embodiments set forth herein. Rather, these exemplary embodiments are provided so that this disclosure will be thorough and complete and will fully convey the scope of this invention to those skilled in the art. In the description, details of well-known features and techniques may be omitted to avoid unnecessarily obscuring the presented embodiments.
  • The present invention may be a system, a method, and/or a computer program product at any possible technical detail level of integration. The computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
  • The computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device. The computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing. A non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing. A computer readable storage medium, as used herein, is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
  • Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network. The network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers. A network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
  • Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, configuration data for integrated circuitry, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++, or the like, and procedural programming languages, such as the “C” programming language or similar programming languages. The computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). In some embodiments, electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
  • Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer readable program instructions.
  • These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
  • The computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the blocks may occur out of the order noted in the Figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts or carry out combinations of special purpose hardware and computer instructions.
  • The following described exemplary embodiments provide a system, method and program product for conflict mitigation. As such, the present embodiment has the capacity to improve the technical field of conflict mitigation by providing real-time response feedback during a conversation that may lead to more conversations with positive results. More specifically, the conflict mitigator program may analyze each person's facial expressions, tone of speech and biometric responses during a real-time conversation in real-time in order to provide positive real-time responses to negative reactions. The real-time responses may be provided to an individual through a message which may be typed on a computing device screen or verbally communicated via a computing device speaker or headphones. The positive responses, provided by the conflict mitigator program, to negative reactions during communications between individuals may foster better and more trusting relationships between parties. A conflict mitigator program may provide real-time feedback for communicators and recipients in the event that there may be a misunderstanding or a mismatch between the recipients internal response and the intent of the communicator.
  • As described previously, sometimes communication between people can break down and leave participants confused as to why they are receiving a reaction they did not anticipate. Even well worded expressions may carry a tone that can negatively impact the recipient. The causes may be that the communicator could have difficulty reading others or have a reduced focus due to stress factors. Another level of communication breakdown can happen even if the communicator does understand the negative impact of their message. A breakdown may occur if the communicator does not know how to respond to the recipient who received the communication negatively. Further, the recipient may also not know how to respond to the negatively received message, which can lead to a degraded conversation and an untrusting relationship.
  • Examples of communicators and recipients may include a doctor and patient relationship, a bully and victim situation, an assisted living support worker and an assisted living resident relationship, a parent and child relationship, or between spouses. A misunderstanding and a proper real-time response may help the individuals in any of the previously mentioned relationships, or any other kind of communicating situation, understand one another and get a more positive result out of the interaction. A better, more positive response and result may lead to stronger relationships and a better understanding between the communicator and the recipient. Assessing sentiment and emotions through voice identifiers, facial expression identifiers and biometric identifiers may assist in understanding how a message is communicated and how a message is received. Therefore, it may be advantageous to, among other things, provide a way to offer real-time feedback to both the communicator and the recipient while providing each with a more positive response in order to foster better relationships.
  • The conflict mitigator program may identify personalities of communicators and recipients and during a conversation may compare the real-time interaction with previous interactions between the same parties. The comparison may identify differences in previous reactions or consistencies in previous reactions whether the reactions are positive or negative. The conflict mitigator program may have a response component that provides the parties with real-time feedback on how a message could have been worded better based on similar conversations that had positive emotive results, personalities of the parties and a history of interactions between the parties.
  • A personality example that the conflict mitigator program may take into consideration is someone who has difficulty with presenting constructive criticism and may need specific advice specific to that person's task. Alternatively, someone who does not handle criticism well might need specific advice on how to respond to criticism. An example of a history of interactions between parties may include a manager who consistently gets negative reactions from specific team members when giving direction. The manager may need feedback on how to solicit the team members advice before giving direction.
  • The conflict mitigator program may provide an individual with specific recommendations for better and more positively phrased real-time communication responses based on real-time communication with another person. The conflict mitigator program may provide real-time responses to individuals by using a cognitive analyzer to analyze both the communicator's and the recipient's tone, biometrics, facial expressions and sentiment. The tone identifier and the facial expression identifier may take data from the communicator while capturing the recipient's facial expression and biometric response. Upon receiving data in the form of communication tone, biometrics and facial expressions between two or more parties, the conflict mitigator program may use the currently received data along with the personalities involved and a history of reactions between the parties to identify exemplar conversations.
  • The cognitive analyzer component of the conflict mitigator program may compare data in order to provide a more positive conversational response by using a database that provides similar and more successful conversational responses. The cognitive analyzer may provide comparisons of conversations that have happened in the past with the current conversation. The response feature may provide the communicator and the receiver of the communication with a better real-time response (e.g., feedback) during the conversation.
  • The conflict mitigator program may provide the person that delivered a negatively received message with an alternative message based on similar but more positive resulting conversations, specific personalities involved and the history of interactions between the parties. Similarly, the conflict mitigator program may also provide the person who received a message that they reacted to in a negative way with a positive response based on similar criteria. When providing a response, the conflict mitigator program may remove identifying data (e.g., individual's names, locations, etc.) so that the better worded responses may provide positive responses without presenting information to the parties (e.g., personality types, history of interactions, etc.). While the conflict mitigator program may analyze personality types and a history of interactions, the information may not be provided in specific responses to communicators and recipients. The conflict mitigator program may be used and be beneficial for first instances of conversations between more than one person and with already developed relationships where the individuals have communicated in multiple instances.
  • One scenario of how the conflict mitigator program may foster a better relationship may be amongst co-workers, for example, Party A often tells jokes about a specific professional group to co-worker Party B. Party A thinks his jokes are funny and help to put people in a good mood in the morning, however, Party A does not know that Party B's spouse is in the specific professional group that Party A jokes about. The conflict mitigator program may detect Party B's increasing annoyance through facial expressions and tone and the conflict mitigator program may then recommend that Party A try a different small talk subject such as sports. Sports may be a topic chosen by the conflict mitigator program if there is a history of positive results amongst co-workers having a positive response to sports talk.
  • Another scenario of how the conflict mitigator program may foster a more trusting relationship may be between a doctor and a patient, for example when a doctor reviews test data with a patient after only one previous meeting between the two. The doctor had ordered blood tests and an ultrasound since the initial physical examination proved inconclusive. The patient is 18 years old and this is the first medical encounter with a doctor other than the patient's pediatrician. The patient is apprehensive to receive the results of the blood tests and the ultrasound since the patient has not previously had any serious conditions. Another apprehension the patient has is that the patient's pediatrician was friendly and was only visited for an occasional cold or sports related abrasions. During the follow-up appointment, the doctor explains to the young patient that the tests were inconclusive so a biopsy or surgery may be necessary. Although the doctor also explains that the patient should wait a month for a re-examination before getting a biopsy or surgery, the words surgery and inconclusive have already made the patient concerned. The 18-year-old patient does not understand what a biopsy is and surgery sounds scary so all the patient can muster up in a response is to say “ok.”
  • The conflict mitigator program may identify the patient's heightened tension through facial expression analysis and biometrics while the doctor's tone is analyzed as confident and even tempered. The conflict mitigator program identifies a misunderstanding (i.e., a mismatch) based on the patient's reaction to the doctor's words. The cognitive analyzer component of the program may compare the mismatch and find similar conversations between a doctor and a younger, more inexperienced, patient from a database where the doctor had to deliver news regarding a continuing diagnosis. The conflict mitigator program may then provide a real-time recommendation to the doctor to slow down and explain the terms to the new patient. Another provided real-time recommendation for the doctor's response may be an adjustment in the phrasing of the information to “While we may determine the need for some further tests, we are confident that the best next step is for you to first come back in a month for a re-examination to see how things many have changed.” The conflict mitigator program may also provide the patient with recommendations, such as to ask for clarification of the terms that were not understood and recommend the patient say “Doctor, please explain what a ‘biopsy’ is and how it is used?”
  • Another scenario of how the conflict mitigator program may foster a more trusting relationship may be between an employee and a resident of an assisted living facility, for example in a situation where an assisted living resident has resided at a facility for a year. The resident does not like physical contact and a newer employee of the facility likes to give friendly hugs when greeting people. The quick affection from the newer employee to the resident may cause the resident to be apprehensive and react negatively to the employee even though the employee believes they are being friendly.
  • The conflict mitigator program may identify the resident's apprehension when interacting with the friendly employee through biometrics and facial expression analysis. The conflict mitigator program may also identify the employee's joyful sentiment and openness which would create a misunderstanding (i.e., a mismatch) with the resident's reaction. The conflict mitigator program may then search a database to find similar scenarios and provide a recommendation of alternative real-time responses. An alternative real-time response for the employee may be to ask the resident permission before doing anything physical such as giving the resident a hug. Further, the conflict mitigator program may recommend that the resident state more directly to the employee about what kind of interaction is preferred.
  • The conflict mitigator program in the above stated examples may generate a real-time response by analyzing past behavior and by analyzing other similarly situated conversation environments. The conflict mitigator program may recognize that even if the communicators are prepared for a conversation, that face-to-face communication may still lead to misunderstandings that are specific to the parties involved and may provide guidance to the communicators in real-time. If there is a mismatch in sentiment during the conversation, the conflict mitigator program may notice that one person may be upset while the other is demonstrating positive characteristics such as joy or empathy. A real-time response may provide the communicators with similar conversations that had more positive results where the participants did not have a negative mood, were not in a heightened response and showed positive sentiments such as empathy, confidence, joy and/or openness. Previous similar conversations that the conflict mitigator program uses may be between the two parties conversing which had positive emotive results or between two different parties which had positive emotive results.
  • Referring to FIG. 1, an exemplary networked computer environment 100 in accordance with one embodiment is depicted. The networked computer environment 100 may include a computer 102 with a processor 104 and a data storage device 106 that is enabled to run a software program 108 and a conflict mitigator program 110 a. The networked computer environment 100 may also include a server 112 that is enabled to run a conflict mitigator program 110 b that may interact with a database 114 and a communication network 116. The networked computer environment 100 may include a plurality of computers 102 and servers 112, only one of which is shown. The communication network 116 may include various types of communication networks, such as a wide area network (WAN), local area network (LAN), a telecommunication network, a wireless network, a public switched network and/or a satellite network. It should be appreciated that FIG. 1 provides only an illustration of one implementation and does not imply any limitations with regard to the environments in which different embodiments may be implemented. Many modifications to the depicted environments may be made based on design and implementation requirements.
  • The client computer 102 may communicate with the server computer 112 via the communications network 116. The communications network 116 may include connections, such as wire, wireless communication links, or fiber optic cables. As will be discussed with reference to FIG. 4, server computer 112 may include internal components 902 a and external components 904 a, respectively, and client computer 102 may include internal components 902 b and external components 904 b, respectively. Server computer 112 may also operate in a cloud computing service model, such as Software as a Service (SaaS), Platform as a Service (PaaS), or Infrastructure as a Service (IaaS). Server 112 may also be located in a cloud computing deployment model, such as a private cloud, community cloud, public cloud, or hybrid cloud. Client computer 102 may be, for example, a mobile device, a telephone, a personal digital assistant, a netbook, a laptop computer, a tablet computer, a desktop computer, or any type of computing devices capable of running a program, accessing a network, and accessing a database 114. According to various implementations of the present embodiment, the conflict mitigator program 110 a, 110 b may interact with a database 114 that may be embedded in various storage devices, such as, but not limited to a computer/mobile device 102, a networked server 112, or a cloud storage service.
  • According to the present embodiment, a user using a client computer 102 or a server computer 112 may use the conflict mitigator program 110 a, 110 b (respectively) to obtain real-time responses to communication with another person where a misunderstanding may be present. The conflict mitigator method is explained in more detail below with respect to FIGS. 2 and 3.
  • Referring now to FIG. 2, a block diagram of the conflict mitigator program components according to at least one embodiment is depicted. The program components 200 may include identifier components such as a tone identifier 202, a biometric identifier 204 and a facial expression identifier 206. The program components 200 may also include a cognitive analyzer 208, a response feature 210 and a logging feature 212.
  • The tone identifier 202 may analyze the sentiment in the communicator's tone of voice (e.g., volume, inflection, sharpness, softness, etc.). The tone identifier 202 may also identify the tone of a digital message by analyzing the words written with other symbols and the use of capital letters. The biometric identifier 204 may measure heightened responses (e.g., changes in blood pressure, heart rate, breathing, pupil dilation, body temperature, etc.). The facial expression identifier 206 may recognize emotions (e.g., joy, fear, duress, anger, etc.).
  • The cognitive analyzer 208 may use computing technology (e.g., Watson Analytics™, Watson Analytics and all Watson Analytics-based trademarks and logos are trademarks or registered trademarks of International Business Machines Corporation and/or its affiliates) to analyze the tone identifiers 202, biometric identifiers 204, and facial expression identifiers 206 that have been inputted into the conflict mitigator program 110 a, 110 b. The cognitive analyzer 208 may analyze the personalities of the communicator and the recipient, the history of reactions between the personalities (e.g., positive or negative reactions) and may identify exemplary conversations that relate to the same subject that is being conversed. The cognitive analyzer 208 may detect patterns of conversations and reactions made by an individual. For example, whether a doctor has had consistent negative or positive responses from the nurse staff. This data may be useful to both the doctor to know how best to talk to co-workers and for the hospital to be able to track the doctor's effectiveness in communicating.
  • The response feature 210 may provide real-time feedback (e.g., through headphones, speakers, text, type written message, etc.) for both the communicator and the recipient based on the cognitive analyzer's identification of exemplar conversations. The response feature 210 may provide both parties with real-time feedback that may be more positive in nature and that might defuse any situation (e.g., alternative reactions and message phrasing). Further, the response feature 210 may also allow the communicators to build trust by communicating through any misunderstandings in real-time. The logging feature 212 may log the tone identifiers 202, biometric identifiers 204, facial expression identifiers 206, cognitive analysis and real-time response to capture the history of conversations and reactions between communicators for use in the present conversation or use in a future conversation. The conflict mitigator program 110 a, 110 b may use a database 114 to store the communications between communicators, the reactions and resolutions. The database 114 may provide better responses (i.e., feedback) as the database obtains more data to pull from. For example, the more conversations and reactions that are added to the database 114 (e.g., Watson) may create more data for the conflict mitigator program 110 a, 110 b to use when providing real-time feedback to communicators.
  • Referring now to FIG. 3, an operational flowchart illustrating the exemplary real-time conflict mitigator process 300 used by the conflict mitigator program 110 a, 110 b according to at least one embodiment is depicted.
  • At 302, the conflict mitigator program 110 a, 110 b receives real-time communication. The communication may be received either verbally or type-written. When the communicator speaks, or sends a message, both the recipient and the conflict mitigator program 110 a, 110 b receive the verbal or type-written message. The conflict mitigator program 110 a, 110 b may receive communication from any individual during a conversation. Verbal communication may be captured using a microphone on a device in an audio file format (e.g., .mp3 and .wav formats). The audio file may be filed in a data repository as an audio message. The type-written communication (i.e., type-written message) may be captured from an email or a text message and may be filed in a data repository.
  • Then at 304, the conflict mitigator program 110 a, 110 b determines if there is a mismatch in communication between the sender and the recipient. A mismatch may be a misunderstanding between the people conversing. A mismatch may occur when the intent of the communicator does not match the resulting emotional response of the recipient. The mismatch may even occur in the form of body contact such as a hug, as presented in the previous example between the employee and the assisted living facility resident. An example of a communication misunderstanding may occur if the communicator makes a statement or asks a question with good intentions and the statement is met with a response from the recipient that is upset, sad or angry. Good intentions may be measured by the communicator's tone of voice, facial expression features and biometric features. Similarly, the recipient's opposite response of sadness or anger may be measured by tone of voice, facial expression features and biometric features. An example of good intentions may show an even-level heart rate while anger may show an elevated heart rate.
  • The tone of voice feature may be captured by a microphone input into a digital audio file format and the conflict mitigator program 110 a, 110 b may identify certain patterns and compare the patterns with known patterns of tone of voice associated with positive and negative reactions. For example, a softer and even tone may convey a calm and nice tone of voice while a sharp and loud tone may convey an angry tone of voice. Facial expression features may be captured through a camera sensor and the conflict mitigator program 110 a, 110 b may conduct an image analysis or pattern matching, or both image analysis and pattern matching, to identify facial features associated with positive and negative reactions. For example, a smile may show a positive facial feature while furrowed eyebrows may convey a negative or confused facial feature. The biometric feature may use known technology to capture body reactions (e.g., body heat, blood flow, heart rate/pulse, dilated pupils, etc.) to comments or physical contact. For example, a LED (light-emitting diode) light may be shined from a device worn on the wrist onto the wrist to detect changes in blood flow. For example, a sharp rise in blood flow during a conversation may convey a more negative reaction.
  • The mismatch may be recognized by the conflict mitigator program 110 a, 110 b by noticing, for example, that the communicator's intent was good, however, the recipient had a negative response to the comment made by the communicator. The tone identifiers 202, biometric identifiers 204 and facial feature identifiers 206 may identify the communicator is positively communicating with good intent through a nice tone, regular heartrate and joyful facial features. The recipient may provide tone identifiers 202, biometric identifiers 204 and facial expression identifiers 206 to show a negative reaction by body temperature rising, facial features that represent a stressful emotion and a distressed tone of voice in a response. This mismatch (e.g., misunderstanding) will cause the conflict mitigator program 110 a, 110 b to search a positive response database (e.g., database 114) containing positive responses for better responses to provide for each party to use. The communicator may be presented with real-time feedback to reword the comment while the recipient may be presented with real-time feedback to ask the communicator to clarify what they meant.
  • For example, a doctor communicates to a young patient that the results to the tests that were run were inconclusive, a biopsy may be needed next month and surgery may follow. The young patient does not fully understand what is being communicated and gets scared of the word biopsy and the word surgery. An older patient in this scenario might not react negatively to this communication from the doctor because an older patient would most likely understand what the doctor is saying. The younger patient, however, does not know what a biopsy is and is frightened at the thought of surgery because the young patient has not been to a doctor for more than a common cold. Therefore, the doctor's good intent with empathetic emotions while communicating to the patient is a mismatch with the patients resulting fear.
  • If the conflict mitigator program 110 a, 110 b determined that there is not a mismatch in communication between the communicator and the recipient at 304, then the conflict mitigator program 110 a, 110 b will log the results in a database (e.g., database 114) at 306. If no mismatch in the tone identifiers 202, biometric identifiers 204, and facial feature identifier 206 between the communicator's intent and the recipient's reaction is present, then the conflict mitigator program 110 a, 110 b may not need to provide any real-time response to parties in the conversation. The conversation may be saved in a database 114 to be marked as successful and the data from the conversation may be used to make the cognitive analyzer 208 component more robust with potential positive responses. For example, if a doctor provides test results to a young patient who understood what the doctor was saying, and the young patient provided positive tone identifiers 202, biometric identifiers 204 and facial expression identifiers 206, then the details of the conversation may be used for a doctor in a similar situation but in an instance where the young patient reacted negatively to the conversation.
  • If the conflict mitigator program 110 a, 110 b determined that there is a mismatch in communication between the communicator and the recipient at 304, then the conflict mitigator program 110 a, 110 b will provide real-time feedback to both the communicator and the recipient at 308. The conflict mitigator program 110 a, 110 b provides real-time feedback for communicators and recipients in the event that there may be a misunderstanding or a mismatch between the recipients internal response and the intent of the communicator. Responses may be in the form of a message on a screen, a voice played through speakers, or through headphones. Responses may also be a combination of an audio message and a type-written message. Real-time feedback may turn negative conversations into positive conversations that yield stronger and more trusting relationships. The real-time component of the conflict mitigator program 110 a, 110 b allows the conversation to be corrected instantly while the conversation is happening. A person could go from having bad feelings associated with a person or conversation in the beginning of conversing to later walking away with good feelings associated with the same person or conversation. The real-time component of the conflict mitigator program 110 a, 110 b may create a more positive environment where the people communicating may get more value and feel better after having the conversation.
  • Continuing from the previous example, since there is a mismatch between the doctor's intent of communication and the patient's emotional response to the communication, the conflict mitigator program 110 a, 110 b may provide real-time responses to both the doctor and the patient. Real-time feedback for the doctor may tell the doctor to explain what a biopsy is to the patient and that nothing further needs to be done between now and the next checkup in a month when more information will be available after more tests are run. The conflict mitigator program 110 a, 110 b may provide real-time feedback for the doctor by searching for similar conversations between doctors and young patients in the positive response database (e.g., database 114), which may be more explanatory in nature by providing definitions to words that an older patient may already know. Alternatively, the conflict mitigator program 110 a, 110 b may provide real-time feedback for the patient to ask the doctor to explain what a biopsy is. Through the dialog that may come from the conflict mitigator program's 110 a, 110 b real-time feedback component, the doctor may earn the trust of the patient and the patient may walk away from the doctor appointment without fear. This more positive interaction may foster better trust and relationships.
  • Then at 310, the conflict mitigator program 110 a, 110 b will log the results in a database 114. The results are logged into a database 114 as stated above at 306. The log may continually collect data and continually grow, providing more responses to more types of situations thereby increasing the effectiveness of the conflict mitigator program 110 a, 110 b over time.
  • It may be appreciated that FIGS. 2 and 3 provide only an illustration of one embodiment and do not imply any limitations with regard to how different embodiments may be implemented. Many modifications to the depicted embodiment(s) may be made based on design and implementation requirements. An alternate embodiment may be where one party receives recommended (i.e., appropriate) responses from the conflict mitigator program 110 a, 110 b. For example, if the communicator has a device running the conflict mitigator program 110 a, 110 b, and was speaking to a recipient who did not have a device running the conflict mitigator program 110 a, 110 b, then the communicator's device may provide the communicator with proper responses based on sentiment analysis of both the communicator and the recipient.
  • FIG. 4 is a block diagram 900 of internal and external components of computers depicted in FIG. 1 in accordance with an illustrative embodiment of the present invention. It should be appreciated that FIG. 4 provides only an illustration of one implementation and does not imply any limitations with regard to the environments in which different embodiments may be implemented. Many modifications to the depicted environments may be made based on design and implementation requirements.
  • Data processing system 902, 904 is representative of any electronic device capable of executing machine-readable program instructions. Data processing system 902, 904 may be representative of a smart phone, a computer system, PDA, or other electronic devices. Examples of computing systems, environments, and/or configurations that may represented by data processing system 902, 904 include, but are not limited to, personal computer systems, server computer systems, thin clients, thick clients, hand-held or laptop devices, multiprocessor systems, microprocessor-based systems, network PCs, minicomputer systems, and distributed cloud computing environments that include any of the above systems or devices.
  • User client computer 102 and network server 112 may include respective sets of internal components 902 a, b and external components 904 a, b illustrated in FIG. 4. Each of the sets of internal components 902 a, b includes one or more processors 906, one or more computer-readable RAMs 908, and one or more computer-readable ROMs 910 on one or more buses 912, and one or more operating systems 914 and one or more computer-readable tangible storage devices 916. The one or more operating systems 914, the software program 108 and the conflict mitigator program 110 a in client computer 102, and the conflict mitigator program 110 b in network server 112, may be stored on one or more computer-readable tangible storage devices 916 for execution by one or more processors 906 via one or more RAMs 908 (which typically include cache memory). In the embodiment illustrated in FIG. 4, each of the computer-readable tangible storage devices 916 is a magnetic disk storage device of an internal hard drive. Alternatively, each of the computer-readable tangible storage devices 916 is a semiconductor storage device such as ROM 910, EPROM, flash memory or any other computer-readable tangible storage device that can store a computer program and digital information.
  • Each set of internal components 902 a, b also includes a R/W drive or interface 918 to read from and write to one or more portable computer-readable tangible storage devices 920 such as a CD-ROM, DVD, memory stick, magnetic tape, magnetic disk, optical disk or semiconductor storage device. A software program, such as the software program 108 and the conflict mitigator program 110 a, 110 b can be stored on one or more of the respective portable computer-readable tangible storage devices 920, read via the respective R/W drive or interface 918, and loaded into the respective hard drive 916.
  • Each set of internal components 902 a, b may also include network adapters (or switch port cards) or interfaces 922 such as a TCP/IP adapter cards, wireless wi-fi interface cards, or 3G or 4G wireless interface cards or other wired or wireless communication links. The software program 108 and the conflict mitigator program 110 a in client computer 102 and the conflict mitigator program 110 b in network server computer 112 can be downloaded from an external computer (e.g., server) via a network (for example, the Internet, a local area network or other, wide area network) and respective network adapters or interfaces 922. From the network adapters (or switch port adaptors) or interfaces 922, the software program 108 and the conflict mitigator program 110 a in client computer 102 and the conflict mitigator program 110 b in network server computer 112 are loaded into the respective hard drive 916. The network may comprise copper wires, optical fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
  • Each of the sets of external components 904 a, b can include a computer display monitor 924, a keyboard 926, and a computer mouse 928. External components 904 a, b can also include touch screens, virtual keyboards, touch pads, pointing devices, and other human interface devices. Each of the sets of internal components 902 a, b also includes device drivers 930 to interface to computer display monitor 924, keyboard 926, and computer mouse 928. The device drivers 930, R/W drive or interface 918, and network adapter or interface 922 comprise hardware and software (stored in storage device 916 and/or ROM 910).
  • It is understood in advance that although this disclosure includes a detailed description on cloud computing, implementation of the teachings recited herein are not limited to a cloud computing environment. Rather, embodiments of the present invention are capable of being implemented in conjunction with any other type of computing environment now known or later developed.
  • Cloud computing is a model of service delivery for enabling convenient, on-demand network access to a shared pool of configurable computing resources (e.g. networks, network bandwidth, servers, processing, memory, storage, applications, virtual machines, and services) that can be rapidly provisioned and released with minimal management effort or interaction with a provider of the service. This cloud model may include at least five characteristics, at least three service models, and at least four deployment models.
  • Characteristics are as follows:
      • On-demand self-service: a cloud consumer can unilaterally provision computing capabilities, such as server time and network storage, as needed automatically without requiring human interaction with the service's provider.
      • Broad network access: capabilities are available over a network and accessed through standard mechanisms that promote use by heterogeneous thin or thick client platforms (e.g., mobile phones, laptops, and PDAs).
      • Resource pooling: the provider's computing resources are pooled to serve multiple consumers using a multi-tenant model, with different physical and virtual resources dynamically assigned and reassigned according to demand. There is a sense of location independence in that the consumer generally has no control or knowledge over the exact location of the provided resources but may be able to specify location at a higher level of abstraction (e.g., country, state, or datacenter).
      • Rapid elasticity: capabilities can be rapidly and elastically provisioned, in some cases automatically, to quickly scale out and rapidly released to quickly scale in. To the consumer, the capabilities available for provisioning often appear to be unlimited and can be purchased in any quantity at any time.
      • Measured service: cloud systems automatically control and optimize resource use by leveraging a metering capability at some level of abstraction appropriate to the type of service (e.g., storage, processing, bandwidth, and active user accounts). Resource usage can be monitored, controlled, and reported providing transparency for both the provider and consumer of the utilized service.
  • Service Models are as follows:
      • Software as a Service (SaaS): the capability provided to the consumer is to use the provider's applications running on a cloud infrastructure. The applications are accessible from various client devices through a thin client interface such as a web browser (e.g., web-based e-mail). The consumer does not manage or control the underlying cloud infrastructure including network, servers, operating systems, storage, or even individual application capabilities, with the possible exception of limited user-specific application configuration settings.
      • Platform as a Service (PaaS): the capability provided to the consumer is to deploy onto the cloud infrastructure consumer-created or acquired applications created using programming languages and tools supported by the provider. The consumer does not manage or control the underlying cloud infrastructure including networks, servers, operating systems, or storage, but has control over the deployed applications and possibly application hosting environment configurations.
      • Infrastructure as a Service (IaaS): the capability provided to the consumer is to provision processing, storage, networks, and other fundamental computing resources where the consumer is able to deploy and run arbitrary software, which can include operating systems and applications. The consumer does not manage or control the underlying cloud infrastructure but has control over operating systems, storage, deployed applications, and possibly limited control of select networking components (e.g., host firewalls).
  • Deployment Models are as follows:
      • Private cloud: the cloud infrastructure is operated solely for an organization. It may be managed by the organization or a third party and may exist on-premises or off-premises.
      • Community cloud: the cloud infrastructure is shared by several organizations and supports a specific community that has shared concerns (e.g., mission, security requirements, policy, and compliance considerations). It may be managed by the organizations or a third party and may exist on-premises or off-premises.
      • Public cloud: the cloud infrastructure is made available to the general public or a large industry group and is owned by an organization selling cloud services.
      • Hybrid cloud: the cloud infrastructure is a composition of two or more clouds (private, community, or public) that remain unique entities but are bound together by standardized or proprietary technology that enables data and application portability (e.g., cloud bursting for load-balancing between clouds).
  • A cloud computing environment is service oriented with a focus on statelessness, low coupling, modularity, and semantic interoperability. At the heart of cloud computing is an infrastructure comprising a network of interconnected nodes.
  • Referring now to FIG. 5, illustrative cloud computing environment 1000 is depicted. As shown, cloud computing environment 1000 comprises one or more cloud computing nodes 100 with which local computing devices used by cloud consumers, such as, for example, personal digital assistant (PDA) or cellular telephone 1000A, desktop computer 1000B, laptop computer 1000C, and/or automobile computer system 1000N may communicate. Nodes 100 may communicate with one another. They may be grouped (not shown) physically or virtually, in one or more networks, such as Private, Community, Public, or Hybrid clouds as described hereinabove, or a combination thereof. This allows cloud computing environment 1000 to offer infrastructure, platforms and/or software as services for which a cloud consumer does not need to maintain resources on a local computing device. It is understood that the types of computing devices 1000A-N shown in FIG. 5 are intended to be illustrative only and that computing nodes 100 and cloud computing environment 1000 can communicate with any type of computerized device over any type of network and/or network addressable connection (e.g., using a web browser).
  • Referring now to FIG. 6, a set of functional abstraction layers 1100 provided by cloud computing environment 1000 is shown. It should be understood in advance that the components, layers, and functions shown in FIG. 6 are intended to be illustrative only and embodiments of the invention are not limited thereto. As depicted, the following layers and corresponding functions are provided:
  • Hardware and software layer 1102 includes hardware and software components. Examples of hardware components include: mainframes 1104; RISC (Reduced Instruction Set Computer) architecture based servers 1106; servers 1108; blade servers 1110; storage devices 1112; and networks and networking components 1114. In some embodiments, software components include network application server software 1116 and database software 1118.
  • Virtualization layer 1120 provides an abstraction layer from which the following examples of virtual entities may be provided: virtual servers 1122; virtual storage 1124; virtual networks 1126, including virtual private networks; virtual applications and operating systems 1128; and virtual clients 1130.
  • In one example, management layer 1132 may provide the functions described below. Resource provisioning 1134 provides dynamic procurement of computing resources and other resources that are utilized to perform tasks within the cloud computing environment. Metering and Pricing 1136 provide cost tracking as resources are utilized within the cloud computing environment, and billing or invoicing for consumption of these resources. In one example, these resources may comprise application software licenses. Security provides identity verification for cloud consumers and tasks, as well as protection for data and other resources. User portal 1138 provides access to the cloud computing environment for consumers and system administrators. Service level management 1140 provides cloud computing resource allocation and management such that required service levels are met. Service Level Agreement (SLA) planning and fulfillment 1142 provide pre-arrangement for, and procurement of, cloud computing resources for which a future requirement is anticipated in accordance with an SLA.
  • Workloads layer 1144 provides examples of functionality for which the cloud computing environment may be utilized. Examples of workloads and functions which may be provided from this layer include: mapping and navigation 1146; software development and lifecycle management 1148; virtual classroom education delivery 1150; data analytics processing 1152; transaction processing 1154; and conflict mitigation 1156. A conflict mitigator program 110 a, 110 b provides real-time feedback for communicators and recipients in the event that there may be a misunderstanding or a mismatch between the recipients internal response and the intent of the communicator.
  • The descriptions of the various embodiments of the present invention have been presented for purposes of illustration, but are not intended to be exhaustive or limited to the embodiments disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope of the described embodiments. The terminology used herein was chosen to best explain the principles of the embodiments, the practical application or technical improvement over technologies found in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein.

Claims (1)

1. A method for mitigating conflicts, the method comprising:
receiving communication by a recipient, wherein the received communication has a communicator sentiment associated with a communicator and a recipient sentiment associated with a recipient;
determining the communicator sentiment associated with the communicator message, wherein the communicator sentiment is measured by a plurality of biometric identifiers and a plurality of facial expression identifiers, wherein the biometric identifiers are selected from a group consisting of body heat, pupil dilation and heart rate, and wherein the facial expression identifiers are captured by a camera on a device or a sensor on a device;
determining the recipient sentiment associated with a received communicator message, wherein the recipient sentiment is measured by a plurality of biometric identifiers and a plurality of facial expression identifiers, wherein the biometric identifiers are selected from the group consisting of body heat, pupil dilation and heart rate, and wherein the facial expression identifiers are captured by a camera on a device or a sensor on a device;
determining that the determined recipient sentiment does not match the determined communicator sentiment in real-time based on the received communication, wherein the unmatched sentiment is a miscommunication between the communicator and the recipient, wherein the communicator makes a statement with good intentions and the statement is received by the recipient as a statement with not good intentions, wherein the plurality of biometric identifiers and the plurality of facial expression identifiers provide a plurality of data to indicate the communicator had good intentions and the recipient received the statement as not good intentions;
determining an appropriate response in real-time based on determining that the communicator sentiment does not match the recipient sentiment, wherein the appropriate response is a response that recognizes the misunderstanding and creates understanding between the communicator and the recipient;
presenting the determined appropriate response in real-time, wherein presenting the appropriate response comprises delivering a message selected from a group consisting of an audio message, a type-written message, and a combination audio message and type-written message, wherein the determined appropriate response is searched for in a positive response database, wherein the positive response database accumulates a plurality of logged conversation responses, wherein the positive response database stores data selected from a group consisting of a plurality of similar previous conversations, a plurality of previous positive conversations that produced a plurality of positive emotive results, a plurality of personalities of the communicator and the recipient and a plurality of historical interactions between the communicator and the recipient, wherein presenting the determined appropriate response comprises presenting the determined appropriate response to the communicator and the recipient; and
logging a conversation result in a database.
US15/700,250 2017-02-07 2017-09-11 Real-time conflict mitigator Abandoned US20180226072A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/700,250 US20180226072A1 (en) 2017-02-07 2017-09-11 Real-time conflict mitigator

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US15/426,394 US10657166B2 (en) 2017-02-07 2017-02-07 Real-time sentiment analysis for conflict mitigation using cognative analytics and identifiers
US15/700,250 US20180226072A1 (en) 2017-02-07 2017-09-11 Real-time conflict mitigator

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US15/426,394 Continuation US10657166B2 (en) 2017-02-07 2017-02-07 Real-time sentiment analysis for conflict mitigation using cognative analytics and identifiers

Publications (1)

Publication Number Publication Date
US20180226072A1 true US20180226072A1 (en) 2018-08-09

Family

ID=63037340

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/426,394 Active 2037-03-17 US10657166B2 (en) 2017-02-07 2017-02-07 Real-time sentiment analysis for conflict mitigation using cognative analytics and identifiers
US15/700,250 Abandoned US20180226072A1 (en) 2017-02-07 2017-09-11 Real-time conflict mitigator

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US15/426,394 Active 2037-03-17 US10657166B2 (en) 2017-02-07 2017-02-07 Real-time sentiment analysis for conflict mitigation using cognative analytics and identifiers

Country Status (1)

Country Link
US (2) US10657166B2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180268505A1 (en) * 2017-03-20 2018-09-20 Next Level Mediation Method of Performing Analysis-Based Conflict Mediation
US11276405B2 (en) * 2020-05-21 2022-03-15 International Business Machines Corporation Inferring sentiment to manage crowded spaces by using unstructured data

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9899038B2 (en) 2016-06-30 2018-02-20 Karen Elaine Khaleghi Electronic notebook system
JP2018167339A (en) * 2017-03-29 2018-11-01 富士通株式会社 Utterance control program, information processor, and utterance control method
US10764534B1 (en) 2017-08-04 2020-09-01 Grammarly, Inc. Artificial intelligence communication assistance in audio-visual composition
US10235998B1 (en) * 2018-02-28 2019-03-19 Karen Elaine Khaleghi Health monitoring system and appliance
US10832659B2 (en) * 2018-08-31 2020-11-10 International Business Machines Corporation Intent authoring using weak supervision and co-training for automated response systems
US10559307B1 (en) 2019-02-13 2020-02-11 Karen Elaine Khaleghi Impaired operator detection and interlock apparatus
US11106875B2 (en) 2019-05-20 2021-08-31 International Business Machines Corporation Evaluation framework for intent authoring processes
US11380306B2 (en) 2019-10-31 2022-07-05 International Business Machines Corporation Iterative intent building utilizing dynamic scheduling of batch utterance expansion methods
WO2022025778A1 (en) * 2020-07-28 2022-02-03 Motorola Solutions, Inc. Device, system and method for resolving misunderstandings on channels
US20220400026A1 (en) * 2021-06-15 2022-12-15 Microsoft Technology Licensing, Llc Retrospection assistant for virtual meetings

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130129076A1 (en) * 2011-11-23 2013-05-23 24/7 Customer, Inc. Interaction management
US20140112556A1 (en) * 2012-10-19 2014-04-24 Sony Computer Entertainment Inc. Multi-modal sensor based emotion recognition and emotional interface
US20140254775A1 (en) * 2013-03-06 2014-09-11 Avaya Inc. System and method for assisting agents of a contact center
US20150195406A1 (en) * 2014-01-08 2015-07-09 Callminer, Inc. Real-time conversational analytics facility
US9191513B1 (en) * 2014-06-06 2015-11-17 Wipro Limited System and method for dynamic job allocation based on acoustic sentiments
US20160127557A1 (en) * 2014-11-05 2016-05-05 Avaya Inc. System and method for agent selection in an enterprise
US20160147731A1 (en) * 2013-12-16 2016-05-26 Whistler Technologies Inc Message sentiment analyzer and feedback
US9380010B2 (en) * 2014-06-03 2016-06-28 International Business Machines Corporation Conversation branching for more efficient resolution
US9646198B2 (en) * 2014-08-08 2017-05-09 International Business Machines Corporation Sentiment analysis in a video conference

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050112535A1 (en) 2000-09-13 2005-05-26 Mcintosh Helen B. Method for enabling conflict resolution
US6904408B1 (en) 2000-10-19 2005-06-07 Mccarthy John Bionet method, system and personalized web content manager responsive to browser viewers' psychological preferences, behavioral responses and physiological stress indicators
US8495503B2 (en) 2002-06-27 2013-07-23 International Business Machines Corporation Indicating the context of a communication
US8337209B2 (en) 2008-08-27 2012-12-25 Ashman Jr Ward Computerized systems and methods for self-awareness and interpersonal relationship skill training and development for improving organizational efficiency
US9105042B2 (en) 2013-02-07 2015-08-11 Verizon Patent And Licensing Inc. Customer sentiment analysis using recorded conversation
US9282066B2 (en) 2013-07-18 2016-03-08 International Business Machines Corporation Targeted message response
US9830475B2 (en) 2014-07-15 2017-11-28 Sagesurfer Inc. Integrated collaboration platform for contextual communication
US10387573B2 (en) * 2015-06-01 2019-08-20 AffectLayer, Inc. Analyzing conversations to automatically identify customer pain points
US20180032126A1 (en) * 2016-08-01 2018-02-01 Yadong Liu Method and system for measuring emotional state
US9812151B1 (en) 2016-11-18 2017-11-07 IPsoft Incorporated Generating communicative behaviors for anthropomorphic virtual agents based on user's affect
US10587565B2 (en) * 2017-02-10 2020-03-10 International Business Machines Corporation Automatic response system based on body language

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130129076A1 (en) * 2011-11-23 2013-05-23 24/7 Customer, Inc. Interaction management
US20140112556A1 (en) * 2012-10-19 2014-04-24 Sony Computer Entertainment Inc. Multi-modal sensor based emotion recognition and emotional interface
US20140254775A1 (en) * 2013-03-06 2014-09-11 Avaya Inc. System and method for assisting agents of a contact center
US20160147731A1 (en) * 2013-12-16 2016-05-26 Whistler Technologies Inc Message sentiment analyzer and feedback
US20150195406A1 (en) * 2014-01-08 2015-07-09 Callminer, Inc. Real-time conversational analytics facility
US9380010B2 (en) * 2014-06-03 2016-06-28 International Business Machines Corporation Conversation branching for more efficient resolution
US9191513B1 (en) * 2014-06-06 2015-11-17 Wipro Limited System and method for dynamic job allocation based on acoustic sentiments
US9646198B2 (en) * 2014-08-08 2017-05-09 International Business Machines Corporation Sentiment analysis in a video conference
US20160127557A1 (en) * 2014-11-05 2016-05-05 Avaya Inc. System and method for agent selection in an enterprise

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180268505A1 (en) * 2017-03-20 2018-09-20 Next Level Mediation Method of Performing Analysis-Based Conflict Mediation
US11276405B2 (en) * 2020-05-21 2022-03-15 International Business Machines Corporation Inferring sentiment to manage crowded spaces by using unstructured data

Also Published As

Publication number Publication date
US10657166B2 (en) 2020-05-19
US20180226070A1 (en) 2018-08-09

Similar Documents

Publication Publication Date Title
US10657166B2 (en) Real-time sentiment analysis for conflict mitigation using cognative analytics and identifiers
US10692606B2 (en) Stress level reduction using haptic feedback
US11132648B2 (en) Cognitive-based enhanced meeting recommendation
US10902842B2 (en) System and method for cognitive intervention on human interactions
US9722965B2 (en) Smartphone indicator for conversation nonproductivity
US10535350B2 (en) Conflict resolution enhancement system
US20200050998A1 (en) Proactive communication channel controller in a collaborative environment
US11039783B2 (en) Automatic cueing system for real-time communication
WO2018150246A1 (en) Cognitive content filtering
US10915928B2 (en) Product solution responsive to problem identification
US20200020447A1 (en) Multi-Level Machine Learning To Detect A Social Media User's Possible Health Issue
US10996741B2 (en) Augmented reality conversation feedback
US20190090772A1 (en) Predicting thought based on neural mapping
US20210181791A1 (en) System, method, and recording medium for predicting cognitive states of a sender of an electronic message
US10373515B2 (en) System and method for cognitive intervention on human interactions
US10726844B2 (en) Smart medical room optimization of speech recognition systems
US10387566B2 (en) Assisting with written communication style based on recipient dress style
US11410110B2 (en) Alert generation based on a cognitive state and a physical state
US20190102737A1 (en) Methods and systems for receiving feedback
US20220108624A1 (en) Reader assistance method and system for comprehension checks
US10825070B2 (en) Problem identification using biometric and social data
US20220084504A1 (en) Artificial intelligence voice response system for speech impaired users
US11615645B2 (en) Automated presentation contributions
US11303649B2 (en) Maintaining electronic communications session continuity during session inactivity
US20230087133A1 (en) User assistance through demonstration

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GORZELA, RICHARD;SILVA, ASIMA;STOCKTON, JAIME M.;SIGNING DATES FROM 20170202 TO 20170203;REEL/FRAME:043541/0639

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION