US20230334827A1 - Methods and systems for providing images for facilitating communication - Google Patents

Methods and systems for providing images for facilitating communication Download PDF

Info

Publication number
US20230334827A1
US20230334827A1 US18/212,349 US202318212349A US2023334827A1 US 20230334827 A1 US20230334827 A1 US 20230334827A1 US 202318212349 A US202318212349 A US 202318212349A US 2023334827 A1 US2023334827 A1 US 2023334827A1
Authority
US
United States
Prior art keywords
information
identified customer
user
generative
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US18/212,349
Inventor
Ahn Truong
Vincent Pham
Fardin Abdi Taghi Abad
Jeremy Goodsitt
Mark Watson
Austin Walters
Kate Key
Reza Farivar
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Capital One Services LLC
Original Assignee
Capital One Services LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Capital One Services LLC filed Critical Capital One Services LLC
Priority to US18/212,349 priority Critical patent/US20230334827A1/en
Assigned to CAPITAL ONE SERVICES, LLC reassignment CAPITAL ONE SERVICES, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ABAD, FARDIN ABDI TAGHI, KEY, KATE, PHAM, VINCENT, GOODSITT, JEREMY, WALTERS, AUSTIN, WATSON, MARK, TRUONG, ANH, FARIVAR, REZA
Publication of US20230334827A1 publication Critical patent/US20230334827A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/53Querying
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/06Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/06Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
    • G10L21/10Transforming into visible information
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B21/00Teaching, or communicating with, the blind, deaf or mute
    • G09B21/009Teaching or communicating with deaf persons
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2354/00Aspects of interface with display user
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/06Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
    • G10L2021/065Aids for the handicapped in understanding

Definitions

  • the present disclosure relates generally to providing images for facilitating communication.
  • aspects of the disclosure include a computer-implemented method for interacting with a user.
  • Identity information for a user can be received.
  • the identity information can be analyzed to identify the user.
  • User information for an identified user can be retrieved, the user information indicating that a voice interacting with the identified user is to be translated into image data to help the identified user communicate with the voice.
  • Translated image data translating voice instructions by the voice can be retrieved.
  • the translated image data can be displayed to the identified user.
  • the translated image data can include photographic image data and/or non-photographic image data.
  • the photographic image data can be generated using a generative adversarial network (GAN).
  • GAN generative adversarial network
  • a sequence of voice instructions can be translated into a sequence of photographic images using the GAN.
  • the translated image data can include pre-defined images representing possible transactions.
  • aspects of the disclosure can include retrieving historical transaction information for the user, and determining a suggested transaction for the user based on the historical transaction information. Aspects of the disclosure can also include: determining the suggested transaction using previous transactions made by other users with demographics similar to the user; determining the suggested transaction using previous transactions made by the user; or determining the suggested transaction using a current location of the user and previous transactions made at the current location; or any combination thereof
  • aspects of the disclosure can include a system for interacting with a user, the system including: a memory storing instructions and a processor.
  • the processor can be configured to: receive physiological information and/or behavioral information for the user representing identifying information for a user; analyze the physiological information and/or the behavioral information for the user to determine an identity of the user; determine a need to provide photographic image data translating voice communication from a voice interacting with the user; display translated photographic image data translating the voice interacting with the user using a generative adversarial network (GAN); and accept instructions from the user.
  • GAN generative adversarial network
  • the translated photographic image data can be displayed in a video format.
  • the processor can also be configured to: play audio data while the translated photographic image data is displayed; and/or display text data while the translated photographic image data is displayed.
  • the processor can also be configured to: receive feedback information comprising body language information for the user; determine if the body language information signifies: an inability to understand the voice, a misunderstanding of the voice, or displeasure with a user experience, or any combination thereof; analyze the body language information in order to identify help information; and store the help information for the user.
  • the processor can be configured to: receive feedback information comprising language spoken by the user; determine if the language spoken by the user includes pre-defined words signifying: an inability to understand the voice, a misunderstanding of the voice; or displeasure with a user experience, or any combination thereof; analyze the language spoken by the user in order to identify help information; and store the help information for the user.
  • the processor can be configured to: receive feedback information from the user, the feedback information indicating difficulty the user has with understanding the transaction option; analyze the feedback information in order to identify help information; and store the help information for the user.
  • the feedback information can include survey information.
  • the processor can be configured to: receive feedback information from the user related to the appearance of a GAN translated photographic image; and store the feedback information for future application.
  • the feedback information can include preferences related to: color, template, outline, font, or font size, or any combination thereof.
  • a device can be provided for communicating with a user.
  • the device can include a memory storing instructions and a processor.
  • the processor can be configured to: display translated photographic image data translating a voice interacting with the user using a generative adversarial network (GAN); and accept instructions from the user.
  • GAN generative adversarial network
  • the processor can also be configured to determine a need to provide the translated photographic image data.
  • the processor can also be configured to: determine multiple transaction options for an identified user based on historical data; and display an image for each transaction option.
  • the processor can also be configured to: determine a probability for each transaction option; and display transaction option images in order of probability.
  • FIG. 1 is a diagram of an illustrative system 100 for interacting with a user, according to some embodiments of the present disclosure.
  • FIG. 2 is another diagram of an illustrative system 200 for interacting with a user, according to some embodiments of the present disclosure.
  • FIG. 3 is a flow diagram 300 showing processing that may occur when providing an insight within the system of FIGS. 1 and/or 2 , according to some embodiments of the present disclosure.
  • FIG. 4 is a flow diagram 400 showing processing that may occur when identifying a user within the system of FIGS. 1 and/or 2 , according to some embodiments of the present disclosure.
  • FIG. 5 illustrates a flow diagram 500 showing processing that may occur when analyzing and storing feedback information within the system of FIGS. 1 and/or 2 , according to some embodiments of the present disclosure.
  • FIG. 6 illustrates an example computer 600 , according to some embodiments of the present disclosure.
  • Embodiments of the present disclosure may help facilitate communication with a user (e.g., in order to help complete a transaction). For example, when people are going to an establishment (e.g., a bank, restaurant/bar, movie theater, theme park, sports venue, music venue, etc.) or visiting an establishment's physical mobile site, web site or mobile device application, it may make it easier for them to communicate and/or complete a transaction if they can better communicate and/or if a relevant transaction were suggested. Examples using a banking institution establishment are discussed below. However, those of ordinary skill in the art will see that the system may also be used by many other types of establishments.
  • an establishment e.g., a bank, restaurant/bar, movie theater, theme park, sports venue, music venue, etc.
  • a banking institution establishment are discussed below. However, those of ordinary skill in the art will see that the system may also be used by many other types of establishments.
  • FIG. 1 is a diagram of an illustrative system 100 for communicating with a user in order to complete a transaction, according to some embodiments of the present disclosure.
  • a client 160 can access (e.g., over a network 150 ) an identification module 110 , a convert voice to image module 115 , a determine historical data module 120 , a display module 130 , and a feedback module 140 .
  • the identification module 110 can identify the user using analyzed identity information.
  • a credit card, photo ID, or other type of identification can be analyzed to identify the user.
  • physiological information and/or behavioral information can be received from the user and provided to an identification system, and used to identify the user.
  • the physiological information can include: iris data; retina data; eye vein data; fingerprint data; hand geometry data; facial data; or finger vein data; or any combination thereof.
  • the behavioral information can include signature information and/or voice information.
  • the user can interact with a system that scans her eye, checks her fingerprint, hand, face or finger, or any combination thereof
  • the user can be asked to provide a signature or talk so that her signature or voice can be identified by the system.
  • the determine historical data module 120 can: review historical information for the identified user, provide needed help information, or suggest one or more transactions based on this analyzed historical transaction data, or any combination thereof.
  • the historical information can include feedback information and/or needed help information (e.g., such as whether or not the user would benefit from use of the convert voice to image module 115 ).
  • the historical information can determine if previous feedback information has been provided for an identified user. For example, it can be determined that the identified user has had trouble understanding the voice of a customer service representative and/or a recorded voice interacting with the user. In this case, the convert voice to image module 115 can be used to help the user understand what the customer service representative and/or the recorded voice are saying. Multiple examples of other types of feedback information are described below in more detail.
  • the historical information can include previous transactions made by the user and/or other users with demographics similar to the user.
  • the historical transaction information may also include information about where the user makes different types of transactions (e.g., at an establishment's physical site, a physical mobile site, a web site, a mobile device application, etc.).
  • possible transactions made by the user can include: opening an account, deposit funds (e.g., cash, check, electronic funds), transfer funds (e.g., within or out of the banking system), withdraw funds (e.g., cash, change in specific denominations ($1 bills, $5 bills, quarters, etc.), view information on an account (e.g., checking, savings, line of credit, mortgage, car loan, other loan), pay a bill (e.g., make a one time payment or set up an auto payment), obtain a cashier's check and/or money order, apply for a mortgage or other loan, get an item notarized, request information for an account issue, etc.
  • deposit funds e.g., cash, check, electronic funds
  • transfer funds e.g., within or out of the banking system
  • withdraw funds e.g., cash, change in specific denominations ($1 bills, $5 bills, quarters, etc.
  • view information on an account e.g., checking, savings, line of credit, mortgage, car loan, other loan
  • historical transaction information for a particular user can indicate that this user usually only uses the banking institution's physical banking machine to withdraw the maximum amount of cash allowed.
  • the banking system can suggest that the user wishes to withdraw the maximum amount of cash allowed.
  • historical transaction information for a particular user can indicate that this user transfers $500 (using the banking institution's mobile application) into her 20 year old daughter's account on the first of every month.
  • this user logs onto the mobile application at the first of the month, it can suggest that the user transfer $500 to her daughter's account.
  • historical transaction information for a particular user can indicate that this user visit's the establishment's web site at the end of the month to pay a car loan.
  • this user can suggest that the user pay the car loan payment.
  • historical transaction information for users in a person's demographic can indicate that it is common for users within this demographic withdraw cash every other Friday (e.g., payday).
  • historical information for users in another person's demographic e.g., men between 18 and 30
  • pay bills e.g., credit card bill, telephone bill
  • the convert voice to image module 115 can translate the voice communicating with the user into one or more images.
  • the translated images can be photo images and/or non-photo images. For example, for elderly people with hearing issues, they may also not be able to read (e.g., in a particular language or in any language). In this case, displaying text may not help and images may be more useful for the user.
  • the convert voice to image module 115 can generate photos using a generative adversarial network (GAN).
  • GAN can be a deep learning model that can translate one or more voice instructions to one or more photo instructions that are easy for the user to understand.
  • the voice instructions can be translated into text. Then, the GAN can synthesize images from the text. More information on the GAN can be found in the following, which are all herein incorporated by reference:
  • the system can translate the voice saying “click on the green button to approve” into text.
  • the GAN can then translate the text into a photo showing a hand pointing to a green button that has “OK” on it.
  • the system can translate “show me your debit card or ID” into a photo showing a debit card and driver's license.
  • the convert voice to image module 115 can generate photo images and/or non-photo images by recognizing key words in the voice that cause certain pre-determined photo images and/or non-photo images to be displayed. For example, if the convert voice to image module 115 recognizes the key words “show ID”, a screen can be displayed showing a photo image or a non-photo image of a debit card, a credit card, and a driver's license.
  • the display module 130 can display one or more images representing one or more suggested transactions.
  • the images can be actual photographs or appear to be photographs (e.g., an image that appears to be a photo of a hand pointing to a green button, the photo generated using a computer).
  • the images can be non-photographic images (e.g., a computer icon or computer image of a hand pointing to a green button).
  • the images can include pictures visualizing the different types of transactions.
  • the display module 130 can also display one or more images (e.g., photo images and/or non-photo images) showing actions the user can take with respect to a suggested transaction (e.g., hand pointing to green button).
  • the display module 130 can also display one or more images showing actions taken by the establishment (e.g., a picture of a receipt when a receipt is being printed).
  • the images can include still images and/or video.
  • some example transaction images can include: deposit funds (e.g., a picture of putting cash into a piggy bank), transfer funds (e.g., an image of moving cash from one picture of a bank to another picture of a bank)), withdraw funds (e.g., an image of taking cash out of a piggy bank), view information on an account (e.g., an image of a piece of paper with account information on it)), pay a bill (e.g., an image of sending an envelope with money in it), obtain a cashier's check and/or money order (e.g., an image of a cashier's check or money order), apply for a mortgage or other loan (e.g., an image of a house, car, etc.), get an item notarized (e.g., an image of a notary stamp), request information for an account issue (e.g., a question mark over small image of sample account summary), ask a question (e.g., a question mark), etc
  • deposit funds e.g
  • the images can be video images.
  • one or more suggested transaction images can be shown as videos that play once, or over and over until the user chooses a transaction.
  • the videos can be shown when a user chooses a particular suggested transaction image.
  • the system can also play audio data while the image data is displayed. For example, data explaining a particular suggested transaction can be played while the image for that suggested transaction is shown.
  • the transaction can be displayed in order of probability (e.g., with the most probably transaction listed first), and any audio data can be played in the same order the transactions are listed on the screen.
  • the system can display text data while the image data is displayed.
  • the text data can be displayed near or on the image.
  • the feedback module can receive feedback information from or for the user.
  • the feedback information can include: body language information for the user, language spoken by the user, or survey information, or any combination thereof.
  • the feedback module can determine if the body language information and/or the language spoken by the user signifies: an inability to understand the voice, a misunderstanding of the voice, or displeasure with a user experience, or any combination thereof.
  • the feedback module can analyze the body language information in order to identify help information.
  • the feedback module can determine any difficulty the user has with understanding a transaction option.
  • the feedback module can also analyze the feedback information in order to identify help information.
  • the feedback module can also help determine any preferences the user has for GAN translated photographic image data.
  • the feedback module can receive feedback information related to the appearance of GAN translated photographic image data.
  • the feedback information can include: body language information for the user, language spoken by the user, or survey information, or any combination thereof. This feedback information can be stored and applied to generate GAN translated photographic image data according to the user's preferences when the user returns.
  • the feedback information can include preferences related to: color, template, outline, font, or font size, or any combination thereof.
  • the preferences of that user can be determined. For example, it can be determined that general users (or users in a certain demographic, or a particular user) tend to like photos better that have: dark colors rather than light colors, no green and/or blue colors, content focused at the center (as opposed to spread-out content).
  • training data can be restricted to generate synthetic images based on the user preferences. For example, using the example feedback provided above, the system can only select synthetic images with dark colors, simple content focused at the center of the screen, a big font size, etc. This information can train the GAN model. Thus, when the user returns to the bank, the system can apply the GAN model.
  • FIG. 2 is another diagram of an illustrative system 200 for providing a suggested transaction, according to some embodiments of the present disclosure.
  • a client 260 can access (e.g., over a network 150 using server(s) 110 ) data representing transactions between customers and establishments stored in a database 120 .
  • the database 120 can store instructions in one or more memories.
  • FIG. 3 is a flow diagram 300 showing processing that may occur within the system of FIG. 1 and/or FIG. 2 , according to some embodiments of the present disclosure.
  • the process can start.
  • step 315 it can be determined if a user can be identified. If not, in step 320 , the process can end. If yes, in step 340 , user information can be found.
  • step 345 it can be determined if historical data for the identified user can be found. If not, the process can move to 362 . If yes, the process can move to step 355 , where suggested transaction options and/or help information can be determined.
  • suggested transaction options can be displayed using image (e.g., photo image and/or non-photo image) data.
  • a voice e.g., live or recorded
  • image e.g., photo image and/or non-photo image
  • instructions from the user can be obtained.
  • feedback from the user can be obtained.
  • the process can end.
  • FIG. 4 is a flow diagram showing additional processing that may occur in step 315 of flow diagram 300 of FIG. 3 , according to some embodiments of the present disclosure.
  • physiological information can be obtained from the user.
  • behavioral information can be obtained from the user.
  • the user can be asked to self-identify.
  • the user can be identified using the physiological, behavioral or self-identified information.
  • FIG. 5 is a flow diagram showing additional processing that may occur in step 365 of flow diagram 300 of FIG. 3 , according to some embodiments of the present disclosure.
  • survey information can be obtained from the user.
  • body language can be obtained from the user.
  • language can be obtained from the user.
  • the survey information, the body language information, and/or the language information can be analyzed.
  • the analyzed survey information, the body language information, and/or the language information can be analyzed as feedback information.
  • the transaction data can include many different properties related to a transaction. This may include: a customer name, a customer ID (e.g., anonymous or not) that allows identification of a person who is making a purchase and what other purchases they have made before and after that transaction, merchant, merchant ID (e.g., anonymous or not), merchant name, location of merchant, amount of the purchase, or how a purchase was made (e.g., in person, online, with APPLE PAY, with card dip, with card swipe, etc.), or any combination thereof.
  • a customer name e.g., anonymous or not
  • merchant ID e.g., anonymous or not
  • merchant name e.g., location of merchant
  • amount of the purchase e.g., amount of the purchase
  • a purchase was made e.g., in person, online, with APPLE PAY, with card dip, with card swipe, etc.
  • the confluence of seeing a multitude of transaction data may provide for powerful insights.
  • Methods described herein may represent processing that occurs within a system for providing an insight about an establishment (e.g., system 100 of FIG. 1 and/or system 200 of FIG. 2 ).
  • the subject matter described herein can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware, including the structural means disclosed in this specification and structural equivalents thereof, or in combinations of them.
  • the subject matter described herein can be implemented as one or more computer program products, such as one or more computer programs tangibly embodied in an information carrier (e.g., in a machine readable storage device), or embodied in a propagated signal, for execution by, or to control the operation of, data processing apparatus (e.g., a programmable processor, a computer, or multiple computers).
  • a computer program (also known as a program, software, software application, or code) can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment.
  • a computer program does not necessarily correspond to a file.
  • a program can be stored in a portion of a file that holds other programs or data, in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code).
  • a computer program can be deployed to be executed on one computer or on multiple computers at one site or distributed across multiple sites and interconnected by a communication network.
  • the processes and logic flows described in this specification can be performed by one or more programmable processors (e.g., processor 600 in FIG. 6 ) executing one or more computer programs to perform functions of the subject matter described herein by operating on input data and generating output.
  • the processes and logic flows can also be performed by, and apparatus of the subject matter described herein can be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
  • FIG. 6 illustrates an example computer 605 , according to some embodiments of the present disclosure.
  • Computer 605 can include a processor 610 suitable for the execution of a computer program, and can include, by way of example, both general and special purpose microprocessors, and any one or more processor of any kind of digital computer.
  • a processor can receive instructions and data from a main memory 630 (e.g., a read only memory or a random access memory or both).
  • Processor 610 can execute instructions and the memory 630 can store instructions and data.
  • a computer can include, or be operatively coupled to receive data from or transfer data to, or both, a storage medium 640 for storing data (e.g., magnetic, magneto optical disks, or optical disks).
  • Information carriers suitable for embodying computer program instructions and data can include all forms of nonvolatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, flash memory device, or magnetic disks.
  • the processor 610 and the memory 630 can be supplemented by, or incorporated in, special purpose logic circuitry.
  • the computer 605 can also include an input/output 620 , a display 650 , and a communications interface 660 .

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Acoustics & Sound (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Computation (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Signal Processing (AREA)
  • User Interface Of Digital Computer (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Machine Translation (AREA)

Abstract

Aspects of the disclosure include computer-implemented methods and systems for providing generative adversarial network (GAN) digital image data. GAN digital image data corresponding to a suggested transaction for an identified customer can be determined.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. application Ser. No. 16/425,347, filed May 29, 2019, the content of which is incorporated herein by reference in its entirety.
  • This application is related to U.S. patent application Ser. No. 16/425,248, which is titled “Methods And Systems For Providing Changes To A Voice Interacting With A User”, filed May 29, 2019, Attorney Docket no. 359025-100174, the content of which is incorporated herein by reference in its entirety.
  • FIELD
  • The present disclosure relates generally to providing images for facilitating communication.
  • BACKGROUND
  • When people are visiting an establishment (e.g., a store or restaurant), they often wish to communicate (e.g., in order to make a transaction). For example, when a person cannot understand either spoken or written communications related to a transaction, help may be needed in order for the person to complete the transaction. There is a need to provide help to make the process easier.
  • SUMMARY
  • Aspects of the disclosure include a computer-implemented method for interacting with a user. Identity information for a user can be received. The identity information can be analyzed to identify the user. User information for an identified user can be retrieved, the user information indicating that a voice interacting with the identified user is to be translated into image data to help the identified user communicate with the voice. Translated image data translating voice instructions by the voice can be retrieved. The translated image data can be displayed to the identified user.
  • Aspects of the disclosure can also include accepting instructions from the identified user based on user interaction with the translated image data. The translated image data can include photographic image data and/or non-photographic image data. The photographic image data can be generated using a generative adversarial network (GAN). A sequence of voice instructions can be translated into a sequence of photographic images using the GAN. The translated image data can include pre-defined images representing possible transactions.
  • Aspects of the disclosure can include retrieving historical transaction information for the user, and determining a suggested transaction for the user based on the historical transaction information. Aspects of the disclosure can also include: determining the suggested transaction using previous transactions made by other users with demographics similar to the user; determining the suggested transaction using previous transactions made by the user; or determining the suggested transaction using a current location of the user and previous transactions made at the current location; or any combination thereof
  • Aspects of the disclosure can include a system for interacting with a user, the system including: a memory storing instructions and a processor. The processor can be configured to: receive physiological information and/or behavioral information for the user representing identifying information for a user; analyze the physiological information and/or the behavioral information for the user to determine an identity of the user; determine a need to provide photographic image data translating voice communication from a voice interacting with the user; display translated photographic image data translating the voice interacting with the user using a generative adversarial network (GAN); and accept instructions from the user.
  • According to some aspects, the translated photographic image data can be displayed in a video format. The processor can also be configured to: play audio data while the translated photographic image data is displayed; and/or display text data while the translated photographic image data is displayed.
  • According to some aspects, the processor can also be configured to: receive feedback information comprising body language information for the user; determine if the body language information signifies: an inability to understand the voice, a misunderstanding of the voice, or displeasure with a user experience, or any combination thereof; analyze the body language information in order to identify help information; and store the help information for the user.
  • According to some aspects, the processor can be configured to: receive feedback information comprising language spoken by the user; determine if the language spoken by the user includes pre-defined words signifying: an inability to understand the voice, a misunderstanding of the voice; or displeasure with a user experience, or any combination thereof; analyze the language spoken by the user in order to identify help information; and store the help information for the user.
  • According to some aspects, the processor can be configured to: receive feedback information from the user, the feedback information indicating difficulty the user has with understanding the transaction option; analyze the feedback information in order to identify help information; and store the help information for the user. The feedback information can include survey information.
  • According to some aspects, the processor can be configured to: receive feedback information from the user related to the appearance of a GAN translated photographic image; and store the feedback information for future application. The feedback information can include preferences related to: color, template, outline, font, or font size, or any combination thereof.
  • According to some aspects, a device can be provided for communicating with a user. The device can include a memory storing instructions and a processor. The processor can be configured to: display translated photographic image data translating a voice interacting with the user using a generative adversarial network (GAN); and accept instructions from the user. The processor can also be configured to determine a need to provide the translated photographic image data. The processor can also be configured to: determine multiple transaction options for an identified user based on historical data; and display an image for each transaction option. The processor can also be configured to: determine a probability for each transaction option; and display transaction option images in order of probability.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Various objectives, features, and advantages of the disclosed subject matter can be more fully appreciated with reference to the following detailed description of the disclosed subject matter when considered in connection with the following drawings, in which like reference numerals identify like elements.
  • FIG. 1 is a diagram of an illustrative system 100 for interacting with a user, according to some embodiments of the present disclosure.
  • FIG. 2 is another diagram of an illustrative system 200 for interacting with a user, according to some embodiments of the present disclosure.
  • FIG. 3 is a flow diagram 300 showing processing that may occur when providing an insight within the system of FIGS. 1 and/or 2 , according to some embodiments of the present disclosure.
  • FIG. 4 is a flow diagram 400 showing processing that may occur when identifying a user within the system of FIGS. 1 and/or 2 , according to some embodiments of the present disclosure.
  • FIG. 5 illustrates a flow diagram 500 showing processing that may occur when analyzing and storing feedback information within the system of FIGS. 1 and/or 2 , according to some embodiments of the present disclosure.
  • FIG. 6 illustrates an example computer 600, according to some embodiments of the present disclosure.
  • The drawings are not necessarily to scale, or inclusive of all elements of a system, emphasis instead generally being placed upon illustrating the concepts, structures, and techniques sought to be protected herein.
  • DETAILED DESCRIPTION OF EMBODIMENTS OF THE INVENTION
  • Embodiments of the present disclosure may help facilitate communication with a user (e.g., in order to help complete a transaction). For example, when people are going to an establishment (e.g., a bank, restaurant/bar, movie theater, theme park, sports venue, music venue, etc.) or visiting an establishment's physical mobile site, web site or mobile device application, it may make it easier for them to communicate and/or complete a transaction if they can better communicate and/or if a relevant transaction were suggested. Examples using a banking institution establishment are discussed below. However, those of ordinary skill in the art will see that the system may also be used by many other types of establishments.
  • FIG. 1 is a diagram of an illustrative system 100 for communicating with a user in order to complete a transaction, according to some embodiments of the present disclosure. As shown in FIG. 1 , a client 160 can access (e.g., over a network 150) an identification module 110, a convert voice to image module 115, a determine historical data module 120, a display module 130, and a feedback module 140.
  • The identification module 110 can identify the user using analyzed identity information. In some aspects of the disclosure, a credit card, photo ID, or other type of identification can be analyzed to identify the user. In other aspects, physiological information and/or behavioral information can be received from the user and provided to an identification system, and used to identify the user. The physiological information can include: iris data; retina data; eye vein data; fingerprint data; hand geometry data; facial data; or finger vein data; or any combination thereof. The behavioral information can include signature information and/or voice information.
  • For example, the user can interact with a system that scans her eye, checks her fingerprint, hand, face or finger, or any combination thereof In addition, the user can be asked to provide a signature or talk so that her signature or voice can be identified by the system.
  • The determine historical data module 120 can: review historical information for the identified user, provide needed help information, or suggest one or more transactions based on this analyzed historical transaction data, or any combination thereof. The historical information can include feedback information and/or needed help information (e.g., such as whether or not the user would benefit from use of the convert voice to image module 115).
  • For example, the historical information can determine if previous feedback information has been provided for an identified user. For example, it can be determined that the identified user has had trouble understanding the voice of a customer service representative and/or a recorded voice interacting with the user. In this case, the convert voice to image module 115 can be used to help the user understand what the customer service representative and/or the recorded voice are saying. Multiple examples of other types of feedback information are described below in more detail.
  • The historical information can include previous transactions made by the user and/or other users with demographics similar to the user. The historical transaction information may also include information about where the user makes different types of transactions (e.g., at an establishment's physical site, a physical mobile site, a web site, a mobile device application, etc.).
  • If the system is a banking system, possible transactions made by the user can include: opening an account, deposit funds (e.g., cash, check, electronic funds), transfer funds (e.g., within or out of the banking system), withdraw funds (e.g., cash, change in specific denominations ($1 bills, $5 bills, quarters, etc.), view information on an account (e.g., checking, savings, line of credit, mortgage, car loan, other loan), pay a bill (e.g., make a one time payment or set up an auto payment), obtain a cashier's check and/or money order, apply for a mortgage or other loan, get an item notarized, request information for an account issue, etc.
  • For example, historical transaction information for a particular user can indicate that this user usually only uses the banking institution's physical banking machine to withdraw the maximum amount of cash allowed. Thus, once the banking system identifies the user as using one of their physical banking machines, the banking system can suggest that the user wishes to withdraw the maximum amount of cash allowed.
  • As another example, historical transaction information for a particular user can indicate that this user transfers $500 (using the banking institution's mobile application) into her 20 year old daughter's account on the first of every month. Thus, if the user logs onto the mobile application at the first of the month, it can suggest that the user transfer $500 to her daughter's account.
  • As a further example, historical transaction information for a particular user can indicate that this user visit's the establishment's web site at the end of the month to pay a car loan. Thus, if the user logs in to the establishment's web site at the end of the month, it can suggest that the user pay the car loan payment.
  • For example, historical transaction information for users in a person's demographic (e.g., men over 60) can indicate that it is common for users within this demographic withdraw cash every other Friday (e.g., payday). As another example, historical information for users in another person's demographic (e.g., men between 18 and 30) can indicate that it is common for these users to electronically transfer funds (e.g., move money from checking to savings account) and pay bills (e.g., credit card bill, telephone bill) every other Friday (e.g., payday).
  • The convert voice to image module 115 can translate the voice communicating with the user into one or more images. The translated images can be photo images and/or non-photo images. For example, for elderly people with hearing issues, they may also not be able to read (e.g., in a particular language or in any language). In this case, displaying text may not help and images may be more useful for the user.
  • In some embodiments, the convert voice to image module 115 can generate photos using a generative adversarial network (GAN). The GAN can be a deep learning model that can translate one or more voice instructions to one or more photo instructions that are easy for the user to understand. In some embodiments, the voice instructions can be translated into text. Then, the GAN can synthesize images from the text. More information on the GAN can be found in the following, which are all herein incorporated by reference:
      • Reed, Scott et al., “Generative Adversarial Text to Image Synthesys,” Proceedings of the 33rd International Conference on Machine Learning (New York, NY, USA, 2016).
      • Goodfellow, Ian et al., “Generative Adversarial Nets,” Proceedings of Advances in Neural Information Processing Systems (NIPS2014) pp. 2672-2680.
      • Apr. 23, 2019 Generative Adversarial Network Wikipedia page (https://en.wikipedia.org/wiki/Generative_adversarial_network).
  • For example, the system can translate the voice saying “click on the green button to approve” into text. The GAN can then translate the text into a photo showing a hand pointing to a green button that has “OK” on it. As another example, the system can translate “show me your debit card or ID” into a photo showing a debit card and driver's license.
  • In some embodiments, the convert voice to image module 115 can generate photo images and/or non-photo images by recognizing key words in the voice that cause certain pre-determined photo images and/or non-photo images to be displayed. For example, if the convert voice to image module 115 recognizes the key words “show ID”, a screen can be displayed showing a photo image or a non-photo image of a debit card, a credit card, and a driver's license.
  • The display module 130 can display one or more images representing one or more suggested transactions. In some embodiments, the images can be actual photographs or appear to be photographs (e.g., an image that appears to be a photo of a hand pointing to a green button, the photo generated using a computer). In some embodiments, the images can be non-photographic images (e.g., a computer icon or computer image of a hand pointing to a green button).
  • Instructions from the user can be accepted based on user interaction with the image. The images can include pictures visualizing the different types of transactions. The display module 130 can also display one or more images (e.g., photo images and/or non-photo images) showing actions the user can take with respect to a suggested transaction (e.g., hand pointing to green button). The display module 130 can also display one or more images showing actions taken by the establishment (e.g., a picture of a receipt when a receipt is being printed). The images can include still images and/or video.
  • If the system is a banking system, some example transaction images can include: deposit funds (e.g., a picture of putting cash into a piggy bank), transfer funds (e.g., an image of moving cash from one picture of a bank to another picture of a bank)), withdraw funds (e.g., an image of taking cash out of a piggy bank), view information on an account (e.g., an image of a piece of paper with account information on it)), pay a bill (e.g., an image of sending an envelope with money in it), obtain a cashier's check and/or money order (e.g., an image of a cashier's check or money order), apply for a mortgage or other loan (e.g., an image of a house, car, etc.), get an item notarized (e.g., an image of a notary stamp), request information for an account issue (e.g., a question mark over small image of sample account summary), ask a question (e.g., a question mark), etc.
  • In some aspects of the disclosure, the images can be video images. For example, one or more suggested transaction images can be shown as videos that play once, or over and over until the user chooses a transaction. As another example, the videos can be shown when a user chooses a particular suggested transaction image.
  • The system can also play audio data while the image data is displayed. For example, data explaining a particular suggested transaction can be played while the image for that suggested transaction is shown.
  • If multiple transactions are suggested, the transaction can be displayed in order of probability (e.g., with the most probably transaction listed first), and any audio data can be played in the same order the transactions are listed on the screen.
  • In addition, the system can display text data while the image data is displayed. The text data can be displayed near or on the image.
  • The feedback module can receive feedback information from or for the user. The feedback information can include: body language information for the user, language spoken by the user, or survey information, or any combination thereof. The feedback module can determine if the body language information and/or the language spoken by the user signifies: an inability to understand the voice, a misunderstanding of the voice, or displeasure with a user experience, or any combination thereof. The feedback module can analyze the body language information in order to identify help information.
  • The feedback module can determine any difficulty the user has with understanding a transaction option. The feedback module can also analyze the feedback information in order to identify help information.
  • The feedback module can also help determine any preferences the user has for GAN translated photographic image data. For example, the feedback module can receive feedback information related to the appearance of GAN translated photographic image data. The feedback information can include: body language information for the user, language spoken by the user, or survey information, or any combination thereof. This feedback information can be stored and applied to generate GAN translated photographic image data according to the user's preferences when the user returns. The feedback information can include preferences related to: color, template, outline, font, or font size, or any combination thereof. Using user feedback on a photo showed to a user, the preferences of that user can be determined. For example, it can be determined that general users (or users in a certain demographic, or a particular user) tend to like photos better that have: dark colors rather than light colors, no green and/or blue colors, content focused at the center (as opposed to spread-out content).
  • In some embodiments, training data can be restricted to generate synthetic images based on the user preferences. For example, using the example feedback provided above, the system can only select synthetic images with dark colors, simple content focused at the center of the screen, a big font size, etc. This information can train the GAN model. Thus, when the user returns to the bank, the system can apply the GAN model.
  • FIG. 2 is another diagram of an illustrative system 200 for providing a suggested transaction, according to some embodiments of the present disclosure. As shown in FIG. 2 , a client 260 can access (e.g., over a network 150 using server(s) 110) data representing transactions between customers and establishments stored in a database 120. The database 120 can store instructions in one or more memories.
  • FIG. 3 is a flow diagram 300 showing processing that may occur within the system of FIG. 1 and/or FIG. 2 , according to some embodiments of the present disclosure. In step 301, the process can start. In step 315, it can be determined if a user can be identified. If not, in step 320, the process can end. If yes, in step 340, user information can be found. In step 345, it can be determined if historical data for the identified user can be found. If not, the process can move to 362. If yes, the process can move to step 355, where suggested transaction options and/or help information can be determined. In 360, suggested transaction options can be displayed using image (e.g., photo image and/or non-photo image) data. In 362, a voice (e.g., live or recorded) interacting with the user can be translated to image (e.g., photo image and/or non-photo image) data. In 365, instructions from the user can be obtained. In 370, feedback from the user can be obtained. In 375, the process can end.
  • FIG. 4 is a flow diagram showing additional processing that may occur in step 315 of flow diagram 300 of FIG. 3 , according to some embodiments of the present disclosure. In 410, physiological information can be obtained from the user. In 420, behavioral information can be obtained from the user. In 430, the user can be asked to self-identify. In 440, the user can be identified using the physiological, behavioral or self-identified information.
  • FIG. 5 is a flow diagram showing additional processing that may occur in step 365 of flow diagram 300 of FIG. 3 , according to some embodiments of the present disclosure. In step 510, survey information can be obtained from the user. In step 520, body language can be obtained from the user. In step 530, language can be obtained from the user. In 540, the survey information, the body language information, and/or the language information can be analyzed. In 550, the analyzed survey information, the body language information, and/or the language information can be analyzed as feedback information.
  • In some aspects of the disclosure the transaction data can include many different properties related to a transaction. This may include: a customer name, a customer ID (e.g., anonymous or not) that allows identification of a person who is making a purchase and what other purchases they have made before and after that transaction, merchant, merchant ID (e.g., anonymous or not), merchant name, location of merchant, amount of the purchase, or how a purchase was made (e.g., in person, online, with APPLE PAY, with card dip, with card swipe, etc.), or any combination thereof. The confluence of seeing a multitude of transaction data may provide for powerful insights.
  • Methods described herein may represent processing that occurs within a system for providing an insight about an establishment (e.g., system 100 of FIG. 1 and/or system 200 of FIG. 2 ). The subject matter described herein can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware, including the structural means disclosed in this specification and structural equivalents thereof, or in combinations of them. The subject matter described herein can be implemented as one or more computer program products, such as one or more computer programs tangibly embodied in an information carrier (e.g., in a machine readable storage device), or embodied in a propagated signal, for execution by, or to control the operation of, data processing apparatus (e.g., a programmable processor, a computer, or multiple computers). A computer program (also known as a program, software, software application, or code) can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program does not necessarily correspond to a file. A program can be stored in a portion of a file that holds other programs or data, in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers at one site or distributed across multiple sites and interconnected by a communication network.
  • The processes and logic flows described in this specification, including the method steps of the subject matter described herein, can be performed by one or more programmable processors (e.g., processor 600 in FIG. 6 ) executing one or more computer programs to perform functions of the subject matter described herein by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus of the subject matter described herein can be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
  • FIG. 6 illustrates an example computer 605, according to some embodiments of the present disclosure. Computer 605 can include a processor 610 suitable for the execution of a computer program, and can include, by way of example, both general and special purpose microprocessors, and any one or more processor of any kind of digital computer. A processor can receive instructions and data from a main memory 630 (e.g., a read only memory or a random access memory or both). Processor 610 can execute instructions and the memory 630 can store instructions and data. A computer can include, or be operatively coupled to receive data from or transfer data to, or both, a storage medium 640 for storing data (e.g., magnetic, magneto optical disks, or optical disks). Information carriers suitable for embodying computer program instructions and data can include all forms of nonvolatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, flash memory device, or magnetic disks. The processor 610 and the memory 630 can be supplemented by, or incorporated in, special purpose logic circuitry. The computer 605 can also include an input/output 620, a display 650, and a communications interface 660.
  • It is to be understood that the disclosed subject matter is not limited in its application to the details of construction and to the arrangements of the components set forth in the following description or illustrated in the drawings. The disclosed subject matter is capable of other embodiments and of being practiced and carried out in various ways. Accordingly, other implementations are within the scope of the following claims. Also, it is to be understood that the phraseology and terminology employed herein are for the purpose of description and should not be regarded as limiting. As such, those skilled in the art will appreciate that the conception, upon which this disclosure is based, may readily be utilized as a basis for the designing of other structures, methods, and systems for carrying out the several purposes of the disclosed subject matter. It is important, therefore, that the claims be regarded as including such equivalent constructions insofar as they do not depart from the spirit and scope of the disclosed subject matter.
  • Although the disclosed subject matter has been described and illustrated in the foregoing exemplary embodiments, it is understood that the present disclosure has been made only by way of example, and that numerous changes in the details of implementation of the disclosed subject matter may be made without departing from the spirit and scope of the disclosed subject matter.
  • In addition, it should be understood that any figures which highlight the functionality and advantages are presented for example purposes only. The disclosed methodology and system are each sufficiently flexible and configurable such that they may be utilized in ways other than that shown. For example, other steps may be provided, or steps may be eliminated, from the described flows, and other components may be added to, or removed from, the described systems. In addition, the order of the steps illustrated or described may be changed.
  • Although the term “at least one” may often be used in the specification, claims and drawings, the terms “a”, “an”, “the”, “said”, etc. also signify “at least one” or “the at least one” in the specification, claims and drawings.
  • Finally, it is the applicant's intent that only claims that include the express language “means for” or “step for” be interpreted under 35 U.S.C. 112(f). Claims that do not expressly include the phrase “means for” or “step for” are not to be interpreted under 35 U.S.C. 112(f).

Claims (21)

1.-21. (canceled)
22. A system for interacting with a customer, comprising:
a non-transitory memory storing instructions; and
a processor that, when executing the instructions, cause the processor to:
determine generative artificial intelligence (AI) user preferences related to generative AI digital image data preferred by an identified customer;
receive a live voice communication of a service representative interacting with the identified customer;
translate the live voice communication into text;
generate an image, using the generative AI, based on a combination of the text and the user preferences; and
transmit the image to the identified customer for display.
23. The system of claim 22, wherein the generative AI is adversarial.
24. The system of claim 22, wherein the instructions further cause the processor to:
generate one or more secondary images, using the generative AI, based on a combination of the text and the user preferences; and
transmit the image and the one or more secondary images to the identified customer for display as a video.
25. The system of claim 22, wherein the instructions further cause the processor to:
play audio data while the image is displayed.
26. The system of claim 22, wherein the instructions further cause the processor to:
display at least a portion of the text while the image is displayed.
27. The system of claim 22, wherein the instructions further cause the processor to:
receive feedback information comprising body language information for the identified customer;
determine if the body language information signifies an inability to understand the live voice, a misunderstanding of the live voice, or displeasure with a user experience, or any combination thereof;
analyze the body language information in order to identify help information; and
store the help information for the identified customer.
28. The system of claim 22, wherein the instructions further cause the processor to:
receive feedback information comprising language spoken by the identified customer;
determine if the language spoken by the user comprises pre-defined words signifying: an inability to understand the live voice, a misunderstanding of the live voice; or displeasure with a user experience, or any combination thereof;
analyze the language spoken by the identified customer in order to identify help information; and
store the help information for the identified customer.
29. The system of claim 22, wherein the instructions further cause the processor to:
receive feedback information from the identified customer, the feedback information indicating difficulty the identified customer has with understanding a transaction option;
analyze the feedback information in order to identify help information; and
store the help information for the identified customer.
30. The system of claim 29, wherein the feedback information comprises survey information.
31. The system of claim 22, wherein the user preferences comprise accessibility information associated with the identified customer.
32. The system of claim 22, wherein the user preferences comprise one or more transactions available to the identified customer.
33. The system of claim 30, wherein the image is associated with the one or more transactions, and wherein the instructions further cause the processor to:
determine a probability for each of the one or more transactions; and
display the one or more transactions in an order based on the probability.
34. The system of claim 30, wherein the user preferences comprise historical transaction data.
35. The system of claim 22, wherein the instructions further cause the processor to:
receive identity information from a customer; and
determine the customer is an identified customer based on the identity information.
36. A computer-implemented method for providing images to a customer interacting with a service representative, comprising:
determining generative artificial intelligence (AI) user preferences related to generative AI digital image data preferred by an identified customer;
receiving a live voice communication of a service representative interacting with the identified customer;
translating the live voice communication into text;
generating an image, using the generative AI, based on a combination of the text and the user preferences; and
transmitting the image to the identified customer for display.
37. The computer-implemented method of claim 34, wherein the generative AI is adversarial.
38. The computer-implemented method of claim 34, further comprising:
generating one or more secondary images, using the generative AI, based on a combination of the text and the user preferences; and
transmitting the image and the one or more secondary images to the identified customer for display as a video.
39. A device for communicating with a customer, comprising:
a memory storing instructions; and
a processor that, when executing the instructions, is configured to:
determine generative artificial intelligence (AI) user preferences related to generative AI digital image data preferred by an identified customer;
receive a live voice communication of a service representative interacting with the identified customer;
translate the live voice communication into text;
generate an image, using the generative AI, based on a combination of the text and the user preferences; and
transmit the image to the identified customer for display.
40. The device of claim 39, wherein the generative AI is adversarial.
41. The device of claim 39, wherein the processor is further configured to:
generate one or more secondary images, using the generative AI, based on a combination of the text and the user preferences; and
transmit the image and the one or more secondary images to the identified customer for display as a video.
US18/212,349 2019-05-29 2023-06-21 Methods and systems for providing images for facilitating communication Pending US20230334827A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/212,349 US20230334827A1 (en) 2019-05-29 2023-06-21 Methods and systems for providing images for facilitating communication

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US16/425,347 US10896686B2 (en) 2019-05-29 2019-05-29 Methods and systems for providing images for facilitating communication
US17/110,882 US11715285B2 (en) 2019-05-29 2020-12-03 Methods and systems for providing images for facilitating communication
US18/212,349 US20230334827A1 (en) 2019-05-29 2023-06-21 Methods and systems for providing images for facilitating communication

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US17/110,882 Continuation US11715285B2 (en) 2019-05-29 2020-12-03 Methods and systems for providing images for facilitating communication

Publications (1)

Publication Number Publication Date
US20230334827A1 true US20230334827A1 (en) 2023-10-19

Family

ID=73551371

Family Applications (3)

Application Number Title Priority Date Filing Date
US16/425,347 Active US10896686B2 (en) 2019-05-29 2019-05-29 Methods and systems for providing images for facilitating communication
US17/110,882 Active 2040-01-08 US11715285B2 (en) 2019-05-29 2020-12-03 Methods and systems for providing images for facilitating communication
US18/212,349 Pending US20230334827A1 (en) 2019-05-29 2023-06-21 Methods and systems for providing images for facilitating communication

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US16/425,347 Active US10896686B2 (en) 2019-05-29 2019-05-29 Methods and systems for providing images for facilitating communication
US17/110,882 Active 2040-01-08 US11715285B2 (en) 2019-05-29 2020-12-03 Methods and systems for providing images for facilitating communication

Country Status (1)

Country Link
US (3) US10896686B2 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7266432B2 (en) * 2019-03-14 2023-04-28 本田技研工業株式会社 AGENT DEVICE, CONTROL METHOD OF AGENT DEVICE, AND PROGRAM
US10896686B2 (en) * 2019-05-29 2021-01-19 Capital One Services, Llc Methods and systems for providing images for facilitating communication
US10878800B2 (en) 2019-05-29 2020-12-29 Capital One Services, Llc Methods and systems for providing changes to a voice interacting with a user

Family Cites Families (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5933805A (en) 1996-12-13 1999-08-03 Intel Corporation Retaining prosody during speech analysis for later playback
GB9824762D0 (en) 1998-11-11 1999-01-06 Ncr Int Inc Self-service terminal
US6377925B1 (en) * 1999-12-16 2002-04-23 Interactive Solutions, Inc. Electronic translator for assisting communications
US6507643B1 (en) * 2000-03-16 2003-01-14 Breveon Incorporated Speech recognition system and method for converting voice mail messages to electronic mail messages
US20020198716A1 (en) * 2001-06-25 2002-12-26 Kurt Zimmerman System and method of improved communication
IL144818A (en) 2001-08-09 2006-08-20 Voicesense Ltd Method and apparatus for speech analysis
US7457757B1 (en) 2002-05-30 2008-11-25 Plantronics, Inc. Intelligibility control for speech communications systems
JP2005346252A (en) * 2004-06-01 2005-12-15 Nec Corp Information transmission system and information transmission method
US7844045B2 (en) 2004-06-16 2010-11-30 Panasonic Corporation Intelligent call routing and call supervision method for call centers
US7133513B1 (en) * 2004-07-21 2006-11-07 Sprint Spectrum L.P. Method and system for transcribing voice content of an on-going teleconference into human-readable notation
US9282377B2 (en) * 2007-05-31 2016-03-08 iCommunicator LLC Apparatuses, methods and systems to provide translations of information into sign language or other formats
US20090043583A1 (en) 2007-08-08 2009-02-12 International Business Machines Corporation Dynamic modification of voice selection based on user specific factors
US20090171670A1 (en) 2007-12-31 2009-07-02 Apple Inc. Systems and methods for altering speech during cellular phone use
US8457688B2 (en) 2009-02-26 2013-06-04 Research In Motion Limited Mobile wireless communications device with voice alteration and related methods
WO2011042543A1 (en) * 2009-10-08 2011-04-14 Sony Corporation Automated language-tutoring method
US8611876B2 (en) * 2009-10-15 2013-12-17 Larry Miller Configurable phone with interactive voice response engine
US20110208523A1 (en) * 2010-02-22 2011-08-25 Kuo Chien-Hua Voice-to-dactylology conversion method and system
CN102939791B (en) * 2010-05-17 2015-09-23 塔塔咨询服务有限公司 For having the hand communication assistor of people of the sense of hearing, speech and dysopia
US20120250039A1 (en) 2011-04-01 2012-10-04 Arthur Austin Ollivierre System and method for presenting information to a user
US9493130B2 (en) 2011-04-22 2016-11-15 Angel A. Penilla Methods and systems for communicating content to connected vehicle users based detected tone/mood in voice input
US20160062987A1 (en) * 2014-08-26 2016-03-03 Ncr Corporation Language independent customer communications
DE112014007207B4 (en) 2014-11-25 2019-12-24 Mitsubishi Electric Corporation Information presentation system
US20160210982A1 (en) 2015-01-16 2016-07-21 Social Microphone, Inc. Method and Apparatus to Enhance Speech Understanding
DK3550858T3 (en) 2015-12-30 2023-06-12 Gn Hearing As A HEAD PORTABLE HEARING AID
US10157626B2 (en) 2016-01-20 2018-12-18 Harman International Industries, Incorporated Voice affect modification
US10014841B2 (en) 2016-09-19 2018-07-03 Nokia Technologies Oy Method and apparatus for controlling audio playback based upon the instrument
US9973627B1 (en) 2017-01-25 2018-05-15 Sorenson Ip Holdings, Llc Selecting audio profiles
US20180277132A1 (en) 2017-03-21 2018-09-27 Rovi Guides, Inc. Systems and methods for increasing language accessability of media content
US20180302686A1 (en) * 2017-04-14 2018-10-18 International Business Machines Corporation Personalizing closed captions for video content
US10521512B2 (en) 2017-05-26 2019-12-31 Bose Corporation Dynamic text-to-speech response from a smart speaker
TWI639114B (en) 2017-08-30 2018-10-21 元鼎音訊股份有限公司 Electronic device with a function of smart voice service and method of adjusting output sound
US10803859B1 (en) 2017-09-05 2020-10-13 Amazon Technologies, Inc. Speech processing for public devices
US10381005B2 (en) 2017-11-28 2019-08-13 Toyota Motor Engineering & Manufacturing North America, Inc. Systems and methods for determining user frustration when using voice control
US10339508B1 (en) 2018-02-12 2019-07-02 Capital One Services, Llc Methods for determining user experience (UX) effectiveness of ATMs
US10333934B1 (en) 2018-05-14 2019-06-25 Capital One Services, Llc Method and system for verifying user identity
US10446158B1 (en) 2018-06-07 2019-10-15 Capital One Services, Llc Method and system for voice input at an ATM without audibly revealing a selected transaction
US10303963B1 (en) 2018-07-09 2019-05-28 Capital One Services, Llc ATM with biometric security
US10581625B1 (en) 2018-11-20 2020-03-03 International Business Machines Corporation Automatically altering the audio of an object during video conferences
US10509949B1 (en) 2019-01-24 2019-12-17 Capital One Services, Llc Method and system for customizing user experience
US10878800B2 (en) 2019-05-29 2020-12-29 Capital One Services, Llc Methods and systems for providing changes to a voice interacting with a user
US10896686B2 (en) * 2019-05-29 2021-01-19 Capital One Services, Llc Methods and systems for providing images for facilitating communication

Also Published As

Publication number Publication date
US20200381004A1 (en) 2020-12-03
US11715285B2 (en) 2023-08-01
US10896686B2 (en) 2021-01-19
US20210090588A1 (en) 2021-03-25

Similar Documents

Publication Publication Date Title
US20230334827A1 (en) Methods and systems for providing images for facilitating communication
US10176466B2 (en) Check cashing automated banking machine
US10861089B2 (en) System and method for customizing real-time applications on a user interface
US10719817B2 (en) Wearable transaction devices
US9286726B2 (en) Mobile information gateway for service provider cooperation
US9665901B2 (en) Mobile information gateway for private customer interaction
US20160098692A1 (en) Method for processing transaction statement inquiries via an atm
US11682062B2 (en) Customized e-commerce tags in realtime multimedia content
US20160098700A1 (en) Method for providing privacy through atm communication to consumer devices
US20230325796A1 (en) Methods for determining user experience (ux) effectiveness of atms
KR100715323B1 (en) Apparatus and method for prohibiting false electronic banking using face recognition technology
US20190385152A1 (en) Single payment validation gateway
CN116018608A (en) Electronic commerce label in multimedia content
US10579780B1 (en) Systems and methods for document authentication
US20210065145A1 (en) Systems and methods for acceptance of payments to a business demand deposit account
CN113283627A (en) Service providing method, device, electronic equipment and machine readable medium
JP6473531B1 (en) Automatic split payment system using face recognition technology
CN110826443B (en) Bill data processing method, system, device and medium based on in vivo detection
KR20230068812A (en) Exchange apparatus linked with fintech
CA2929205A1 (en) Wearable transaction devices

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: CAPITAL ONE SERVICES, LLC, VIRGINIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TRUONG, ANH;PHAM, VINCENT;ABAD, FARDIN ABDI TAGHI;AND OTHERS;SIGNING DATES FROM 20190228 TO 20190528;REEL/FRAME:064352/0106