US20210264429A1 - Reducing false positive fraud alerts for card-present financial transactions - Google Patents

Reducing false positive fraud alerts for card-present financial transactions Download PDF

Info

Publication number
US20210264429A1
US20210264429A1 US15/466,002 US201715466002A US2021264429A1 US 20210264429 A1 US20210264429 A1 US 20210264429A1 US 201715466002 A US201715466002 A US 201715466002A US 2021264429 A1 US2021264429 A1 US 2021264429A1
Authority
US
United States
Prior art keywords
data
transaction
location
geographic location
fraud
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/466,002
Inventor
Timothy Kramme
Elizabeth Flowers
Reena Batra
Miriam Valero
Puneit Dua
Shanna L. Phillips
Russell Ruestman
Bradley A. Craig
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
State Farm Mutual Automobile Insurance Co
Original Assignee
State Farm Mutual Automobile Insurance Co
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by State Farm Mutual Automobile Insurance Co filed Critical State Farm Mutual Automobile Insurance Co
Priority to US15/466,002 priority Critical patent/US20210264429A1/en
Assigned to STATE FARM MUTUAL AUTOMOBILE INSURANCE COMPANY reassignment STATE FARM MUTUAL AUTOMOBILE INSURANCE COMPANY ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BATRA, REENA, KRAMME, TIMOTHY, Craig, Bradley A., FLOWERS, ELIZABETH, VALERO, MIRIAM, DUA, PUNEIT, PHILLIPS, SHANNA L., RUESTMAN, RUSSELL
Publication of US20210264429A1 publication Critical patent/US20210264429A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/018Certifying business or products
    • G06Q30/0185Product, service or business identity fraud
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models
    • G06N5/046Forward inferencing; Production systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/08Payment architectures
    • G06Q20/10Payment architectures specially adapted for electronic funds transfer [EFT] systems; specially adapted for home banking systems
    • G06Q20/102Bill distribution or payments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/08Payment architectures
    • G06Q20/20Point-of-sale [POS] network systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/22Payment schemes or models
    • G06Q20/24Credit schemes, i.e. "pay after"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/30Payment architectures, schemes or protocols characterised by the use of specific devices or networks
    • G06Q20/32Payment architectures, schemes or protocols characterised by the use of specific devices or networks using wireless devices
    • G06Q20/322Aspects of commerce using mobile devices [M-devices]
    • G06Q20/3224Transactions dependent on location of M-devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/30Payment architectures, schemes or protocols characterised by the use of specific devices or networks
    • G06Q20/34Payment architectures, schemes or protocols characterised by the use of specific devices or networks using cards, e.g. integrated circuit [IC] cards or magnetic cards
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/38Payment protocols; Details thereof
    • G06Q20/40Authorisation, e.g. identification of payer or payee, verification of customer or shop credentials; Review and approval of payers, e.g. check credit lines or negative lists
    • G06Q20/401Transaction verification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/38Payment protocols; Details thereof
    • G06Q20/40Authorisation, e.g. identification of payer or payee, verification of customer or shop credentials; Review and approval of payers, e.g. check credit lines or negative lists
    • G06Q20/401Transaction verification
    • G06Q20/4016Transaction verification involving fraud or risk level assessment in transaction processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/38Payment protocols; Details thereof
    • G06Q20/40Authorisation, e.g. identification of payer or payee, verification of customer or shop credentials; Review and approval of payers, e.g. check credit lines or negative lists
    • G06Q20/407Cancellation of a transaction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/38Payment protocols; Details thereof
    • G06Q20/40Authorisation, e.g. identification of payer or payee, verification of customer or shop credentials; Review and approval of payers, e.g. check credit lines or negative lists
    • G06Q20/409Device specific authentication in transaction processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0207Discounts or incentives, e.g. coupons or rebates
    • G06Q30/0225Avoiding frauds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/01Dynamic search techniques; Heuristics; Dynamic trees; Branch-and-bound
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0241Advertisements
    • G06Q30/0248Avoiding fraud
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L2209/00Additional information or applications relating to cryptographic mechanisms or cryptographic arrangements for secret or secure communication H04L9/00
    • H04L2209/56Financial cryptography, e.g. electronic payment or e-cash

Definitions

  • the present disclosure generally relates to financial fraud and, more specifically, to processing techniques for reducing false positive fraud alerts.
  • Financial fraud in its many forms, is a problem of enormous magnitude and scope, causing billions of dollars in economic losses and impacting many millions of people.
  • Types of financial fraud include use of a lost or stolen card, account takeover, skimming, chargeback (“friendly”) fraud, counterfeiting, forgeries and application (e.g., loan application) fraud, to name just a few.
  • friendly chargeback
  • counterfeiting forgeries
  • application e.g., loan application
  • the present embodiments may, inter alia, use new processing techniques to reduce false positive fraud alerts.
  • fraud alerts may be generated, or fraud alerts based upon various other triggers (e.g., presence of a large transaction, presence of a transaction initiated in a different state or country, cardholder reporting of unrecognized or fraudulent charges, etc.) may be either confirmed or ruled out (e.g., identified as a false positive), using location information.
  • a method of reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions is implemented in one or more servers.
  • the method may include: (1) determining, by one or more processors of the one or more servers, that a fraud alert exists for a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account; (2) determining, by the one or more processors, a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determining, by the one or more processors, a time of the financial transaction; (4) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction; (5) determining, by the one or more
  • a computer system for reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions includes a location database configured to store geolocation data indicating geographic locations of authorized cardholders over time, one or more processors, and a non-transitory memory.
  • the memory stores instructions that, when executed by the one or more processors, cause the one or more processors to: (1) determine that a fraud alert exists for a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account; (2) determine a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determine a time of the financial transaction; (4) determine, based upon first geolocation data stored in the location database and indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction; (5) determine that the second geographic location corresponds to the first geographic location; and/or (6) in response to determining that the second geographic location corresponds to the first geographic location, mark the fraud alert as a false positive such that no fraud alert is sent to the authorized cardholder in connection with the
  • a computer-implemented method of preventing fraudulent card-present financial transactions is implemented in one or more servers.
  • the method may include: (1) receiving, by one or more processors of the one or more servers, a request to authorize a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly being entered into by an authorized cardholder associated with the debit or credit card account; (2) determining, by the one or more processors, a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at a time of the financial transaction; (4) determining, by the one or more processors, that the second geographic location does not correspond to the first geographic location; and/or (5) in response to determining that the second geographic location
  • FIG. 1 depicts an exemplary environment in which techniques for fraud detection, verification and/or classification may be implemented, according to one embodiment.
  • FIG. 2 depicts an exemplary process flow for machine learning of fraud detection, verification and/or classification rules, according to one embodiment.
  • FIGS. 3A-3F depict exemplary process flows for machine learning of particular types of fraud detection, verification and/or classification rules, according to different embodiments.
  • FIGS. 4A-4F depict exemplary factors and algorithms that may be used in connection with various fraud detection, verification and/or classification rule sets, according to different embodiments.
  • FIGS. 5 and 6 illustrate exemplary computer-implemented methods of using customer data to determine that geolocation-based fraud alerts are false positives, according to different embodiments.
  • FIGS. 7 through 10 illustrate exemplary computer-implemented methods of using information about the locations of authorized cardholders to prevent false positive fraud alerts, or to block potentially fraudulent financial transactions, according to different embodiments.
  • FIG. 11 depicts an exemplary computer system in which the techniques described herein may be implemented, according to one embodiment.
  • detecting” or “determining” fraud may be used herein to refer to initially flagging fraudulent (or potentially fraudulent) activity, to verifying/confirming that suspect/flagged activity was indeed fraudulent, or generally to both.
  • the systems and techniques described herein may be used, for example, to identify, prevent and/or quantify/measure instances of lost or stolen card use, account takeover, counterfeiting, skimming, chargeback (“friendly”) fraud, collusive merchant fraud, application (e.g., loan application) fraud, mortgage fraud, and/or one or more other types of fraud relating to existing and/or potential financial transactions and/or accounts.
  • application e.g., loan application
  • a fraud detection and/or classification system may analyze data relating to a number of existing or potential financial accounts.
  • the analysis/processing may be performed in batch processing operations, or substantially in real-time (e.g., as the data is generated and/or as financial transactions occur, etc.), and the data may be obtained from a variety of sources based upon the particular embodiment and/or scenario.
  • data from financial account records may be analyzed, along with data indicating online activity of an account holder, location data (e.g., global positioning satellite (GPS) data from a smartphone or vehicle of the account holder) and/or other data, to determine whether a particular financial transaction was fraudulent or likely fraudulent.
  • the analysis may be performed automatically after the transaction has been made, or may be performed in response to a person or algorithm flagging the transaction as a potentially fraudulent one, for example.
  • GPS global positioning satellite
  • the analysis may include determining whether the account holder has expressed interest in the object (e.g., product or service) of the transaction or the merchant, and/or determining whether the transaction is consistent with spending patterns associated with the account holder (e.g., spending patterns identified using the account holder's transaction records), for example.
  • the account holder e.g. multiple credit or debit card holders
  • accuracy may be improved by identifying spending patterns at the individual level rather than, or in addition to, at the aggregate account level.
  • a maximum amount of money typically spent in a single transaction may be determined for each of two cardholders listed on a single account, and the maximum amount for the cardholder who purportedly made a particular purchase may be compared to the purchase amount to determine whether fraud is suspected.
  • the locations of authorized cardholders may be analyzed, in conjunction with the locations at which cards were presented to a merchant or merchant device (if a card-present transaction) or the locations of computing devices via which card information was entered (if an online transaction), to determine whether a fraud alert is likely a false positive.
  • locations may be analyzed to determine whether to block a transaction that is currently in-process (e.g., by issuing a fraud alert to the merchant or card issuer, or by not clearing the transaction, etc.).
  • problems that have beset the field of fraud detection, classification and/or prevention in the past may be greatly mitigated or eliminated.
  • information that has conventionally been overlooked or ignored may be used to more accurately detect, prevent and/or classify fraud, and/or to reduce false positive fraud alerts.
  • a significant amount of time may be saved by removing the need for manual investigations, or by reducing the number of instances where manual investigations are required.
  • FIG. 1 depicts an exemplary environment 10 in which techniques for fraud detection and/or classification may be implemented, according to one embodiment.
  • the environment 10 may include an anti-fraud services system (AFSS) 12 , a financial account management system (FAMS) 14 , a card network computing system 16 , a number of cardholder computing devices 20 , a number of merchant computing systems 22 , a number of other sources 24 , and a network 26 .
  • AFSS anti-fraud services system
  • FAMS financial account management system
  • the environment 10 may include one or more additional financial account management systems and/or card network computing systems, and/or one or more of the cardholder computing devices 20 may instead be a computing device of a holder of a non-card account (e.g., a checking, savings or loan account) or an applicant for a new account (e.g., a new loan account).
  • the environment 10 may include a computing system of one or more acquiring/merchant banks, and some or all of the communications with merchant computing systems 22 described below may instead be with the acquiring bank(s).
  • FAMS 14 may be associated with (e.g., owned and/or maintained by) a bank or other financial entity.
  • FAMS 14 may be a bank that acts as a card issuer associated with a particular type of card network (e.g., VISA®, Mastercard®, etc.), and/or an entity that provides loans (e.g., mortgage, home equity, vehicle, etc.), saving/checking account services, and/or other financial services to customers.
  • FAMS 14 may maintain an account records database 30 that stores various kinds of account information, including account holder information (e.g., names, addresses, etc.) and data indicative of financial transactions made in connection with each account (e.g., dates, amounts and merchants for credit or debit card transactions, dates and amounts for customer deposits and withdrawals, etc.).
  • account records database 30 may store account information for some or all of the cardholders associated with cardholder computing devices 20 , for example. While shown in FIG. 1 as a single entity within FAMS 14 , it is understood that account records database 30 may, in some embodiments, be distributed across multiple databases and/or multiple physical/hardware memories, and/or may be wholly or partially external to (e.g., remote from) FAMS 14 .
  • AFSS 12 may generally provide services that help to detect and/or classify fraudulent activity in connection with existing and/or potential (e.g., applied for) financial accounts, such as the accounts managed by FAMS 14 .
  • AFSS 12 is included within FAMS 14 .
  • AFSS 12 may include a network interface 32 , a memory 34 , and a fraud detection/classification unit 36 .
  • Network interface 32 may include hardware, firmware and/or software configured to enable AFSS 12 to wirelessly exchange electronic data with one or more other components of environment 10 via network 26 .
  • network interface 32 may include an Ethernet port, a modem, a router, and/or one or more other ports and/or transceivers for one or more other wired and/or wireless communication technologies.
  • Memory 34 may be a computer-readable, non-transitory storage unit or device, or collection of units/devices, and may include persistent (e.g., hard disk) and/or non-persistent memory components. Memory 34 may store instructions that are executable on one or more processors of AFSS 12 (not shown in FIG. 1 ) to perform various operations, including the instructions of various software applications and data generated and/or used by such applications.
  • Card network computing system 16 may be a computing system (e.g., one or more servers) of a credit and/or debit card network entity, such as VISA® or Mastercard®, for example. In some embodiments and/or scenarios where the card network entity also acts as the issuer (e.g., American Express® or Discover®), card network computing system 16 may include FAMS 14 . Card network computing system 16 may provide various services to FAMS 14 and/or AFSS 12 . For example, card network computing system 16 may provide electronic updates to chargeback rules, fraud scores for particular customers and/or transactions, and so on.
  • Each of cardholder computing devices 20 may be a computing device of a respective holder of a credit or debit card account managed by FAMS 14 .
  • one or more of cardholder computing devices 20 may be desktop computers, laptop computers, tablet computers, smartphones, smart watches, and so on.
  • the cardholders e.g., credit or debit card account holders
  • cardholder computing devices 20 may instead be computing devices of other types of customers or potential customers, such as holders of non-card-based accounts, or individuals who have submitted an online application for a loan, etc., as discussed further below.
  • the environment 10 may omit card network computing system 16 .
  • Each of merchant computing systems 22 may include one or more computing devices associated with a particular provider of products and/or services.
  • some or all of merchant computing systems 22 may include servers associated with online retailers.
  • some or all of merchant computing systems 22 may include point-of-sale terminal devices providing credit and/or debit card payment processing features for “card present” transactions.
  • AFSS 12 detects and/or classifies activity not related to customer purchases (e.g., if AFSS 12 only detects loan application fraud, etc.)
  • the environment 10 may omit merchant computing systems 22 .
  • the other sources 24 may include computing devices and/or systems associated with sources of one or more other types of information.
  • other sources 24 may include vehicle telematics systems (e.g., installed in vehicles of cardholders associated with cardholder computing devices 20 ), one or more Internet service providers (ISPs) (e.g., ISPs providing Internet access to some or all cardholders), “smart home” system devices (e.g., installed in homes of some or all cardholders), and/or other systems/devices.
  • ISPs Internet service providers
  • smart home system devices e.g., installed in homes of some or all cardholders
  • the environment 10 does not include the other sources 24 .
  • Network 26 may communicatively couple some or all of the components shown in FIG. 1 .
  • FAMS 14 may use network 26 to communicate with AFSS 12 , card network computing system 16 , cardholder computing devices 20 and/or merchant computing systems 22 .
  • AFSS 12 may use network 26 to communicate with FAMS 14 , card network computing system 16 , cardholder computing devices 20 , merchant computing systems 22 and/or one or more of the other sources 24 .
  • network 26 may include multiple communication networks of one or more types (e.g., one or more wired and/or wireless local area networks (LANs), and/or one or more wired and/or wireless wide area networks (WANs) such as the Internet).
  • LANs local area networks
  • WANs wide area networks
  • network 26 may use partially or entirely distinct network components to support communications between different endpoints or computing devices, such as wireless communication or data transmission over one or more radio frequency links and/or wireless communication channels.
  • the portion(s) of network 26 used for communications between FAMS 14 and AFSS 12 may be the same as, or different than, the portion(s) of network 26 used for communications between FAMS 14 and one or more of cardholder computing devices 20 over one or more radio links or wireless communication channels, or between AFSS 12 and one or more of the other sources 24 , etc.
  • Those skilled in the art will appreciate different types of networks that are appropriate for network 26 , depending upon, for example, how AFSS 12 , FAMS 14 and/or other components of environment 10 are localized or distributed across a relatively large geographic area.
  • fraud detection/classification unit 36 of AFSS 12 may detect fraudulent activity, confirm whether suspected or reported fraudulent activity is truly fraudulent, and/or classify fraudulent or suspected fraudulent activity. For example, fraud detection/classification unit 36 may analyze each transaction stored in account records database 30 to determine whether that transaction is, or potentially is, fraudulent. Alternatively, fraud detection/classification unit 36 may analyze only those transactions that were flagged as possibly being fraudulent (e.g., by a cardholder calling in to report an unauthorized and/or unrecognized transaction, or by FAMS 14 or AFSS 12 generating a preliminary fraud alert after applying an initial set of rules to a transaction, etc.).
  • Fraud detection/classification unit 36 may also, or instead, analyze location information associated with potential transactions (e.g., GPS or other data indicating cardholder location, transaction data indicating a merchant location for a card-present transaction, etc.), and issue a pre-transaction alert or otherwise prevent a transaction from being fully executed. Fraud detection/classification unit 36 may also, or instead, support additional functionality, such as that described below in connection with the various components of fraud detection/classification unit 36 shown in FIG. 1 .
  • location information associated with potential transactions e.g., GPS or other data indicating cardholder location, transaction data indicating a merchant location for a card-present transaction, etc.
  • Fraud detection/classification unit 36 may also, or instead, support additional functionality, such as that described below in connection with the various components of fraud detection/classification unit 36 shown in FIG. 1 .
  • fraud detection/classification unit 36 may include a machine learning (ML) rule generator 40 , an external data collection unit 42 , a behavior analysis unit 44 , a dispute resolution unit 46 , a chargeback analysis unit 50 , an image analysis unit 52 , a classification unit 54 , and/or a notification unit 56 .
  • fraud detection/classification unit 36 may include more, fewer and/or different components/units than those shown in FIG. 1 .
  • each of ML rule generator 40 , external data collection unit 42 , behavior analysis unit 44 , dispute resolution unit 46 , chargeback analysis unit 50 , image analysis unit 52 , classification unit 54 , notification unit 56 , and/or other units or components of fraud detection/classification unit 36 may be a software component stored in memory 34 and implemented by one or more processors of one or more computing devices (e.g., servers) included in AFSS 12 .
  • ML rule generator 40 may generally analyze various types of data to generate and/or update fraud detection and/or classification rules to be applied by fraud detection/classification unit 36 and stored in an ML rules database 58 . As discussed in further detail below, the rules may be used to detect and/or classify a single type or category of fraudulent activity, or may be used broadly in connection with multiple types or categories of fraudulent activity.
  • ML rule generator 40 may implement any suitable type or types of machine learning. For example, ML rule generator 40 may implement supervised learning techniques, such as decision trees, regression-based models, support vector machines (SVMs) and/or neural networks, and/or unsupervised learning techniques such as Dirichlet process mixture models and/or k-means clustering.
  • SVMs support vector machines
  • ML rules database 58 may, in some embodiments, be distributed across multiple databases and/or multiple physical/hardware memories, and/or may be wholly or partially external to (e.g., remote from) AFSS 12 .
  • External data collection unit 42 may generally collect, via network interface 32 and/or from sources internal to AFSS 12 , information from various sources (e.g., FAMS 14 , cardholder computing devices 20 , other sources 24 , etc.), and provide that data to other portions of AFSS 12 as needed (e.g., to ML rule generator 40 to generate and/or update rules, and/or to behavior analysis unit 44 , dispute resolution unit 46 , chargeback analysis unit 50 , image analysis unit 52 and/or classification unit 54 to detect and/or classify fraudulent activity). Some data may be collected indirectly. For example, FAMS 14 may collect transaction data from merchant computing systems 22 (and/or from acquiring banks associated with one or more of merchant computing systems 22 ), and external data collection unit 42 may then collect that data from the account records database 30 of FAMS 14 .
  • sources e.g., FAMS 14 , cardholder computing devices 20 , other sources 24 , etc.
  • Some data may be collected indirectly.
  • FAMS 14 may collect transaction data from merchant computing systems 22 (and/or from
  • ML rules database 58 Once an initial set of rules has been generated and stored in ML rules database 58 , those rules may dictate some or all of the types of data gathered by external data collection unit 42 . In some embodiments, however, external data collection unit 42 collects a broad set of data types that may or may not be relevant to fraud determination or classification, and ML rule generator 40 continually analyzes that data to determine which data types are most predictive of fraud and/or fraud type/class.
  • Behavior analysis unit 44 may generally analyze cardholder-related (or other customer-related) information to identify patterns of behavior, which may then be used by fraud detection/classification unit 36 to detect and/or classify fraudulent activity. For example, behavior analysis unit 44 may analyze information obtained from account records database 30 to identify spending patterns associated with different cardholders. The operation of behavior analysis unit 44 , including the types of information analyzed and the ways in which that information is used to arrive at a result (e.g., a pattern of behavior), may be dictated by the rules stored in ML rules database 58 .
  • Data indicative of the behavior patterns identified by behavior analysis unit 44 may be stored in an account holder behaviors database 60 , for example. While shown in FIG. 1 as a single entity within AFSS 12 , it is understood that account holder behaviors database 60 may, in some embodiments, be distributed across multiple databases and/or multiple physical/hardware memories, and/or may be wholly or partially external to (e.g., remote from) AFSS 12 . In one embodiment, for example, account holder behaviors database 60 may be included within account records database 30 .
  • the environment 10 may not include account holder behaviors database 60 , and behavior patterns may be only identified by behavior analysis unit 44 “on the fly” as needed by fraud detection/classification unit 36 (e.g., when needed to analyze a transaction in view of past spending patterns of a particular cardholder, etc.).
  • behavior analysis unit 44 may separately analyze the transactions associated with each account holder, even if more than one account holder exists for a particular account. For example, behavior analysis unit 44 may independently analyze the transactions of each cardholder for a credit or debit card account in which each spouse has been issued a credit or debit card in his or her name. Fraud detection/classification unit 36 may then utilize the individual spending patterns when detecting and/or classifying fraud. In one embodiment where fraud detection/classification unit 36 utilizes a dollar amount threshold to detect likely fraudulent transactions, for example, a first threshold may be used for transactions made by a first cardholder listed on an account, and a higher, second threshold may be used for transactions made by a second cardholder listed on the account. Further examples are provided below in connection with FIG.
  • fraud detection and/or classification may be made more precise than would be the case if spending patterns were only identified at the aggregate level (e.g., using a single dollar amount threshold, regardless of which cardholder made a particular transaction).
  • Dispute resolution unit 46 may generally analyze financial transaction data and/or other information to automatically generate queries for cardholders or other customers. For example, dispute resolution unit 46 may analyze information obtained from account records database 30 . The generated queries may be designed to help fraud detection/classification unit 36 determine whether a particular transaction was fraudulent, or estimate a probability that the transaction was fraudulent, etc. Dispute resolution unit 46 may also process responses from cardholders/customers, and automatically generate additional queries based upon those responses. Examples of the operation of dispute resolution unit 46 are provided below in connection with FIGS. 4E and 9 , according to various embodiments.
  • Chargeback analysis unit 50 may generally analyze financial transaction and/or other information to identify transactions that are good candidates for chargeback payments. For example, chargeback analysis unit 50 may analyze information obtained from account records database 30 to determine whether there is a relatively high probability that the merchant (or an acquiring bank) should be responsible for a chargeback payment to a card issuer associated with FAMS 14 . The operation of chargeback analysis unit 50 , including the types of information analyzed and the ways in which that information is used to arrive at a result (e.g., flagging a transaction as a chargeback candidate), may be dictated by the rules stored in ML rules database 58 .
  • ML rule generator 40 may make use of chargeback rules obtained from a card network entity (e.g., from card network computing system 16 ), and stored in chargeback rules database 62 , to generate and/or update the rules applied by chargeback analysis unit 50 . Examples of the operation of chargeback analysis unit 50 are provided below in connection with FIGS. 4B and 7 , according to various embodiments.
  • transactions flagged by chargeback analysis unit 50 are subject to further, manual review using the chargeback rules stored in chargeback rules database 62 .
  • chargeback analysis unit 50 (or another component of fraud detection/classification unit not shown in FIG. 1 ) automatically, with little or no manual input/assistance, applies the chargeback rules from chargeback rules database 62 for each flagged transaction. While shown in FIG. 1 as a single entity within AFSS 12 , it is understood that chargeback rules database 62 may, in some embodiments, be distributed across multiple databases and/or multiple physical/hardware memories, and/or may be wholly or partially external to (e.g., remote from) AFSS 12 .
  • Image analysis unit 52 may generally analyze image data corresponding to physical documents to identify fraudulent (e.g., counterfeit and/or forged) documents, and/or to flag potentially fraudulent documents for further (e.g., manual) review. For example, image analysis unit 52 may analyze information obtained from merchant computing systems 22 to determine whether there is a relatively high probability that documents presented to the merchants (e.g., personal checks, identification cards, etc.) are fraudulent. Image analysis unit 52 may be configured to analyze only a single type of document, or multiple types of documents. The operation of image analysis unit 52 , including the image characteristics analyzed and the ways in which the characteristics may be used to arrive at a result (e.g., flagging a document as potentially fraudulent), may be dictated by the rules stored in ML rules database 58 . Examples of the operation of image analysis unit 52 are provided below in connection with FIGS. 4F and 10 , according to various embodiments.
  • Classification unit 54 may generally analyze broad categories of data from various sources (e.g., account records database 30 , cardholder computing devices 20 , merchant computing systems 22 , and/or other sources 24 ) to categorize/classify types of suspected fraudulent financial activity. Classification unit 54 may classify fraudulent activity only within a particular subset of fraudulent financial activity (e.g., classifying debit and/or credit card transactions as involving a potential case of counterfeiting, skimming, lost/stolen card use, chargeback fraud, etc.), or may classify fraudulent financial activity across a broader spectrum (e.g., including types of identity theft not necessarily tied to a single financial transaction, such as application fraud).
  • sources e.g., account records database 30 , cardholder computing devices 20 , merchant computing systems 22 , and/or other sources 24 .
  • Classification unit 54 may classify fraudulent activity only within a particular subset of fraudulent financial activity (e.g., classifying debit and/or credit card transactions as involving a potential case of counterfeiting, skimming,
  • classification unit 54 classifies suspected fraudulent activity in connection with a particular account or transaction in response to being notified of suspect activity (e.g., notified by another component of fraud detection/classification unit 36 , or by a manual user input, etc.). In other embodiments, classification unit 54 itself (or another component of fraud detection/classification unit 36 ) identifies suspect activity before classification unit 54 classifies that activity. Examples of the operation of classification unit 54 are provided below in connection with FIGS. 4C and 11 , according to various embodiments.
  • Notification unit 56 may generally provide alerts, confirmations, and/or other notifications to various individuals (e.g., customers, bank employees associated with FAMS 14 , third party employees associated with AFSS 12 , etc.). For example, notification unit 56 may generate a notification message stating that a fraud alert associated with a particular transaction is a false positive, and cause network interface 32 to send the message to a computer terminal or to FAMS 14 for display to a system user. As another example, notification unit 56 may cause network interface 32 to send other flagged transactions and/or documents (e.g., chargeback candidates identified by chargeback analysis unit 50 , documents that image analysis unit 52 has identified as potentially fraudulent, etc.) to a computer terminal or FAMS 14 for display to a system user.
  • individuals e.g., customers, bank employees associated with FAMS 14 , third party employees associated with AFSS 12 , etc.
  • notification unit 56 may generate a notification message stating that a fraud alert associated with a particular transaction is a false positive, and cause network interface 32 to send the message
  • notification unit 56 may cause network interface 32 to send FAMS 14 and/or one of merchant computing systems 22 an alert indicating that a transaction that is in-process should be terminated due to suspected fraud.
  • notification unit 56 may cause network interface 32 to send queries generated by dispute resolution unit 46 to various ones of cardholder computing devices 20 for display to cardholders.
  • ML rule generator 40 may generate and/or update rules that are used for one or more of a variety of different purposes relating to fraud detection and/or classification.
  • FIG. 2 depicts one generalized, example process flow 80 for machine learning that may be implemented by ML rule generator 40 , and possibly one or more other components of fraud detection/classification unit 36 .
  • multi-account data 82 may represent data associated with multiple financial accounts, each with one or more account holders.
  • the financial accounts may be existing or potential accounts, and the account holders may include holders of accounts and/or potential holders of potential accounts.
  • the multi-account data 82 may include existing and/or applied-for credit card accounts, debit card accounts, savings accounts, checking accounts, investment accounts, loan accounts, etc.
  • the multi-account data 82 may include one or more different types of information obtained (e.g., by external data collection unit 42 of FIG. 1 ) from one or more of FAMS 14 , cardholder computing devices 20 , merchant computing systems 22 , and/or other sources 24 .
  • the multi-account data 82 may include transaction data (e.g., transaction dates, amounts, locations, etc.) from account records database 30 of FAMS 14 , data indicative of Internet Protocol (IP) addresses of cardholder computing devices 20 and/or devices in merchant computing systems 22 , Internet browsing and/or search history data from cardholder computing devices 20 (or from an ISP computer system included in other sources 24 , etc.), vehicle telematics data from telematics systems of cardholder vehicles, home occupancy and/or usage data (e.g., smart appliance data) from smart home systems of cardholders, autonomous or smart vehicle data, vehicle navigation system data, mobile device data, mobile device and/or vehicle GPS data, and/or one or more other types of data.
  • transaction data e.g., transaction dates, amounts, locations, etc.
  • IP Internet Protocol
  • ISP computer system included in other sources 24 e.g., etc.
  • vehicle telematics data from telematics systems of cardholder vehicles
  • home occupancy and/or usage data e.g., smart appliance
  • the multi-account data 82 only includes data that account holders or potential account holders have expressly consented to share with an entity associated with FAMS 14 and/or AFSS 12 (e.g., in exchange for fraud protection services). In certain other embodiments, however, express consent is only needed for certain types of information, such as browsing history information, vehicle telematics data, etc.
  • the multi-account data 82 may be associated with multiple fraud determination labels.
  • the labels may simply reflect whether or not fraud existed (e.g., “fraud” or “no fraud”), or may also indicate a type or class of fraud (e.g., “counterfeiting,” “lost or stolen card use,” etc.), for example.
  • each of a number of data sets in the multi-account data 82 is associated with such a label, and includes data relating to a particular financial transaction, financial account, loan application, etc., for which the fraud determination was made (e.g., after a manual and/or automated fraud investigation).
  • the labels may include final fraud determinations that were made via earlier iterations of the process flow 80 , and/or external to the process flow 80 .
  • a first data set associated with a “card present” credit card transaction may include data describing that transaction (e.g., from account records database 30 ) and data indicative of the cardholder's online browsing activity (e.g., from one of cardholder computing devices 20 ) for the 15 days immediately preceding the transaction, and be labeled “confirmed fraud.”
  • a second data set, associated with another “card present” transaction may include the same general types of data but be labeled “no fraud,” and so on.
  • the same data may appear in, or be used by, two or more of the data sets. If the two “card present” transactions described above are both associated with the same account, for example, and if the second transaction occurred less than 15 days after the first transaction, some of the same online activity data may be shared by the first and second data sets.
  • the multi-account data 82 may be analyzed to generate fraud detection and/or classification rules (e.g., to be stored in ML rules database 58 ). Any suitable type of supervised machine learning program/technique(s) may be used, such as SVMs, neural networks, logistic regression, etc.
  • process stage 84 may serve to identify which type(s) of data is/are probative of whether fraud has occurred (and/or the type/category of fraud that may have occurred), and to determine the data values and/or combinations that are probative of whether fraud has occurred (and/or the type/category of fraud that may have occurred).
  • process stage 84 may learn that certain spending patterns within a threshold time of a transaction tend to indicate that the cardholder made the transaction (e.g., thereby indicating that fraud has not occurred, or that a fraud report is itself fraudulent or mistaken, etc.), that certain types of online searches by a cardholder (e.g., including a descriptor of a product purchased in the transaction, or a name of the merchant, etc.) tend to indicate that the cardholder made the transaction, that the cardholder's distance from the site of a “card present” transaction (e.g., as determined from GPS information provided by the cardholder's smartphone, wearable electronics, or vehicle) relates to the probability of fraudulent activity according to a particular equation, and so on.
  • Other specific examples of such rules, and how those rules may be generated, are discussed below in connection with FIGS. 3A-3F and 4A-4F , according to various embodiments
  • the rules generated or updated at process stage 84 may be applied to first account data 90 associated with a particular account and customer(s) (e.g., a customer associated with a particular one of computing devices 20 ).
  • the types of data included in first account data 90 may depend upon which types of data were determined, by process stage 84 , to be relevant to a fraud determination.
  • the first account data 90 may include the amount and date of one or more transactions, as well as data indicative of visited web sites (e.g., Uniform Resource Locators (URLs) and/or content of visited websites, etc.).
  • the first account data 90 may include information obtained (e.g., by external data collection unit 42 ) from one or more of FAMS 14 , one of cardholder computing devices 20 associated with the customer holding the first account, one or more of merchant computing systems 22 , and/or one or more of other sources 24 , for example.
  • Process stage 86 may output various different types of information, depending upon the embodiment and/or scenario. For example, depending upon the content of first account data 90 and the rules generated or updated at process stage 84 , process stage 86 may generate data indicating that a particular financial transaction associated with first account data 90 is, or is not, fraudulent or potentially fraudulent. Alternatively, or additionally, process stage 86 may generate data indicating a particular classification for fraudulent or suspected fraudulent activity (e.g., a fraudulent transaction) associated with first account data 90 .
  • a particular classification for fraudulent or suspected fraudulent activity e.g., a fraudulent transaction
  • further analysis may be performed at an additional stage, shown in dashed lines in FIG. 2 as process stage 92 .
  • the additional analysis may then be used to make a final fraud determination (e.g., a final decision on whether fraud occurred, and/or on the type of fraud that occurred) at process stage 94 .
  • process stage 92 is omitted from process flow 80 , and process stage 94 merely represents the output of process stage 86 .
  • the final determination made at process stage 94 along with the first account data 90 used to make that determination, may be fed back into process stage 84 to provide additional labeled data for purposes of updating the rules.
  • the process flow 80 includes more, fewer and/or different stages, such as any of those discussed elsewhere herein (e.g., in connection with FIGS. 3A-3F ).
  • process stages 84 and 86 may be combined.
  • the multi-account data 82 may be unlabeled rather than labeled (or the labels may be ignored), and the combined process stage 84 , 86 may use unsupervised learning techniques (e.g., clustering techniques) to classify anomalous/outlier financial transactions, accounts, applications, etc., as “suspect” and needing further analysis.
  • unsupervised learning techniques e.g., clustering techniques
  • FIGS. 3A-3F generally correspond to embodiments in which supervised machine learning techniques are used, other embodiments may instead use unsupervised machine learning techniques, as noted above.
  • fraud detection/classification unit 36 may be configured to implement only one of the process flows of FIGS. 3A-3F , or may be configured to implement two or more (e.g., all) of the process flows shown in FIGS. 3A-3F .
  • multi-customer online activity data 102 may represent data associated with the online activities of a number (e.g., thousands) of customers (e.g., credit or debit cardholders, checking or saving account holders, etc.).
  • the multi-customer online activity data 102 may include data indicating actions that the customers took, and/or web sites visited by the customers, while the customers were connected to the Internet via web browsers (e.g., executing on respective ones of cardholder computing devices 20 ).
  • the multi-customer online activity data 102 may include URLs of, and/or content (e.g., text) within, web sites visited by customers, search terms entered by customers using search engine tools, search results presented to customers by search engine tools, indications of interactive controls (e.g., virtual buttons) selected by customers on various web pages, and so on.
  • content e.g., text
  • search engine tools e.g., search terms entered by customers using search engine tools
  • search results presented to customers by search engine tools e.g., indications of interactive controls (e.g., virtual buttons) selected by customers on various web pages, and so on.
  • the multi-customer online activity data 102 may include data obtained (e.g., by external data collection unit 42 of FIG. 1 ) from cardholder computing devices 20 , from one or more ISPs of other sources 24 , and/or from a third party aggregator of such information, for example.
  • the multi-customer online activity data 102 may only include data that customers have expressly consented to share with an entity associated with FAMS 14 and/or AFSS 12 (e.g., in exchange for fraud protection services or other benefits, such as discounts).
  • the multi-customer online account data 102 may be associated with multiple fraud determination labels.
  • each label may be associated with a data set that includes not only the corresponding portion of multi-customer online activity data 102 , but also one or more other types of data, such as transaction data (e.g., transaction dates, amounts, locations, etc.) for each customer from account records database 30 of FAMS 14 , data indicative of IP addresses of cardholder computing devices 20 and/or devices in merchant computing systems 22 , Internet browsing and/or search history data from cardholder computing devices 20 (or from an ISP computer system included in other sources 24 , etc.), vehicle telematics data from telematics systems of other sources 24 , home occupancy and/or usage data (e.g., smart appliance data) from smart home systems of other sources 24 , and so on.
  • transaction data e.g., transaction dates, amounts, locations, etc.
  • data indicative of IP addresses of cardholder computing devices 20 and/or devices in merchant computing systems 22 e.g., Internet browsing and/or search history data from
  • the labels may include final fraud determinations that were made via earlier iterations of the process flow 100 , and/or external to the process flow 100 .
  • Multi-customer online account data 102 may include many (e.g., thousands) of positively and negatively labeled data sets.
  • the multi-customer online activity data 102 may be analyzed to generate fraud detection rules (e.g., to be stored in ML rules database 58 ).
  • fraud detection rules e.g., to be stored in ML rules database 58 .
  • any suitable type of supervised machine learning program/technique(s) may be used.
  • process stage 104 may serve to identify which type(s) of online activity data is/are probative of whether fraud has occurred, and to determine the data values and/or combinations that are probative of whether fraud has occurred.
  • the fraud detection rules may not only detect fraud, but also classify fraud (e.g., as described below in connection with FIG. 3C ), in some embodiments.
  • the rules generated or updated at process stage 104 may be applied to first customer online activity data 110 .
  • the first customer online activity data 110 may be associated with a particular customer, such as a customer associated with a particular one of computing devices 20 , for example.
  • the types of data included in first customer online activity data 110 may depend upon which types of online activity data were determined, by process stage 104 , to be relevant to a fraud determination.
  • the first customer online activity data 110 may include information obtained (e.g., by external data collection unit 42 ) from one of cardholder computing devices 20 (i.e., the device associated with the first customer), and/or from an ISP of other sources 24 .
  • Process stage 106 may output various different types of information, depending upon the embodiment and/or scenario. For example, depending upon the content of first customer online activity data 110 and the rules, process stage 106 may generate data indicating that a particular financial transaction associated with the first customer is, or is not, fraudulent or potentially fraudulent. Alternatively, or additionally, process stage 106 may generate data indicating a particular classification of fraudulent or potentially fraudulent activity associated with first customer online activity data 110 .
  • further analysis e.g., a manual review, or further automated review using additional data sources, etc.
  • the additional analysis may then be used to make a final fraud determination (e.g., a final decision on whether fraud occurred, and/or on the type of fraud that occurred) at process stage 114 .
  • process stage 112 is omitted from process flow 100 , and process stage 114 merely represents the output of process stage 106 .
  • the final determination made at process stage 114 may be fed back into process stage 104 to provide additional labeled data for purposes of updating the rules.
  • a preliminary fraud determination made at process stage 106 is also fed back into process stage 104 , to allow the machine learning program to determine and improve upon past performance/accuracy.
  • an exemplary process flow 120 may generally be used to identify the financial transactions for which chargebacks (e.g., post-transaction payments from merchants, or acquiring/merchant banks, back to the issuer to return proceeds from transactions) are appropriate.
  • chargebacks e.g., post-transaction payments from merchants, or acquiring/merchant banks, back to the issuer to return proceeds from transactions
  • multi-account transaction data 122 may represent data associated with the financial transactions involving the accounts of a number (e.g., thousands) of credit or debit cardholders.
  • the multi-account transaction data 122 may include information such as transaction dates, transaction amounts, merchant names (and/or aliases) associated with the transaction, information relating to how the card information was collected by the merchant (e.g., by swiping, an EMV chip reader, manual entry of the card number, etc.), geographic locations of “card present” transactions, and so on.
  • the multi-account transaction data 122 may include data obtained (e.g., by external data collection unit 42 of FIG. 1 ) from merchant computing systems 22 and/or from acquiring/merchant banks associated with those merchants, for example.
  • the multi-account transaction data 122 may be associated with multiple chargeback outcome labels.
  • each label may be associated with a data set that includes the corresponding portion of multi-account transaction data 122 .
  • the outcome labels may include final chargeback determinations that were made (in connection with the transactions represented in multi-account transaction data 122 ) via earlier iterations of the process flow 120 , and/or external to the process flow 120 .
  • Multi-account transaction data 122 may include many (e.g., thousands) of positively and negatively labeled data sets.
  • the multi-account transaction data 122 may be analyzed to generate chargeback candidate detection rules (e.g., to be stored in ML rules database 58 ).
  • chargeback candidate detection rules e.g., to be stored in ML rules database 58 .
  • any suitable type of supervised machine learning program/technique(s) may be used.
  • process stage 124 may serve to identify which type(s) of transaction data is/are probative of whether, under the full chargeback rules of the card network entity, a chargeback is appropriate for a given transaction.
  • Process stage 124 may also determine the transaction data values and/or combinations that are probative of whether a chargeback is appropriate for the transaction.
  • the rules generated or updated at process stage 124 may be applied to first account transaction data 130 to determine whether a transaction associated with the first account is a “good” chargeback candidate.
  • process stage 126 may, instead of applying the full chargeback rules of the card network entity (which may be quite lengthy and complex) to the facts surrounding the transaction, use various factors and algorithms developed at process stage 124 to determine whether there exists a relatively high probability that a chargeback would be appropriate for the transaction if the full chargeback rules were applied.
  • the process stage 126 may calculate a percentage probability that the transaction is one in which a chargeback is appropriate, for example.
  • the first account transaction data 130 may be associated with the account of a particular cardholder or cardholders, such as a cardholder associated with a particular one of cardholder computing devices 20 , for example.
  • the types of data included in first account transaction data 130 may depend upon which types of transaction-related data were determined, by process stage 124 , to be relevant to a chargeback candidate determination.
  • the first account transaction data 130 may include information obtained (e.g., by external data collection unit 42 ) from one of merchant computing systems 22 (e.g., the computing system of the merchant involved in the transaction being analyzed) and/or from an acquiring/merchant bank associated with that merchant.
  • the first account transaction data 130 may also include information about one or more other transactions associated with the first account (e.g., data pertaining to other transactions occurring shortly before and/or after the transaction at issue). Some specific examples of rules that may be generated by process stage 124 , and applied at process stage 126 , are described below in connection with FIG. 4B .
  • Process stage 126 may output information indicating whether the particular transaction represented by first account transaction data 130 is a “good” candidate for chargeback detection. For example, process stage 126 may output a percentage probability, calculated according to the rules generated or updated at process stage 124 , that the transaction is one in which a chargeback is appropriate. As another example, process stage 126 may output a binary indicator of whether the transaction is, or is not, a strong/likely chargeback candidate (e.g., by comparing the percentage probability to a threshold probability).
  • the full chargeback rules of the card network entity may be applied at a process stage 132 .
  • Process stage 132 may include manual application of the full chargeback rules, and/or automated application of the full chargeback rules, in various different embodiments.
  • a final chargeback determination may be made at a process stage 134 .
  • the final determination made at process stage 134 along with the first account transaction data 130 (and any other data) used to make that determination, may be fed back into process stage 124 to provide additional labeled data for purposes of updating the rules.
  • the indication of whether the transaction is a good chargeback candidate generated at process stage 126 may also be fed back into process stage 124 , to allow the machine learning program to determine and improve upon past performance/accuracy.
  • an exemplary process flow 140 may generally be used to classify instances of suspected or potential fraud.
  • the process flow 140 may represent ongoing, real-time or batch processing of a large amount of data associated with a large number of potential and/or existing financial accounts (e.g., all accounts associated with a particular bank, or all accounts opting in to a fraud protection program, etc.).
  • the process flow 140 may be used to initially flag situations for closer investigation, and provide one or more classifications of the type(s) of fraud potentially at issue in order to narrow or otherwise facilitate the investigation.
  • the process flow 140 may be used to provide a narrower classification (e.g., “skimming”) when a broader class of fraud (e.g., credit card fraud) is already suspected.
  • a narrower classification e.g., “skimming”
  • multi-account data 142 may represent data associated with financial accounts of a number (e.g., thousands) of account holders.
  • the financial accounts may be existing or potential accounts, and the account holders may include holders of accounts and/or potential holders of potential accounts.
  • the multi-account data 142 may include existing and/or applied-for credit card accounts, debit card accounts, savings accounts, checking accounts, investment accounts, loan accounts, etc.
  • the multi-account data 142 may include one or more different types of information obtained (e.g., by external data collection unit 42 of FIG. 1 ) from one or more of FAMS 14 , cardholder computing devices 20 , merchant computing systems 22 , and/or other sources 24 .
  • the multi-account data 142 may include transaction data (e.g., transaction dates, amounts, locations, etc.) from account records database 30 of FAMS 14 , data indicative of IP addresses of cardholder computing devices 20 and/or devices in merchant computing systems 22 , Internet browsing and/or search history data from cardholder computing devices 20 (or from an ISP computer system included in other sources 24 , etc.), vehicle telematics data from telematics systems of cardholder vehicles, home occupancy and/or usage data (e.g., smart appliance data) from smart home systems of cardholders, and/or one or more other types of data.
  • Some or all data within multi-account data 142 may be information that account holders or potential account holders have expressly consented to share with an entity associated with FAMS 14 and/or AFSS 12 (e.g., in exchange for fraud protection services).
  • the multi-account data 142 may be associated with multiple fraud determination labels, each indicating a type or class of fraud (e.g., “counterfeiting,” “lost or stolen card use,” “skimming,” “chargeback fraud,” “application fraud,” etc.), or indicating a lack of fraud, for example.
  • each of a number of data sets in the multi-account data 142 is associated with at least one such classification/label, and includes data relating to a particular financial transaction, financial account, loan application, etc., for which the fraud classification or classifications was/were made (e.g., after a previous iteration of process flow 140 , or after another manual and/or automated fraud investigation).
  • Multi-account data 142 may include many (e.g., thousands) of data sets labeled with various known fraud classifications.
  • the multi-account data 142 may be analyzed to generate fraud classification rules (e.g., to be stored in ML rules database 58 ).
  • fraud classification rules e.g., to be stored in ML rules database 58 .
  • any suitable type of supervised machine learning program/technique(s) may be used.
  • process stage 144 may serve to identify which type(s) of transaction data is/are probative of the particular type of fraud (if any) that has occurred.
  • Process stage 144 may also determine the data values and/or combinations that are probative of the particular type of fraud (if any) that has occurred.
  • the rules generated or updated at process stage 144 may be applied to first account data 150 .
  • the first account data 150 may be associated with a particular account and a particular customer (e.g., a cardholder associated with a particular one of computing devices 20 ).
  • the types of data included in first account data 150 may depend upon which types of data were determined, by process stage 144 , to be relevant to fraud classification.
  • the first account data 150 may include information obtained (e.g., by external data collection unit 42 ) from one or more of FAMS 14 , one of cardholder computing devices 20 (i.e., the device associated with the customer holding or applying for the first account), one or more of merchant computing systems 22 , and/or one or more of other sources 24 .
  • Process stage 146 may output data (e.g., a message or code) that is used to classify suspected fraudulent activity (in connection with the account associated with first account data 150 ) at a process stage 152 .
  • process stage 152 may assign a classification of “counterfeiting” if process stage 146 determined that the first account data 150 indicated a number of circumstances that, according to the rules generated at process stage 144 , are known to be correlated with counterfeiting activity (e.g., two “card present” transactions occurring in different states within the same one-hour time period, etc.).
  • two or more classifications may concurrently be assigned to first account data 150 .
  • process stage 146 may determine a set of probabilities for a set of two or more potential types of fraud, and process stage 152 may assign each classification, with each respective probability, to first account data 150 . Moreover, in some embodiments and scenarios, process stage 152 may assign a classification that corresponds to an absence of any suspected fraud (e.g., “no fraud”).
  • process stage 152 may assign a classification that corresponds to an absence of any suspected fraud (e.g., “no fraud”).
  • process stage 154 if process stage 152 assigned a classification other than one indicating the absence of suspected fraud, the first account data 150 , and/or other information associated with the account and the suspected class of fraud, may be analyzed in depth to make a final fraud determination at a process stage 156 .
  • the fraud classification may be used to facilitate the analysis at process stage 154 , with process stage 154 including manual and/or automated fraud detection techniques. For example, personnel associated with AFSS 12 may use the fraud classification(s) to inform their strategy and/or focus with respect to conducting an in-depth fraud investigation.
  • the additional analysis at process stage 154 may then result in a final fraud determination at process stage 156 .
  • the final determination may indicate both whether fraud occurred and, if so, the class(es)/type(s) of fraud that occurred.
  • the final determination made at process stage 156 and information used to make that determination (e.g., the first account data 150 and potentially other data), may be fed back into process stage 144 to provide additional labeled data for purposes of updating the rules.
  • the (preliminary) fraud classification made at process stage 152 may also be fed back into process stage 144 to help the machine learning program identify instances in which the preliminary classifications at process stage 152 were incorrect. Process stage 144 may then update the fraud classification rules in ways that seek to prevent or reduce such instances in the future.
  • an exemplary process flow 160 may generally be used to detect application fraud.
  • “Application fraud” may generally refer to fraud in connection with the application for any type of financial account, loan and/or line of credit (e.g., mortgage loan, vehicle loan, small business loan, payday loan, home equity line of credit, credit card account, debit card account, checking account, savings account, investment account, etc.).
  • the application may be for non-financial purposes, such as an application for membership in a particular group or institution, for example.
  • multi-applicant search history data 162 may represent data associated with the Internet search history of a number (e.g., thousands) of applicants.
  • the multi-applicant search history data 162 may include search terms entered by the applicants using online search engine tools, for example, and/or the results of such searches (e.g., URLs, titles and/or contents of search results), for example.
  • the multi-applicant search history data 162 may include data obtained (e.g., by external data collection unit 42 of FIG. 1 ) from cardholder computing devices 20 , from one or more ISPs of other sources 24 , and/or from a third party aggregator of such information, for example.
  • the multi-applicant search history data 162 only includes data that the applicants have expressly consented to share with an entity associated with FAMS 14 and/or AFSS 12 (e.g., in exchange for consideration of their applications).
  • the multi-applicant search history data 162 may be associated with multiple fraud determination labels.
  • each label may be associated with a data set that corresponds to an application submitted by a particular applicant, where the data set includes the corresponding portion of multi-applicant search history data 162 (e.g., the search terms and/or results associated with the particular application).
  • the labels may include final fraud determinations that were made via earlier iterations of the process flow 160 , and/or external to the process flow 160 .
  • Multi-applicant search history data 162 may include many (e.g., thousands) of positively and negatively labeled data sets.
  • the multi-applicant search history data 162 may be analyzed to generate application fraud detection rules (e.g., to be stored in ML rules database 58 ).
  • application fraud detection rules e.g., to be stored in ML rules database 58 .
  • any suitable type of supervised machine learning program/technique(s) may be used.
  • process stage 164 may serve to identify which type(s) of Internet search-related data is/are probative of whether application fraud has occurred, and to determine the data values and/or combinations that are probative of whether application fraud has occurred.
  • first applicant search history data 170 may be associated with a particular application and a particular applicant (e.g., a person associated with a particular one of computing devices 20 ), for example.
  • the types of data included in first applicant search history data 170 may depend upon which types of Internet search-related data were determined, by process stage 164 , to be relevant to a fraud determination.
  • the first applicant search history data 170 may include information obtained (e.g., by external data collection unit 42 ) from one of computing devices 20 (i.e., the device associated with the first applicant), and/or from an ISP of other sources 24 , for example.
  • Process stage 166 may output information indicating whether fraud is suspected in connection with the application corresponding to first applicant search history data 170 .
  • process stage 166 may output a percentage probability, calculated according to the rules generated or updated at process stage 164 , that the application was fraudulently made (e.g., by someone other than the purported applicant or an authorized representative thereof).
  • process stage 166 may output a binary indicator of whether the application likely was, or likely was not, fraudulently made (e.g., by comparing a percentage probability to a threshold probability).
  • further analysis e.g., a manual review, or further automated review using additional data sources, etc.
  • the additional analysis may then be used to make a final fraud determination (e.g., a final decision on whether application fraud occurred) at process stage 174 .
  • process stage 172 is omitted from process flow 160 , and process stage 174 merely represents the output of process stage 166 .
  • the final determination made at process stage 174 along with the first applicant search history data 170 (and any other data) used to make that determination, may be fed back into process stage 164 to provide additional labeled data for purposes of updating the rules.
  • a preliminary fraud determination made at process stage 166 is also fed back into process stage 164 , to allow the machine learning program to determine and improve upon past performance/accuracy.
  • an exemplary process flow 180 may generally be used to facilitate the resolution of fraud disputes (or potential disputes) with customers/account holders.
  • the process flow 180 may be used to determine whether a reportedly unauthorized or fraudulent transaction (e.g., one that the account holder reported as such when looking at his or her account statement) was indeed unauthorized or fraudulent.
  • the process flow 180 may also, or instead, be used to determine whether an “unrecognized” transaction (i.e., one that the account holder does not recall, but does not necessarily report as fraudulent) was unauthorized or fraudulent.
  • multi-account data 182 may represent data associated with financial accounts of a number (e.g., thousands) of account holders.
  • the multi-account data 182 may include data associated with financial transactions relating to credit card accounts, debit card accounts, savings accounts, checking accounts, etc.
  • FIG. 3E will be described with reference to an embodiment in which the accounts are credit card accounts.
  • the multi-account data 182 may include transaction data (e.g., transaction dates, amounts, locations, etc.) obtained from FAMS 14 (e.g., by external data collection unit 42 of FIG. 1 ). In some embodiments, however, the multi-account data 182 also includes information obtained from cardholder computing devices 20 , merchant computing systems 22 , and/or other sources 24 .
  • transaction data e.g., transaction dates, amounts, locations, etc.
  • FAMS 14 e.g., by external data collection unit 42 of FIG. 1
  • the multi-account data 182 also includes information obtained from cardholder computing devices 20 , merchant computing systems 22 , and/or other sources 24 .
  • the multi-account data 182 may include, in addition to transaction data from account records database 30 of FAMS 14 , data indicative of IP addresses of cardholder computing devices 20 and/or devices in merchant computing systems 22 , Internet browsing and/or search history data from cardholder computing devices 20 (or from an ISP computer system included in other sources 24 , etc.), vehicle telematics data from telematics systems of cardholder vehicles, home occupancy and/or usage data (e.g., smart appliance data) from smart home systems of cardholders, autonomous vehicle data, smart vehicle data, mobile device data, vehicle or mobile device GPS data, and/or one or more other types of data.
  • Some or all data within multi-account data 182 may be information that account holders or potential account holders have expressly consented to share with an entity associated with FAMS 14 and/or AFSS 12 (e.g., in exchange for fraud protection services).
  • the multi-account data 182 may be associated with multiple fraud determination labels (e.g., “fraud” and “no fraud,” and/or more complex labels that indicate type/class, such as “lost/stolen card use,” etc.).
  • each label may be associated with a data set that includes the corresponding portion of multi-account data 182 .
  • the labels may include final fraud determinations that were made via earlier iterations of the process flow 180 , and/or external to the process flow 180 .
  • Multi-account data 182 may include many (e.g., thousands) of positively and negatively labeled data sets.
  • the multi-account data 182 may be analyzed to generate query generation rules (e.g., to be stored in ML rules database 58 ).
  • query generation rules e.g., to be stored in ML rules database 58 .
  • any suitable type of supervised machine learning program/technique(s) may be used.
  • process stage 184 may serve to identify which types of information are probative of whether fraud has occurred, and to craft rules that formulate queries to ascertain such information based upon account data.
  • process stage 184 may determine that, for a suspect “card present” transaction, a verified, non-fraudulent “card present” transaction within 10 miles and 3 hours of the suspect transaction is probative of whether the suspect transaction was fraudulent. Based upon this finding, process stage 184 may also generate a rule specifying that a cardholder should be queried as to whether he/she can confirm making each “card present” transaction within 10 miles and 3 hours of the suspect transaction.
  • process stage 184 may determine that a merchant using a billing alias different from its legal and/or commonly-known name (e.g., by at least some threshold level of similarity, as measured by number of similar characters, order of characters, etc.) is probative of whether the cardholder authorized a transaction associated with that billing alias. Based upon this finding, process stage 184 may generate a rule specifying that a cardholder should be queried as to whether he/she is aware of a billing alias used for a suspect transaction if that billing alias is sufficiently different from the legal/common name of the merchant.
  • the rules generated or updated at process stage 184 may be applied to first account data 190 .
  • the first account data 190 may be associated with a particular cardholder, such as a cardholder associated with a particular one of cardholder computing devices 20 , for example.
  • the types of data included in first account data 190 may depend upon which types of data were determined, by process stage 184 , to be relevant to developing dispute resolution queries.
  • Process stage 186 may generate a set of one or more queries in accordance with the rules and the contents of first account data.
  • the generated queries may be sent to the cardholder in one or more of various ways, such as sending the queries via SMS text message and/or email, and/or via a web browser or dedicated application executing on the one of cardholder computing devices 20 that is associated with the cardholder, for example.
  • responses to the queries are received from the cardholder (e.g., via inputs made by the cardholder via the web browser or application, or a responsive SMS text message or email, etc.).
  • the rules generated or updated at process stage 184 specify the manner in which follow-up queries should be generated based upon the responses received at process stage 194 , and process stages 192 and 194 may be repeated multiple times.
  • process stage 196 further analysis (e.g., a manual review, or further automated review using additional data sources, etc.) that makes use of the received responses is performed at an additional stage, shown in dashed lines in FIG. 3E as process stage 196 .
  • the additional analysis may then be used to make a final fraud determination (e.g., a final decision on whether fraud occurred, and/or on the type of fraud that occurred) at process stage 198 .
  • process stage 196 is omitted from process flow 180
  • process stage 198 is based upon information from the cardholder. For example, the questions generated at process stage 192 may “jog” the cardholder's memory, and cause him or her to indicate that the transaction at issue was authorized.
  • an exemplary process flow 200 may generally be used to detect fraud relating to documents, such as counterfeit and/or forged documents.
  • the process flow 200 may be used in connection with various kinds of documents, such as checks (e.g., personal checks, cashier's checks, etc.), money orders, treasury bills, identification documents (e.g., social security cards, driver's licenses, passports, birth certificates, etc.), certification documents, and so on.
  • checks e.g., personal checks, cashier's checks, etc.
  • money orders e.g., money orders, treasury bills, identification documents (e.g., social security cards, driver's licenses, passports, birth certificates, etc.), certification documents, and so on.
  • identification documents e.g., social security cards, driver's licenses, passports, birth certificates, etc.
  • multi-document image data 202 may represent digital images of a number (e.g., thousands) of physical documents of one or more types.
  • the multi-document image data 202 may include images in one or more formats, such as raster formats (e.g., JPEG, TIFF, GIF, BMP, PNG, etc.) and/or vector formats (e.g., CGM, SVG, etc.), for example.
  • the multi-document image data 202 may include data obtained (e.g., by external data collection unit 42 of FIG. 1 ) from merchant computing systems 22 (e.g., point-of-sale devices with cameras for document identification) and/or from FAMS 14 (e.g., images of personal checks), for example.
  • the multi-document image data 202 may only include data representing images that customers (or other individuals associated with the documents) have expressly consented to share (e.g., as a prerequisite to making a purchase, or in exchange for fraud protection services, etc.).
  • the multi-document image data 202 may be associated with multiple fraud determination labels.
  • each label may be associated with data representing a digital image of a particular document.
  • the labels may include final fraud determinations (e.g., “fraud” or “no fraud,” or more complex labels such as “forgery,” “counterfeit,” “forgery—signature,” “counterfeit—angular line offset(s) outside tolerance,” etc.) that were made via earlier iterations of the process flow 200 , and/or external to the process flow 200 .
  • Multi-document image data 202 may include many (e.g., thousands) of positively and negatively labeled data sets.
  • the multi-document image data 202 may be analyzed to generate document fraud detection rules (e.g., to be stored in ML rules database 58 ).
  • document fraud detection rules e.g., to be stored in ML rules database 58 .
  • process stage 204 may serve to identify which characteristics of a document are probative of whether the document is counterfeit, and to determine the ranges, tolerances, etc., that are probative of whether the document is counterfeit.
  • process stage 204 also, or instead, identifies which characteristics of information entered in document fields are probative of whether the document was forged (e.g., drafted or populated by someone other than the person purported to have drafted or populated the document).
  • the rules generated or updated at process stage 204 may be applied to first document image data 210 .
  • the first document image data 210 may be digital image data corresponding to a particular, physical document.
  • the first document image data 210 may include information obtained (e.g., by external data collection unit 42 ) from one of merchant computing systems 22 (e.g., for real-time verification of an identification or other document presented during or prior to a sale), or from FAMS 14 (e.g., for real-time or batch-processing verification of a personal check prior to clearing the check), for example.
  • Some specific examples of rules that may be generated by process stage 204 , and applied at process stage 206 are described below in connection with FIG. 4F .
  • Process stage 206 may output information indicating whether fraud is suspected in connection with the document corresponding to first document image data 210 .
  • process stage 206 may output two percentage probabilities calculated according to the rules generated or updated at process stage 204 , with the first indicating the likelihood that the document is counterfeit and the second indicating the likelihood that the document includes forged content.
  • process stage 206 may output binary indicators of whether the document likely is, or likely is not, counterfeit and/or includes forged content (e.g., by comparing percentage probabilities to threshold probabilities).
  • further analysis may be performed at a process stage 212 .
  • the additional analysis may then be used to make a final fraud determination (e.g., a final decision on whether the document is fraudulent) at process stage 214 .
  • the process stage 206 may act as a filter, and flag only those documents having a relatively high probability of being fraudulent. In this manner, a considerably smaller amount of human and/or processing resources may be consumed at process stage 212 .
  • the final determination made at process stage 214 may be fed back into process stage 204 to provide additional labeled data for purposes of updating the rules.
  • a preliminary fraud determination made at process stage 206 may also be fed back into process stage 204 , to allow the machine learning program to determine and improve upon past performance/accuracy.
  • FIGS. 4A-4F depict exemplary factors and algorithms that may be used in connection with various fraud detection and/or classification rules, according to different embodiments. It is noted that the rule sets corresponding to FIGS. 4A-4F are purely for purposes of illustration and are not limiting. Particularly in embodiments where machine learning is utilized, for example, the algorithms and/or factors may be far more complex, and/or less intuitive, than some or all of the examples shown in FIGS. 4A-4F .
  • an exemplary rule set 220 may use various factors relating to online activity of a cardholder to detect fraud in connection with a particular credit or debit card transaction.
  • the rule set 220 may correspond to a particular embodiment and scenario in which the transaction at issue is a “card present” transaction, and in which the rule set 220 seeks to determine whether the cardholder made or otherwise authorized the transaction.
  • the rule set 220 may be incorporated into a review process that is generally applied to all transactions, a review process applied only to those transactions that were flagged by a preliminary fraud alert, or a review process applied only after a cardholder reports the transaction as unauthorized, for example.
  • the factors considered under the rule set 220 may include a number of interest-based factors 222 and a number of location-based factors 224 .
  • the interest-based factors 222 may relate to the cardholder's interest (or non-interest) in a product or service purchased via the transaction, and/or the merchant providing the product or service, while the location-based factors 224 may relate to the cardholder's location or probable location.
  • the interest-based factors 222 may include: (1) whether the cardholder searched online for the specific product or service purchased via the transaction at issue (e.g., by determining whether search terms entered by the cardholder included the name of the product or service involved in the transaction, or included a description of the product or service, etc.); (2) whether the cardholder visited a website associated with the merchant (e.g., by comparing URLs of web sites visited by the cardholder to a known URL of the merchant's website, or by searching the contents of websites visited by the cardholder for the merchant's name, etc.); (3) whether the cardholder endorsed the merchant, or the product or service provided by the merchant, via a social media account of the cardholder (e.g., by determining whether the cardholder “liked” the merchant, product or service via his or her Facebook® account, etc.); (4) whether the cardholder visited a website associated with a competitor of the merchant (e.g., by comparing URLs of web sites visited by the cardholder to known URLs of known
  • the location-based factors 224 may include: (1) whether the cardholder “checked in” to a flight having a destination near the location where the transaction was initiated (e.g., by determining whether the cardholder checked in to a flight having a destination at the city in which the transaction occurred, or within a threshold number of miles of the city in which the transaction occurred, etc.); (2) whether the cardholder visited a website associated with a place near (or in) which the transaction was initiated (e.g., by comparing URLs of web sites visited by the cardholder to URLs of websites known to be associated with particular areas, and/or by searching the contents of websites visited by the cardholder for location or area names, etc.); and/or (3) whether the cardholder endorsed a place near (or in) which the transaction was initiated via a social media account of the cardholder (e.g., by determining whether the cardholder “liked” the geographic area, attraction or other place via his or her Facebook® account, etc.).
  • the location-based factors 224 may include: (1) whether the cardholder
  • the data indicative of whether the circumstance corresponding to each of interest-based factors 222 and/or location-based factors 224 is present/true for a particular cardholder may be included in the first customer online activity data 110 described above in connection with FIG. 3A .
  • external data collection unit 42 of FIG. 1 may obtain the search terms, URLs, user online selections, etc., needed to determine whether the various factors exist, from the cardholder's computing device (e.g., one of cardholder computing devices 20 ) and/or from an ISP of other sources 24 .
  • each of the interest-based factors 222 and location-based factors 224 may be associated with a particular score or weighting value.
  • a total score may be calculated based upon which factors are, or are not, present (e.g., add 94 points if it is determined that the cardholder searched for the particular lawnmower model that was purchased, add another 80 points if the transaction was a “card present” transaction in the Chicago suburb of Joliet and the cardholder checked in to a flight to Chicago just prior to the transaction, etc.).
  • certain factors may instead be associated with negative scores (e.g., minus 80 if the cardholder checked in to a flight with a destination at least 200 miles from the site of the transaction and within one day of the transaction, etc.).
  • certain factors may be associated with metrics or algorithms that determine how heavily those factors are weighed.
  • search terms entered by the cardholder may be used to calculate a “need score” X (e.g., where X is based upon frequency of certain search terms being used, the amount of time spent clicking through search results, the magnitude and/or urgency of a problem indicated by the search terms, etc.), with X then being used to calculate a score equal to 0.2X.
  • a threshold e.g., a threshold of +100
  • a probability calculated based upon the total score e.g., a threshold of +100
  • a probability calculated based upon the total score e.g., a threshold of +100
  • FIG. 4A it can be seen that larger scores generally correspond to a greater probability that the transaction was made or authorized by the cardholder. If the transaction is being automatically reviewed (e.g., to determine whether a fraud alert is appropriate, without any initial input from the cardholder), this may mean that a lower score corresponds to a higher probability of fraud. Conversely, if the cardholder had reported the transaction as being fraudulent, a higher score may correspond to
  • the rule set 220 may also include one or more other types of factors not necessarily based upon online activities of the cardholder (e.g., whether GPS of the cardholder's smartphone or vehicle indicates that he or she was in that area shortly before or after the transaction, etc.), and/or may omit either interest-based factors 222 or location-based factors 224 .
  • an exemplary rule set 230 may use various factors relating to a transaction between a cardholder and a merchant to determine whether the transaction should be flagged as a candidate for a chargeback (e.g., to determine whether the transaction should be reviewed under a full set of chargeback rules associated with the appropriate card network entity).
  • the rule set 230 may correspond to a particular embodiment and scenario in which the transaction at issue is a “card present” transaction.
  • the factors considered under the rule set 230 may include: (1) whether an EMV chip card was not inserted in a point-of-sale EMV chip reader device of the merchant; (2) whether a non-EMV card was not swiped in a point-of-sale device of the merchant; (3) whether the card is past its expiration date; (4) whether the transaction is for the same amount and/or date as another transaction involving the same card and merchant (e.g., by analyzing other transactions involving the same account and merchant within a particular time span); and/or (2) whether the transaction is for greater than a threshold amount.
  • a threshold amount For example, one of merchant computing systems 22 of FIG.
  • the factors considered under rule set 230 may include more, fewer and/or different factors than those shown in FIG. 4B . It is noted that, in some embodiments, one or more factors may simply relate to the desirability (e.g., from a card issuer perspective) of further reviewing whether a chargeback is appropriate, without necessarily relating to the likelihood that a chargeback is appropriate.
  • each of the factors may be associated with a particular score or weighting value.
  • a total score may be calculated based upon which factors are, or are not, present (e.g., add 62 points if it is determined that the transaction has the same amount and date as another transaction occurring close in time and involving the same card and merchant).
  • certain factors may instead be associated with negative scores, and/or certain factors may be associated with metrics or algorithms that determine how heavily those factors are weighed.
  • the rule set 230 may then output the total score, a normalized total score, an indication of whether the total score exceeded a threshold, a probability calculated based upon the total score, and/or some other indicator or measure of the likelihood that a chargeback is appropriate for the transaction.
  • a normalized total score an indication of whether the total score exceeded a threshold
  • a probability calculated based upon the total score and/or some other indicator or measure of the likelihood that a chargeback is appropriate for the transaction.
  • FIG. 4B it can be seen that larger scores generally correspond to a greater probability that a chargeback is appropriate.
  • an exemplary rule set 240 (e.g., generated at process stage 144 of FIG. 3C ) may use a diverse array of factors to classify the type(s) of fraudulent activity, if any, that is/are suspected to be associated with an event or series of events.
  • the rule set 240 may correspond to a particular embodiment and scenario in which the event at issue is a financial transaction involving a debit or credit card.
  • the rule set 240 may classify fraudulent activity with respect to specific other types of events (e.g., loan applications), or may detect a variety of different event types (e.g., various types of financial transactions, loan or credit applications, etc.) and broadly classify fraudulent activity in connection with the detected event types (e.g., lost/stolen card use, application fraud, etc.).
  • specific other types of events e.g., loan applications
  • event types e.g., various types of financial transactions, loan or credit applications, etc.
  • broadly classify fraudulent activity in connection with the detected event types e.g., lost/stolen card use, application fraud, etc.
  • each potential classification may be associated with a number of factors probative of whether that type/class of fraud has occurred.
  • the rule set 240 may include counterfeit factors 242 (e.g., factors indicating that a counterfeit card was used for the transaction), account takeover factors 244 (e.g., factors indicating that the transaction resulted from an unauthorized person gaining online access to the credit or debit card account itself, via phishing, malware or other means), chargeback fraud factors 246 (e.g., factors indicating that the cardholder made or otherwise authorized a purchase that the cardholder later contested) and skimming factors 248 (e.g., factors indicating that the card information used for the transaction was obtained via a skimming card reader device illegally installed in an ATM, gas station pump or other location).
  • the rule set 240 may also, or instead, include factors corresponding to one or more other fraud classifications (e.g., forgery, lost/
  • the counterfeit factors 242 may include: (1) whether the suspect transaction and another, contemporaneous transaction (e.g., occurring within one hour, etc.) in another state are both “card present” transactions; and/or (2) if the suspect transaction is a “card present” transaction, whether the card (if an EMV chip card) was not inserted in an EMV chip card reader.
  • the counterfeit factors 242 may include more, fewer and/or different factors than those shown in FIG. 4C .
  • the account takeover factors 244 may include: (1) whether the debit or credit card account password was changed within the 10 days prior to the transaction; and/or (2) whether the transaction was originated from an IP address not associated with the cardholder.
  • external data collection unit 42 may retrieve password change information from account records database 30 of FIG. 1 , which may log all password update activity, and/or may retrieve IP address information from one of merchant computing systems 22 (e.g., the computing system of the merchant involved in the transaction).
  • the account takeover factors 244 may include more, fewer and/or different factors than those shown in FIG. 4C .
  • the chargeback fraud factors 246 may include: (1) whether the cardholder had searched online for the product or service purchased via the transaction; and/or (2) whether the cardholder had visited a website associated with the merchant involved in the transaction.
  • external data collection unit 42 of FIG. 1 may retrieve online search information (e.g., search terms and/or results) and/or URLs from the one of cardholder computing devices 20 that is associated with the cardholder, and/or from an ISP (of other sources 24 ) used by the cardholder.
  • the chargeback fraud factors 246 may include more, fewer and/or different factors than those shown in FIG. 4C .
  • the skimming factors 248 may include: (1) the number (X) of earlier transactions in which the card used for the transaction at issue was used at an ATM machine or a gas station pump within the 10 days prior to the transaction at issue; and/or (2) whether the transaction at issue originated from an IP address not associated with the cardholder.
  • external data collection unit 42 of FIG. 1 may retrieve transaction data indicating that certain past purchases were made using gas station pump card readers, and/or indicating that the card was used for one or more ATM withdrawals, from account records database 30 , and/or may retrieve the originating IP address from the one of merchant computing systems 22 associated with the merchant involved in the transaction at issue.
  • the skimming factors 248 may include more, fewer and/or different factors than those shown in FIG. 4C .
  • the data indicative of whether the circumstance corresponding to each of counterfeit factors 242 , account takeover factors 244 , chargeback fraud factors 246 and/or skimming factors 248 is present/true for a particular transaction may be included in the first account data 150 described above in connection with FIG. 3C , for example.
  • each of the counterfeit factors 242 , account takeover factors 244 , chargeback fraud factors 246 and skimming factors 248 may be associated with a particular score or weighting value.
  • the factors for each classification may be used to calculate a total score specific to that classification.
  • a counterfeit score may be calculated based upon which of factors 242 are, or are not, present
  • an account takeover score may be calculated based upon which of factors 244 are, or are not, present, and so on.
  • certain factors may instead be associated with negative scores, and/or certain factors (e.g., the first of skimming factors 248 shown in FIG. 4C ) may be associated with metrics or algorithms that determine how heavily those factors are weighed.
  • the rule set 240 may output the total score, a normalized total score, an indication of whether the total score exceeded a threshold, a probability calculated based upon the total score, and/or some other indicator or measure of the likelihood that fraud of that particular type/class occurred in connection with the transaction.
  • the classification at process stage 152 may be the classification having the highest score and/or probability under rule set 240 , or may include the score and/or probability for each classification, the top three classifications, etc.
  • an exemplary rule set 260 may use online search information (e.g., search terms, search results, clicked/selected search results, etc.) to detect whether an application was fraudulent (e.g., not populated and/or submitted by the purported applicant).
  • the rule set 260 may have been generated at process stage 164 of FIG. 3D , for example.
  • the rule set 260 may be incorporated into a review process that is generally applied to all applications received by a particular entity or anti-fraud service, or a review process applied only to those applications that were flagged by a preliminary fraud alert, for example.
  • the factors considered under the rule set 260 may generally be probative of whether the person that submitted the application (e.g., via a web browser, a dedicated application, as an email attachment, by snail mail, etc.) had performed one or more online searches indicating that he or she was trying to learn more about the purported applicant in order to populate particular fields of the application (e.g., a “home address” field, “employment history” fields, etc.).
  • the “purported applicant” may be a person whose name appears in a name and/or signature field of the application, for example.
  • the factors of exemplary rule set 260 may include: (1) whether the applicant used search terms that included the name of the purported applicant; (2) whether the search terms also included the words “address” or “residence” (and possibly other synonyms or near-synonyms); and/or (3) whether the search terms also included the words “employer,” “job” and/or “career” (and possibly other synonyms or near-synonyms).
  • the rule set 260 may include more, fewer and/or different factors than those shown in FIG. 4D .
  • the rule set 260 may include one or more factors relating to which search results appeared and/or were selected (e.g., “clicked” on after appearing on a user interface) by the applicant.
  • the data indicative of whether the circumstances corresponding to the factors of rule set 260 are present/true for a particular applicant may be included in the first applicant search history data 170 described above in connection with FIG. 3D .
  • external data collection unit 42 of FIG. 1 may obtain the search terms, search results, search result user selections, etc., needed to determine whether the various factors exist, from the applicant's computing device (e.g., similar to one of cardholder computing devices 20 ) and/or from an ISP of other sources 24 . Access to such information may be made a condition of having the application be considered, for example.
  • each of the factors of rule set 260 may be associated with a particular score or weighting value. A total score may then be calculated based upon which factors are, or are not, present. In some embodiments, certain factors may instead be associated with negative scores, and/or certain factors may be associated with metrics or algorithms that determine how heavily those factors are weighed.
  • the rule set 260 may then output the total score, a normalized total score, an indication of whether the total score exceeded a threshold, a probability calculated based upon the total score, and/or some other indicator or measure of the existence or likelihood of application fraud.
  • a normalized total score an indication of whether the total score exceeded a threshold
  • a probability calculated based upon the total score and/or some other indicator or measure of the existence or likelihood of application fraud.
  • larger scores may generally correspond to a greater probability that the application was not populated and/or submitted by the purported applicant.
  • a flow diagram illustrates at least a portion of a process flow 270 implementing an exemplary rule set for fraud dispute, or potential fraud dispute, resolution (e.g., a rule set generated at process stage 184 of FIG. 3E ).
  • the process flow 270 may be used to help resolve a dispute over a contested transaction, or to help a customer recall an unrecognized transaction, for example.
  • FIG. 4E illustrates a process flow, rather than just a set of factors, in order to better illustrate an example process for generating queries based upon the generated rules, according to one embodiment.
  • the process flow 270 may correspond to a particular embodiment and scenario in which the transaction subject to dispute or potential dispute is a credit or debit card transaction.
  • the rule set may specify that a process stage 272 determines whether the transaction was a “card present” transaction. If not, the rule set may specify that the flow proceed directly to a process stage 280 . If so, however, the rule set may specify that the flow instead proceeds to a process stage 274 .
  • the rule set may also specify that process stage 274 determines whether at least one other transaction associated with the cardholder's account occurred within some threshold number of hours (X) of the transaction at issue. If not, the rule set may specify that the flow proceeds directly to process stage 280 . If so, however, the rule set may specify that the flow instead proceeds to a process stage 276 .
  • Process stage 276 may generate one or more location-related queries using transaction data associated with the cardholder's account.
  • the queries may ask, for example, whether the cardholder was in (or near) one or more particular geographic areas or locations at various times. If the transaction at issue occurred in San Francisco, for example, with a first other “card present” transaction occurring in Santa Rosa four hours earlier and a second other “card present” transaction occurring in San Jose two hours later, process stage 276 may generate one or more queries asking whether the cardholder made or authorized the earlier and/or later transactions, and/or whether the cardholder traveled on a route from Santa Rosa to San Jose that passed through San Francisco, etc.
  • the location-related queries are generated based upon data associated with events or circumstances other than transactions. For example, if the transaction at issue occurred in Sarasota, Fla., and the data considered under the rule set indicates that the cardholder checked in to a flight to Tampa, process stage 276 may generate one or more queries asking whether the cardholder completed the flight, where the cardholder went after landing in Tampa, etc.
  • the rule set may also specify that process stage 280 determines whether the transaction at issue is associated with a billing alias that is dissimilar to the name of the merchant involved in the transaction.
  • the computing system of the merchant e.g., one of merchant computing systems 22 of FIG. 1
  • the determination at process stage 280 may use the billing alias to identify a legal and/or common name of the merchant (e.g., using a relational database stored in AFSS 12 or FAMS 14 ), and determine that there is at least some threshold level of dissimilarity (e.g., based upon difference of characters, character ordering, etc.) between the billing alias and the merchant name.
  • a relational database stored in AFSS 12 or FAMS 14
  • Process stage 282 may generate a query relating to the billing alias that was presented to the cardholder. For example, the query may ask whether the cardholder is aware that the billing alias is used by that particular merchant. In some embodiments, process stage 282 may instead generate a message that simply informs the cardholder that the billing alias corresponds to the merchant, without posing a question.
  • the rule set may specify that process stage 284 generates one or more default queries.
  • one default query may ask whether the cardholder lent his or her card to a friend or family member around the time of the transaction.
  • process stage 284 may be omitted from process flow 270 .
  • the queries (and possibly non-query messages) generated in process flow 270 may serve to help the cardholder recall whether the transaction was made or authorized, and/or process flow 270 may prompt the cardholder for responses that are considered by others (e.g., personnel of an entity associated with FAMS 14 of FIG. 1 ) to determine whether the transaction was likely fraudulent.
  • process flow 270 may include a number of iterative stages in which responses are received from the cardholder (e.g., from the respective one of cardholder computing devices 20 in FIG. 1 ) and used to generate additional, more detailed questions for the cardholder. For example, if a first query asks whether the cardholder recalls personally making another “card present” transaction that occurred at a nearby time and place, and the cardholder responds “no,” a new query may be generated asking whether the cardholder recalls personally making the next closest transaction (in terms of time and/or location).
  • an exemplary rule set 290 may use various factors relating to an imaged (e.g., photographed or scanned) physical document to determine whether the document should be flagged as a candidate for a more in-depth (e.g., manual) analysis/review for fraud purposes.
  • the rule set 290 may correspond to a particular embodiment and scenario in which the document is one that includes at least a signature field (e.g., a personal check, a driver's license, etc.).
  • the factors considered under the rule set 290 may include a number of counterfeit factors 292 and a number of forgery factors 294 , each of which may be evaluated by image analysis unit 52 of FIG. 1 using one or more image processing techniques.
  • the counterfeit factors 292 may relate to the look, presentation, format and/or structure of the document, while the forgery factors 294 may relate to the substance, style or format of information entered in one or more fields of the document.
  • the counterfeit factors 292 may include: (1) whether one or more absolute or relative dimensions and/or angles of the document, or of lines, illustrations, patterns, etc. shown on the document (excluding user-entered contents in fields such as the signature line), are outside one or more predetermined tolerances; (2) whether one or more colors on the document are outside a predetermined tolerance (e.g., color/frequency range); (3) whether one or more line thicknesses of the document (excluding user-entered field contents) are outside one or more predetermined tolerances; and/or (4) whether one or more fonts on the document (excluding user-entered field contents) are outside one or more predetermined tolerances.
  • predetermined tolerance e.g., color/frequency range
  • image analysis unit 52 may determine whether the ratio of the document length to the document width is within 0.1% of an expected value. As another example, image analysis unit 52 may determine whether horizontal and vertical lines on the document are within 0.3 degrees of the horizontal and vertical edges of the document, respectively. As yet another example, image analysis unit 52 may determine whether a font used for a field descriptor or other text on the document matches an expected font (e.g., by meeting a similarity threshold measured in any suitable manner). In other embodiments, the counterfeit factors 292 may include more, fewer and/or different factors than those shown in FIG. 4F .
  • the forgery factors 294 may include: (1) whether a signature entered in a signature field of the document match is outside a predetermined tolerance (e.g., using any suitable signature recognition technique); (2) whether handwriting entered in one or more fields of the document is outside a predetermined tolerance (e.g., by applying a suitable handwriting recognition technique); and/or (3) whether the format of information entered by a user in one or more fields does not match an expected format (e.g., using “9.12.16” rather than the expected “9/12/2016,” as established based upon other documents known to have been populated and/or submitted by the purported applicant).
  • the forgery factors 294 may include more, fewer and/or different factors than those shown in FIG. 4F .
  • the data indicative of whether the circumstances corresponding to counterfeit factors 292 and/or forgery factors 294 are present/true for a particular document may be included in the first document image data 210 described above in connection with FIG. 3F .
  • each of the counterfeit factors 292 and forgery factors 294 may be associated with a particular score or weighting value.
  • a total score may be calculated based upon which factors are, or are not, present.
  • certain factors may instead be associated with negative scores, and/or certain factors may be associated with metrics or algorithms that determine how heavily those factors are weighed.
  • the rule set 290 may then output the total score, a normalized total score, an indication of whether the total score exceeded a threshold, a probability calculated based upon the total score, and/or some other indicator or measure of the likelihood that the document is fraudulent.
  • the rule set 290 may output a separate total score, normalized score, probability, or other metric, for each of counterfeit factors 292 and forgery factors 294 , with the counterfeit metric indicating the likelihood that the document is a counterfeit and the forgery metric indicating the likelihood that the document was fraudulently populated by someone other than the purported person (e.g., by someone other than the person corresponding to the name, signature, address, etc. on the document).
  • the counterfeit metric indicating the likelihood that the document is a counterfeit
  • the forgery metric indicating the likelihood that the document was fraudulently populated by someone other than the purported person (e.g., by someone other than the person corresponding to the name, signature, address, etc. on the document).
  • the rule set 290 also includes one or more other types of factors not shown in FIG. 4F , and/or omits either counterfeit factors 292 or forgery factors 294 .
  • FIGS. 5-10 depict flow diagrams of various exemplary computer-implemented methods that may be implemented by one or more components of AFSS 12 of FIG. 1 .
  • AFSS 12 implements all of the methods corresponding to FIGS. 5-10 .
  • AFSS 12 implements only a subset (e.g., one, two, etc.) of the methods corresponding to FIGS. 5-10 .
  • Each of the methods described below may be implemented by fraud detection/classification unit 36 of FIG. 1 , for example.
  • FIG. 5 illustrates an exemplary computer-implemented method 300 of using customer data to determine that geolocation-based fraud alerts are false positives.
  • the method 300 may include, via one or more processors and/or transceivers (or a trained machine learning program), determining whether the electronic fraud alert is geolocation based (block 302 ), and if so, with customer permission, then receiving customer data (block 304 ), such as via wireless communication or data transmission over one or more radio links or wireless communication channels.
  • the customer data may be collected or generated via various processors, transceivers, or sensors associated with mobile devices, smart homes, and/or smart vehicles.
  • the customer data may indicate or be associated with telematics, online activity, browsing activity, IP address, credit card, customer location, and/or financial transaction data.
  • the method 300 may include determining whether two or more of the customer data sources include customer data indicating or confirming that the customer is traveling (block 306 ). If so, the method 300 may include determining whether the current customer location corresponds to the financial transaction location (block 308 ). If so, the method 300 may include not transmitting the electronic fraud alert to the customer's mobile device and/or flagging the fraud alert as a false positive; and if not, then transmitting the electronic fraud alert to the customer's mobile device (block 310 ).
  • FIG. 6 illustrates another exemplary computer-implemented method 320 of using customer data to determine whether geolocation-based fraud alerts are false positives.
  • the method 320 may include, via one or more processors and/or transceivers, receiving customer data (such as mobile device, smart home, smart vehicle, telematics, online activity, IP address, credit card, location, and/or financial transaction data) (block 322 ), such as via wireless communication or data transmission over one or more radio links or wireless communication channels.
  • the method 320 may include determining whether the customer is traveling using (or based upon processor analysis of) the customer data (block 324 ), such as by identifying that the current GPS location of the customer's mobile device and/or vehicle is outside of the customer's home address county or city.
  • the method 320 may include receiving an electronic fraud alert associated with the customer (block 326 ), such as an alert associated with a potentially unauthorized financial transaction being charged to the customer.
  • the method 320 may include determining whether the reason the electronic fraud alert was generated was location-based (block 328 ), such as processor or machine learning program determining that a location associated with the financial transaction does not match a home address location or home city of the customer. If so, and if the customer is determined to be traveling, the method 320 may include determining whether the current customer GPS or other location, such as determined from mobile device, IP address, or vehicle location, corresponds to the financial transaction location (block 330 ).
  • the method 320 may further include not transmitting the electronic fraud alert to the customer's mobile device, and/or flagging the fraud alert as a false positive; and if not, then transmitting the electronic fraud alert to the customer's mobile device (block 332 ).
  • a computer-implemented method of using customer data to determine that geolocation-based fraud alerts are false positives may be provided.
  • the method may include (1) determining, via the one or more processors, if an electronic fraud alert is a geolocation-based fraud alert (or otherwise generated based upon an unexpected or abnormal transaction location), such as by inputting the fraud alert and/or financial transaction underlying data into a machine learning program trained to identify geolocation-based fraud alerts; (2) if the electronic fraud alert is geolocation-based, then retrieving or receiving (with customer permission or affirmative consent), via the one or more processors and/or transceivers, two or more sources of customer data over one or more radio frequency links; (3) determining, via the one or more processors, if the customer data from two or more sources indicate or confirm that the customer is traveling (such as not currently at their home address or within a predetermined distance of their home address); (4) if the customer data indicates that the customer is traveling, then determining, via the one or more processors, whether a current customer location indicated by the
  • the method may further include receiving, via one or more processors and/or transceivers, transaction data associated with a financial transaction over a wireless communication channel; and inputting, via the one or more processors, the transaction data into a rules-engine to identify the financial transaction as potentially fraudulent and generate an electronic fraud alert.
  • the customer data may be collected or generated by a mobile device and/or mobile device sensors, and include one or more current or past GPS locations.
  • the customer data may be collected or generated by a vehicle controller or processor and/or vehicle-mounted sensors, and include one or more current or past GPS locations.
  • the customer data may be collected or generated by a smart home controller and/or home-mounted sensors, and include data indicating whether or not a home of the customer is presently occupied or vacant, and/or how long the home has been vacant.
  • the customer data may include an IP address of a customer computing device, and include one or more current or past GPS locations.
  • the customer data may include online, browsing, and/or social media activity received from a customer computing device. Additionally or alternatively, the customer data may include vehicle telematics data that includes one or more past or current GPS locations.
  • the method may include marking, via the one or more processors, the electronic fraud alert as verified and transmitting the electronic fraud alert to a customer mobile device to facilitate sending only confirmed fraud alerts to customers.
  • the fraud alert may be determined to be geolocation-based when a financial transaction location is not within a predetermined distance of a customer home address. Additionally or alternatively, the fraud alert may be determined to be geolocation-based when a financial transaction location does not correspond to normal travel activity or locations frequented by, or associated with, the customer.
  • a computer system configured to use customer data to determine that geolocation-based fraud alerts are false positives.
  • the computer system may include one or more processors and/or transceivers configured to: determine if an electronic fraud alert is a geolocation-based fraud alert (or otherwise generated based upon an unexpected or abnormal transaction location); if the electronic fraud alert is geolocation-based, then retrieve or receive (with customer permission or affirmative consent) via wireless communication or data transmission two or more sources of customer data over one or more radio frequency links or wireless communication channels; determine if the customer data from two or more sources indicate or confirm that the customer is traveling (such as not currently at their home address or within a predetermined distance of their home address); if the customer data indicates that the customer is traveling, then determine whether a current customer location indicated by the customer data retrieved matches, or corresponds to, the transaction location; and/or if the current customer location corresponds to the transaction location, then mark the electronic fraud alert as a false positive and not transmit the electronic fraud alert to a customer mobile device to reduce an amount of false
  • FIGS. 7 through 10 illustrate exemplary computer-implemented methods that use information about the locations of authorized cardholders (e.g., the primary cardholder, or another individual listed on the account) to prevent false positive fraud alerts, or to block potentially fraudulent financial transactions.
  • FIGS. 7 and 8 correspond to card-present financial transactions (e.g., where an individual swipes the card or inserts the card in a chip reader, or where a merchant does so on behalf of that individual after being handed the card)
  • FIGS. 9 and 10 correspond to online financial transactions (e.g., where an individual types card information into a website page configured to accept such information in connection with a desired transaction).
  • the methods of FIGS. 7 through 10 may be implemented by a card issuer/bank or by another entity (e.g., by an entity associated with FAMS 14 or AFSS 12 of FIG. 1 ), for example.
  • the methods of FIGS. 7 and 10 may provide the benefit of avoiding unnecessary network communications and/or computer processing associated with false positive fraud alerts, while the methods of FIGS. 8 and 9 may provide the benefit of avoiding fraudulent transactions in the first instance, without requiring more cumbersome and/or time-consuming techniques (e.g., sending an authorization code to the authorized cardholder to verify a suspicious transaction). Moreover, all of the methods in FIGS. 7 through 10 may more accurately detect fraud by using the geographic location of the authorized cardholder in conjunction with the location of the transaction or the location of the computer used to enter card information.
  • FIGS. 7 through 10 may detect and/or prevent fraudulent transactions even when the physical debit or credit card has been stolen (as opposed to only copying down the card number or “skimming,” etc.).
  • the fraud alert may be an electronic fraud alert generated by the device or system implementing the method 400 , or received from another device or system (e.g., from a card issuer system, such as FAMS 14 of FIG. 1 , via a network such as network 26 of FIG. 1 ), for example.
  • the financial transaction may be a card-present transaction that is associated with a debit or credit card account, and was purportedly entered into by an authorized cardholder associated with the account.
  • a first geographic location, at which information associated with the account was obtained, may be determined (block 404 ).
  • the information associated with the account may have been obtained by swiping or inserting the card in a device (e.g., part of one of merchant computing systems 22 of FIG. 1 ) in connection with the financial transaction, for example.
  • the first geographic location may be identified based upon location information included in a field of transaction data associated with the financial transaction, such as transaction data that was retrieved from an account records database (e.g., database 30 of FIG. 1 ).
  • block 404 may occur prior to block 402 , in which case block 402 may include comparing the first geographic location to a set of one or more locations known to be typical or expected for the authorized cardholder (e.g., a home address, city and/or state), and/or may include generating the fraud alert in response to determining that the first geographic location does not correspond to (e.g., is not at, or not within a threshold distance of) the set of typical/expected locations.
  • a set of one or more locations known to be typical or expected for the authorized cardholder e.g., a home address, city and/or state
  • generating the fraud alert in response to determining that the first geographic location does not correspond to (e.g., is not at, or not within a threshold distance of) the set of typical/expected locations.
  • the time of the financial transaction may also be determined (block 406 ).
  • the time may be identified based upon time information (e.g., a time stamp) included in a particular field of transaction data associated with the financial transaction, such as the transaction data described above.
  • the geolocation data may be time-stamped data received from a third party server with the express consent of the authorized cardholder, or retrieved from a database in which the data was stored (with the cardholder's express consent) after being received from a mobile device of the cardholder, for example.
  • the geolocation data may include GPS data (e.g., collected by a smartphone or other mobile device of the authorized cardholder), data indicating identifiers and/or signal strengths of WiFi access points that were near the cardholder (e.g., collected by a smartphone or other mobile device of the authorized cardholder), data indicating that the authorized cardholder had “checked in” at a particular location (e.g., via a social media or other application), data indicating that the authorized cardholder used a smart appliance at a known location (e.g., at the cardholder's home), and/or other types of data indicative of the cardholder's locations at particular times.
  • GPS data e.g., collected by a smartphone or other mobile device of the authorized cardholder
  • data indicating identifiers and/or signal strengths of WiFi access points that were near the cardholder e.g., collected by a smartphone or other mobile device of the authorized cardholder
  • data indicating that the authorized cardholder had “checked in” at a particular location e.g.,
  • the location of the cardholder at the time of the financial transaction may be determined by matching a time-stamp to the time determined at block 406 , using a location with a time-stamp that corresponds to a nearest time (e.g., so long as that time is within some threshold time of the time determined at block 406 ), or in another suitable manner.
  • the second geographic location corresponds to the first geographic location (block 410 ). For example, it may be determined at block 410 that the first and second geographic locations are the same (e.g., the same city), are within a same geographic territory (e.g., cities within the same state), or are within a threshold distance of each other (e.g., cities or more precise locations within 50 miles of each other, 100 miles of each other, etc.).
  • the first and second geographic locations are the same (e.g., the same city), are within a same geographic territory (e.g., cities within the same state), or are within a threshold distance of each other (e.g., cities or more precise locations within 50 miles of each other, 100 miles of each other, etc.).
  • the fraud alert may be marked as a false positive (block 412 ), such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction.
  • a “verified” flag or field associated with the fraud alert may be set to a value of “0” or “false” at block 412 , and a notification unit (e.g., notification unit 56 of FIG. 1 ) may decide not to send the fraud alert to the authorized cardholder's mobile device and/or other computing device (e.g., as an email or text message) based upon the flag or field value.
  • a request to authorize a financial transaction may be received (block 422 ).
  • the financial transaction may be a card-present transaction that is associated with a debit or credit card account, and is purportedly being entered into by an authorized cardholder associated with the account.
  • the financial transaction in the method 420 , is one that has not yet been fully executed.
  • the request may have been automatically or manually generated by the card issuer when deciding whether to clear the transaction, or automatically or manually generated by a merchant shortly after receiving credit card information (e.g., by a swipe or insertion of the card), for example.
  • the request may include the credit card information (e.g., credit card number, expiration date and/or security code) and/or other information relating to the financial transaction.
  • a first geographic location, at which information associated with the account was obtained (e.g., by swiping or inserting the card in connection with the financial transaction), may be determined (block 424 ).
  • Block 424 may be similar to block 404 of the method 400 , for example. In some embodiments and/or scenarios, however, the first geographic location is determined by identifying the location as specified in the request received at block 422 .
  • the geolocation data and/or the source of such data may be similar to that described above in connection with block 408 of the method 400 , for example.
  • the geolocation data may not be time-stamped, or such time stamps may exist but not be utilized.
  • the second geographic location may be the current location of the authorized cardholder.
  • the second geographic location does not correspond to the first geographic location (block 428 ). For example, it may be determined at block 428 that the first and second geographic locations are not the same (e.g., not the same city), are not within a same geographic territory (e.g., not cities within the same state), or are not within a threshold distance of each other (e.g., not cities or other, more specific locations within 50 miles of each other, 100 miles of each other, etc.).
  • the financial transaction may be prevented from being executed (block 430 ). If the method 420 is implemented by a computing system of the card issuer, for example, block 430 may include not clearing the financial transaction. As another example, a merchant terminal (e.g., part of one of merchant computing systems 22 of FIG. 1 ) sending the request received at block 422 (or that is otherwise associated with the financial transaction) may be sent a fraud alert indicating that the transaction may be fraudulent and/or should not be completed. In yet another example embodiment, the fraud alert may be sent to a computing system of the card issuer (e.g., if the request received at block 422 was received from such a computing system).
  • a computing system of the card issuer e.g., if the request received at block 422 was received from such a computing system.
  • a request to authorize a financial transaction may be received (block 442 ).
  • the financial transaction may be an online transaction that is associated with a debit or credit card account, and is purportedly being entered into by an authorized cardholder associated with the account.
  • the financial transaction in the method 440 , is one that has not yet been fully executed.
  • the request may have been automatically or manually generated by the card issuer when deciding whether to clear the transaction, or automatically or manually generated by a merchant shortly after receiving credit card information (e.g., shortly after the merchant computing system received credit card information that was manually entered by a person purporting to be the authorized cardholder), for example.
  • the request may include the credit card information (e.g., credit card number, expiration date and/or security code) and/or other information relating to the financial transaction.
  • a computing device at which information associated with the card account (e.g., the card number, expiration date, and/or three- or four-digit security code) was entered in connection with the financial transaction may be identified (block 444 ).
  • the computing device may be identified by receiving an IP address of the computing device from the computing system of the merchant associated with the financial transaction (either directly, or via the card issuer), for example.
  • a first geographic location, at which the computing device identified at block 444 resides, may be determined (block 446 ).
  • the first geographic location is determined by using the IP address of the computing device.
  • the IP address itself may indicate physical location (at a high level of generality), or the IP address may be used as a key to a location database that relates IP addresses to more specific physical locations.
  • a computing system implementing the method 440 may, as a part of its fraud prevention services, ask cardholders to voluntarily register any fixed-location computers (e.g., desktop computers) that they expect to use for online purchases, with the registration process including sending (from each such computer) a message specifying the physical location of the computer.
  • the first geographic location is determined by identifying a location specified in the request received at block 442 , or in another suitable manner.
  • the geolocation data and/or the source of such data may be similar to that described above in connection with block 408 of the method 400 , for example.
  • the geolocation data may not be time-stamped, or such time stamps may exist but not be utilized.
  • the second geographic location may be the current location of the authorized cardholder.
  • Block 450 may be similar to block 428 of the method 420 , for example.
  • the financial transaction may be prevented from being executed (block 452 ).
  • Block 452 may be similar to block 430 of the method 420 , for example.
  • the fraud alert may be an electronic fraud alert generated by the device or system implementing the method 460 , or received from another device or system (e.g., from a card issuer system, such as FAMS 14 of FIG. 1 , via a network such as network 26 of FIG. 1 ), for example.
  • the financial transaction may be an online transaction that is associated with a debit or credit card account, and is purportedly entered into by an authorized cardholder associated with the account.
  • a computing device at which information associated with the card account (e.g., the card number, expiration date, and/or three- or four-digit security code) was entered in connection with the financial transaction may be identified (block 464 ).
  • Block 464 may be similar to block 444 of the method 440 , for example.
  • a first geographic location, at which the computing device identified at block 464 resides, may be determined (block 466 ).
  • the first geographic location may be identified based upon an IP address, of the computing device, that may be specified in a particular field of transaction data that is retrieved from an account records database (e.g., database 30 of FIG. 1 ).
  • the first geographic location itself may be specified in such a field.
  • block 466 may occur prior to block 462 , in which case block 462 may include comparing the first geographic location to a set of one or more locations known to be typical or expected for the authorized cardholder (e.g., a home address, city and/or state), and/or may include generating the fraud alert in response to determining that the first geographic location does not correspond to (e.g., is not at, or not within a threshold distance of) the set of typical/expected locations.
  • a set of one or more locations known to be typical or expected for the authorized cardholder e.g., a home address, city and/or state
  • generating the fraud alert in response to determining that the first geographic location does not correspond to (e.g., is not at, or not within a threshold distance of) the set of typical/expected locations.
  • the time of the financial transaction may also be determined (block 468 ).
  • the time may be identified based upon time information (e.g., a time stamp) included in a particular field of transaction data associated with the financial transaction, such as transaction data that is retrieved from an account records database (e.g., database 30 of FIG. 1 ).
  • time information e.g., a time stamp
  • Block 470 may be similar to block 408 of the method 400 , for example.
  • Block 472 may be similar to block 410 of the method 400 , for example.
  • the fraud alert may be marked as a false positive (block 474 ), such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction.
  • Block 474 may be similar to block 412 of the method 400 , for example.
  • FIG. 11 depicts an exemplary computer system 500 in which the techniques described herein may be implemented, according to one embodiment.
  • the computer system 500 of FIG. 11 may include a computing device in the form of a computer 510 .
  • Components of the computer 510 may include, but are not limited to, a processing unit 520 , a system memory 530 , and a system bus 521 that couples various system components including the system memory 530 to the processing unit 520 .
  • the system bus 521 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, or a local bus, and may use any suitable bus architecture.
  • such architectures include the Industry Standard Architecture (ISA) bus, Micro Channel Architecture (MCA) bus, Enhanced ISA (EISA) bus, Video Electronics Standards Association (VESA) local bus, and Peripheral Component Interconnect (PCI) bus (also known as Mezzanine bus).
  • ISA Industry Standard Architecture
  • MCA Micro Channel Architecture
  • EISA Enhanced ISA
  • VESA Video Electronics Standards Association
  • PCI Peripheral Component Interconnect
  • Computer 510 may include a variety of computer-readable media.
  • Computer-readable media may be any available media that can be accessed by computer 510 and may include both volatile and nonvolatile media, and both removable and non-removable media.
  • Computer-readable media may comprise computer storage media and communication media.
  • Computer storage media may include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data.
  • Computer storage media may include, but is not limited to, RAM, ROM, EEPROM, FLASH memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by computer 510 .
  • Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and may include any information delivery media.
  • modulated data signal means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
  • communication media may include wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, radio frequency (RF), infrared and other wireless media. Combinations of any of the above are also included within the scope of computer-readable media.
  • the system memory 530 may include computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 531 and random access memory (RAM) 532 .
  • ROM read only memory
  • RAM random access memory
  • BIOS basic input/output system
  • RAM 532 typically contains data and/or program modules that are immediately accessible to, and/or presently being operated on, by processing unit 520 .
  • FIG. 11 illustrates operating system 534 , application programs 535 , other program modules 536 , and program data 537 .
  • the computer 510 may also include other removable/non-removable, volatile/nonvolatile computer storage media.
  • FIG. 11 illustrates a hard disk drive 541 that reads from or writes to non-removable, nonvolatile magnetic media, a magnetic disk drive 551 that reads from or writes to a removable, nonvolatile magnetic disk 552 , and an optical disk drive 555 that reads from or writes to a removable, nonvolatile optical disk 556 such as a CD ROM or other optical media.
  • removable/non-removable, volatile/nonvolatile computer storage media that can be used in the exemplary operating environment include, but are not limited to, magnetic tape cassettes, flash memory cards, digital versatile disks, digital video tape, solid state RAM, solid state ROM, and the like.
  • the hard disk drive 541 may be connected to the system bus 521 through a non-removable memory interface such as interface 540
  • magnetic disk drive 551 and optical disk drive 555 may be connected to the system bus 521 by a removable memory interface, such as interface 550 .
  • the drives and their associated computer storage media discussed above and illustrated in FIG. 11 provide storage of computer-readable instructions, data structures, program modules and other data for the computer 510 .
  • hard disk drive 541 is illustrated as storing operating system 544 , application programs 545 , other program modules 546 , and program data 547 .
  • operating system 544 application programs 545 , other program modules 546 , and program data 547 are given different numbers here to illustrate that, at a minimum, they are different copies.
  • a user may enter commands and information into the computer 510 through input devices such as cursor control device 561 (e.g., a mouse, trackball, touch pad, etc.) and keyboard 562 .
  • cursor control device 561 e.g., a mouse, trackball, touch pad, etc.
  • keyboard 562 e.g., a mouse, trackball, touch pad, etc.
  • a monitor 591 or other type of display device is also connected to the system bus 521 via an interface, such as a video interface 590 .
  • computers may also include other peripheral output devices such as printer 596 , which may be connected through an output peripheral interface 595 .
  • the computer 510 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 580 .
  • the remote computer 580 may be a personal computer, a server, a router, a network PC, a peer device or other common network node, and may include many or all of the elements described above relative to the computer 510 , although only a memory storage device 581 has been illustrated in FIG. 11 .
  • the logical connections depicted in FIG. 11 include a local area network (LAN) 571 and a wide area network (WAN) 573 , but may also include other networks.
  • LAN local area network
  • WAN wide area network
  • Such networking environments are commonplace in hospitals, offices, enterprise-wide computer networks, intranets and the Internet.
  • the computer 510 When used in a LAN networking environment, the computer 510 is connected to the LAN 571 through a network interface or adapter 570 .
  • the computer 510 may include a modem 572 or other means for establishing communications over the WAN 573 , such as the Internet.
  • the modem 572 which may be internal or external, may be connected to the system bus 521 via the input interface 560 , or other appropriate mechanism.
  • the communications connections 570 , 572 which allow the device to communicate with other devices, are an example of communication media, as discussed above.
  • program modules depicted relative to the computer 510 may be stored in the remote memory storage device 581 .
  • FIG. 11 illustrates remote application programs 585 as residing on memory device 581 .
  • the techniques for detecting and/or classifying fraud described above may be implemented in part or in their entirety within a computer system such as the computer system 500 illustrated in FIG. 11 .
  • the computer 510 may be included in AFSS 12 of FIG. 1 , for example, and/or the remote application programs 585 may include one or more applications of either FAMS 14 , one of cardholder computing device 20 , one of merchant computing systems 22 , or a computing device of other sources 24 .
  • the functionality of fraud detection/classification unit 36 of FIG. 1 may be implemented by one or more of application programs 535 and/or other program modules 536 .
  • hard disk drive 541 e.g., as program data 547
  • magnetic disk 552 and/or optical disk drive 555 and/or the data retrieved by fraud detection/classification unit 36 of FIG. 1 may be stored in hard disk drive 541 (e.g., as program data 547 ) and/or RAM 532 (e.g., as program data 537 ).
  • a computer-implemented method of using customer data to determine that geolocation-based fraud alerts are false positives may be implemented in one or more servers or other computing devices.
  • the method may include (1) determining, by one or more processors, that an electronic fraud alert is a geolocation-based fraud alert generated based upon an unexpected or abnormal transaction location; (2) in response to determining that the electronic fraud alert is a geolocation-based fraud alert, obtaining, by the one or more processors and via one or more radio frequency links, customer data from two or more sources; (3) determining, by the one or more processors, that the customer data from the two or more sources indicates that a customer is traveling; (4) in response to determining that the customer data indicates that the customer is traveling, determining, by the one or more processors, that a customer location indicated by the customer data corresponds to the transaction location; and/or (5) in response to determining that the customer location corresponds to the transaction location, (i) marking, by the one or more processors, the electronic fraud alert as a false positive and (ii)
  • determining that the electronic fraud alert is a geolocation-based fraud alert may include inputting, by the one or more processors, one or both of (i) the electronic fraud alert, and (ii) transaction data corresponding to a financial transaction associated with the electronic fraud alert, into a machine learning program that is trained to identify geolocation-based fraud alerts.
  • the method may further include obtaining, by the one or more processors and via a wireless communication channel, transaction data corresponding to a financial transaction associated with the electronic fraud alert, and/or inputting, by the one or more processors, the transaction data into a rules engine to identify the financial transaction as potentially fraudulent and generate the electronic fraud alert.
  • the customer data may be collected or generated by one or both of (i) the mobile device and (ii) one or more sensors of the mobile device, and/or may include one or more current or past GPS locations.
  • the customer data may be collected or generated by one or both of (i) a vehicle controller or processor and (ii) one or more vehicle-mounted sensors, and/or may include one or more current or past GPS locations.
  • the customer data may be collected or generated by one or both of (i) a smart home controller and (ii) one or more home-mounted sensors, and/or may include data indicating one or both of (i) whether a home of the customer is presently occupied or vacant and (ii) how long the home of the customer has been vacant.
  • the customer data may include an IP address of a customer computing device, and/or may include one or more current or past GPS locations. Additionally or alternatively, the customer data may include one or more of (i) online data received from a customer computing device, (ii) browsing data received from the customer computing device, or (iii) social media activity data received from the customer computing device. Additionally or alternatively, the customer data may include vehicle telematics data that includes one or more past or current GPS locations.
  • the method may include: determining, by the one or more processors, that another electronic fraud alert is a geolocation-based fraud alert generated based upon another unexpected or abnormal transaction location; in response to determining that the other electronic fraud alert is a geolocation-based fraud alert, obtaining, by the one or more processors and via one or more other radio frequency links, additional customer data from two or more other sources; determining, by the one or more processors, that the additional customer data from the two or more other sources indicates that another customer is traveling; in response to determining that the additional customer data indicates that the other customer is traveling, determining, by the one or more processors, that another customer location indicated by the additional customer data does not correspond to the other transaction location; and/or in response to determining that the other customer location does not correspond to the other transaction location, (i) marking, by the one or more processors, the other electronic fraud alert as verified and (ii) causing, by the one or more processors, the electronic fraud alert to be transmitted to a mobile device of the other customer to facilitate sending only confirmed fraud
  • determining that the electronic fraud alert is a geolocation-based fraud alert may include determining that the transaction location is not within a predetermined distance of a customer home address. Additionally or alternatively, determining that the electronic fraud alert is a geolocation-based fraud alert may include determining that the transaction location does not correspond to travel activity or locations associated with the customer.
  • a computer-implemented method of reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions may be implemented in one or more servers or other computing devices.
  • the method may include: (1) determining, by one or more processors of the one or more servers, that a fraud alert exists for a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account; (2) determining, by the one or more processors, a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determining, by the one or more processors, a time of the financial transaction; (4) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction;
  • determining the first geographic location may occur prior to determining that the fraud alert exists, and determining that the fraud alert exists may include comparing the first geographic location to a set of one or more typical locations of the authorized cardholder, and/or generating the fraud alert in response to determining that the first geographic location does not correspond to the set of one or more typical locations.
  • the method may further include retrieving, by the one or more processors and from an account records database, transaction data associated with the financial transaction, and/or determining the first geographic location may include identifying the first geographic location based upon location information included in a first field of the transaction data.
  • determining the time of the financial transaction may include identifying the time of the financial transaction based upon time information included in a second field of the transaction data. Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a third party server.
  • determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a mobile device of the authorized cardholder, storing the geolocation data in a database, and/or retrieving the geolocation data from the database. Additionally or alternatively, receiving the geolocation data from a mobile device of the authorized cardholder may include receiving GPS location data from the mobile device.
  • determining that the second geographic location corresponds to the first geographic location may include determining that the first geographic location and the second geographic location are within a threshold distance of each other. Additionally or alternatively, determining that the second geographic location corresponds to the first geographic location may include determining that the first geographic location and the second geographic location are within a same geographic territory.
  • a computer-implemented method of preventing fraudulent card-present financial transactions may be implemented in one or more servers.
  • the method may include: (1) receiving, by one or more processors of the one or more servers, a request to authorize a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly being entered into by an authorized cardholder associated with the debit or credit card account; (2) determining, by the one or more processors, a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at a time of the financial transaction; (4) determining, by the one or more processors, that the second geographic location does not correspond to the first geographic location; and/or (5) in response to determining that the second
  • determining a first geographic location may include identifying a first geographic location specified in the request. Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a third party server.
  • preventing the financial transaction from being executed may include one or both of causing a fraud alert to be sent to a merchant terminal associated with the financial transaction, and causing a fraud alert to be sent to a computing system of a card issuer associated with the debit or credit card account.
  • a computer-implemented method of preventing fraudulent online financial transactions may be implemented in one or more servers.
  • the method may include: (1) receiving, by one or more processors of the one or more servers, a request to authorize a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is an online transaction purportedly being entered into by an authorized cardholder associated with the debit or credit card account; (2) identifying, by the one or more processors, a computing device at which information associated with the debit or credit card account was entered in connection with the financial transaction; (3) determining, by the one or more processors, a first geographic location at which the computing device resides; (4) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at a time of the financial transaction; (5) determining, by the one or more processors, that the second geographic location does not correspond to the first geographic location; and/or (6) in response to
  • receiving the request to authorize the financial transaction may include receiving the request to authorize the financial transaction from a computing system of a merchant associated with the financial transaction.
  • identifying the computing device at which information associated with the debit or credit card account was entered may include receiving an IP address of the computing device from the computing system of the merchant associated with the financial transaction.
  • determining the first geographic location may include determining the first geographic location by using the IP address as a key to a location database. Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a third party server.
  • determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a mobile device of the authorized cardholder, storing the geolocation data in a database, and/or retrieving the geolocation data from the database. Additionally or alternatively, receiving the geolocation data from a mobile device of the authorized cardholder may include receiving GPS location data from the mobile device.
  • determining that the second geographic location does not correspond to the first geographic location may include one or both of determining that the first geographic location and the second geographic location are not within a threshold distance of each other, and/or determining that the first geographic location and the second geographic location are not within a same geographic territory.
  • a computer-implemented method of reducing false positives among geolocation-based fraud alerts issued in connection with online financial transactions may be implemented in one or more servers.
  • the method may include: (1) determining, by one or more processors of the one or more servers, that a fraud alert exists for a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is an online transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account; (2) identifying, by the one or more processors, a computing device at which information associated with the debit or credit card account was entered in connection with the financial transaction; (3) determining, by the one or more processors, a first geographic location at which the computing device resides; (4) determining, by the one or more processors, a time of the financial transaction; (5) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction
  • determining the first geographic location may occur prior to determining that the fraud alert exists, and determining that the fraud alert exists may include comparing the first geographic location to a set of one or more typical locations of the authorized cardholder, and/or generating the fraud alert in response to determining that the first geographic location does not correspond to the set of one or more typical locations.
  • the method may further include retrieving, by the one or more processors and from an account records database, transaction data associated with the financial transaction, determining the first geographic location may include identifying the first geographic location based upon location information included in a first field of the transaction data, and/or determining the time of the financial transaction may include identifying the time of the financial transaction based upon time information included in a second field of the transaction data.
  • determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a third party server. Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving GPS location data from a mobile device of the authorized cardholder, storing the GPS location data in a database, and/or retrieving the GPS location data from the database.
  • determining that the second geographic location corresponds to the first geographic location may include determining that the first geographic location and the second geographic location are within a threshold distance of each other. Additionally or alternatively, determining that the second geographic location corresponds to the first geographic location may include determining that the first geographic location and the second geographic location are within a same geographic territory.
  • the memory may store instructions that, when executed by the one or more processors, cause the computer system to: (1) determine that an electronic fraud alert is a geolocation-based fraud alert generated based upon an unexpected or abnormal transaction location; (2) in response to determining that the electronic fraud alert is a geolocation-based fraud alert, obtain, via one or more radio frequency links, customer data from two or more sources; (3) determine that the customer data from the two or more sources indicates that a customer is traveling; (4) in response to determining that the customer data indicates that the customer is traveling, determine that a customer location indicated by the customer data corresponds to the transaction location; and/or (5) in response to determining that the customer location corresponds to the transaction location, (i) mark the electronic fraud alert as a false positive and (ii) cause the electronic fraud alert to not be transmitted to a mobile device of the customer in order to reduce an amount of false positives that are transmitted to customers
  • the instructions may cause the computing system to determine that the electronic fraud alert is a geolocation-based fraud alert at least by inputting one or both of (i) the electronic fraud alert, and (ii) transaction data corresponding to a financial transaction associated with the electronic fraud alert, into a machine learning program that is trained to identify geolocation-based fraud alerts.
  • the instructions may further cause the computing system to obtain, via a wireless communication channel, transaction data corresponding to a financial transaction associated with the electronic fraud alert, and/or input the transaction data into a rules engine to identify the financial transaction as potentially fraudulent and generate the electronic fraud alert.
  • the customer data may include vehicle telematics data that includes one or more past or current GPS locations.
  • a computer system for reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions may include a location database configured to store geolocation data indicating geographic locations of authorized cardholders over time, one or more processors, and a non-transitory memory.
  • the memory may store instructions that, when executed by the one or more processors, cause the one or more processors to: (1) determine that a fraud alert exists for a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account; (2) determine a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determine a time of the financial transaction; (4) determine, based upon first geolocation data stored in the location database and indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction; (5) determine that the second geographic location corresponds to the first geographic location; and/or (6) in response to determining that the second geographic location corresponds to the first geographic location, mark the fraud alert as a false positive such that no fraud alert is sent to the authorized cardholder in connection with
  • the instructions may further cause the one or more processors to retrieve, from an account records database, transaction data associated with the financial transaction, the instructions may cause the one or more processors to determine the first geographic location at least by identifying the first geographic location based upon location information included in a first field of the transaction data, and/or the instructions may cause the one or more processors to determine the time of the financial transaction at least by identifying the time of the financial transaction based upon time information included in a second field of the transaction data.
  • the instructions may cause the one or more processors to determine that the authorized cardholder was at the second geographic location at the time of the financial transaction at least by receiving the first geolocation data from a mobile device of the authorized cardholder, storing the first geolocation data in the location database, and/or retrieving the first geolocation data from the location database. Additionally or alternatively, receiving the first geolocation data may include GPS location data.
  • the instructions may cause the one or more processors to determine that the second geographic location corresponds to the first geographic location by at least one of determining that the first geographic location and the second geographic location are within a threshold distance of each other, or determining that the first geographic location and the second geographic location are within a same geographic territory.
  • a computer system for preventing fraudulent online financial transactions may include a location database configured to store geolocation data indicating geographic locations of authorized cardholders over time, one or more processors, and a non-transitory memory.
  • the memory stores instructions that, when executed by the one or more processors, may cause the one or more processors to: (1) receive a request to authorize a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is an online transaction purportedly being entered into by an authorized cardholder associated with the debit or credit card account; (2) identify a computing device at which information associated with the debit or credit card account was entered in connection with the financial transaction; (3) determine a first geographic location at which the computing device resides; (4) determine, based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at a time of the financial transaction; (5) determine that the second geographic location does not correspond to the first geographic location; and/or (6) in response to determining that the second
  • the instructions may cause the one or more processors to receive the request to authorize the financial transaction from a computing system of a merchant associated with the financial transaction. Additionally or alternatively, the instructions may cause the one or more processors to identify the computing device at which information associated with the debit or credit card account was entered at least by receiving an IP address of the computing device from the computing system of the merchant associated with the financial transaction.
  • the instructions may cause the one or more processors to determine the first geographic location at least by determining the first geographic location by using the IP address as a key to a location database. Additionally or alternatively, the instructions may cause the one or more processors to determine that the authorized cardholder was at the second geographic location at the time of the financial transaction at least by receiving the geolocation data from either (i) a third party server; or (ii) a mobile device of the authorized cardholder.
  • the instructions may cause the one or more processors to determine that the second geographic location does not correspond to the first geographic location at least by one or both of determining that the first geographic location and the second geographic location are not within a threshold distance of each other, and determining that the first geographic location and the second geographic location are not within a same geographic territory.
  • a non-transitory, computer-readable medium stores instructions that, when executed by one or more processors, may cause the one or more processors to: (1) determine that an electronic fraud alert is a geolocation-based fraud alert generated based upon an unexpected or abnormal transaction location; (2) in response to determining that the electronic fraud alert is a geolocation-based fraud alert, obtain, via one or more radio frequency links, customer data from two or more sources; (3) determine that the customer data from the two or more sources indicates that a customer is traveling; (4) in response to determining that the customer data indicates that the customer is traveling, determine that a customer location indicated by the customer data corresponds to the transaction location; and/or (5) in response to determining that the customer location corresponds to the transaction location, (i) mark the electronic fraud alert as a false positive and (ii) cause the electronic fraud alert to not be transmitted to a mobile device of the customer in order to reduce an amount of false positives that are transmitted to customers.
  • the non-transitory, computer-readable medium may store instructions that include additional, fewer or alternative
  • the instructions may cause the computing system to determine that the electronic fraud alert is a geolocation-based fraud alert at least by inputting one or both of (i) the electronic fraud alert, and (ii) transaction data corresponding to a financial transaction associated with the electronic fraud alert, into a machine learning program that is trained to identify geolocation-based fraud alerts.
  • the instructions may further cause the computing system to obtain, via a wireless communication channel, transaction data corresponding to a financial transaction associated with the electronic fraud alert, and/or input the transaction data into a rules engine to identify the financial transaction as potentially fraudulent and generate the electronic fraud alert.
  • the customer data may include vehicle telematics data that includes one or more past or current GPS locations.

Abstract

In a computer-implemented method of reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions, it may be determined that a fraud alert exists for a card-present financial transaction. A first geographic location, at which information associated with a debit or credit card account was obtained by swiping or inserting a card, may be determined, along with a time of the transaction. Based upon geolocation data indicating one or more geographic locations of the authorized cardholder, it may be determined that the authorized cardholder was at a second geographic location at the time of the financial transaction. If the second geographic location corresponds to the first geographic location, the fraud alert may be marked as a false positive, such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This claims the benefit of U.S. Patent Application No. 62/313,196, filed on Mar. 25, 2016 and entitled “Reducing Financial Fraud Using Machine Learning and Other Techniques,” U.S. Patent Application No. 62/318,423, filed on Apr. 5, 2016 and entitled “Reducing Financial Fraud Using Machine Learning and Other Techniques,” U.S. Patent Application No. 62/331,530, filed on May 4, 2016 and entitled “Reducing Financial Fraud Using Machine Learning and Other Techniques,” and U.S. Patent Application No. 62/365,699, filed on Jul. 22, 2016 and entitled “Detecting and/or Preventing Financial Fraud Using Geolocation Data,” the disclosures of which are hereby incorporated herein by reference in their entireties.
  • FIELD OF THE DISCLOSURE
  • The present disclosure generally relates to financial fraud and, more specifically, to processing techniques for reducing false positive fraud alerts.
  • BACKGROUND
  • Financial fraud, in its many forms, is a problem of enormous magnitude and scope, causing billions of dollars in economic losses and impacting many millions of people. Types of financial fraud include use of a lost or stolen card, account takeover, skimming, chargeback (“friendly”) fraud, counterfeiting, forgeries and application (e.g., loan application) fraud, to name just a few. The problem only continues to grow as various technological advances, intended to improve convenience and efficiency in the marketplace, provide new opportunities for bad actors. For example, an ever-increasing amount of fraud may be linked to online transactions made via the Internet.
  • Various software applications have been developed to detect potentially fraudulent transactions. For example, dollar amounts and geographic locations have generally been used to flag particular credit or debit card transactions, with cardholders then being contacted by employees of the card issuer to determine whether the transactions were indeed fraudulent. To ensure that most instances of fraud are captured, however, such techniques generally have a low threshold for triggering a fraud alert. As a result, numerous fraud alerts are false positives. The prevalence of false positives leads to a large cost in terms of the drain on human resources (e.g., calling customers to discuss each suspect transaction, and/or other manual investigation techniques), and considerable distraction or annoyance for cardholders. To provide a solution to these shortcomings in the field of automated fraud detection, innovative processing techniques capable of reducing false positives are needed.
  • BRIEF SUMMARY
  • The present embodiments may, inter alia, use new processing techniques to reduce false positive fraud alerts. For example, fraud alerts may be generated, or fraud alerts based upon various other triggers (e.g., presence of a large transaction, presence of a transaction initiated in a different state or country, cardholder reporting of unrecognized or fraudulent charges, etc.) may be either confirmed or ruled out (e.g., identified as a false positive), using location information.
  • In one embodiment, a method of reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions is implemented in one or more servers. The method may include: (1) determining, by one or more processors of the one or more servers, that a fraud alert exists for a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account; (2) determining, by the one or more processors, a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determining, by the one or more processors, a time of the financial transaction; (4) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction; (5) determining, by the one or more processors, that the second geographic location corresponds to the first geographic location; and/or (6) in response to determining that the second geographic location corresponds to the first geographic location, marking, by the one or more processors, the fraud alert as a false positive such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction. The method may include additional, less, or alternate actions, including those discussed elsewhere herein.
  • In another embodiment, a computer system for reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions includes a location database configured to store geolocation data indicating geographic locations of authorized cardholders over time, one or more processors, and a non-transitory memory. The memory stores instructions that, when executed by the one or more processors, cause the one or more processors to: (1) determine that a fraud alert exists for a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account; (2) determine a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determine a time of the financial transaction; (4) determine, based upon first geolocation data stored in the location database and indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction; (5) determine that the second geographic location corresponds to the first geographic location; and/or (6) in response to determining that the second geographic location corresponds to the first geographic location, mark the fraud alert as a false positive such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction. The computer system may include additional, less, or alternate functionality, including that discussed elsewhere herein.
  • In another embodiment, a computer-implemented method of preventing fraudulent card-present financial transactions is implemented in one or more servers. The method may include: (1) receiving, by one or more processors of the one or more servers, a request to authorize a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly being entered into by an authorized cardholder associated with the debit or credit card account; (2) determining, by the one or more processors, a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at a time of the financial transaction; (4) determining, by the one or more processors, that the second geographic location does not correspond to the first geographic location; and/or (5) in response to determining that the second geographic location does not correspond to the first geographic location, preventing, by the one or more processors, the financial transaction from being executed. The method may include additional, less, or alternate actions, including those discussed elsewhere herein.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The Figures described below depict various aspects of the systems and methods disclosed herein. It should be understood that each Figure depicts an embodiment of a particular aspect of the disclosed systems and methods, and that each of the Figures is intended to accord with a possible embodiment thereof.
  • FIG. 1 depicts an exemplary environment in which techniques for fraud detection, verification and/or classification may be implemented, according to one embodiment.
  • FIG. 2 depicts an exemplary process flow for machine learning of fraud detection, verification and/or classification rules, according to one embodiment.
  • FIGS. 3A-3F depict exemplary process flows for machine learning of particular types of fraud detection, verification and/or classification rules, according to different embodiments.
  • FIGS. 4A-4F depict exemplary factors and algorithms that may be used in connection with various fraud detection, verification and/or classification rule sets, according to different embodiments.
  • FIGS. 5 and 6 illustrate exemplary computer-implemented methods of using customer data to determine that geolocation-based fraud alerts are false positives, according to different embodiments.
  • FIGS. 7 through 10 illustrate exemplary computer-implemented methods of using information about the locations of authorized cardholders to prevent false positive fraud alerts, or to block potentially fraudulent financial transactions, according to different embodiments.
  • FIG. 11 depicts an exemplary computer system in which the techniques described herein may be implemented, according to one embodiment.
  • DETAILED DESCRIPTION
  • I. Exemplary Fraud Detection and/or Classification
  • The embodiments described herein relate to, inter alia, wholly or partially automated detection, verification and/or classification of financial fraud. For ease of explanation, and unless otherwise clearly indicated by the context of usage, “detecting” or “determining” fraud may be used herein to refer to initially flagging fraudulent (or potentially fraudulent) activity, to verifying/confirming that suspect/flagged activity was indeed fraudulent, or generally to both. The systems and techniques described herein may be used, for example, to identify, prevent and/or quantify/measure instances of lost or stolen card use, account takeover, counterfeiting, skimming, chargeback (“friendly”) fraud, collusive merchant fraud, application (e.g., loan application) fraud, mortgage fraud, and/or one or more other types of fraud relating to existing and/or potential financial transactions and/or accounts. Moreover, those skilled in the art will appreciate that at least some of the technical advancements described below (and/or shown in the accompanying figures) are not necessarily restricted to the financial field.
  • In some embodiments, a fraud detection and/or classification system may analyze data relating to a number of existing or potential financial accounts. The analysis/processing may be performed in batch processing operations, or substantially in real-time (e.g., as the data is generated and/or as financial transactions occur, etc.), and the data may be obtained from a variety of sources based upon the particular embodiment and/or scenario. In one embodiment, for example, data from financial account records may be analyzed, along with data indicating online activity of an account holder, location data (e.g., global positioning satellite (GPS) data from a smartphone or vehicle of the account holder) and/or other data, to determine whether a particular financial transaction was fraudulent or likely fraudulent. The analysis may be performed automatically after the transaction has been made, or may be performed in response to a person or algorithm flagging the transaction as a potentially fraudulent one, for example.
  • The analysis may include determining whether the account holder has expressed interest in the object (e.g., product or service) of the transaction or the merchant, and/or determining whether the transaction is consistent with spending patterns associated with the account holder (e.g., spending patterns identified using the account holder's transaction records), for example. In the case of multiple account holders (e.g. multiple credit or debit card holders), accuracy may be improved by identifying spending patterns at the individual level rather than, or in addition to, at the aggregate account level. For example, a maximum amount of money typically spent in a single transaction (e.g., over the course of a one-month window, etc.) may be determined for each of two cardholders listed on a single account, and the maximum amount for the cardholder who purportedly made a particular purchase may be compared to the purchase amount to determine whether fraud is suspected.
  • In another exemplary embodiment, the locations of authorized cardholders may be analyzed, in conjunction with the locations at which cards were presented to a merchant or merchant device (if a card-present transaction) or the locations of computing devices via which card information was entered (if an online transaction), to determine whether a fraud alert is likely a false positive. Alternatively, such locations may be analyzed to determine whether to block a transaction that is currently in-process (e.g., by issuing a fraud alert to the merchant or card issuer, or by not clearing the transaction, etc.).
  • By replacing conventional processing techniques with one or more of the processing techniques described herein, problems that have beset the field of fraud detection, classification and/or prevention in the past may be greatly mitigated or eliminated. For example, information that has conventionally been overlooked or ignored may be used to more accurately detect, prevent and/or classify fraud, and/or to reduce false positive fraud alerts. As another example, a significant amount of time may be saved by removing the need for manual investigations, or by reducing the number of instances where manual investigations are required.
  • II. Exemplary Environment for Implementing Fraud Detection and/or Classification Processing Techniques
  • FIG. 1 depicts an exemplary environment 10 in which techniques for fraud detection and/or classification may be implemented, according to one embodiment. The environment 10 may include an anti-fraud services system (AFSS) 12, a financial account management system (FAMS) 14, a card network computing system 16, a number of cardholder computing devices 20, a number of merchant computing systems 22, a number of other sources 24, and a network 26. It is noted that, in other embodiments and/or scenarios, the environment 10 may include more, fewer and/or different components than those shown in FIG. 1, such as any of those discussed elsewhere herein. For example, the environment 10 may include one or more additional financial account management systems and/or card network computing systems, and/or one or more of the cardholder computing devices 20 may instead be a computing device of a holder of a non-card account (e.g., a checking, savings or loan account) or an applicant for a new account (e.g., a new loan account). As another example, the environment 10 may include a computing system of one or more acquiring/merchant banks, and some or all of the communications with merchant computing systems 22 described below may instead be with the acquiring bank(s).
  • FAMS 14 may be associated with (e.g., owned and/or maintained by) a bank or other financial entity. For example, FAMS 14 may be a bank that acts as a card issuer associated with a particular type of card network (e.g., VISA®, Mastercard®, etc.), and/or an entity that provides loans (e.g., mortgage, home equity, vehicle, etc.), saving/checking account services, and/or other financial services to customers. FAMS 14 may maintain an account records database 30 that stores various kinds of account information, including account holder information (e.g., names, addresses, etc.) and data indicative of financial transactions made in connection with each account (e.g., dates, amounts and merchants for credit or debit card transactions, dates and amounts for customer deposits and withdrawals, etc.). Account records database 30 may store account information for some or all of the cardholders associated with cardholder computing devices 20, for example. While shown in FIG. 1 as a single entity within FAMS 14, it is understood that account records database 30 may, in some embodiments, be distributed across multiple databases and/or multiple physical/hardware memories, and/or may be wholly or partially external to (e.g., remote from) FAMS 14.
  • AFSS 12 may generally provide services that help to detect and/or classify fraudulent activity in connection with existing and/or potential (e.g., applied for) financial accounts, such as the accounts managed by FAMS 14. In some embodiments, AFSS 12 is included within FAMS 14. As seen in FIG. 1, AFSS 12 may include a network interface 32, a memory 34, and a fraud detection/classification unit 36.
  • Network interface 32 may include hardware, firmware and/or software configured to enable AFSS 12 to wirelessly exchange electronic data with one or more other components of environment 10 via network 26. For example, network interface 32 may include an Ethernet port, a modem, a router, and/or one or more other ports and/or transceivers for one or more other wired and/or wireless communication technologies.
  • Memory 34 may be a computer-readable, non-transitory storage unit or device, or collection of units/devices, and may include persistent (e.g., hard disk) and/or non-persistent memory components. Memory 34 may store instructions that are executable on one or more processors of AFSS 12 (not shown in FIG. 1) to perform various operations, including the instructions of various software applications and data generated and/or used by such applications.
  • Card network computing system 16 may be a computing system (e.g., one or more servers) of a credit and/or debit card network entity, such as VISA® or Mastercard®, for example. In some embodiments and/or scenarios where the card network entity also acts as the issuer (e.g., American Express® or Discover®), card network computing system 16 may include FAMS 14. Card network computing system 16 may provide various services to FAMS 14 and/or AFSS 12. For example, card network computing system 16 may provide electronic updates to chargeback rules, fraud scores for particular customers and/or transactions, and so on.
  • Each of cardholder computing devices 20 may be a computing device of a respective holder of a credit or debit card account managed by FAMS 14. For example, one or more of cardholder computing devices 20 may be desktop computers, laptop computers, tablet computers, smartphones, smart watches, and so on. The cardholders (e.g., credit or debit card account holders) may use cardholder computing devices 20 to access (e.g., view, modify, etc.) their account information stored in account records database 30 online via network 26. In some embodiments where AFSS 12 detects and/or classifies activity not related to credit or debit card fraud (e.g., a fraudulent application for a home equity loan, etc.), cardholder computing devices 20 may instead be computing devices of other types of customers or potential customers, such as holders of non-card-based accounts, or individuals who have submitted an online application for a loan, etc., as discussed further below. In some of these embodiments, the environment 10 may omit card network computing system 16.
  • Each of merchant computing systems 22 may include one or more computing devices associated with a particular provider of products and/or services. For example, some or all of merchant computing systems 22 may include servers associated with online retailers. Alternatively, or additionally, some or all of merchant computing systems 22 may include point-of-sale terminal devices providing credit and/or debit card payment processing features for “card present” transactions. In some embodiments where AFSS 12 detects and/or classifies activity not related to customer purchases (e.g., if AFSS 12 only detects loan application fraud, etc.), the environment 10 may omit merchant computing systems 22.
  • The other sources 24 may include computing devices and/or systems associated with sources of one or more other types of information. For example, other sources 24 may include vehicle telematics systems (e.g., installed in vehicles of cardholders associated with cardholder computing devices 20), one or more Internet service providers (ISPs) (e.g., ISPs providing Internet access to some or all cardholders), “smart home” system devices (e.g., installed in homes of some or all cardholders), and/or other systems/devices. In some embodiments, the environment 10 does not include the other sources 24.
  • Network 26 may communicatively couple some or all of the components shown in FIG. 1. For example, FAMS 14 may use network 26 to communicate with AFSS 12, card network computing system 16, cardholder computing devices 20 and/or merchant computing systems 22. As another example, AFSS 12 may use network 26 to communicate with FAMS 14, card network computing system 16, cardholder computing devices 20, merchant computing systems 22 and/or one or more of the other sources 24. While shown as a single entity in FIG. 1, network 26 may include multiple communication networks of one or more types (e.g., one or more wired and/or wireless local area networks (LANs), and/or one or more wired and/or wireless wide area networks (WANs) such as the Internet). Moreover, network 26 may use partially or entirely distinct network components to support communications between different endpoints or computing devices, such as wireless communication or data transmission over one or more radio frequency links and/or wireless communication channels. For example, the portion(s) of network 26 used for communications between FAMS 14 and AFSS 12 may be the same as, or different than, the portion(s) of network 26 used for communications between FAMS 14 and one or more of cardholder computing devices 20 over one or more radio links or wireless communication channels, or between AFSS 12 and one or more of the other sources 24, etc. Those skilled in the art will appreciate different types of networks that are appropriate for network 26, depending upon, for example, how AFSS 12, FAMS 14 and/or other components of environment 10 are localized or distributed across a relatively large geographic area.
  • Generally, fraud detection/classification unit 36 of AFSS 12 may detect fraudulent activity, confirm whether suspected or reported fraudulent activity is truly fraudulent, and/or classify fraudulent or suspected fraudulent activity. For example, fraud detection/classification unit 36 may analyze each transaction stored in account records database 30 to determine whether that transaction is, or potentially is, fraudulent. Alternatively, fraud detection/classification unit 36 may analyze only those transactions that were flagged as possibly being fraudulent (e.g., by a cardholder calling in to report an unauthorized and/or unrecognized transaction, or by FAMS 14 or AFSS 12 generating a preliminary fraud alert after applying an initial set of rules to a transaction, etc.). Fraud detection/classification unit 36 may also, or instead, analyze location information associated with potential transactions (e.g., GPS or other data indicating cardholder location, transaction data indicating a merchant location for a card-present transaction, etc.), and issue a pre-transaction alert or otherwise prevent a transaction from being fully executed. Fraud detection/classification unit 36 may also, or instead, support additional functionality, such as that described below in connection with the various components of fraud detection/classification unit 36 shown in FIG. 1.
  • As seen in FIG. 1, fraud detection/classification unit 36 may include a machine learning (ML) rule generator 40, an external data collection unit 42, a behavior analysis unit 44, a dispute resolution unit 46, a chargeback analysis unit 50, an image analysis unit 52, a classification unit 54, and/or a notification unit 56. In other embodiments, fraud detection/classification unit 36 may include more, fewer and/or different components/units than those shown in FIG. 1. In some embodiments, each of ML rule generator 40, external data collection unit 42, behavior analysis unit 44, dispute resolution unit 46, chargeback analysis unit 50, image analysis unit 52, classification unit 54, notification unit 56, and/or other units or components of fraud detection/classification unit 36 may be a software component stored in memory 34 and implemented by one or more processors of one or more computing devices (e.g., servers) included in AFSS 12.
  • ML rule generator 40 may generally analyze various types of data to generate and/or update fraud detection and/or classification rules to be applied by fraud detection/classification unit 36 and stored in an ML rules database 58. As discussed in further detail below, the rules may be used to detect and/or classify a single type or category of fraudulent activity, or may be used broadly in connection with multiple types or categories of fraudulent activity. ML rule generator 40 may implement any suitable type or types of machine learning. For example, ML rule generator 40 may implement supervised learning techniques, such as decision trees, regression-based models, support vector machines (SVMs) and/or neural networks, and/or unsupervised learning techniques such as Dirichlet process mixture models and/or k-means clustering. Other machine learning techniques are also possible, such as techniques utilizing Bayesian networks, “deep learning” techniques, and so on. While shown in FIG. 1 as a single entity within AFSS 12, it is understood that ML rules database 58 may, in some embodiments, be distributed across multiple databases and/or multiple physical/hardware memories, and/or may be wholly or partially external to (e.g., remote from) AFSS 12.
  • External data collection unit 42 may generally collect, via network interface 32 and/or from sources internal to AFSS 12, information from various sources (e.g., FAMS 14, cardholder computing devices 20, other sources 24, etc.), and provide that data to other portions of AFSS 12 as needed (e.g., to ML rule generator 40 to generate and/or update rules, and/or to behavior analysis unit 44, dispute resolution unit 46, chargeback analysis unit 50, image analysis unit 52 and/or classification unit 54 to detect and/or classify fraudulent activity). Some data may be collected indirectly. For example, FAMS 14 may collect transaction data from merchant computing systems 22 (and/or from acquiring banks associated with one or more of merchant computing systems 22), and external data collection unit 42 may then collect that data from the account records database 30 of FAMS 14.
  • Once an initial set of rules has been generated and stored in ML rules database 58, those rules may dictate some or all of the types of data gathered by external data collection unit 42. In some embodiments, however, external data collection unit 42 collects a broad set of data types that may or may not be relevant to fraud determination or classification, and ML rule generator 40 continually analyzes that data to determine which data types are most predictive of fraud and/or fraud type/class.
  • Behavior analysis unit 44 may generally analyze cardholder-related (or other customer-related) information to identify patterns of behavior, which may then be used by fraud detection/classification unit 36 to detect and/or classify fraudulent activity. For example, behavior analysis unit 44 may analyze information obtained from account records database 30 to identify spending patterns associated with different cardholders. The operation of behavior analysis unit 44, including the types of information analyzed and the ways in which that information is used to arrive at a result (e.g., a pattern of behavior), may be dictated by the rules stored in ML rules database 58.
  • Data indicative of the behavior patterns identified by behavior analysis unit 44 may be stored in an account holder behaviors database 60, for example. While shown in FIG. 1 as a single entity within AFSS 12, it is understood that account holder behaviors database 60 may, in some embodiments, be distributed across multiple databases and/or multiple physical/hardware memories, and/or may be wholly or partially external to (e.g., remote from) AFSS 12. In one embodiment, for example, account holder behaviors database 60 may be included within account records database 30. In still other embodiments, the environment 10 may not include account holder behaviors database 60, and behavior patterns may be only identified by behavior analysis unit 44 “on the fly” as needed by fraud detection/classification unit 36 (e.g., when needed to analyze a transaction in view of past spending patterns of a particular cardholder, etc.).
  • In some embodiments, behavior analysis unit 44 may separately analyze the transactions associated with each account holder, even if more than one account holder exists for a particular account. For example, behavior analysis unit 44 may independently analyze the transactions of each cardholder for a credit or debit card account in which each spouse has been issued a credit or debit card in his or her name. Fraud detection/classification unit 36 may then utilize the individual spending patterns when detecting and/or classifying fraud. In one embodiment where fraud detection/classification unit 36 utilizes a dollar amount threshold to detect likely fraudulent transactions, for example, a first threshold may be used for transactions made by a first cardholder listed on an account, and a higher, second threshold may be used for transactions made by a second cardholder listed on the account. Further examples are provided below in connection with FIG. 6, according to various embodiments. In this manner, fraud detection and/or classification may be made more precise than would be the case if spending patterns were only identified at the aggregate level (e.g., using a single dollar amount threshold, regardless of which cardholder made a particular transaction).
  • Dispute resolution unit 46 may generally analyze financial transaction data and/or other information to automatically generate queries for cardholders or other customers. For example, dispute resolution unit 46 may analyze information obtained from account records database 30. The generated queries may be designed to help fraud detection/classification unit 36 determine whether a particular transaction was fraudulent, or estimate a probability that the transaction was fraudulent, etc. Dispute resolution unit 46 may also process responses from cardholders/customers, and automatically generate additional queries based upon those responses. Examples of the operation of dispute resolution unit 46 are provided below in connection with FIGS. 4E and 9, according to various embodiments.
  • Chargeback analysis unit 50 may generally analyze financial transaction and/or other information to identify transactions that are good candidates for chargeback payments. For example, chargeback analysis unit 50 may analyze information obtained from account records database 30 to determine whether there is a relatively high probability that the merchant (or an acquiring bank) should be responsible for a chargeback payment to a card issuer associated with FAMS 14. The operation of chargeback analysis unit 50, including the types of information analyzed and the ways in which that information is used to arrive at a result (e.g., flagging a transaction as a chargeback candidate), may be dictated by the rules stored in ML rules database 58. ML rule generator 40 may make use of chargeback rules obtained from a card network entity (e.g., from card network computing system 16), and stored in chargeback rules database 62, to generate and/or update the rules applied by chargeback analysis unit 50. Examples of the operation of chargeback analysis unit 50 are provided below in connection with FIGS. 4B and 7, according to various embodiments.
  • In some embodiments, transactions flagged by chargeback analysis unit 50 are subject to further, manual review using the chargeback rules stored in chargeback rules database 62. In other embodiments, chargeback analysis unit 50 (or another component of fraud detection/classification unit not shown in FIG. 1) automatically, with little or no manual input/assistance, applies the chargeback rules from chargeback rules database 62 for each flagged transaction. While shown in FIG. 1 as a single entity within AFSS 12, it is understood that chargeback rules database 62 may, in some embodiments, be distributed across multiple databases and/or multiple physical/hardware memories, and/or may be wholly or partially external to (e.g., remote from) AFSS 12.
  • Image analysis unit 52 may generally analyze image data corresponding to physical documents to identify fraudulent (e.g., counterfeit and/or forged) documents, and/or to flag potentially fraudulent documents for further (e.g., manual) review. For example, image analysis unit 52 may analyze information obtained from merchant computing systems 22 to determine whether there is a relatively high probability that documents presented to the merchants (e.g., personal checks, identification cards, etc.) are fraudulent. Image analysis unit 52 may be configured to analyze only a single type of document, or multiple types of documents. The operation of image analysis unit 52, including the image characteristics analyzed and the ways in which the characteristics may be used to arrive at a result (e.g., flagging a document as potentially fraudulent), may be dictated by the rules stored in ML rules database 58. Examples of the operation of image analysis unit 52 are provided below in connection with FIGS. 4F and 10, according to various embodiments.
  • Classification unit 54 may generally analyze broad categories of data from various sources (e.g., account records database 30, cardholder computing devices 20, merchant computing systems 22, and/or other sources 24) to categorize/classify types of suspected fraudulent financial activity. Classification unit 54 may classify fraudulent activity only within a particular subset of fraudulent financial activity (e.g., classifying debit and/or credit card transactions as involving a potential case of counterfeiting, skimming, lost/stolen card use, chargeback fraud, etc.), or may classify fraudulent financial activity across a broader spectrum (e.g., including types of identity theft not necessarily tied to a single financial transaction, such as application fraud). In some embodiments, classification unit 54 classifies suspected fraudulent activity in connection with a particular account or transaction in response to being notified of suspect activity (e.g., notified by another component of fraud detection/classification unit 36, or by a manual user input, etc.). In other embodiments, classification unit 54 itself (or another component of fraud detection/classification unit 36) identifies suspect activity before classification unit 54 classifies that activity. Examples of the operation of classification unit 54 are provided below in connection with FIGS. 4C and 11, according to various embodiments.
  • Notification unit 56 may generally provide alerts, confirmations, and/or other notifications to various individuals (e.g., customers, bank employees associated with FAMS 14, third party employees associated with AFSS 12, etc.). For example, notification unit 56 may generate a notification message stating that a fraud alert associated with a particular transaction is a false positive, and cause network interface 32 to send the message to a computer terminal or to FAMS 14 for display to a system user. As another example, notification unit 56 may cause network interface 32 to send other flagged transactions and/or documents (e.g., chargeback candidates identified by chargeback analysis unit 50, documents that image analysis unit 52 has identified as potentially fraudulent, etc.) to a computer terminal or FAMS 14 for display to a system user. As still another example, notification unit 56 may cause network interface 32 to send FAMS 14 and/or one of merchant computing systems 22 an alert indicating that a transaction that is in-process should be terminated due to suspected fraud. As yet another example, notification unit 56 may cause network interface 32 to send queries generated by dispute resolution unit 46 to various ones of cardholder computing devices 20 for display to cardholders.
  • The operation of various components of the environment 10 shown in FIG. 1, according to different embodiments and/or scenarios, will be described further below in connection with the remaining figures.
  • III. Exemplary Process Flows for Machine Learning of Fraud Detection and/or Classification Rules
  • As discussed above, ML rule generator 40 may generate and/or update rules that are used for one or more of a variety of different purposes relating to fraud detection and/or classification. FIG. 2 depicts one generalized, example process flow 80 for machine learning that may be implemented by ML rule generator 40, and possibly one or more other components of fraud detection/classification unit 36.
  • In the process flow 80, multi-account data 82 may represent data associated with multiple financial accounts, each with one or more account holders. The financial accounts may be existing or potential accounts, and the account holders may include holders of accounts and/or potential holders of potential accounts. For example, the multi-account data 82 may include existing and/or applied-for credit card accounts, debit card accounts, savings accounts, checking accounts, investment accounts, loan accounts, etc.
  • Depending upon the embodiment, the multi-account data 82 may include one or more different types of information obtained (e.g., by external data collection unit 42 of FIG. 1) from one or more of FAMS 14, cardholder computing devices 20, merchant computing systems 22, and/or other sources 24. For example, the multi-account data 82 may include transaction data (e.g., transaction dates, amounts, locations, etc.) from account records database 30 of FAMS 14, data indicative of Internet Protocol (IP) addresses of cardholder computing devices 20 and/or devices in merchant computing systems 22, Internet browsing and/or search history data from cardholder computing devices 20 (or from an ISP computer system included in other sources 24, etc.), vehicle telematics data from telematics systems of cardholder vehicles, home occupancy and/or usage data (e.g., smart appliance data) from smart home systems of cardholders, autonomous or smart vehicle data, vehicle navigation system data, mobile device data, mobile device and/or vehicle GPS data, and/or one or more other types of data. In some embodiments, the multi-account data 82 only includes data that account holders or potential account holders have expressly consented to share with an entity associated with FAMS 14 and/or AFSS 12 (e.g., in exchange for fraud protection services). In certain other embodiments, however, express consent is only needed for certain types of information, such as browsing history information, vehicle telematics data, etc.
  • The multi-account data 82 may be associated with multiple fraud determination labels. The labels may simply reflect whether or not fraud existed (e.g., “fraud” or “no fraud”), or may also indicate a type or class of fraud (e.g., “counterfeiting,” “lost or stolen card use,” etc.), for example. In one embodiment, each of a number of data sets in the multi-account data 82 is associated with such a label, and includes data relating to a particular financial transaction, financial account, loan application, etc., for which the fraud determination was made (e.g., after a manual and/or automated fraud investigation). The labels may include final fraud determinations that were made via earlier iterations of the process flow 80, and/or external to the process flow 80.
  • To provide a more detailed example, a first data set associated with a “card present” credit card transaction may include data describing that transaction (e.g., from account records database 30) and data indicative of the cardholder's online browsing activity (e.g., from one of cardholder computing devices 20) for the 15 days immediately preceding the transaction, and be labeled “confirmed fraud.” A second data set, associated with another “card present” transaction (for the same account, or for a different account), may include the same general types of data but be labeled “no fraud,” and so on. In some embodiments and/or scenarios, the same data may appear in, or be used by, two or more of the data sets. If the two “card present” transactions described above are both associated with the same account, for example, and if the second transaction occurred less than 15 days after the first transaction, some of the same online activity data may be shared by the first and second data sets.
  • At a process stage 84, the multi-account data 82 may be analyzed to generate fraud detection and/or classification rules (e.g., to be stored in ML rules database 58). Any suitable type of supervised machine learning program/technique(s) may be used, such as SVMs, neural networks, logistic regression, etc. Generally, process stage 84 may serve to identify which type(s) of data is/are probative of whether fraud has occurred (and/or the type/category of fraud that may have occurred), and to determine the data values and/or combinations that are probative of whether fraud has occurred (and/or the type/category of fraud that may have occurred). By analyzing many (e.g., thousands) of positively and negatively labeled data sets in the multi-account data 82, for example, process stage 84 may learn that certain spending patterns within a threshold time of a transaction tend to indicate that the cardholder made the transaction (e.g., thereby indicating that fraud has not occurred, or that a fraud report is itself fraudulent or mistaken, etc.), that certain types of online searches by a cardholder (e.g., including a descriptor of a product purchased in the transaction, or a name of the merchant, etc.) tend to indicate that the cardholder made the transaction, that the cardholder's distance from the site of a “card present” transaction (e.g., as determined from GPS information provided by the cardholder's smartphone, wearable electronics, or vehicle) relates to the probability of fraudulent activity according to a particular equation, and so on. Other specific examples of such rules, and how those rules may be generated, are discussed below in connection with FIGS. 3A-3F and 4A-4F, according to various embodiments.
  • At process stage 86, the rules generated or updated at process stage 84 may be applied to first account data 90 associated with a particular account and customer(s) (e.g., a customer associated with a particular one of computing devices 20). The types of data included in first account data 90 may depend upon which types of data were determined, by process stage 84, to be relevant to a fraud determination. For example, if the rules give weight to the amount and date of a financial transaction when determining whether the transaction is fraudulent, and also give weight to whether the account holder visits a particular type of website, then the first account data 90 may include the amount and date of one or more transactions, as well as data indicative of visited web sites (e.g., Uniform Resource Locators (URLs) and/or content of visited websites, etc.). The first account data 90 may include information obtained (e.g., by external data collection unit 42) from one or more of FAMS 14, one of cardholder computing devices 20 associated with the customer holding the first account, one or more of merchant computing systems 22, and/or one or more of other sources 24, for example.
  • Process stage 86 may output various different types of information, depending upon the embodiment and/or scenario. For example, depending upon the content of first account data 90 and the rules generated or updated at process stage 84, process stage 86 may generate data indicating that a particular financial transaction associated with first account data 90 is, or is not, fraudulent or potentially fraudulent. Alternatively, or additionally, process stage 86 may generate data indicating a particular classification for fraudulent or suspected fraudulent activity (e.g., a fraudulent transaction) associated with first account data 90.
  • In some embodiments, further analysis (e.g., a manual review, or further automated review using additional data sources, etc.) may be performed at an additional stage, shown in dashed lines in FIG. 2 as process stage 92. The additional analysis may then be used to make a final fraud determination (e.g., a final decision on whether fraud occurred, and/or on the type of fraud that occurred) at process stage 94. In other embodiments, process stage 92 is omitted from process flow 80, and process stage 94 merely represents the output of process stage 86. The final determination made at process stage 94, along with the first account data 90 used to make that determination, may be fed back into process stage 84 to provide additional labeled data for purposes of updating the rules.
  • In some embodiments, the process flow 80 includes more, fewer and/or different stages, such as any of those discussed elsewhere herein (e.g., in connection with FIGS. 3A-3F). In one alternative embodiment, process stages 84 and 86 may be combined. For example, the multi-account data 82 may be unlabeled rather than labeled (or the labels may be ignored), and the combined process stage 84, 86 may use unsupervised learning techniques (e.g., clustering techniques) to classify anomalous/outlier financial transactions, accounts, applications, etc., as “suspect” and needing further analysis.
  • More specific, machine learning-based process flows generally corresponding to process flow 80 of FIG. 2 will now be described with reference to FIGS. 3A-3F. It is noted, however, that other process flows are also within the scope of the invention described herein. Moreover, while FIGS. 3A-3F generally correspond to embodiments in which supervised machine learning techniques are used, other embodiments may instead use unsupervised machine learning techniques, as noted above. In various different embodiments, fraud detection/classification unit 36 may be configured to implement only one of the process flows of FIGS. 3A-3F, or may be configured to implement two or more (e.g., all) of the process flows shown in FIGS. 3A-3F.
  • A. Exemplary Process Flow for Machine Learning of Fraud Detection Rules Using Online Activity Data
  • Referring first to FIG. 3A, an exemplary process flow 100 may generally be used to detect fraud using customer online activity data. In the process flow 100, multi-customer online activity data 102 may represent data associated with the online activities of a number (e.g., thousands) of customers (e.g., credit or debit cardholders, checking or saving account holders, etc.). The multi-customer online activity data 102 may include data indicating actions that the customers took, and/or web sites visited by the customers, while the customers were connected to the Internet via web browsers (e.g., executing on respective ones of cardholder computing devices 20). For example, the multi-customer online activity data 102 may include URLs of, and/or content (e.g., text) within, web sites visited by customers, search terms entered by customers using search engine tools, search results presented to customers by search engine tools, indications of interactive controls (e.g., virtual buttons) selected by customers on various web pages, and so on.
  • The multi-customer online activity data 102 may include data obtained (e.g., by external data collection unit 42 of FIG. 1) from cardholder computing devices 20, from one or more ISPs of other sources 24, and/or from a third party aggregator of such information, for example. In some embodiments, the multi-customer online activity data 102 may only include data that customers have expressly consented to share with an entity associated with FAMS 14 and/or AFSS 12 (e.g., in exchange for fraud protection services or other benefits, such as discounts).
  • As described above in connection with multi-account data 82 of process flow 80, the multi-customer online account data 102 may be associated with multiple fraud determination labels. In some embodiments, each label may be associated with a data set that includes not only the corresponding portion of multi-customer online activity data 102, but also one or more other types of data, such as transaction data (e.g., transaction dates, amounts, locations, etc.) for each customer from account records database 30 of FAMS 14, data indicative of IP addresses of cardholder computing devices 20 and/or devices in merchant computing systems 22, Internet browsing and/or search history data from cardholder computing devices 20 (or from an ISP computer system included in other sources 24, etc.), vehicle telematics data from telematics systems of other sources 24, home occupancy and/or usage data (e.g., smart appliance data) from smart home systems of other sources 24, and so on. The labels may include final fraud determinations that were made via earlier iterations of the process flow 100, and/or external to the process flow 100. Multi-customer online account data 102 may include many (e.g., thousands) of positively and negatively labeled data sets.
  • At a process stage 104, the multi-customer online activity data 102 may be analyzed to generate fraud detection rules (e.g., to be stored in ML rules database 58). As described above in connection with process stage 84 of process flow 80, any suitable type of supervised machine learning program/technique(s) may be used. Generally, process stage 104 may serve to identify which type(s) of online activity data is/are probative of whether fraud has occurred, and to determine the data values and/or combinations that are probative of whether fraud has occurred. While not shown in FIG. 3A, the fraud detection rules may not only detect fraud, but also classify fraud (e.g., as described below in connection with FIG. 3C), in some embodiments.
  • At process stage 106, the rules generated or updated at process stage 104 may be applied to first customer online activity data 110. The first customer online activity data 110 may be associated with a particular customer, such as a customer associated with a particular one of computing devices 20, for example. The types of data included in first customer online activity data 110 may depend upon which types of online activity data were determined, by process stage 104, to be relevant to a fraud determination. For example, the first customer online activity data 110 may include information obtained (e.g., by external data collection unit 42) from one of cardholder computing devices 20 (i.e., the device associated with the first customer), and/or from an ISP of other sources 24. Some specific examples of rules that may be generated by process stage 104, and applied at process stage 106, are described below in connection with FIG. 4A.
  • Process stage 106 may output various different types of information, depending upon the embodiment and/or scenario. For example, depending upon the content of first customer online activity data 110 and the rules, process stage 106 may generate data indicating that a particular financial transaction associated with the first customer is, or is not, fraudulent or potentially fraudulent. Alternatively, or additionally, process stage 106 may generate data indicating a particular classification of fraudulent or potentially fraudulent activity associated with first customer online activity data 110.
  • In some embodiments, further analysis (e.g., a manual review, or further automated review using additional data sources, etc.) is performed at an additional stage, shown in dashed lines in FIG. 3A as process stage 112. The additional analysis may then be used to make a final fraud determination (e.g., a final decision on whether fraud occurred, and/or on the type of fraud that occurred) at process stage 114. In other embodiments, process stage 112 is omitted from process flow 100, and process stage 114 merely represents the output of process stage 106.
  • The final determination made at process stage 114, along with the first customer online activity data 110 (and any other data) used to make that determination, may be fed back into process stage 104 to provide additional labeled data for purposes of updating the rules. In some embodiments, a preliminary fraud determination made at process stage 106 is also fed back into process stage 104, to allow the machine learning program to determine and improve upon past performance/accuracy.
  • B. Exemplary Process Flow for Machine Learning of Chargeback Candidate Detection Rules
  • Referring next to FIG. 3B, an exemplary process flow 120 may generally be used to identify the financial transactions for which chargebacks (e.g., post-transaction payments from merchants, or acquiring/merchant banks, back to the issuer to return proceeds from transactions) are appropriate. In the process flow 120, multi-account transaction data 122 may represent data associated with the financial transactions involving the accounts of a number (e.g., thousands) of credit or debit cardholders. The multi-account transaction data 122 may include information such as transaction dates, transaction amounts, merchant names (and/or aliases) associated with the transaction, information relating to how the card information was collected by the merchant (e.g., by swiping, an EMV chip reader, manual entry of the card number, etc.), geographic locations of “card present” transactions, and so on. The multi-account transaction data 122 may include data obtained (e.g., by external data collection unit 42 of FIG. 1) from merchant computing systems 22 and/or from acquiring/merchant banks associated with those merchants, for example.
  • Similar to the labels described above in connection with multi-account data 82 of process flow 80, the multi-account transaction data 122 may be associated with multiple chargeback outcome labels. For example, each label may be associated with a data set that includes the corresponding portion of multi-account transaction data 122. The outcome labels may include final chargeback determinations that were made (in connection with the transactions represented in multi-account transaction data 122) via earlier iterations of the process flow 120, and/or external to the process flow 120. Multi-account transaction data 122 may include many (e.g., thousands) of positively and negatively labeled data sets.
  • At a process stage 124, the multi-account transaction data 122 may be analyzed to generate chargeback candidate detection rules (e.g., to be stored in ML rules database 58). As described above in connection with process stage 84 of process flow 80, any suitable type of supervised machine learning program/technique(s) may be used. Generally, process stage 124 may serve to identify which type(s) of transaction data is/are probative of whether, under the full chargeback rules of the card network entity, a chargeback is appropriate for a given transaction. Process stage 124 may also determine the transaction data values and/or combinations that are probative of whether a chargeback is appropriate for the transaction.
  • At a process stage 126, the rules generated or updated at process stage 124 may be applied to first account transaction data 130 to determine whether a transaction associated with the first account is a “good” chargeback candidate. Put differently, process stage 126 may, instead of applying the full chargeback rules of the card network entity (which may be quite lengthy and complex) to the facts surrounding the transaction, use various factors and algorithms developed at process stage 124 to determine whether there exists a relatively high probability that a chargeback would be appropriate for the transaction if the full chargeback rules were applied. The process stage 126 may calculate a percentage probability that the transaction is one in which a chargeback is appropriate, for example.
  • The first account transaction data 130 may be associated with the account of a particular cardholder or cardholders, such as a cardholder associated with a particular one of cardholder computing devices 20, for example. The types of data included in first account transaction data 130 may depend upon which types of transaction-related data were determined, by process stage 124, to be relevant to a chargeback candidate determination. For example, the first account transaction data 130 may include information obtained (e.g., by external data collection unit 42) from one of merchant computing systems 22 (e.g., the computing system of the merchant involved in the transaction being analyzed) and/or from an acquiring/merchant bank associated with that merchant. The first account transaction data 130 may also include information about one or more other transactions associated with the first account (e.g., data pertaining to other transactions occurring shortly before and/or after the transaction at issue). Some specific examples of rules that may be generated by process stage 124, and applied at process stage 126, are described below in connection with FIG. 4B.
  • Process stage 126 may output information indicating whether the particular transaction represented by first account transaction data 130 is a “good” candidate for chargeback detection. For example, process stage 126 may output a percentage probability, calculated according to the rules generated or updated at process stage 124, that the transaction is one in which a chargeback is appropriate. As another example, process stage 126 may output a binary indicator of whether the transaction is, or is not, a strong/likely chargeback candidate (e.g., by comparing the percentage probability to a threshold probability).
  • If the transaction is identified as a chargeback candidate at process stage 126, the full chargeback rules of the card network entity may be applied at a process stage 132. Process stage 132 may include manual application of the full chargeback rules, and/or automated application of the full chargeback rules, in various different embodiments. Based upon the analysis at process stage 132, a final chargeback determination may be made at a process stage 134. The final determination made at process stage 134, along with the first account transaction data 130 (and any other data) used to make that determination, may be fed back into process stage 124 to provide additional labeled data for purposes of updating the rules. In some embodiments, the indication of whether the transaction is a good chargeback candidate generated at process stage 126 may also be fed back into process stage 124, to allow the machine learning program to determine and improve upon past performance/accuracy.
  • C. Exemplary Process Flow for Machine Learning of Fraud Classification Rules
  • Referring now to FIG. 3C, an exemplary process flow 140 may generally be used to classify instances of suspected or potential fraud. For example, the process flow 140 may represent ongoing, real-time or batch processing of a large amount of data associated with a large number of potential and/or existing financial accounts (e.g., all accounts associated with a particular bank, or all accounts opting in to a fraud protection program, etc.). In this manner, the process flow 140 may be used to initially flag situations for closer investigation, and provide one or more classifications of the type(s) of fraud potentially at issue in order to narrow or otherwise facilitate the investigation. In other embodiments, the process flow 140 may be used to provide a narrower classification (e.g., “skimming”) when a broader class of fraud (e.g., credit card fraud) is already suspected.
  • In the process flow 140, multi-account data 142 may represent data associated with financial accounts of a number (e.g., thousands) of account holders. The financial accounts may be existing or potential accounts, and the account holders may include holders of accounts and/or potential holders of potential accounts. For example, the multi-account data 142 may include existing and/or applied-for credit card accounts, debit card accounts, savings accounts, checking accounts, investment accounts, loan accounts, etc.
  • Depending upon the embodiment, the multi-account data 142 may include one or more different types of information obtained (e.g., by external data collection unit 42 of FIG. 1) from one or more of FAMS 14, cardholder computing devices 20, merchant computing systems 22, and/or other sources 24. For example, the multi-account data 142 may include transaction data (e.g., transaction dates, amounts, locations, etc.) from account records database 30 of FAMS 14, data indicative of IP addresses of cardholder computing devices 20 and/or devices in merchant computing systems 22, Internet browsing and/or search history data from cardholder computing devices 20 (or from an ISP computer system included in other sources 24, etc.), vehicle telematics data from telematics systems of cardholder vehicles, home occupancy and/or usage data (e.g., smart appliance data) from smart home systems of cardholders, and/or one or more other types of data. Some or all data within multi-account data 142 may be information that account holders or potential account holders have expressly consented to share with an entity associated with FAMS 14 and/or AFSS 12 (e.g., in exchange for fraud protection services).
  • The multi-account data 142 may be associated with multiple fraud determination labels, each indicating a type or class of fraud (e.g., “counterfeiting,” “lost or stolen card use,” “skimming,” “chargeback fraud,” “application fraud,” etc.), or indicating a lack of fraud, for example. In one embodiment, each of a number of data sets in the multi-account data 142 is associated with at least one such classification/label, and includes data relating to a particular financial transaction, financial account, loan application, etc., for which the fraud classification or classifications was/were made (e.g., after a previous iteration of process flow 140, or after another manual and/or automated fraud investigation). Multi-account data 142 may include many (e.g., thousands) of data sets labeled with various known fraud classifications.
  • At a process stage 144, the multi-account data 142 may be analyzed to generate fraud classification rules (e.g., to be stored in ML rules database 58). As described above in connection with process stage 84 of process flow 80, any suitable type of supervised machine learning program/technique(s) may be used. Generally, process stage 144 may serve to identify which type(s) of transaction data is/are probative of the particular type of fraud (if any) that has occurred. Process stage 144 may also determine the data values and/or combinations that are probative of the particular type of fraud (if any) that has occurred.
  • At a process stage 146, the rules generated or updated at process stage 144 may be applied to first account data 150. The first account data 150 may be associated with a particular account and a particular customer (e.g., a cardholder associated with a particular one of computing devices 20). The types of data included in first account data 150 may depend upon which types of data were determined, by process stage 144, to be relevant to fraud classification. For example, the first account data 150 may include information obtained (e.g., by external data collection unit 42) from one or more of FAMS 14, one of cardholder computing devices 20 (i.e., the device associated with the customer holding or applying for the first account), one or more of merchant computing systems 22, and/or one or more of other sources 24. Some specific examples of rules that may be generated by process stage 144, and applied at process stage 146, are described below in connection with FIG. 4C.
  • Process stage 146 may output data (e.g., a message or code) that is used to classify suspected fraudulent activity (in connection with the account associated with first account data 150) at a process stage 152. For example, process stage 152 may assign a classification of “counterfeiting” if process stage 146 determined that the first account data 150 indicated a number of circumstances that, according to the rules generated at process stage 144, are known to be correlated with counterfeiting activity (e.g., two “card present” transactions occurring in different states within the same one-hour time period, etc.). In some embodiments and/or scenarios, two or more classifications may concurrently be assigned to first account data 150. For example, process stage 146 may determine a set of probabilities for a set of two or more potential types of fraud, and process stage 152 may assign each classification, with each respective probability, to first account data 150. Moreover, in some embodiments and scenarios, process stage 152 may assign a classification that corresponds to an absence of any suspected fraud (e.g., “no fraud”).
  • At a process stage 154, if process stage 152 assigned a classification other than one indicating the absence of suspected fraud, the first account data 150, and/or other information associated with the account and the suspected class of fraud, may be analyzed in depth to make a final fraud determination at a process stage 156. Generally, the fraud classification may be used to facilitate the analysis at process stage 154, with process stage 154 including manual and/or automated fraud detection techniques. For example, personnel associated with AFSS 12 may use the fraud classification(s) to inform their strategy and/or focus with respect to conducting an in-depth fraud investigation.
  • The additional analysis at process stage 154 may then result in a final fraud determination at process stage 156. The final determination may indicate both whether fraud occurred and, if so, the class(es)/type(s) of fraud that occurred. The final determination made at process stage 156, and information used to make that determination (e.g., the first account data 150 and potentially other data), may be fed back into process stage 144 to provide additional labeled data for purposes of updating the rules. In some embodiments, the (preliminary) fraud classification made at process stage 152 may also be fed back into process stage 144 to help the machine learning program identify instances in which the preliminary classifications at process stage 152 were incorrect. Process stage 144 may then update the fraud classification rules in ways that seek to prevent or reduce such instances in the future.
  • D. Exemplary Process Flow for Machine Learning of Application Fraud Detection Rules
  • Referring now to FIG. 3D, an exemplary process flow 160 may generally be used to detect application fraud. “Application fraud” may generally refer to fraud in connection with the application for any type of financial account, loan and/or line of credit (e.g., mortgage loan, vehicle loan, small business loan, payday loan, home equity line of credit, credit card account, debit card account, checking account, savings account, investment account, etc.). In some embodiments and/or scenarios, however, the application may be for non-financial purposes, such as an application for membership in a particular group or institution, for example.
  • In the process flow 160, multi-applicant search history data 162 may represent data associated with the Internet search history of a number (e.g., thousands) of applicants. The multi-applicant search history data 162 may include search terms entered by the applicants using online search engine tools, for example, and/or the results of such searches (e.g., URLs, titles and/or contents of search results), for example.
  • The multi-applicant search history data 162 may include data obtained (e.g., by external data collection unit 42 of FIG. 1) from cardholder computing devices 20, from one or more ISPs of other sources 24, and/or from a third party aggregator of such information, for example. In some embodiments, the multi-applicant search history data 162 only includes data that the applicants have expressly consented to share with an entity associated with FAMS 14 and/or AFSS 12 (e.g., in exchange for consideration of their applications).
  • As described above in connection with multi-account data 82 of process flow 80, the multi-applicant search history data 162 may be associated with multiple fraud determination labels. In some embodiments, each label may be associated with a data set that corresponds to an application submitted by a particular applicant, where the data set includes the corresponding portion of multi-applicant search history data 162 (e.g., the search terms and/or results associated with the particular application). The labels may include final fraud determinations that were made via earlier iterations of the process flow 160, and/or external to the process flow 160. Multi-applicant search history data 162 may include many (e.g., thousands) of positively and negatively labeled data sets.
  • At a process stage 164, the multi-applicant search history data 162 may be analyzed to generate application fraud detection rules (e.g., to be stored in ML rules database 58). As described above in connection with process stage 84 of process flow 80, any suitable type of supervised machine learning program/technique(s) may be used. Generally, process stage 164 may serve to identify which type(s) of Internet search-related data is/are probative of whether application fraud has occurred, and to determine the data values and/or combinations that are probative of whether application fraud has occurred.
  • At process stage 166, the rules generated or updated at process stage 164 may be applied to first applicant search history data 170. The first applicant search history data 170 may be associated with a particular application and a particular applicant (e.g., a person associated with a particular one of computing devices 20), for example. The types of data included in first applicant search history data 170 may depend upon which types of Internet search-related data were determined, by process stage 164, to be relevant to a fraud determination. The first applicant search history data 170 may include information obtained (e.g., by external data collection unit 42) from one of computing devices 20 (i.e., the device associated with the first applicant), and/or from an ISP of other sources 24, for example. Some specific examples of rules that may be generated by process stage 164, and applied at process stage 166, are described below in connection with FIG. 4D.
  • Process stage 166 may output information indicating whether fraud is suspected in connection with the application corresponding to first applicant search history data 170. For example, process stage 166 may output a percentage probability, calculated according to the rules generated or updated at process stage 164, that the application was fraudulently made (e.g., by someone other than the purported applicant or an authorized representative thereof). As another example, process stage 166 may output a binary indicator of whether the application likely was, or likely was not, fraudulently made (e.g., by comparing a percentage probability to a threshold probability).
  • In some embodiments, further analysis (e.g., a manual review, or further automated review using additional data sources, etc.) is performed at an additional stage, shown in dashed lines in FIG. 3D as process stage 172. The additional analysis may then be used to make a final fraud determination (e.g., a final decision on whether application fraud occurred) at process stage 174. In other embodiments, process stage 172 is omitted from process flow 160, and process stage 174 merely represents the output of process stage 166. The final determination made at process stage 174, along with the first applicant search history data 170 (and any other data) used to make that determination, may be fed back into process stage 164 to provide additional labeled data for purposes of updating the rules. In some embodiments, a preliminary fraud determination made at process stage 166 is also fed back into process stage 164, to allow the machine learning program to determine and improve upon past performance/accuracy.
  • E. Exemplary Process Flow for Machine Learning of Fraud Dispute Resolution Rules
  • Referring now to FIG. 3E, an exemplary process flow 180 may generally be used to facilitate the resolution of fraud disputes (or potential disputes) with customers/account holders. For example, the process flow 180 may be used to determine whether a reportedly unauthorized or fraudulent transaction (e.g., one that the account holder reported as such when looking at his or her account statement) was indeed unauthorized or fraudulent. In some embodiments, the process flow 180 may also, or instead, be used to determine whether an “unrecognized” transaction (i.e., one that the account holder does not recall, but does not necessarily report as fraudulent) was unauthorized or fraudulent.
  • In the process flow 180, multi-account data 182 may represent data associated with financial accounts of a number (e.g., thousands) of account holders. For example, the multi-account data 182 may include data associated with financial transactions relating to credit card accounts, debit card accounts, savings accounts, checking accounts, etc. For ease of explanation, FIG. 3E will be described with reference to an embodiment in which the accounts are credit card accounts.
  • In one embodiment, the multi-account data 182 may include transaction data (e.g., transaction dates, amounts, locations, etc.) obtained from FAMS 14 (e.g., by external data collection unit 42 of FIG. 1). In some embodiments, however, the multi-account data 182 also includes information obtained from cardholder computing devices 20, merchant computing systems 22, and/or other sources 24. For example, the multi-account data 182 may include, in addition to transaction data from account records database 30 of FAMS 14, data indicative of IP addresses of cardholder computing devices 20 and/or devices in merchant computing systems 22, Internet browsing and/or search history data from cardholder computing devices 20 (or from an ISP computer system included in other sources 24, etc.), vehicle telematics data from telematics systems of cardholder vehicles, home occupancy and/or usage data (e.g., smart appliance data) from smart home systems of cardholders, autonomous vehicle data, smart vehicle data, mobile device data, vehicle or mobile device GPS data, and/or one or more other types of data. Some or all data within multi-account data 182 may be information that account holders or potential account holders have expressly consented to share with an entity associated with FAMS 14 and/or AFSS 12 (e.g., in exchange for fraud protection services).
  • As described above in connection with multi-account data 82 of process flow 80, the multi-account data 182 may be associated with multiple fraud determination labels (e.g., “fraud” and “no fraud,” and/or more complex labels that indicate type/class, such as “lost/stolen card use,” etc.). In some embodiments, each label may be associated with a data set that includes the corresponding portion of multi-account data 182. The labels may include final fraud determinations that were made via earlier iterations of the process flow 180, and/or external to the process flow 180. Multi-account data 182 may include many (e.g., thousands) of positively and negatively labeled data sets.
  • At a process stage 184, the multi-account data 182 may be analyzed to generate query generation rules (e.g., to be stored in ML rules database 58). As described above in connection with process stage 84 of process flow 80, any suitable type of supervised machine learning program/technique(s) may be used. Generally, process stage 184 may serve to identify which types of information are probative of whether fraud has occurred, and to craft rules that formulate queries to ascertain such information based upon account data.
  • For example, process stage 184 may determine that, for a suspect “card present” transaction, a verified, non-fraudulent “card present” transaction within 10 miles and 3 hours of the suspect transaction is probative of whether the suspect transaction was fraudulent. Based upon this finding, process stage 184 may also generate a rule specifying that a cardholder should be queried as to whether he/she can confirm making each “card present” transaction within 10 miles and 3 hours of the suspect transaction. As another example, process stage 184 may determine that a merchant using a billing alias different from its legal and/or commonly-known name (e.g., by at least some threshold level of similarity, as measured by number of similar characters, order of characters, etc.) is probative of whether the cardholder authorized a transaction associated with that billing alias. Based upon this finding, process stage 184 may generate a rule specifying that a cardholder should be queried as to whether he/she is aware of a billing alias used for a suspect transaction if that billing alias is sufficiently different from the legal/common name of the merchant.
  • At process stage 186, the rules generated or updated at process stage 184 may be applied to first account data 190. The first account data 190 may be associated with a particular cardholder, such as a cardholder associated with a particular one of cardholder computing devices 20, for example. The types of data included in first account data 190 may depend upon which types of data were determined, by process stage 184, to be relevant to developing dispute resolution queries. Process stage 186 may generate a set of one or more queries in accordance with the rules and the contents of first account data. Some specific examples of rules that may be generated by process stage 184 and applied at process stage 186, and the queries that may be generated as a result, are described below in connection with FIG. 4E.
  • At a process stage 192, the generated queries may be sent to the cardholder in one or more of various ways, such as sending the queries via SMS text message and/or email, and/or via a web browser or dedicated application executing on the one of cardholder computing devices 20 that is associated with the cardholder, for example. At a process stage 194, responses to the queries are received from the cardholder (e.g., via inputs made by the cardholder via the web browser or application, or a responsive SMS text message or email, etc.). In some embodiments, the rules generated or updated at process stage 184 specify the manner in which follow-up queries should be generated based upon the responses received at process stage 194, and process stages 192 and 194 may be repeated multiple times.
  • In some embodiments, further analysis (e.g., a manual review, or further automated review using additional data sources, etc.) that makes use of the received responses is performed at an additional stage, shown in dashed lines in FIG. 3E as process stage 196. The additional analysis may then be used to make a final fraud determination (e.g., a final decision on whether fraud occurred, and/or on the type of fraud that occurred) at process stage 198. In other embodiments, process stage 196 is omitted from process flow 180, and process stage 198 is based upon information from the cardholder. For example, the questions generated at process stage 192 may “jog” the cardholder's memory, and cause him or her to indicate that the transaction at issue was authorized. The final determination made at process stage 198, along with the first account data 110 (and any other data used at process stage 196), the queries generated at process stage 186 and/or the responses received at process stage 194, may be fed back into process stage 184 to provide additional labeled data for purposes of updating the rules.
  • F. Exemplary Process Flow for Machine Learning of Document Fraud Detection Rules
  • Referring now to FIG. 3F, an exemplary process flow 200 may generally be used to detect fraud relating to documents, such as counterfeit and/or forged documents. The process flow 200 may be used in connection with various kinds of documents, such as checks (e.g., personal checks, cashier's checks, etc.), money orders, treasury bills, identification documents (e.g., social security cards, driver's licenses, passports, birth certificates, etc.), certification documents, and so on.
  • In the process flow 200, multi-document image data 202 may represent digital images of a number (e.g., thousands) of physical documents of one or more types. The multi-document image data 202 may include images in one or more formats, such as raster formats (e.g., JPEG, TIFF, GIF, BMP, PNG, etc.) and/or vector formats (e.g., CGM, SVG, etc.), for example. The multi-document image data 202 may include data obtained (e.g., by external data collection unit 42 of FIG. 1) from merchant computing systems 22 (e.g., point-of-sale devices with cameras for document identification) and/or from FAMS 14 (e.g., images of personal checks), for example. In some embodiments, the multi-document image data 202 may only include data representing images that customers (or other individuals associated with the documents) have expressly consented to share (e.g., as a prerequisite to making a purchase, or in exchange for fraud protection services, etc.).
  • As described above in connection with multi-account data 82 of process flow 80, the multi-document image data 202 may be associated with multiple fraud determination labels. In some embodiments, each label may be associated with data representing a digital image of a particular document. The labels may include final fraud determinations (e.g., “fraud” or “no fraud,” or more complex labels such as “forgery,” “counterfeit,” “forgery—signature,” “counterfeit—angular line offset(s) outside tolerance,” etc.) that were made via earlier iterations of the process flow 200, and/or external to the process flow 200. Multi-document image data 202 may include many (e.g., thousands) of positively and negatively labeled data sets.
  • At a process stage 204, the multi-document image data 202 may be analyzed to generate document fraud detection rules (e.g., to be stored in ML rules database 58). As described above in connection with process stage 84 of process flow 80, any suitable type of supervised machine learning program/technique(s) may be used. Generally, process stage 204 may serve to identify which characteristics of a document are probative of whether the document is counterfeit, and to determine the ranges, tolerances, etc., that are probative of whether the document is counterfeit. In some embodiments, process stage 204 also, or instead, identifies which characteristics of information entered in document fields are probative of whether the document was forged (e.g., drafted or populated by someone other than the person purported to have drafted or populated the document).
  • At process stage 206, the rules generated or updated at process stage 204 may be applied to first document image data 210. The first document image data 210 may be digital image data corresponding to a particular, physical document. The first document image data 210 may include information obtained (e.g., by external data collection unit 42) from one of merchant computing systems 22 (e.g., for real-time verification of an identification or other document presented during or prior to a sale), or from FAMS 14 (e.g., for real-time or batch-processing verification of a personal check prior to clearing the check), for example. Some specific examples of rules that may be generated by process stage 204, and applied at process stage 206, are described below in connection with FIG. 4F.
  • Process stage 206 may output information indicating whether fraud is suspected in connection with the document corresponding to first document image data 210. For example, process stage 206 may output two percentage probabilities calculated according to the rules generated or updated at process stage 204, with the first indicating the likelihood that the document is counterfeit and the second indicating the likelihood that the document includes forged content. As another example, process stage 206 may output binary indicators of whether the document likely is, or likely is not, counterfeit and/or includes forged content (e.g., by comparing percentage probabilities to threshold probabilities).
  • In some embodiments, further analysis (e.g., a manual review, or further automated review using additional data sources, etc.) may be performed at a process stage 212. The additional analysis may then be used to make a final fraud determination (e.g., a final decision on whether the document is fraudulent) at process stage 214. For example, the process stage 206 may act as a filter, and flag only those documents having a relatively high probability of being fraudulent. In this manner, a considerably smaller amount of human and/or processing resources may be consumed at process stage 212.
  • The final determination made at process stage 214, along with the first document image data 210 used to make that determination, may be fed back into process stage 204 to provide additional labeled data for purposes of updating the rules. In some embodiments, a preliminary fraud determination made at process stage 206 may also be fed back into process stage 204, to allow the machine learning program to determine and improve upon past performance/accuracy.
  • IV. Exemplary Rules for Fraud Detection and/or Classification
  • FIGS. 4A-4F depict exemplary factors and algorithms that may be used in connection with various fraud detection and/or classification rules, according to different embodiments. It is noted that the rule sets corresponding to FIGS. 4A-4F are purely for purposes of illustration and are not limiting. Particularly in embodiments where machine learning is utilized, for example, the algorithms and/or factors may be far more complex, and/or less intuitive, than some or all of the examples shown in FIGS. 4A-4F.
  • A. Exemplary Fraud Detection Rule Set Using Online Activity
  • Referring first to FIG. 4A, an exemplary rule set 220 (e.g., generated at process stage 104 of FIG. 3A) may use various factors relating to online activity of a cardholder to detect fraud in connection with a particular credit or debit card transaction. The rule set 220 may correspond to a particular embodiment and scenario in which the transaction at issue is a “card present” transaction, and in which the rule set 220 seeks to determine whether the cardholder made or otherwise authorized the transaction. The rule set 220 may be incorporated into a review process that is generally applied to all transactions, a review process applied only to those transactions that were flagged by a preliminary fraud alert, or a review process applied only after a cardholder reports the transaction as unauthorized, for example.
  • The factors considered under the rule set 220 may include a number of interest-based factors 222 and a number of location-based factors 224. The interest-based factors 222 may relate to the cardholder's interest (or non-interest) in a product or service purchased via the transaction, and/or the merchant providing the product or service, while the location-based factors 224 may relate to the cardholder's location or probable location.
  • As seen in FIG. 4A, the interest-based factors 222 may include: (1) whether the cardholder searched online for the specific product or service purchased via the transaction at issue (e.g., by determining whether search terms entered by the cardholder included the name of the product or service involved in the transaction, or included a description of the product or service, etc.); (2) whether the cardholder visited a website associated with the merchant (e.g., by comparing URLs of web sites visited by the cardholder to a known URL of the merchant's website, or by searching the contents of websites visited by the cardholder for the merchant's name, etc.); (3) whether the cardholder endorsed the merchant, or the product or service provided by the merchant, via a social media account of the cardholder (e.g., by determining whether the cardholder “liked” the merchant, product or service via his or her Facebook® account, etc.); (4) whether the cardholder visited a website associated with a competitor of the merchant (e.g., by comparing URLs of web sites visited by the cardholder to known URLs of known competitors' websites, or by searching the contents of websites visited by the cardholder for the competitors' names, etc.); (5) whether the cardholder searched online for a different product or service in the same price range as the transaction amount (e.g., by analyzing search terms and/or results, and/or by analyzing URLs or contents of websites visited by the cardholder and comparing prices of products/services, etc.); and/or (6) whether the cardholder entered search terms indicative of the cardholder's need for the product or service (e.g., by determining that the cardholder entered search terms including “pipe leak” prior to the purchase of new plumbing hardware, or “computer repair” prior to the purchase of a new hard drive, etc.). In other embodiments, the interest-based factors 222 may include more, fewer and/or different factors than those shown in FIG. 4A.
  • As is also seen in FIG. 4A, the location-based factors 224 may include: (1) whether the cardholder “checked in” to a flight having a destination near the location where the transaction was initiated (e.g., by determining whether the cardholder checked in to a flight having a destination at the city in which the transaction occurred, or within a threshold number of miles of the city in which the transaction occurred, etc.); (2) whether the cardholder visited a website associated with a place near (or in) which the transaction was initiated (e.g., by comparing URLs of web sites visited by the cardholder to URLs of websites known to be associated with particular areas, and/or by searching the contents of websites visited by the cardholder for location or area names, etc.); and/or (3) whether the cardholder endorsed a place near (or in) which the transaction was initiated via a social media account of the cardholder (e.g., by determining whether the cardholder “liked” the geographic area, attraction or other place via his or her Facebook® account, etc.). In other embodiments, the location-based factors 224 may include more, fewer and/or different factors than those shown in FIG. 4A.
  • Generally, the data indicative of whether the circumstance corresponding to each of interest-based factors 222 and/or location-based factors 224 is present/true for a particular cardholder may be included in the first customer online activity data 110 described above in connection with FIG. 3A. For example, external data collection unit 42 of FIG. 1 may obtain the search terms, URLs, user online selections, etc., needed to determine whether the various factors exist, from the cardholder's computing device (e.g., one of cardholder computing devices 20) and/or from an ISP of other sources 24.
  • As is also seen in FIG. 4A, each of the interest-based factors 222 and location-based factors 224 may be associated with a particular score or weighting value. In the rule set 220 shown in FIG. 4A, a total score may be calculated based upon which factors are, or are not, present (e.g., add 94 points if it is determined that the cardholder searched for the particular lawnmower model that was purchased, add another 80 points if the transaction was a “card present” transaction in the Chicago suburb of Joliet and the cardholder checked in to a flight to Chicago just prior to the transaction, etc.).
  • In some embodiments, certain factors may instead be associated with negative scores (e.g., minus 80 if the cardholder checked in to a flight with a destination at least 200 miles from the site of the transaction and within one day of the transaction, etc.). Moreover, certain factors may be associated with metrics or algorithms that determine how heavily those factors are weighed. As indicated in FIG. 4A, for example, search terms entered by the cardholder may be used to calculate a “need score” X (e.g., where X is based upon frequency of certain search terms being used, the amount of time spent clicking through search results, the magnitude and/or urgency of a problem indicated by the search terms, etc.), with X then being used to calculate a score equal to 0.2X.
  • The rule set 220 may then output the total score (e.g., 94+80=+174), a normalized total score, an indication of whether the total score exceeded a threshold (e.g., a threshold of +100), a probability calculated based upon the total score, and/or some other indicator or measure of the existence or likelihood of fraud. In the example shown in FIG. 4A, it can be seen that larger scores generally correspond to a greater probability that the transaction was made or authorized by the cardholder. If the transaction is being automatically reviewed (e.g., to determine whether a fraud alert is appropriate, without any initial input from the cardholder), this may mean that a lower score corresponds to a higher probability of fraud. Conversely, if the cardholder had reported the transaction as being fraudulent, a higher score may correspond to a higher probability of fraud (i.e., fraud on the part of the cardholder).
  • In some embodiments, the rule set 220 may also include one or more other types of factors not necessarily based upon online activities of the cardholder (e.g., whether GPS of the cardholder's smartphone or vehicle indicates that he or she was in that area shortly before or after the transaction, etc.), and/or may omit either interest-based factors 222 or location-based factors 224.
  • B. Exemplary Chargeback Candidate Detection Rule Set
  • Referring next to FIG. 4B, an exemplary rule set 230 (e.g., generated at process stage 124 of FIG. 3B) may use various factors relating to a transaction between a cardholder and a merchant to determine whether the transaction should be flagged as a candidate for a chargeback (e.g., to determine whether the transaction should be reviewed under a full set of chargeback rules associated with the appropriate card network entity). The rule set 230 may correspond to a particular embodiment and scenario in which the transaction at issue is a “card present” transaction.
  • As seen in FIG. 4B, the factors considered under the rule set 230 may include: (1) whether an EMV chip card was not inserted in a point-of-sale EMV chip reader device of the merchant; (2) whether a non-EMV card was not swiped in a point-of-sale device of the merchant; (3) whether the card is past its expiration date; (4) whether the transaction is for the same amount and/or date as another transaction involving the same card and merchant (e.g., by analyzing other transactions involving the same account and merchant within a particular time span); and/or (2) whether the transaction is for greater than a threshold amount. For example, one of merchant computing systems 22 of FIG. 1 (or an acquiring/merchant bank) may provide transaction details that include the amounts, dates, etc., to FAMS 14 for storage in account records database 30, and external data collection unit 42 may then retrieve that information from account records database 30. Generally, the data indicative of whether the circumstance corresponding to each of the factors is present/true for a particular transaction may be included in the first account transaction data 130 described above in connection with FIG. 3B. In other embodiments, the factors considered under rule set 230 may include more, fewer and/or different factors than those shown in FIG. 4B. It is noted that, in some embodiments, one or more factors may simply relate to the desirability (e.g., from a card issuer perspective) of further reviewing whether a chargeback is appropriate, without necessarily relating to the likelihood that a chargeback is appropriate.
  • As is also seen in FIG. 4B, each of the factors may be associated with a particular score or weighting value. A total score may be calculated based upon which factors are, or are not, present (e.g., add 62 points if it is determined that the transaction has the same amount and date as another transaction occurring close in time and involving the same card and merchant). In some embodiments, certain factors may instead be associated with negative scores, and/or certain factors may be associated with metrics or algorithms that determine how heavily those factors are weighed.
  • The rule set 230 may then output the total score, a normalized total score, an indication of whether the total score exceeded a threshold, a probability calculated based upon the total score, and/or some other indicator or measure of the likelihood that a chargeback is appropriate for the transaction. In the example shown in FIG. 4B, it can be seen that larger scores generally correspond to a greater probability that a chargeback is appropriate.
  • C. Exemplary Fraud Classification Rule Set
  • Referring now to FIG. 4C, an exemplary rule set 240 (e.g., generated at process stage 144 of FIG. 3C) may use a diverse array of factors to classify the type(s) of fraudulent activity, if any, that is/are suspected to be associated with an event or series of events. The rule set 240 may correspond to a particular embodiment and scenario in which the event at issue is a financial transaction involving a debit or credit card. In other embodiments and/or scenarios, however, the rule set 240 may classify fraudulent activity with respect to specific other types of events (e.g., loan applications), or may detect a variety of different event types (e.g., various types of financial transactions, loan or credit applications, etc.) and broadly classify fraudulent activity in connection with the detected event types (e.g., lost/stolen card use, application fraud, etc.).
  • In one embodiment, each potential classification (with the possible exception of “no fraud”) may be associated with a number of factors probative of whether that type/class of fraud has occurred. As seen in FIG. 4C, for example, the rule set 240 may include counterfeit factors 242 (e.g., factors indicating that a counterfeit card was used for the transaction), account takeover factors 244 (e.g., factors indicating that the transaction resulted from an unauthorized person gaining online access to the credit or debit card account itself, via phishing, malware or other means), chargeback fraud factors 246 (e.g., factors indicating that the cardholder made or otherwise authorized a purchase that the cardholder later contested) and skimming factors 248 (e.g., factors indicating that the card information used for the transaction was obtained via a skimming card reader device illegally installed in an ATM, gas station pump or other location). In other embodiments, the rule set 240 may also, or instead, include factors corresponding to one or more other fraud classifications (e.g., forgery, lost/stolen card use, etc.).
  • As seen in FIG. 4C, the counterfeit factors 242 may include: (1) whether the suspect transaction and another, contemporaneous transaction (e.g., occurring within one hour, etc.) in another state are both “card present” transactions; and/or (2) if the suspect transaction is a “card present” transaction, whether the card (if an EMV chip card) was not inserted in an EMV chip card reader. For example, one or more of merchant computing systems 22 of FIG. 1 (or one or more acquiring/merchant banks) may provide transaction details that include whether the transaction was “card present,” whether the card was inserted in an EMV chip card reader, etc., to FAMS 14 for storage in account records database 30, and external data collection unit 42 may then retrieve that information from account records database 30. In other embodiments, the counterfeit factors 242 may include more, fewer and/or different factors than those shown in FIG. 4C.
  • The account takeover factors 244 may include: (1) whether the debit or credit card account password was changed within the 10 days prior to the transaction; and/or (2) whether the transaction was originated from an IP address not associated with the cardholder. For example, external data collection unit 42 may retrieve password change information from account records database 30 of FIG. 1, which may log all password update activity, and/or may retrieve IP address information from one of merchant computing systems 22 (e.g., the computing system of the merchant involved in the transaction). In other embodiments, the account takeover factors 244 may include more, fewer and/or different factors than those shown in FIG. 4C.
  • The chargeback fraud factors 246 may include: (1) whether the cardholder had searched online for the product or service purchased via the transaction; and/or (2) whether the cardholder had visited a website associated with the merchant involved in the transaction. For example, external data collection unit 42 of FIG. 1 may retrieve online search information (e.g., search terms and/or results) and/or URLs from the one of cardholder computing devices 20 that is associated with the cardholder, and/or from an ISP (of other sources 24) used by the cardholder. In other embodiments, the chargeback fraud factors 246 may include more, fewer and/or different factors than those shown in FIG. 4C.
  • The skimming factors 248 may include: (1) the number (X) of earlier transactions in which the card used for the transaction at issue was used at an ATM machine or a gas station pump within the 10 days prior to the transaction at issue; and/or (2) whether the transaction at issue originated from an IP address not associated with the cardholder. For example, external data collection unit 42 of FIG. 1 may retrieve transaction data indicating that certain past purchases were made using gas station pump card readers, and/or indicating that the card was used for one or more ATM withdrawals, from account records database 30, and/or may retrieve the originating IP address from the one of merchant computing systems 22 associated with the merchant involved in the transaction at issue. In other embodiments, the skimming factors 248 may include more, fewer and/or different factors than those shown in FIG. 4C.
  • Generally, the data indicative of whether the circumstance corresponding to each of counterfeit factors 242, account takeover factors 244, chargeback fraud factors 246 and/or skimming factors 248 is present/true for a particular transaction may be included in the first account data 150 described above in connection with FIG. 3C, for example.
  • As is also seen in FIG. 4C, each of the counterfeit factors 242, account takeover factors 244, chargeback fraud factors 246 and skimming factors 248 may be associated with a particular score or weighting value. The factors for each classification (counterfeit, account takeover, chargeback fraud, skimming) may be used to calculate a total score specific to that classification. In the rule set 240 shown in FIG. 4C, for example, a counterfeit score may be calculated based upon which of factors 242 are, or are not, present, an account takeover score may be calculated based upon which of factors 244 are, or are not, present, and so on. In some embodiments, certain factors may instead be associated with negative scores, and/or certain factors (e.g., the first of skimming factors 248 shown in FIG. 4C) may be associated with metrics or algorithms that determine how heavily those factors are weighed.
  • For each classification/category, the rule set 240 may output the total score, a normalized total score, an indication of whether the total score exceeded a threshold, a probability calculated based upon the total score, and/or some other indicator or measure of the likelihood that fraud of that particular type/class occurred in connection with the transaction. In the example shown in FIG. 4C, it can be seen that larger scores generally correspond to a greater probability that the respective classification is accurate. Referring back to FIG. 3C, the classification at process stage 152 may be the classification having the highest score and/or probability under rule set 240, or may include the score and/or probability for each classification, the top three classifications, etc.
  • D. Exemplary Application Fraud Detection Rule Set
  • Referring now to FIG. 4D, an exemplary rule set 260 may use online search information (e.g., search terms, search results, clicked/selected search results, etc.) to detect whether an application was fraudulent (e.g., not populated and/or submitted by the purported applicant). The rule set 260 may have been generated at process stage 164 of FIG. 3D, for example. The rule set 260 may be incorporated into a review process that is generally applied to all applications received by a particular entity or anti-fraud service, or a review process applied only to those applications that were flagged by a preliminary fraud alert, for example.
  • The factors considered under the rule set 260 may generally be probative of whether the person that submitted the application (e.g., via a web browser, a dedicated application, as an email attachment, by snail mail, etc.) had performed one or more online searches indicating that he or she was trying to learn more about the purported applicant in order to populate particular fields of the application (e.g., a “home address” field, “employment history” fields, etc.). The “purported applicant” may be a person whose name appears in a name and/or signature field of the application, for example.
  • As seen in FIG. 4D, the factors of exemplary rule set 260 may include: (1) whether the applicant used search terms that included the name of the purported applicant; (2) whether the search terms also included the words “address” or “residence” (and possibly other synonyms or near-synonyms); and/or (3) whether the search terms also included the words “employer,” “job” and/or “career” (and possibly other synonyms or near-synonyms). In other embodiments, the rule set 260 may include more, fewer and/or different factors than those shown in FIG. 4D. For example, the rule set 260 may include one or more factors relating to which search results appeared and/or were selected (e.g., “clicked” on after appearing on a user interface) by the applicant.
  • Generally, the data indicative of whether the circumstances corresponding to the factors of rule set 260 are present/true for a particular applicant may be included in the first applicant search history data 170 described above in connection with FIG. 3D. For example, external data collection unit 42 of FIG. 1 may obtain the search terms, search results, search result user selections, etc., needed to determine whether the various factors exist, from the applicant's computing device (e.g., similar to one of cardholder computing devices 20) and/or from an ISP of other sources 24. Access to such information may be made a condition of having the application be considered, for example.
  • As is also seen in FIG. 4D, each of the factors of rule set 260 may be associated with a particular score or weighting value. A total score may then be calculated based upon which factors are, or are not, present. In some embodiments, certain factors may instead be associated with negative scores, and/or certain factors may be associated with metrics or algorithms that determine how heavily those factors are weighed.
  • The rule set 260 may then output the total score, a normalized total score, an indication of whether the total score exceeded a threshold, a probability calculated based upon the total score, and/or some other indicator or measure of the existence or likelihood of application fraud. In the example shown in FIG. 4D, it can be seen that larger scores may generally correspond to a greater probability that the application was not populated and/or submitted by the purported applicant.
  • E. Exemplary Fraud Dispute Resolution Rule Set
  • Referring now to FIG. 4E, a flow diagram illustrates at least a portion of a process flow 270 implementing an exemplary rule set for fraud dispute, or potential fraud dispute, resolution (e.g., a rule set generated at process stage 184 of FIG. 3E). The process flow 270 may be used to help resolve a dispute over a contested transaction, or to help a customer recall an unrecognized transaction, for example. FIG. 4E illustrates a process flow, rather than just a set of factors, in order to better illustrate an example process for generating queries based upon the generated rules, according to one embodiment. The process flow 270 may correspond to a particular embodiment and scenario in which the transaction subject to dispute or potential dispute is a credit or debit card transaction.
  • In the exemplary process flow 270, the rule set may specify that a process stage 272 determines whether the transaction was a “card present” transaction. If not, the rule set may specify that the flow proceed directly to a process stage 280. If so, however, the rule set may specify that the flow instead proceeds to a process stage 274.
  • The rule set may also specify that process stage 274 determines whether at least one other transaction associated with the cardholder's account occurred within some threshold number of hours (X) of the transaction at issue. If not, the rule set may specify that the flow proceeds directly to process stage 280. If so, however, the rule set may specify that the flow instead proceeds to a process stage 276.
  • Process stage 276 may generate one or more location-related queries using transaction data associated with the cardholder's account. The queries may ask, for example, whether the cardholder was in (or near) one or more particular geographic areas or locations at various times. If the transaction at issue occurred in San Francisco, for example, with a first other “card present” transaction occurring in Santa Rosa four hours earlier and a second other “card present” transaction occurring in San Jose two hours later, process stage 276 may generate one or more queries asking whether the cardholder made or authorized the earlier and/or later transactions, and/or whether the cardholder traveled on a route from Santa Rosa to San Jose that passed through San Francisco, etc.
  • In some embodiments, the location-related queries are generated based upon data associated with events or circumstances other than transactions. For example, if the transaction at issue occurred in Sarasota, Fla., and the data considered under the rule set indicates that the cardholder checked in to a flight to Tampa, process stage 276 may generate one or more queries asking whether the cardholder completed the flight, where the cardholder went after landing in Tampa, etc.
  • The rule set may also specify that process stage 280 determines whether the transaction at issue is associated with a billing alias that is dissimilar to the name of the merchant involved in the transaction. For example, the computing system of the merchant (e.g., one of merchant computing systems 22 of FIG. 1) may have sent to FAMS 14 a transaction record that identified the merchant by the alias, and was presented to the cardholder as an online or paper account statement. The determination at process stage 280 may use the billing alias to identify a legal and/or common name of the merchant (e.g., using a relational database stored in AFSS 12 or FAMS 14), and determine that there is at least some threshold level of dissimilarity (e.g., based upon difference of characters, character ordering, etc.) between the billing alias and the merchant name.
  • If the billing alias and merchant name are not sufficiently dissimilar, the rule set may specify that the flow proceeds directly to a process stage 284. If sufficiently dissimilar, however, the rule set may specify that the flow instead proceeds to a process stage 282. Process stage 282 may generate a query relating to the billing alias that was presented to the cardholder. For example, the query may ask whether the cardholder is aware that the billing alias is used by that particular merchant. In some embodiments, process stage 282 may instead generate a message that simply informs the cardholder that the billing alias corresponds to the merchant, without posing a question.
  • The rule set may specify that process stage 284 generates one or more default queries. For example, one default query may ask whether the cardholder lent his or her card to a friend or family member around the time of the transaction. In some embodiments and/or scenarios, process stage 284 may be omitted from process flow 270. Generally, the queries (and possibly non-query messages) generated in process flow 270 may serve to help the cardholder recall whether the transaction was made or authorized, and/or process flow 270 may prompt the cardholder for responses that are considered by others (e.g., personnel of an entity associated with FAMS 14 of FIG. 1) to determine whether the transaction was likely fraudulent.
  • Although not shown in FIG. 4E, in some embodiments process flow 270 may include a number of iterative stages in which responses are received from the cardholder (e.g., from the respective one of cardholder computing devices 20 in FIG. 1) and used to generate additional, more detailed questions for the cardholder. For example, if a first query asks whether the cardholder recalls personally making another “card present” transaction that occurred at a nearby time and place, and the cardholder responds “no,” a new query may be generated asking whether the cardholder recalls personally making the next closest transaction (in terms of time and/or location).
  • F. Exemplary Document Fraud Detection Rule Set
  • Referring next to FIG. 4F, an exemplary rule set 290 (e.g., generated at process stage 204 of FIG. 3F) may use various factors relating to an imaged (e.g., photographed or scanned) physical document to determine whether the document should be flagged as a candidate for a more in-depth (e.g., manual) analysis/review for fraud purposes. The rule set 290 may correspond to a particular embodiment and scenario in which the document is one that includes at least a signature field (e.g., a personal check, a driver's license, etc.).
  • The factors considered under the rule set 290 may include a number of counterfeit factors 292 and a number of forgery factors 294, each of which may be evaluated by image analysis unit 52 of FIG. 1 using one or more image processing techniques. The counterfeit factors 292 may relate to the look, presentation, format and/or structure of the document, while the forgery factors 294 may relate to the substance, style or format of information entered in one or more fields of the document.
  • As seen in FIG. 4F, the counterfeit factors 292 may include: (1) whether one or more absolute or relative dimensions and/or angles of the document, or of lines, illustrations, patterns, etc. shown on the document (excluding user-entered contents in fields such as the signature line), are outside one or more predetermined tolerances; (2) whether one or more colors on the document are outside a predetermined tolerance (e.g., color/frequency range); (3) whether one or more line thicknesses of the document (excluding user-entered field contents) are outside one or more predetermined tolerances; and/or (4) whether one or more fonts on the document (excluding user-entered field contents) are outside one or more predetermined tolerances. For example, image analysis unit 52 may determine whether the ratio of the document length to the document width is within 0.1% of an expected value. As another example, image analysis unit 52 may determine whether horizontal and vertical lines on the document are within 0.3 degrees of the horizontal and vertical edges of the document, respectively. As yet another example, image analysis unit 52 may determine whether a font used for a field descriptor or other text on the document matches an expected font (e.g., by meeting a similarity threshold measured in any suitable manner). In other embodiments, the counterfeit factors 292 may include more, fewer and/or different factors than those shown in FIG. 4F.
  • The forgery factors 294 may include: (1) whether a signature entered in a signature field of the document match is outside a predetermined tolerance (e.g., using any suitable signature recognition technique); (2) whether handwriting entered in one or more fields of the document is outside a predetermined tolerance (e.g., by applying a suitable handwriting recognition technique); and/or (3) whether the format of information entered by a user in one or more fields does not match an expected format (e.g., using “9.12.16” rather than the expected “9/12/2016,” as established based upon other documents known to have been populated and/or submitted by the purported applicant). In other embodiments, the forgery factors 294 may include more, fewer and/or different factors than those shown in FIG. 4F.
  • Generally, the data indicative of whether the circumstances corresponding to counterfeit factors 292 and/or forgery factors 294 are present/true for a particular document may be included in the first document image data 210 described above in connection with FIG. 3F.
  • As is also seen in FIG. 4F, each of the counterfeit factors 292 and forgery factors 294 may be associated with a particular score or weighting value. In the rule set 290 shown in FIG. 4F, a total score may be calculated based upon which factors are, or are not, present. In some embodiments, certain factors may instead be associated with negative scores, and/or certain factors may be associated with metrics or algorithms that determine how heavily those factors are weighed.
  • The rule set 290 may then output the total score, a normalized total score, an indication of whether the total score exceeded a threshold, a probability calculated based upon the total score, and/or some other indicator or measure of the likelihood that the document is fraudulent. Alternatively, the rule set 290 may output a separate total score, normalized score, probability, or other metric, for each of counterfeit factors 292 and forgery factors 294, with the counterfeit metric indicating the likelihood that the document is a counterfeit and the forgery metric indicating the likelihood that the document was fraudulently populated by someone other than the purported person (e.g., by someone other than the person corresponding to the name, signature, address, etc. on the document). In the example shown in FIG. 4F, it can be seen that larger scores generally correspond to a greater probability that the document is fraudulent. In some embodiments, the rule set 290 also includes one or more other types of factors not shown in FIG. 4F, and/or omits either counterfeit factors 292 or forgery factors 294.
  • V. Exemplary Methods for Fraud Detection & Classification
  • FIGS. 5-10 depict flow diagrams of various exemplary computer-implemented methods that may be implemented by one or more components of AFSS 12 of FIG. 1. In one embodiment, AFSS 12 implements all of the methods corresponding to FIGS. 5-10. In other embodiments, AFSS 12 implements only a subset (e.g., one, two, etc.) of the methods corresponding to FIGS. 5-10. Each of the methods described below may be implemented by fraud detection/classification unit 36 of FIG. 1, for example.
  • A. Exemplary False Positive Identification
  • FIG. 5 illustrates an exemplary computer-implemented method 300 of using customer data to determine that geolocation-based fraud alerts are false positives. The method 300 may include, via one or more processors and/or transceivers (or a trained machine learning program), determining whether the electronic fraud alert is geolocation based (block 302), and if so, with customer permission, then receiving customer data (block 304), such as via wireless communication or data transmission over one or more radio links or wireless communication channels. The customer data may be collected or generated via various processors, transceivers, or sensors associated with mobile devices, smart homes, and/or smart vehicles. The customer data may indicate or be associated with telematics, online activity, browsing activity, IP address, credit card, customer location, and/or financial transaction data.
  • The method 300 may include determining whether two or more of the customer data sources include customer data indicating or confirming that the customer is traveling (block 306). If so, the method 300 may include determining whether the current customer location corresponds to the financial transaction location (block 308). If so, the method 300 may include not transmitting the electronic fraud alert to the customer's mobile device and/or flagging the fraud alert as a false positive; and if not, then transmitting the electronic fraud alert to the customer's mobile device (block 310).
  • FIG. 6 illustrates another exemplary computer-implemented method 320 of using customer data to determine whether geolocation-based fraud alerts are false positives. The method 320 may include, via one or more processors and/or transceivers, receiving customer data (such as mobile device, smart home, smart vehicle, telematics, online activity, IP address, credit card, location, and/or financial transaction data) (block 322), such as via wireless communication or data transmission over one or more radio links or wireless communication channels. The method 320 may include determining whether the customer is traveling using (or based upon processor analysis of) the customer data (block 324), such as by identifying that the current GPS location of the customer's mobile device and/or vehicle is outside of the customer's home address county or city. The method 320 may include receiving an electronic fraud alert associated with the customer (block 326), such as an alert associated with a potentially unauthorized financial transaction being charged to the customer. The method 320 may include determining whether the reason the electronic fraud alert was generated was location-based (block 328), such as processor or machine learning program determining that a location associated with the financial transaction does not match a home address location or home city of the customer. If so, and if the customer is determined to be traveling, the method 320 may include determining whether the current customer GPS or other location, such as determined from mobile device, IP address, or vehicle location, corresponds to the financial transaction location (block 330). If so, the method 320 may further include not transmitting the electronic fraud alert to the customer's mobile device, and/or flagging the fraud alert as a false positive; and if not, then transmitting the electronic fraud alert to the customer's mobile device (block 332).
  • In one aspect, a computer-implemented method of using customer data to determine that geolocation-based fraud alerts are false positives may be provided. The method may include (1) determining, via the one or more processors, if an electronic fraud alert is a geolocation-based fraud alert (or otherwise generated based upon an unexpected or abnormal transaction location), such as by inputting the fraud alert and/or financial transaction underlying data into a machine learning program trained to identify geolocation-based fraud alerts; (2) if the electronic fraud alert is geolocation-based, then retrieving or receiving (with customer permission or affirmative consent), via the one or more processors and/or transceivers, two or more sources of customer data over one or more radio frequency links; (3) determining, via the one or more processors, if the customer data from two or more sources indicate or confirm that the customer is traveling (such as not currently at their home address or within a predetermined distance of their home address); (4) if the customer data indicates that the customer is traveling, then determining, via the one or more processors, whether a current customer location indicated by the customer data retrieved matches, or corresponds to, the transaction location; and/or (5) if the current customer location corresponds to the transaction location, then marking, via the one or more processors, the electronic fraud alert as a false positive and not transmitting the electronic fraud alert to a customer mobile device to reduce an amount of false positives that are transmitted to customers.
  • The method may further include receiving, via one or more processors and/or transceivers, transaction data associated with a financial transaction over a wireless communication channel; and inputting, via the one or more processors, the transaction data into a rules-engine to identify the financial transaction as potentially fraudulent and generate an electronic fraud alert. The customer data may be collected or generated by a mobile device and/or mobile device sensors, and include one or more current or past GPS locations.
  • The customer data may be collected or generated by a vehicle controller or processor and/or vehicle-mounted sensors, and include one or more current or past GPS locations. The customer data may be collected or generated by a smart home controller and/or home-mounted sensors, and include data indicating whether or not a home of the customer is presently occupied or vacant, and/or how long the home has been vacant. The customer data may include an IP address of a customer computing device, and include one or more current or past GPS locations. The customer data may include online, browsing, and/or social media activity received from a customer computing device. Additionally or alternatively, the customer data may include vehicle telematics data that includes one or more past or current GPS locations.
  • If the current customer location does not match or correspond to the transaction location, then the method may include marking, via the one or more processors, the electronic fraud alert as verified and transmitting the electronic fraud alert to a customer mobile device to facilitate sending only confirmed fraud alerts to customers. The fraud alert may be determined to be geolocation-based when a financial transaction location is not within a predetermined distance of a customer home address. Additionally or alternatively, the fraud alert may be determined to be geolocation-based when a financial transaction location does not correspond to normal travel activity or locations frequented by, or associated with, the customer.
  • In another aspect, a computer system configured to use customer data to determine that geolocation-based fraud alerts are false positives may be provided. The computer system may include one or more processors and/or transceivers configured to: determine if an electronic fraud alert is a geolocation-based fraud alert (or otherwise generated based upon an unexpected or abnormal transaction location); if the electronic fraud alert is geolocation-based, then retrieve or receive (with customer permission or affirmative consent) via wireless communication or data transmission two or more sources of customer data over one or more radio frequency links or wireless communication channels; determine if the customer data from two or more sources indicate or confirm that the customer is traveling (such as not currently at their home address or within a predetermined distance of their home address); if the customer data indicates that the customer is traveling, then determine whether a current customer location indicated by the customer data retrieved matches, or corresponds to, the transaction location; and/or if the current customer location corresponds to the transaction location, then mark the electronic fraud alert as a false positive and not transmit the electronic fraud alert to a customer mobile device to reduce an amount of false positives that are transmitted to customers.
  • B. Exemplary Use of Cardholder Location to Identify or Prevent Fraud
  • FIGS. 7 through 10 illustrate exemplary computer-implemented methods that use information about the locations of authorized cardholders (e.g., the primary cardholder, or another individual listed on the account) to prevent false positive fraud alerts, or to block potentially fraudulent financial transactions. FIGS. 7 and 8 correspond to card-present financial transactions (e.g., where an individual swipes the card or inserts the card in a chip reader, or where a merchant does so on behalf of that individual after being handed the card), and FIGS. 9 and 10 correspond to online financial transactions (e.g., where an individual types card information into a website page configured to accept such information in connection with a desired transaction). The methods of FIGS. 7 through 10 may be implemented by a card issuer/bank or by another entity (e.g., by an entity associated with FAMS 14 or AFSS 12 of FIG. 1), for example.
  • Generally, the methods of FIGS. 7 and 10 may provide the benefit of avoiding unnecessary network communications and/or computer processing associated with false positive fraud alerts, while the methods of FIGS. 8 and 9 may provide the benefit of avoiding fraudulent transactions in the first instance, without requiring more cumbersome and/or time-consuming techniques (e.g., sending an authorization code to the authorized cardholder to verify a suspicious transaction). Moreover, all of the methods in FIGS. 7 through 10 may more accurately detect fraud by using the geographic location of the authorized cardholder in conjunction with the location of the transaction or the location of the computer used to enter card information. For example, while a $30 gas purchase 50 miles from the authorized cardholder's home may not look suspicious to a fraud detection algorithm, the transaction may become much more suspect if the cardholder is 10 miles away from that gas station at the time of the purchase. Similarly, an online purchase for an item that the authorized cardholder has often bought in the past may not look suspicious to some fraud detection algorithms, but may become much more suspect if the authorized cardholder was, at the time the card information was entered at a computer, located a significant distance away from that computer. Further, the methods of FIGS. 7 through 10 may detect and/or prevent fraudulent transactions even when the physical debit or credit card has been stolen (as opposed to only copying down the card number or “skimming,” etc.).
  • Referring first to FIG. 7, in one exemplary computer-implemented method 400, it may be determined that a fraud alert exists for a financial transaction (block 402). The fraud alert may be an electronic fraud alert generated by the device or system implementing the method 400, or received from another device or system (e.g., from a card issuer system, such as FAMS 14 of FIG. 1, via a network such as network 26 of FIG. 1), for example. The financial transaction may be a card-present transaction that is associated with a debit or credit card account, and was purportedly entered into by an authorized cardholder associated with the account.
  • A first geographic location, at which information associated with the account was obtained, may be determined (block 404). The information associated with the account may have been obtained by swiping or inserting the card in a device (e.g., part of one of merchant computing systems 22 of FIG. 1) in connection with the financial transaction, for example. In some embodiments and/or scenarios, the first geographic location may be identified based upon location information included in a field of transaction data associated with the financial transaction, such as transaction data that was retrieved from an account records database (e.g., database 30 of FIG. 1).
  • In some embodiments and/or scenarios, block 404 may occur prior to block 402, in which case block 402 may include comparing the first geographic location to a set of one or more locations known to be typical or expected for the authorized cardholder (e.g., a home address, city and/or state), and/or may include generating the fraud alert in response to determining that the first geographic location does not correspond to (e.g., is not at, or not within a threshold distance of) the set of typical/expected locations.
  • The time of the financial transaction may also be determined (block 406). In some embodiments and/or scenarios, the time may be identified based upon time information (e.g., a time stamp) included in a particular field of transaction data associated with the financial transaction, such as the transaction data described above.
  • It may also be determined, based upon geolocation data indicating one or more geographic locations of the authorized cardholder (e.g., over a period of time), that the authorized cardholder was at a second geographic location at the time of the financial transaction (block 408). The geolocation data may be time-stamped data received from a third party server with the express consent of the authorized cardholder, or retrieved from a database in which the data was stored (with the cardholder's express consent) after being received from a mobile device of the cardholder, for example. The geolocation data may include GPS data (e.g., collected by a smartphone or other mobile device of the authorized cardholder), data indicating identifiers and/or signal strengths of WiFi access points that were near the cardholder (e.g., collected by a smartphone or other mobile device of the authorized cardholder), data indicating that the authorized cardholder had “checked in” at a particular location (e.g., via a social media or other application), data indicating that the authorized cardholder used a smart appliance at a known location (e.g., at the cardholder's home), and/or other types of data indicative of the cardholder's locations at particular times. The location of the cardholder at the time of the financial transaction may be determined by matching a time-stamp to the time determined at block 406, using a location with a time-stamp that corresponds to a nearest time (e.g., so long as that time is within some threshold time of the time determined at block 406), or in another suitable manner.
  • It may then be determined that the second geographic location corresponds to the first geographic location (block 410). For example, it may be determined at block 410 that the first and second geographic locations are the same (e.g., the same city), are within a same geographic territory (e.g., cities within the same state), or are within a threshold distance of each other (e.g., cities or more precise locations within 50 miles of each other, 100 miles of each other, etc.).
  • In response to the determination at block 410, the fraud alert may be marked as a false positive (block 412), such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction. For example, a “verified” flag or field associated with the fraud alert may be set to a value of “0” or “false” at block 412, and a notification unit (e.g., notification unit 56 of FIG. 1) may decide not to send the fraud alert to the authorized cardholder's mobile device and/or other computing device (e.g., as an email or text message) based upon the flag or field value.
  • Referring next to FIG. 8, in an exemplary computer-implemented method 420, a request to authorize a financial transaction may be received (block 422). The financial transaction may be a card-present transaction that is associated with a debit or credit card account, and is purportedly being entered into by an authorized cardholder associated with the account. As will be understood from the description that follows, the financial transaction, in the method 420, is one that has not yet been fully executed. The request may have been automatically or manually generated by the card issuer when deciding whether to clear the transaction, or automatically or manually generated by a merchant shortly after receiving credit card information (e.g., by a swipe or insertion of the card), for example. The request may include the credit card information (e.g., credit card number, expiration date and/or security code) and/or other information relating to the financial transaction.
  • A first geographic location, at which information associated with the account was obtained (e.g., by swiping or inserting the card in connection with the financial transaction), may be determined (block 424). Block 424 may be similar to block 404 of the method 400, for example. In some embodiments and/or scenarios, however, the first geographic location is determined by identifying the location as specified in the request received at block 422.
  • It may also be determined, based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction (block 426). The geolocation data and/or the source of such data may be similar to that described above in connection with block 408 of the method 400, for example. In some embodiments and/or scenarios, however, the geolocation data may not be time-stamped, or such time stamps may exist but not be utilized. For example, it may be known that the financial transaction is currently in process, and therefore the second geographic location may be the current location of the authorized cardholder.
  • It may then be determined that the second geographic location does not correspond to the first geographic location (block 428). For example, it may be determined at block 428 that the first and second geographic locations are not the same (e.g., not the same city), are not within a same geographic territory (e.g., not cities within the same state), or are not within a threshold distance of each other (e.g., not cities or other, more specific locations within 50 miles of each other, 100 miles of each other, etc.).
  • In response to the determination at block 428, the financial transaction may be prevented from being executed (block 430). If the method 420 is implemented by a computing system of the card issuer, for example, block 430 may include not clearing the financial transaction. As another example, a merchant terminal (e.g., part of one of merchant computing systems 22 of FIG. 1) sending the request received at block 422 (or that is otherwise associated with the financial transaction) may be sent a fraud alert indicating that the transaction may be fraudulent and/or should not be completed. In yet another example embodiment, the fraud alert may be sent to a computing system of the card issuer (e.g., if the request received at block 422 was received from such a computing system).
  • Referring next to FIG. 9, in an exemplary computer-implemented method 440, a request to authorize a financial transaction may be received (block 442). The financial transaction may be an online transaction that is associated with a debit or credit card account, and is purportedly being entered into by an authorized cardholder associated with the account. As with the method 420, the financial transaction, in the method 440, is one that has not yet been fully executed. The request may have been automatically or manually generated by the card issuer when deciding whether to clear the transaction, or automatically or manually generated by a merchant shortly after receiving credit card information (e.g., shortly after the merchant computing system received credit card information that was manually entered by a person purporting to be the authorized cardholder), for example. The request may include the credit card information (e.g., credit card number, expiration date and/or security code) and/or other information relating to the financial transaction.
  • A computing device at which information associated with the card account (e.g., the card number, expiration date, and/or three- or four-digit security code) was entered in connection with the financial transaction may be identified (block 444). The computing device may be identified by receiving an IP address of the computing device from the computing system of the merchant associated with the financial transaction (either directly, or via the card issuer), for example.
  • A first geographic location, at which the computing device identified at block 444 resides, may be determined (block 446). In some embodiments and/or scenarios, the first geographic location is determined by using the IP address of the computing device. For example, the IP address itself may indicate physical location (at a high level of generality), or the IP address may be used as a key to a location database that relates IP addresses to more specific physical locations. With respect to the latter embodiment, for instance, a computing system implementing the method 440 may, as a part of its fraud prevention services, ask cardholders to voluntarily register any fixed-location computers (e.g., desktop computers) that they expect to use for online purchases, with the registration process including sending (from each such computer) a message specifying the physical location of the computer. In still other embodiments and/or scenarios, the first geographic location is determined by identifying a location specified in the request received at block 442, or in another suitable manner.
  • It may also be determined, based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction (block 448). The geolocation data and/or the source of such data may be similar to that described above in connection with block 408 of the method 400, for example. In some embodiments and/or scenarios, however, the geolocation data may not be time-stamped, or such time stamps may exist but not be utilized. For example, it may be known that the financial transaction is currently in process, and therefore the second geographic location may be the current location of the authorized cardholder.
  • It may then be determined that the second geographic location does not correspond to the first geographic location (block 450). Block 450 may be similar to block 428 of the method 420, for example. In response to the determination at block 450, the financial transaction may be prevented from being executed (block 452). Block 452 may be similar to block 430 of the method 420, for example.
  • Referring next to FIG. 10, in an exemplary computer-implemented method 460, it may be determined that a fraud alert exists for a financial transaction (block 462). The fraud alert may be an electronic fraud alert generated by the device or system implementing the method 460, or received from another device or system (e.g., from a card issuer system, such as FAMS 14 of FIG. 1, via a network such as network 26 of FIG. 1), for example. The financial transaction may be an online transaction that is associated with a debit or credit card account, and is purportedly entered into by an authorized cardholder associated with the account.
  • A computing device at which information associated with the card account (e.g., the card number, expiration date, and/or three- or four-digit security code) was entered in connection with the financial transaction may be identified (block 464). Block 464 may be similar to block 444 of the method 440, for example.
  • A first geographic location, at which the computing device identified at block 464 resides, may be determined (block 466). In some embodiments and/or scenarios, the first geographic location may be identified based upon an IP address, of the computing device, that may be specified in a particular field of transaction data that is retrieved from an account records database (e.g., database 30 of FIG. 1). In other embodiments and/or scenarios, the first geographic location itself may be specified in such a field.
  • In some embodiments and/or scenarios, block 466 may occur prior to block 462, in which case block 462 may include comparing the first geographic location to a set of one or more locations known to be typical or expected for the authorized cardholder (e.g., a home address, city and/or state), and/or may include generating the fraud alert in response to determining that the first geographic location does not correspond to (e.g., is not at, or not within a threshold distance of) the set of typical/expected locations.
  • The time of the financial transaction may also be determined (block 468). In some embodiments and/or scenarios, the time may be identified based upon time information (e.g., a time stamp) included in a particular field of transaction data associated with the financial transaction, such as transaction data that is retrieved from an account records database (e.g., database 30 of FIG. 1).
  • It may also be determined, based upon geolocation data indicating one or more geographic locations of the authorized cardholder (e.g., over a period of time), that the authorized cardholder was at a second geographic location at the time of the financial transaction (block 470). Block 470 may be similar to block 408 of the method 400, for example.
  • It may then be determined that the second geographic location corresponds to the first geographic location (block 472). Block 472 may be similar to block 410 of the method 400, for example. In response to the determination at block 472, the fraud alert may be marked as a false positive (block 474), such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction. Block 474 may be similar to block 412 of the method 400, for example.
  • VI. Exemplary System for Fraud Detection & Classification
  • FIG. 11 depicts an exemplary computer system 500 in which the techniques described herein may be implemented, according to one embodiment. The computer system 500 of FIG. 11 may include a computing device in the form of a computer 510. Components of the computer 510 may include, but are not limited to, a processing unit 520, a system memory 530, and a system bus 521 that couples various system components including the system memory 530 to the processing unit 520. The system bus 521 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, or a local bus, and may use any suitable bus architecture. By way of example, and not limitation, such architectures include the Industry Standard Architecture (ISA) bus, Micro Channel Architecture (MCA) bus, Enhanced ISA (EISA) bus, Video Electronics Standards Association (VESA) local bus, and Peripheral Component Interconnect (PCI) bus (also known as Mezzanine bus).
  • Computer 510 may include a variety of computer-readable media. Computer-readable media may be any available media that can be accessed by computer 510 and may include both volatile and nonvolatile media, and both removable and non-removable media. By way of example, and not limitation, computer-readable media may comprise computer storage media and communication media. Computer storage media may include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data. Computer storage media may include, but is not limited to, RAM, ROM, EEPROM, FLASH memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by computer 510.
  • Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and may include any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media may include wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, radio frequency (RF), infrared and other wireless media. Combinations of any of the above are also included within the scope of computer-readable media.
  • The system memory 530 may include computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 531 and random access memory (RAM) 532. A basic input/output system 533 (BIOS), containing the basic routines that help to transfer information between elements within computer 510, such as during start-up, is typically stored in ROM 531. RAM 532 typically contains data and/or program modules that are immediately accessible to, and/or presently being operated on, by processing unit 520. By way of example, and not limitation, FIG. 11 illustrates operating system 534, application programs 535, other program modules 536, and program data 537.
  • The computer 510 may also include other removable/non-removable, volatile/nonvolatile computer storage media. By way of example only, FIG. 11 illustrates a hard disk drive 541 that reads from or writes to non-removable, nonvolatile magnetic media, a magnetic disk drive 551 that reads from or writes to a removable, nonvolatile magnetic disk 552, and an optical disk drive 555 that reads from or writes to a removable, nonvolatile optical disk 556 such as a CD ROM or other optical media. Other removable/non-removable, volatile/nonvolatile computer storage media that can be used in the exemplary operating environment include, but are not limited to, magnetic tape cassettes, flash memory cards, digital versatile disks, digital video tape, solid state RAM, solid state ROM, and the like. The hard disk drive 541 may be connected to the system bus 521 through a non-removable memory interface such as interface 540, and magnetic disk drive 551 and optical disk drive 555 may be connected to the system bus 521 by a removable memory interface, such as interface 550.
  • The drives and their associated computer storage media discussed above and illustrated in FIG. 11 provide storage of computer-readable instructions, data structures, program modules and other data for the computer 510. In FIG. 11, for example, hard disk drive 541 is illustrated as storing operating system 544, application programs 545, other program modules 546, and program data 547. Note that these components can either be the same as or different from operating system 534, application programs 535, other program modules 536, and program data 537. Operating system 544, application programs 545, other program modules 546, and program data 547 are given different numbers here to illustrate that, at a minimum, they are different copies. A user may enter commands and information into the computer 510 through input devices such as cursor control device 561 (e.g., a mouse, trackball, touch pad, etc.) and keyboard 562. A monitor 591 or other type of display device is also connected to the system bus 521 via an interface, such as a video interface 590. In addition to the monitor, computers may also include other peripheral output devices such as printer 596, which may be connected through an output peripheral interface 595.
  • The computer 510 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 580. The remote computer 580 may be a personal computer, a server, a router, a network PC, a peer device or other common network node, and may include many or all of the elements described above relative to the computer 510, although only a memory storage device 581 has been illustrated in FIG. 11. The logical connections depicted in FIG. 11 include a local area network (LAN) 571 and a wide area network (WAN) 573, but may also include other networks. Such networking environments are commonplace in hospitals, offices, enterprise-wide computer networks, intranets and the Internet.
  • When used in a LAN networking environment, the computer 510 is connected to the LAN 571 through a network interface or adapter 570. When used in a WAN networking environment, the computer 510 may include a modem 572 or other means for establishing communications over the WAN 573, such as the Internet. The modem 572, which may be internal or external, may be connected to the system bus 521 via the input interface 560, or other appropriate mechanism. The communications connections 570, 572, which allow the device to communicate with other devices, are an example of communication media, as discussed above. In a networked environment, program modules depicted relative to the computer 510, or portions thereof, may be stored in the remote memory storage device 581. By way of example, and not limitation, FIG. 11 illustrates remote application programs 585 as residing on memory device 581.
  • The techniques for detecting and/or classifying fraud described above may be implemented in part or in their entirety within a computer system such as the computer system 500 illustrated in FIG. 11. The computer 510 may be included in AFSS 12 of FIG. 1, for example, and/or the remote application programs 585 may include one or more applications of either FAMS 14, one of cardholder computing device 20, one of merchant computing systems 22, or a computing device of other sources 24. Moreover, the functionality of fraud detection/classification unit 36 of FIG. 1 may be implemented by one or more of application programs 535 and/or other program modules 536. As another example, ML rules database 58, account holder behaviors database 60 and/or chargeback rules database 62 of FIG. 1 may be stored in hard disk drive 541 (e.g., as program data 547), magnetic disk 552 and/or optical disk drive 555, and/or the data retrieved by fraud detection/classification unit 36 of FIG. 1 may be stored in hard disk drive 541 (e.g., as program data 547) and/or RAM 532 (e.g., as program data 537).
  • VII. Exemplary Method Embodiments
  • In one aspect, a computer-implemented method of using customer data to determine that geolocation-based fraud alerts are false positives may be implemented in one or more servers or other computing devices. The method may include (1) determining, by one or more processors, that an electronic fraud alert is a geolocation-based fraud alert generated based upon an unexpected or abnormal transaction location; (2) in response to determining that the electronic fraud alert is a geolocation-based fraud alert, obtaining, by the one or more processors and via one or more radio frequency links, customer data from two or more sources; (3) determining, by the one or more processors, that the customer data from the two or more sources indicates that a customer is traveling; (4) in response to determining that the customer data indicates that the customer is traveling, determining, by the one or more processors, that a customer location indicated by the customer data corresponds to the transaction location; and/or (5) in response to determining that the customer location corresponds to the transaction location, (i) marking, by the one or more processors, the electronic fraud alert as a false positive and (ii) causing, by the one or more processors, the electronic fraud alert to not be transmitted to a mobile device of the customer in order to reduce an amount of false positives that are transmitted to customers. The method may include additional, fewer or alternative actions, such as any of those discussed elsewhere herein.
  • For instance, determining that the electronic fraud alert is a geolocation-based fraud alert may include inputting, by the one or more processors, one or both of (i) the electronic fraud alert, and (ii) transaction data corresponding to a financial transaction associated with the electronic fraud alert, into a machine learning program that is trained to identify geolocation-based fraud alerts.
  • Additionally or alternatively, the method may further include obtaining, by the one or more processors and via a wireless communication channel, transaction data corresponding to a financial transaction associated with the electronic fraud alert, and/or inputting, by the one or more processors, the transaction data into a rules engine to identify the financial transaction as potentially fraudulent and generate the electronic fraud alert.
  • Additionally or alternatively, the customer data may be collected or generated by one or both of (i) the mobile device and (ii) one or more sensors of the mobile device, and/or may include one or more current or past GPS locations.
  • Additionally or alternatively, the customer data may be collected or generated by one or both of (i) a vehicle controller or processor and (ii) one or more vehicle-mounted sensors, and/or may include one or more current or past GPS locations.
  • Additionally or alternatively, the customer data may be collected or generated by one or both of (i) a smart home controller and (ii) one or more home-mounted sensors, and/or may include data indicating one or both of (i) whether a home of the customer is presently occupied or vacant and (ii) how long the home of the customer has been vacant.
  • Additionally or alternatively, the customer data may include an IP address of a customer computing device, and/or may include one or more current or past GPS locations. Additionally or alternatively, the customer data may include one or more of (i) online data received from a customer computing device, (ii) browsing data received from the customer computing device, or (iii) social media activity data received from the customer computing device. Additionally or alternatively, the customer data may include vehicle telematics data that includes one or more past or current GPS locations.
  • Additionally or alternatively, the method may include: determining, by the one or more processors, that another electronic fraud alert is a geolocation-based fraud alert generated based upon another unexpected or abnormal transaction location; in response to determining that the other electronic fraud alert is a geolocation-based fraud alert, obtaining, by the one or more processors and via one or more other radio frequency links, additional customer data from two or more other sources; determining, by the one or more processors, that the additional customer data from the two or more other sources indicates that another customer is traveling; in response to determining that the additional customer data indicates that the other customer is traveling, determining, by the one or more processors, that another customer location indicated by the additional customer data does not correspond to the other transaction location; and/or in response to determining that the other customer location does not correspond to the other transaction location, (i) marking, by the one or more processors, the other electronic fraud alert as verified and (ii) causing, by the one or more processors, the electronic fraud alert to be transmitted to a mobile device of the other customer to facilitate sending only confirmed fraud alerts to customers.
  • Additionally or alternatively, determining that the electronic fraud alert is a geolocation-based fraud alert may include determining that the transaction location is not within a predetermined distance of a customer home address. Additionally or alternatively, determining that the electronic fraud alert is a geolocation-based fraud alert may include determining that the transaction location does not correspond to travel activity or locations associated with the customer.
  • In another aspect, a computer-implemented method of reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions may be implemented in one or more servers or other computing devices. The method may include: (1) determining, by one or more processors of the one or more servers, that a fraud alert exists for a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account; (2) determining, by the one or more processors, a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determining, by the one or more processors, a time of the financial transaction; (4) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction; (5) determining, by the one or more processors, that the second geographic location corresponds to the first geographic location; and/or (6) in response to determining that the second geographic location corresponds to the first geographic location, marking, by the one or more processors, the fraud alert as a false positive such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction. The method may include additional, fewer or alternative actions, such as any of those discussed elsewhere herein.
  • For instance, determining the first geographic location may occur prior to determining that the fraud alert exists, and determining that the fraud alert exists may include comparing the first geographic location to a set of one or more typical locations of the authorized cardholder, and/or generating the fraud alert in response to determining that the first geographic location does not correspond to the set of one or more typical locations.
  • Additionally or alternatively, the method may further include retrieving, by the one or more processors and from an account records database, transaction data associated with the financial transaction, and/or determining the first geographic location may include identifying the first geographic location based upon location information included in a first field of the transaction data.
  • Additionally or alternatively, determining the time of the financial transaction may include identifying the time of the financial transaction based upon time information included in a second field of the transaction data. Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a third party server.
  • Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a mobile device of the authorized cardholder, storing the geolocation data in a database, and/or retrieving the geolocation data from the database. Additionally or alternatively, receiving the geolocation data from a mobile device of the authorized cardholder may include receiving GPS location data from the mobile device.
  • Additionally or alternatively, determining that the second geographic location corresponds to the first geographic location may include determining that the first geographic location and the second geographic location are within a threshold distance of each other. Additionally or alternatively, determining that the second geographic location corresponds to the first geographic location may include determining that the first geographic location and the second geographic location are within a same geographic territory.
  • In another aspect, a computer-implemented method of preventing fraudulent card-present financial transactions may be implemented in one or more servers. The method may include: (1) receiving, by one or more processors of the one or more servers, a request to authorize a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly being entered into by an authorized cardholder associated with the debit or credit card account; (2) determining, by the one or more processors, a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at a time of the financial transaction; (4) determining, by the one or more processors, that the second geographic location does not correspond to the first geographic location; and/or (5) in response to determining that the second geographic location does not correspond to the first geographic location, preventing, by the one or more processors, the financial transaction from being executed. The method may include additional, fewer or alternative actions, such as any of those discussed elsewhere herein.
  • For instance, determining a first geographic location may include identifying a first geographic location specified in the request. Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a third party server.
  • Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a mobile device of the authorized cardholder, storing the geolocation data in a database, and/or retrieving the geolocation data from the database. Additionally or alternatively, determining that the second geographic location does not correspond to the first geographic location may include one or both of determining that the first geographic location and the second geographic location are not within a threshold distance of each other, and determining that the first geographic location and the second geographic location are not within a same geographic territory.
  • Additionally or alternatively, preventing the financial transaction from being executed may include one or both of causing a fraud alert to be sent to a merchant terminal associated with the financial transaction, and causing a fraud alert to be sent to a computing system of a card issuer associated with the debit or credit card account.
  • In another aspect, a computer-implemented method of preventing fraudulent online financial transactions may be implemented in one or more servers. The method may include: (1) receiving, by one or more processors of the one or more servers, a request to authorize a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is an online transaction purportedly being entered into by an authorized cardholder associated with the debit or credit card account; (2) identifying, by the one or more processors, a computing device at which information associated with the debit or credit card account was entered in connection with the financial transaction; (3) determining, by the one or more processors, a first geographic location at which the computing device resides; (4) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at a time of the financial transaction; (5) determining, by the one or more processors, that the second geographic location does not correspond to the first geographic location; and/or (6) in response to determining that the second geographic location does not correspond to the first geographic location, preventing, by the one or more processors, the financial transaction from being executed. The method may include additional, fewer or alternative actions, such as any of those discussed elsewhere herein.
  • For instance, receiving the request to authorize the financial transaction may include receiving the request to authorize the financial transaction from a computing system of a merchant associated with the financial transaction. Additionally or alternatively, identifying the computing device at which information associated with the debit or credit card account was entered may include receiving an IP address of the computing device from the computing system of the merchant associated with the financial transaction.
  • Additionally or alternatively, determining the first geographic location may include determining the first geographic location by using the IP address as a key to a location database. Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a third party server.
  • Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a mobile device of the authorized cardholder, storing the geolocation data in a database, and/or retrieving the geolocation data from the database. Additionally or alternatively, receiving the geolocation data from a mobile device of the authorized cardholder may include receiving GPS location data from the mobile device.
  • Additionally or alternatively, determining that the second geographic location does not correspond to the first geographic location may include one or both of determining that the first geographic location and the second geographic location are not within a threshold distance of each other, and/or determining that the first geographic location and the second geographic location are not within a same geographic territory.
  • In another aspect, a computer-implemented method of reducing false positives among geolocation-based fraud alerts issued in connection with online financial transactions may be implemented in one or more servers. The method may include: (1) determining, by one or more processors of the one or more servers, that a fraud alert exists for a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is an online transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account; (2) identifying, by the one or more processors, a computing device at which information associated with the debit or credit card account was entered in connection with the financial transaction; (3) determining, by the one or more processors, a first geographic location at which the computing device resides; (4) determining, by the one or more processors, a time of the financial transaction; (5) determining, by the one or more processors and based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction; (6) determining, by the one or more processors, that the second geographic location corresponds to the first geographic location; and/or (7) in response to determining that the second geographic location corresponds to the first geographic location, marking, by the one or more processors, the fraud alert as a false positive such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction. The method may include additional, fewer or alternative actions, such as any of those discussed elsewhere herein.
  • For instance, determining the first geographic location may occur prior to determining that the fraud alert exists, and determining that the fraud alert exists may include comparing the first geographic location to a set of one or more typical locations of the authorized cardholder, and/or generating the fraud alert in response to determining that the first geographic location does not correspond to the set of one or more typical locations.
  • Additionally or alternatively, the method may further include retrieving, by the one or more processors and from an account records database, transaction data associated with the financial transaction, determining the first geographic location may include identifying the first geographic location based upon location information included in a first field of the transaction data, and/or determining the time of the financial transaction may include identifying the time of the financial transaction based upon time information included in a second field of the transaction data.
  • Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving the geolocation data from a third party server. Additionally or alternatively, determining that the authorized cardholder was at the second geographic location at the time of the financial transaction may include receiving GPS location data from a mobile device of the authorized cardholder, storing the GPS location data in a database, and/or retrieving the GPS location data from the database.
  • Additionally or alternatively, determining that the second geographic location corresponds to the first geographic location may include determining that the first geographic location and the second geographic location are within a threshold distance of each other. Additionally or alternatively, determining that the second geographic location corresponds to the first geographic location may include determining that the first geographic location and the second geographic location are within a same geographic territory.
  • VIII. Exemplary System Embodiments
  • In one aspect, a computer system configured to use customer data to determine that geolocation-based fraud alerts are false positives may include one or more processors and a memory. The memory may store instructions that, when executed by the one or more processors, cause the computer system to: (1) determine that an electronic fraud alert is a geolocation-based fraud alert generated based upon an unexpected or abnormal transaction location; (2) in response to determining that the electronic fraud alert is a geolocation-based fraud alert, obtain, via one or more radio frequency links, customer data from two or more sources; (3) determine that the customer data from the two or more sources indicates that a customer is traveling; (4) in response to determining that the customer data indicates that the customer is traveling, determine that a customer location indicated by the customer data corresponds to the transaction location; and/or (5) in response to determining that the customer location corresponds to the transaction location, (i) mark the electronic fraud alert as a false positive and (ii) cause the electronic fraud alert to not be transmitted to a mobile device of the customer in order to reduce an amount of false positives that are transmitted to customers. The system may include additional, fewer or alternative components, configurations and/or functionality, such as any of those discussed elsewhere herein.
  • For instance, the instructions may cause the computing system to determine that the electronic fraud alert is a geolocation-based fraud alert at least by inputting one or both of (i) the electronic fraud alert, and (ii) transaction data corresponding to a financial transaction associated with the electronic fraud alert, into a machine learning program that is trained to identify geolocation-based fraud alerts.
  • Additionally or alternatively, the instructions may further cause the computing system to obtain, via a wireless communication channel, transaction data corresponding to a financial transaction associated with the electronic fraud alert, and/or input the transaction data into a rules engine to identify the financial transaction as potentially fraudulent and generate the electronic fraud alert. Additionally or alternatively, the customer data may include vehicle telematics data that includes one or more past or current GPS locations.
  • In another aspect, a computer system for reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions may include a location database configured to store geolocation data indicating geographic locations of authorized cardholders over time, one or more processors, and a non-transitory memory. The memory may store instructions that, when executed by the one or more processors, cause the one or more processors to: (1) determine that a fraud alert exists for a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account; (2) determine a first geographic location at which information associated with the debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with the financial transaction; (3) determine a time of the financial transaction; (4) determine, based upon first geolocation data stored in the location database and indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at the time of the financial transaction; (5) determine that the second geographic location corresponds to the first geographic location; and/or (6) in response to determining that the second geographic location corresponds to the first geographic location, mark the fraud alert as a false positive such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction. The system may include additional, fewer or alternative components, configurations and/or functionality, such as any of those discussed elsewhere herein.
  • For instance, the instructions may further cause the one or more processors to retrieve, from an account records database, transaction data associated with the financial transaction, the instructions may cause the one or more processors to determine the first geographic location at least by identifying the first geographic location based upon location information included in a first field of the transaction data, and/or the instructions may cause the one or more processors to determine the time of the financial transaction at least by identifying the time of the financial transaction based upon time information included in a second field of the transaction data.
  • Additionally or alternatively, the instructions may cause the one or more processors to determine that the authorized cardholder was at the second geographic location at the time of the financial transaction at least by receiving the first geolocation data from a mobile device of the authorized cardholder, storing the first geolocation data in the location database, and/or retrieving the first geolocation data from the location database. Additionally or alternatively, receiving the first geolocation data may include GPS location data.
  • Additionally or alternatively, the instructions may cause the one or more processors to determine that the second geographic location corresponds to the first geographic location by at least one of determining that the first geographic location and the second geographic location are within a threshold distance of each other, or determining that the first geographic location and the second geographic location are within a same geographic territory.
  • In another aspect, a computer system for preventing fraudulent online financial transactions may include a location database configured to store geolocation data indicating geographic locations of authorized cardholders over time, one or more processors, and a non-transitory memory. The memory stores instructions that, when executed by the one or more processors, may cause the one or more processors to: (1) receive a request to authorize a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is an online transaction purportedly being entered into by an authorized cardholder associated with the debit or credit card account; (2) identify a computing device at which information associated with the debit or credit card account was entered in connection with the financial transaction; (3) determine a first geographic location at which the computing device resides; (4) determine, based upon geolocation data indicating one or more geographic locations of the authorized cardholder, that the authorized cardholder was at a second geographic location at a time of the financial transaction; (5) determine that the second geographic location does not correspond to the first geographic location; and/or (6) in response to determining that the second geographic location does not correspond to the first geographic location, prevent the financial transaction from being executed. The system may include additional, fewer or alternative components, configurations and/or functionality, such as any of those discussed elsewhere herein.
  • For instance, the instructions may cause the one or more processors to receive the request to authorize the financial transaction from a computing system of a merchant associated with the financial transaction. Additionally or alternatively, the instructions may cause the one or more processors to identify the computing device at which information associated with the debit or credit card account was entered at least by receiving an IP address of the computing device from the computing system of the merchant associated with the financial transaction.
  • Additionally or alternatively, the instructions may cause the one or more processors to determine the first geographic location at least by determining the first geographic location by using the IP address as a key to a location database. Additionally or alternatively, the instructions may cause the one or more processors to determine that the authorized cardholder was at the second geographic location at the time of the financial transaction at least by receiving the geolocation data from either (i) a third party server; or (ii) a mobile device of the authorized cardholder.
  • Additionally or alternatively, the instructions may cause the one or more processors to determine that the second geographic location does not correspond to the first geographic location at least by one or both of determining that the first geographic location and the second geographic location are not within a threshold distance of each other, and determining that the first geographic location and the second geographic location are not within a same geographic territory.
  • IX. Exemplary Computer-Readable Medium Embodiments
  • In one aspect, a non-transitory, computer-readable medium stores instructions that, when executed by one or more processors, may cause the one or more processors to: (1) determine that an electronic fraud alert is a geolocation-based fraud alert generated based upon an unexpected or abnormal transaction location; (2) in response to determining that the electronic fraud alert is a geolocation-based fraud alert, obtain, via one or more radio frequency links, customer data from two or more sources; (3) determine that the customer data from the two or more sources indicates that a customer is traveling; (4) in response to determining that the customer data indicates that the customer is traveling, determine that a customer location indicated by the customer data corresponds to the transaction location; and/or (5) in response to determining that the customer location corresponds to the transaction location, (i) mark the electronic fraud alert as a false positive and (ii) cause the electronic fraud alert to not be transmitted to a mobile device of the customer in order to reduce an amount of false positives that are transmitted to customers. The non-transitory, computer-readable medium may store instructions that include additional, fewer or alternative functions, such as any of those discussed elsewhere herein.
  • For instance, the instructions may cause the computing system to determine that the electronic fraud alert is a geolocation-based fraud alert at least by inputting one or both of (i) the electronic fraud alert, and (ii) transaction data corresponding to a financial transaction associated with the electronic fraud alert, into a machine learning program that is trained to identify geolocation-based fraud alerts.
  • Additionally or alternatively, the instructions may further cause the computing system to obtain, via a wireless communication channel, transaction data corresponding to a financial transaction associated with the electronic fraud alert, and/or input the transaction data into a rules engine to identify the financial transaction as potentially fraudulent and generate the electronic fraud alert. Additionally or alternatively, the customer data may include vehicle telematics data that includes one or more past or current GPS locations.
  • X. Additional Considerations
  • The following additional considerations apply to the foregoing discussion. Throughout this specification, plural instances may implement operations or structures described as a single instance. Although individual operations of one or more methods are illustrated and described as separate operations, one or more of the individual operations may be performed concurrently, and nothing requires that the operations be performed in the order illustrated. These and other variations, modifications, additions, and improvements fall within the scope of the subject matter herein.

Claims (23)

1. A computer-implemented method, implemented in one or more servers, of reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions, the method comprising:
determining, by one or more processors of the one or more servers, a first geographic location at which information associated with a debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account;
determining, by the one or more processors, that a fraud alert exists for the financial transaction, at least in part by
comparing the first geographic location to a set of one or more typical locations of the authorized cardholder, and
generating the fraud alert in response to determining that the first geographic location does not correspond to the set of one or more typical locations; and
based at least partly on generating the fraud alert:
determining, by the one or more processors, a time of the financial transaction,
receiving, by the one or more processors, geolocation data indicating one or more location-based factors including an indication of a flight to which the authorized cardholder checked in,
determining, by the one or more processors, that the authorized cardholder was at a second geographic location at the time of the financial transaction, the second geographic location including a destination of the flight to which the authorized cardholder checked in,
determining, by the one or more processors, that the second geographic location is within the threshold distance of the first geographic location,
receiving additional geolocation data from a mobile device of the authorized cardholder, the additional geolocation data including GPS location data; and
in response to determining that the second geographic location is within the threshold distance of the first geographic location and based at least in part on the additional geolocation data, marking, by the one or more processors, the fraud alert as a false positive such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction.
2. (canceled)
3. The computer-implemented method of claim 1, further comprising:
retrieving, by the one or more processors and from an account records database, transaction data associated with the financial transaction,
wherein determining the first geographic location includes identifying the first geographic location based upon location information included in a first field of the transaction data.
4. The computer-implemented method of claim 3, wherein determining the time of the financial transaction includes identifying the time of the financial transaction based upon time information included in a second field of the transaction data.
5. The computer-implemented method of claim 1, wherein determining that the authorized cardholder was at the second geographic location at the time of the financial transaction includes:
receiving the geolocation data from a third-party server.
6. The computer-implemented method of claim 1, wherein determining that the authorized cardholder was at the second geographic location at the time of the financial transaction includes:
storing the additional geolocation data in a database; and
retrieving the additional geolocation data from the database.
7. The computer-implemented method of claim 1, wherein receiving the additional geolocation data from a mobile device of the authorized cardholder includes:
receiving, from the mobile device, an indication of a location at which the authorized card holder checked in on social media.
8. The computer-implemented method of claim 1, wherein
the second geographic location includes a destination city of the flight to which the authorized cardholder checked in.
9. The computer-implemented method of claim 1, wherein determining that the second geographic location corresponds to the first geographic location includes:
determining that the first geographic location and the second geographic location are within a same geographic territory.
10. A computer system for reducing false positives among geolocation-based fraud alerts issued in connection with card-present financial transactions, the computer system comprising:
a location database configured to store geolocation data indicating geographic locations of authorized cardholders over time;
one or more processors; and
a non-transitory memory storing instructions that, when executed by the one or more processors, cause the one or more processors to
determine a first geographic location at which information associated with a debit or credit card account was obtained by swiping or inserting a debit or credit card in connection with a financial transaction, wherein the financial transaction (i) is associated with a debit or credit card account and (ii) is a card-present transaction purportedly entered into by an authorized cardholder associated with the debit or credit card account,
determine that a fraud alert exists for the financial transaction, at least in part by
comparing the first geographic location to a set of one or more typical locations of the authorized cardholder, and
generating the fraud alert in response to determining that the first geographic location does not correspond to the set of one or more typical locations, and
based at least partly on generating the fraud alert:
determine a time of the financial transaction,
receive the geolocation data indicating one or more location-based factors including an indication of a flight to which the authorized cardholder checked in,
determine that the authorized cardholder was at a second geographic location at the time of the financial transaction, the second geographic location including a destination of the flight to which the authorized cardholder checked in,
determine that the second geographic location is within a threshold distance of the first geographic location,
receive additional geolocation data from a mobile device of the authorized cardholder, the additional geolocation data including GPS location data; and
in response to determining that the second geographic location is within the threshold distance of the first geographic location and based at least in part on the additional geolocation data, mark the fraud alert as a false positive such that no fraud alert is sent to the authorized cardholder in connection with the financial transaction.
11. The computer system of claim 10, wherein the instructions further cause the one or more processors to:
retrieve, from an account records database, transaction data associated with the financial transaction,
wherein the instructions cause the one or more processors to determine the first geographic location at least by identifying the first geographic location based upon location information included in a first field of the transaction data, and
wherein the instructions cause the one or more processors to determine the time of the financial transaction at least by identifying the time of the financial transaction based upon time information included in a second field of the transaction data.
12. The computer system of claim 10, wherein the instructions cause the one or more processors to determine that the authorized cardholder was at the second geographic location at the time of the financial transaction at least by:
storing the additional geolocation data in the location database; and
retrieving the additional geolocation data from the location database.
13. The computer system of claim 10, wherein receiving the additional geolocation data includes receiving, from the mobile device, an indication of a location at which the authorized cardholder checked in on social media.
14. The computer system of claim 12, wherein the second geographic location includes a destination city of the flight to which the authorized cardholder checked in.
15.-20. (canceled)
21. A method comprising:
determining a first location of a requested card-present transaction associated with a card of an authorized cardholder, the card comprising at least one of a debit card or a credit card;
determining that the first location is greater than a first threshold distance from a second location of the authorized cardholder, the second location being a home address of the authorized cardholder;
based at least in part on determining that the first location is greater than the first threshold distance from the second location, identifying a third location of the authorized cardholder at a time of the card-present transaction, wherein the third location is identified based at least in part on geolocation data indicating that the third location is a destination of a flight to which the authorized cardholder checked in;
determining that the third location is greater than a second threshold distance from the first location;
receiving additional geolocation data from a mobile device of the authorized cardholder, the additional geolocation data including GPS location data; and
transmitting a fraud alert to a computing system associated with the authorized cardholder based at least in part on determining that the third location is greater than the second threshold distance from the first location and further based at least in part on the additional geolocation data.
22. (canceled)
23. The method of claim 21, wherein the geolocation data indicates that the authorized cardholder visited a website associated with the third location.
24. The method of claim 21, wherein the geolocation data indicates that the third location comprises a place that was endorsed by the authorized cardholder via a social media account.
25. The method of claim 21, further comprising:
receiving, from a merchant computing system, a request to authorize the card-present transaction,
wherein transmitting the fraud alert comprises transmitting, to the merchant computing system, a message indicating that the card-present transaction is unauthorized.
26. The method of claim 21, further comprising:
generating the fraud alert.
27. The method of claim 21, wherein the geolocation data comprises online activity data of the authorized cardholder.
28. The method of claim 21, wherein the the third location includes a destination city of the flight to which the authorized card holder checked in.
US15/466,002 2016-03-25 2017-03-22 Reducing false positive fraud alerts for card-present financial transactions Abandoned US20210264429A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/466,002 US20210264429A1 (en) 2016-03-25 2017-03-22 Reducing false positive fraud alerts for card-present financial transactions

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US201662313196P 2016-03-25 2016-03-25
US201662318423P 2016-04-05 2016-04-05
US201662331530P 2016-05-04 2016-05-04
US201662365699P 2016-07-22 2016-07-22
US15/466,002 US20210264429A1 (en) 2016-03-25 2017-03-22 Reducing false positive fraud alerts for card-present financial transactions

Publications (1)

Publication Number Publication Date
US20210264429A1 true US20210264429A1 (en) 2021-08-26

Family

ID=73019959

Family Applications (27)

Application Number Title Priority Date Filing Date
US15/465,880 Abandoned US20210264458A1 (en) 2016-03-25 2017-03-22 Preempting or resolving fraud disputes relating to introductory offer expirations
US15/465,981 Abandoned US20220122071A1 (en) 2016-03-25 2017-03-22 Identifying fraudulent instruments and identification
US15/465,827 Active 2037-07-03 US10832248B1 (en) 2016-03-25 2017-03-22 Reducing false positives using customer data and machine learning
US15/465,874 Active 2038-04-22 US10825028B1 (en) 2016-03-25 2017-03-22 Identifying fraudulent online applications
US15/465,858 Active 2039-07-10 US11037159B1 (en) 2016-03-25 2017-03-22 Identifying chargeback scenarios based upon non-compliant merchant computer terminals
US15/465,871 Abandoned US20210158355A1 (en) 2016-03-25 2017-03-22 Preempting or resolving fraud disputes relating to billing aliases
US15/466,014 Active US11334894B1 (en) 2016-03-25 2017-03-22 Identifying false positive geolocation-based fraud alerts
US15/465,977 Active 2037-10-21 US10949852B1 (en) 2016-03-25 2017-03-22 Document-based fraud detection
US15/465,842 Active 2038-09-23 US11170375B1 (en) 2016-03-25 2017-03-22 Automated fraud classification using machine learning
US15/466,009 Abandoned US20210065186A1 (en) 2016-03-25 2017-03-22 Reducing false positive fraud alerts for online financial transactions
US15/465,832 Active 2037-06-12 US10872339B1 (en) 2016-03-25 2017-03-22 Reducing false positives using customer feedback and machine learning
US15/465,856 Abandoned US20210374753A1 (en) 2016-03-25 2017-03-22 Identifying potential chargeback scenarios using machine learning
US15/466,002 Abandoned US20210264429A1 (en) 2016-03-25 2017-03-22 Reducing false positive fraud alerts for card-present financial transactions
US15/465,868 Abandoned US20210374764A1 (en) 2016-03-25 2017-03-22 Facilitating fraud dispute resolution using machine learning
US16/540,505 Active US10949854B1 (en) 2016-03-25 2019-08-14 Reducing false positives using customer feedback and machine learning
US15/931,560 Active US11004079B1 (en) 2016-03-25 2020-05-13 Identifying chargeback scenarios based upon non-compliant merchant computer terminals
US16/899,486 Active US11049109B1 (en) 2016-03-25 2020-06-11 Reducing false positives using customer data and machine learning
US16/913,814 Active US11348122B1 (en) 2016-03-25 2020-06-26 Identifying fraudulent online applications
US17/078,744 Active US11687937B1 (en) 2016-03-25 2020-10-23 Reducing false positives using customer data and machine learning
US17/080,476 Active US11687938B1 (en) 2016-03-25 2020-10-26 Reducing false positives using customer feedback and machine learning
US17/134,901 Active US11699158B1 (en) 2016-03-25 2020-12-28 Reducing false positive fraud alerts for online financial transactions
US17/745,541 Pending US20220351216A1 (en) 2016-03-25 2022-05-16 Identifying false positive geolocation-based fraud alerts
US17/827,015 Active US11741480B2 (en) 2016-03-25 2022-05-27 Identifying fraudulent online applications
US17/993,758 Pending US20230088436A1 (en) 2016-03-25 2022-11-23 Reducing false positives using customer feedback and machine learning
US18/207,069 Pending US20230316285A1 (en) 2016-03-25 2023-06-07 Reducing false positives using customer feedback and machine learning
US18/207,081 Pending US20230316286A1 (en) 2016-03-25 2023-06-07 Reducing false positive fraud alerts for online financial transactions
US18/222,199 Pending US20230394500A1 (en) 2016-03-25 2023-07-14 Identifying fraudulent online applications

Family Applications Before (12)

Application Number Title Priority Date Filing Date
US15/465,880 Abandoned US20210264458A1 (en) 2016-03-25 2017-03-22 Preempting or resolving fraud disputes relating to introductory offer expirations
US15/465,981 Abandoned US20220122071A1 (en) 2016-03-25 2017-03-22 Identifying fraudulent instruments and identification
US15/465,827 Active 2037-07-03 US10832248B1 (en) 2016-03-25 2017-03-22 Reducing false positives using customer data and machine learning
US15/465,874 Active 2038-04-22 US10825028B1 (en) 2016-03-25 2017-03-22 Identifying fraudulent online applications
US15/465,858 Active 2039-07-10 US11037159B1 (en) 2016-03-25 2017-03-22 Identifying chargeback scenarios based upon non-compliant merchant computer terminals
US15/465,871 Abandoned US20210158355A1 (en) 2016-03-25 2017-03-22 Preempting or resolving fraud disputes relating to billing aliases
US15/466,014 Active US11334894B1 (en) 2016-03-25 2017-03-22 Identifying false positive geolocation-based fraud alerts
US15/465,977 Active 2037-10-21 US10949852B1 (en) 2016-03-25 2017-03-22 Document-based fraud detection
US15/465,842 Active 2038-09-23 US11170375B1 (en) 2016-03-25 2017-03-22 Automated fraud classification using machine learning
US15/466,009 Abandoned US20210065186A1 (en) 2016-03-25 2017-03-22 Reducing false positive fraud alerts for online financial transactions
US15/465,832 Active 2037-06-12 US10872339B1 (en) 2016-03-25 2017-03-22 Reducing false positives using customer feedback and machine learning
US15/465,856 Abandoned US20210374753A1 (en) 2016-03-25 2017-03-22 Identifying potential chargeback scenarios using machine learning

Family Applications After (14)

Application Number Title Priority Date Filing Date
US15/465,868 Abandoned US20210374764A1 (en) 2016-03-25 2017-03-22 Facilitating fraud dispute resolution using machine learning
US16/540,505 Active US10949854B1 (en) 2016-03-25 2019-08-14 Reducing false positives using customer feedback and machine learning
US15/931,560 Active US11004079B1 (en) 2016-03-25 2020-05-13 Identifying chargeback scenarios based upon non-compliant merchant computer terminals
US16/899,486 Active US11049109B1 (en) 2016-03-25 2020-06-11 Reducing false positives using customer data and machine learning
US16/913,814 Active US11348122B1 (en) 2016-03-25 2020-06-26 Identifying fraudulent online applications
US17/078,744 Active US11687937B1 (en) 2016-03-25 2020-10-23 Reducing false positives using customer data and machine learning
US17/080,476 Active US11687938B1 (en) 2016-03-25 2020-10-26 Reducing false positives using customer feedback and machine learning
US17/134,901 Active US11699158B1 (en) 2016-03-25 2020-12-28 Reducing false positive fraud alerts for online financial transactions
US17/745,541 Pending US20220351216A1 (en) 2016-03-25 2022-05-16 Identifying false positive geolocation-based fraud alerts
US17/827,015 Active US11741480B2 (en) 2016-03-25 2022-05-27 Identifying fraudulent online applications
US17/993,758 Pending US20230088436A1 (en) 2016-03-25 2022-11-23 Reducing false positives using customer feedback and machine learning
US18/207,069 Pending US20230316285A1 (en) 2016-03-25 2023-06-07 Reducing false positives using customer feedback and machine learning
US18/207,081 Pending US20230316286A1 (en) 2016-03-25 2023-06-07 Reducing false positive fraud alerts for online financial transactions
US18/222,199 Pending US20230394500A1 (en) 2016-03-25 2023-07-14 Identifying fraudulent online applications

Country Status (1)

Country Link
US (27) US20210264458A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11170375B1 (en) 2016-03-25 2021-11-09 State Farm Mutual Automobile Insurance Company Automated fraud classification using machine learning
WO2023081617A1 (en) * 2021-11-04 2023-05-11 Capital One Services, Llc Systems and methods for enhancing transaction data

Families Citing this family (79)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11429675B2 (en) * 2018-06-20 2022-08-30 Mongodb, Inc. Systems and methods for managing transactional operation
US11074537B2 (en) * 2015-12-29 2021-07-27 Workfusion, Inc. Candidate answer fraud for worker assessment
US11210670B2 (en) * 2017-02-28 2021-12-28 Early Warning Services, Llc Authentication and security for mobile-device transactions
US11232364B2 (en) * 2017-04-03 2022-01-25 DataVisor, Inc. Automated rule recommendation engine
US11354749B2 (en) * 2017-04-27 2022-06-07 Sap Se Computing device for machine learning based risk analysis
US10699295B1 (en) * 2017-05-05 2020-06-30 Wells Fargo Bank, N.A. Fraudulent content detector using augmented reality platforms
CN110637321A (en) * 2017-05-16 2019-12-31 维萨国际服务协会 Dynamic claims submission system
SG11201911608WA (en) * 2017-06-13 2020-01-30 Visa Int Service Ass Method, system, and computer program product for controlling transaction velocity limits
US11507953B1 (en) * 2017-11-16 2022-11-22 Worldpay, Llc Systems and methods for optimizing transaction conversion rate using machine learning
US11574204B2 (en) * 2017-12-06 2023-02-07 Accenture Global Solutions Limited Integrity evaluation of unstructured processes using artificial intelligence (AI) techniques
US11829866B1 (en) * 2017-12-27 2023-11-28 Intuit Inc. System and method for hierarchical deep semi-supervised embeddings for dynamic targeted anomaly detection
US11288672B2 (en) * 2017-12-28 2022-03-29 Paypal, Inc. Machine learning engine for fraud detection following link selection
US11457042B1 (en) * 2018-02-27 2022-09-27 Wells Fargo Bank, N.A. Multi-tiered system for detecting and reducing unauthorized network access
US10275613B1 (en) * 2018-04-20 2019-04-30 Capital One Services, Llc Identity breach notification and remediation
US11049112B2 (en) * 2018-05-24 2021-06-29 Paypal, Inc. Systems and methods facilitating chargebacks in electronic transactions
US11250345B1 (en) * 2018-06-08 2022-02-15 Intuit Inc. Methods for identifying transactions with user location information
WO2019245559A2 (en) * 2018-06-21 2019-12-26 Visa International Service Association System and method for detecting and preventing "friendly fraud"
US11144990B1 (en) 2018-06-29 2021-10-12 Square, Inc. Credit offers based on non-transactional data
FR3083356B1 (en) * 2018-06-29 2020-09-11 Ingenico Group PROCESS FOR CARRYING OUT A TRANSACTION, TERMINAL, SERVER AND CORRESPONDING COMPUTER PROGRAM
US11539716B2 (en) * 2018-07-31 2022-12-27 DataVisor, Inc. Online user behavior analysis service backed by deep learning models trained on shared digital information
WO2020105198A1 (en) * 2018-11-21 2020-05-28 孝一 西郷 Advertisement device
US20200167772A1 (en) 2018-11-24 2020-05-28 Walmart Apollo, Llc System and method for detecting signature forgeries
US11157913B2 (en) * 2018-12-28 2021-10-26 Mastercard International Incorporated Systems and methods for improved detection of network fraud events
US11410047B2 (en) * 2018-12-31 2022-08-09 Paypal, Inc. Transaction anomaly detection using artificial intelligence techniques
US11605085B2 (en) * 2019-01-24 2023-03-14 Walmart Apollo, Llc Methods and apparatus for fraud detection
US11355103B2 (en) * 2019-01-28 2022-06-07 Pindrop Security, Inc. Unsupervised keyword spotting and word discovery for fraud analytics
US20200258147A1 (en) * 2019-02-13 2020-08-13 Yuh-Shen Song Intelligent alert system
US20200273039A1 (en) * 2019-02-25 2020-08-27 Jpmorgan Chase Bank, N.A. Systems and methods for automated fraud-type identification and decisioning
US20200279266A1 (en) 2019-03-01 2020-09-03 Mastercard Technologies Canada ULC Multi-page online application origination (oao) service for fraud prevention systems
US11842593B2 (en) * 2019-03-14 2023-12-12 IDEMIA National Security Solutions LLC Systems and methods for detection of counterfeit documents
US11164178B2 (en) * 2019-04-04 2021-11-02 Comenity Llc Adding a credit account to a mobile wallet to make a transaction when the physical card associated with the credit account is unavailable
US10963885B2 (en) * 2019-04-04 2021-03-30 Paypal, Inc. Systems and methods for using machine learning to predict events associated with transactions
US11803925B1 (en) * 2019-04-16 2023-10-31 Danielle Hutchinson System and method for selecting a dispute resolution process
US11741465B2 (en) * 2019-05-02 2023-08-29 Mastercard International Incorporated Systems and methods for generating pre-chargeback dispute records
US10861593B1 (en) * 2019-05-29 2020-12-08 Capital One Service, LLC Utilizing a machine learning model to identify unhealthy online user behavior and to cause healthy physical user behavior
US11276124B2 (en) * 2019-07-02 2022-03-15 Sap Se Machine learning-based techniques for detecting payroll fraud
US11113689B2 (en) * 2019-07-03 2021-09-07 Sap Se Transaction policy audit
US10706423B1 (en) * 2019-07-09 2020-07-07 Capital One Services, Llc Systems and methods for mitigating fraudulent transactions
US11393023B1 (en) 2019-07-19 2022-07-19 Block, Inc. Adaptive multi-stage user interface for credit offers
US11288673B1 (en) * 2019-07-29 2022-03-29 Intuit Inc. Online fraud detection using machine learning models
US11468272B2 (en) 2019-08-15 2022-10-11 Visa International Service Association Method, system, and computer program product for detecting fraudulent interactions
US11631082B2 (en) 2019-09-20 2023-04-18 Walmart Apollo, Llc Methods and apparatus for payment transfer fraud detection
EP4038527A4 (en) * 2019-10-01 2023-10-18 Mastercard Technologies Canada ULC Feature encoding in online application origination (oao) service for a fraud prevention system
US11157914B2 (en) * 2019-12-13 2021-10-26 Visa International Service Association Method, system, and computer program product for processing a potentially fraudulent transaction
US11947643B2 (en) * 2019-12-26 2024-04-02 Rakuten Group, Inc. Fraud detection system, fraud detection method, and program
US11853982B1 (en) 2020-01-30 2023-12-26 Freedom Financial Network, LLC User dashboard for enabling user participation with account management services
US11645543B2 (en) * 2020-01-30 2023-05-09 Visa International Service Association System, method, and computer program product for implementing a generative adversarial network to determine activations
US20210248512A1 (en) * 2020-02-06 2021-08-12 Signal Financial Technologies Llc Intelligent machine learning recommendation platform
US20210295379A1 (en) * 2020-03-17 2021-09-23 Com Olho It Private Limited System and method for detecting fraudulent advertisement traffic
US20210312451A1 (en) * 2020-04-01 2021-10-07 Mastercard International Incorporated Systems and methods for modeling and classification of fraudulent transactions
US11551228B2 (en) * 2020-04-08 2023-01-10 Capital One Services, Llc Localized account freeze for fraudulent transactions
US11687936B2 (en) * 2020-04-27 2023-06-27 SOURCE Ltd. System and method for managing chargeback risk
US11367008B2 (en) * 2020-05-01 2022-06-21 Cognitive Ops Inc. Artificial intelligence techniques for improving efficiency
US20210383392A1 (en) * 2020-06-05 2021-12-09 Capital One Services, Llc Systems and methods for fraud detection and prevention
US11580561B1 (en) * 2020-07-02 2023-02-14 American Express Travel Related Services Company, Inc. Detecting security breaches with watchdog transaction accounts
US11895264B2 (en) * 2020-07-02 2024-02-06 Pindrop Security, Inc. Fraud importance system
US11455324B2 (en) * 2020-10-23 2022-09-27 Settle Smart Ltd. Method for determining relevant search results
US20220148001A1 (en) * 2020-11-06 2022-05-12 Capital One Services, Llc Patching security vulnerabilities using machine learning
US20220172142A1 (en) * 2020-11-30 2022-06-02 Sandra K. Johnson System and methodology for dynamic accounts in dynamic lightweight personalized analytics
US11836727B1 (en) 2020-12-04 2023-12-05 Wells Fargo Bank, N.A. Location based transaction authentication
US20220207352A1 (en) * 2020-12-30 2022-06-30 Capital One Services, Llc Methods and systems for generating recommendations for counterfactual explanations of computer alerts that are automatically detected by a machine learning algorithm
US11887172B2 (en) 2021-01-29 2024-01-30 Walmart Apollo, Llc Methods and apparatus for electronic detection of fraudulent transactions using machine learning processes
US20220255791A1 (en) * 2021-02-08 2022-08-11 Verizon Patent And Licensing Inc. Systems and methods for reducing a quantity of false positives associated with rule-based alarms
US20220253856A1 (en) * 2021-02-11 2022-08-11 The Toronto-Dominion Bank System and method for machine learning based detection of fraud
US11595446B2 (en) * 2021-04-19 2023-02-28 Tekion Corp Identifying suspicious entries in a document management system
US20220358508A1 (en) * 2021-05-08 2022-11-10 Mastercard International Incorporated Methods and systems for predicting account-level risk scores of cardholders
US20220366513A1 (en) * 2021-05-14 2022-11-17 Jpmorgan Chase Bank, N.A. Method and apparatus for check fraud detection through check image analysis
US20230033368A1 (en) * 2021-07-28 2023-02-02 Capital One Services, Llc Transaction Based Authentication with Item-Level Data
US11875335B2 (en) 2021-08-10 2024-01-16 Capital One Services, Llc Determining merchant enforced transaction rules
US11605126B1 (en) 2021-11-29 2023-03-14 Wells Fargo Bank, N.A. Detecting fraud in credit applications
US20230177512A1 (en) * 2021-12-08 2023-06-08 Chime Financial, Inc. Generating a fraud prediction utilizing a fraud-prediction machine-learning model
US20230186308A1 (en) * 2021-12-09 2023-06-15 Chime Financial, Inc. Utilizing a fraud prediction machine-learning model to intelligently generate fraud predictions for network transactions
US20230196369A1 (en) * 2021-12-20 2023-06-22 International Business Machines Corporation Identifying suspicious behavior based on patterns of digital identification documents
US20230230088A1 (en) * 2022-01-06 2023-07-20 Socure, Inc. Method and System of Predictive Document Verification and Machine Learning Therefor
WO2023168222A1 (en) * 2022-03-03 2023-09-07 Worldpay, Llc Systems and methods for predictive analysis of electronic transaction representment data using machine learning
US20230298028A1 (en) * 2022-03-18 2023-09-21 Fidelity Information Services, Llc Analyzing a transaction in a payment processing system
US20230306433A1 (en) * 2022-03-24 2023-09-28 Bank Of America Corporation Cognitive Identification of Credit Reporting Disputes and Dispute Resolution Using Quantum Computing
US20230325841A1 (en) * 2022-04-07 2023-10-12 Gen Digital Inc. Systems and methods for detecting websites that perpetrate at least one of scams or frauds
US20230334505A1 (en) * 2022-04-19 2023-10-19 Capital One Services, Llc Processing of customer messages to avoid unneccessary fraud disputes

Family Cites Families (299)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5774882A (en) 1992-03-12 1998-06-30 Keen; Regina D. Credit approval system
US7251624B1 (en) 1992-09-08 2007-07-31 Fair Isaac Corporation Score based decisioning
US5819226A (en) 1992-09-08 1998-10-06 Hnc Software Inc. Fraud detection using predictive modeling
US5444763A (en) 1993-06-17 1995-08-22 Research In Motion Limited Translation and connection device for radio frequency point of sale transaction systems
US5748780A (en) * 1994-04-07 1998-05-05 Stolfo; Salvatore J. Method and apparatus for imaging, image processing and data compression
US5467341A (en) 1994-04-14 1995-11-14 Toshiba America Information Systems, Inc. Apparatus and method for alerting computer users in a wireless LAN of a service area transition
US5708422A (en) 1995-05-31 1998-01-13 At&T Transaction authorization and alert system
US5825863A (en) 1995-12-11 1998-10-20 Walker Asset Management Limited Partnership Prepaid limited usage calling card
US8162125B1 (en) 1996-05-29 2012-04-24 Cummins-Allison Corp. Apparatus and system for imaging currency bills and financial documents and method for using the same
US6094643A (en) 1996-06-14 2000-07-25 Card Alert Services, Inc. System for detecting counterfeit financial card fraud
US6018723A (en) 1997-05-27 2000-01-25 Visa International Service Association Method and apparatus for pattern generation
US6119103A (en) 1997-05-27 2000-09-12 Visa International Service Association Financial risk prediction systems and methods therefor
US7403922B1 (en) 1997-07-28 2008-07-22 Cybersource Corporation Method and apparatus for evaluating fraud risk in an electronic commerce transaction
US6269169B1 (en) 1998-07-17 2001-07-31 Imaging Automation, Inc. Secure document reader and method therefor
US6170744B1 (en) 1998-09-24 2001-01-09 Payformance Corporation Self-authenticating negotiable documents
US6301579B1 (en) 1998-10-20 2001-10-09 Silicon Graphics, Inc. Method, system, and computer program product for visualizing a data structure
US7058597B1 (en) 1998-12-04 2006-06-06 Digital River, Inc. Apparatus and method for adaptive fraud screening for electronic commerce transactions
US6430539B1 (en) 1999-05-06 2002-08-06 Hnc Software Predictive modeling of consumer financial behavior
US6533168B1 (en) 1999-05-27 2003-03-18 Peter N. Ching Method and apparatus for computer-readable purchase receipts using multi-dimensional bar codes
US6437812B1 (en) 1999-06-30 2002-08-20 Cerebrus Solutions Limited Graphical user interface and method for displaying hierarchically structured information
US6215358B1 (en) 1999-09-16 2001-04-10 Samsung Electronics Co., Ltd. Temperature compensated bias network for a power amplifier and method of operation
US7735721B1 (en) 1999-11-30 2010-06-15 Diebold Self-Service Systems Division Of Diebold, Incorporated Method of evaluating checks deposited into a cash dispensing automated banking machine
US7428984B1 (en) 1999-11-30 2008-09-30 Diebold Self-Service Systems Division Of Diebold, Incorporated Method and system of evaluating checks deposited into a cash dispensing automated banking machine
US7494052B1 (en) 1999-11-30 2009-02-24 Diebold Self-Service Systems Division Of Diebold, Incorporated Method of evaluating checks deposited into a cash dispensing automated banking machine
US7377425B1 (en) 1999-11-30 2008-05-27 Diebold Self-Service Systems Division Of Diebold, Incorporated Method and system of evaluating checks deposited into a cash dispensing automated banking machine
US7263506B2 (en) 2000-04-06 2007-08-28 Fair Isaac Corporation Identification and management of fraudulent credit/debit card purchases at merchant ecommerce sites
GB0029229D0 (en) 2000-11-30 2001-01-17 Unisys Corp Counter measures for irregularities in financial transactions
US20020147694A1 (en) 2001-01-31 2002-10-10 Dempsey Derek M. Retraining trainable data classifiers
US7089592B2 (en) * 2001-03-15 2006-08-08 Brighterion, Inc. Systems and methods for dynamic detection and prevention of electronic fraud
US7676430B2 (en) 2001-05-09 2010-03-09 Lenovo (Singapore) Ptd. Ltd. System and method for installing a remote credit card authorization on a system with a TCPA complaint chipset
US7865427B2 (en) 2001-05-30 2011-01-04 Cybersource Corporation Method and apparatus for evaluating fraud risk in an electronic commerce transaction
US7707108B2 (en) 2002-01-31 2010-04-27 International Business Machines Corporation Detection of unauthorized account transactions
US20030182194A1 (en) 2002-02-06 2003-09-25 Mark Choey Method and system of transaction card fraud mitigation utilizing location based services
US20030172036A1 (en) 2002-03-05 2003-09-11 Idan Feigenbaum Online financial transaction veracity assurance mechanism
DE60311904D1 (en) 2002-03-15 2007-04-05 Computer Sciences Corp Methods and apparatus for analyzing writing in documents
US7356516B2 (en) 2002-06-13 2008-04-08 Visa U.S.A. Inc. Method and system for facilitating electronic dispute resolution
US6960990B2 (en) 2002-10-31 2005-11-01 General Motors Corporation Telematics vehicle security system and method
US7870078B2 (en) 2002-11-01 2011-01-11 Id Insight Incorporated System, method and computer program product for assessing risk of identity theft
US7143095B2 (en) 2002-12-31 2006-11-28 American Express Travel Related Services Company, Inc. Method and system for implementing and managing an enterprise identity management for distributed security
US7548886B2 (en) 2003-06-12 2009-06-16 International Business Machines Corporation System and method for early detection and prevention of identity theft
US20110225064A1 (en) 2003-09-02 2011-09-15 Augustine Fou Methods and systems for using universally unique item identifiers
US20050097046A1 (en) 2003-10-30 2005-05-05 Singfield Joy S. Wireless electronic check deposit scanning and cashing machine with web-based online account cash management computer application system
US20050097051A1 (en) * 2003-11-05 2005-05-05 Madill Robert P.Jr. Fraud potential indicator graphical interface
US6948656B2 (en) 2003-12-23 2005-09-27 First Data Corporation System with GPS to manage risk of financial transactions
US20050137982A1 (en) 2003-12-23 2005-06-23 Leslie Michelassi Systems and methods for determining a reconcilement result
US7716135B2 (en) 2004-01-29 2010-05-11 International Business Machines Corporation Incremental compliance environment, an enterprise-wide system for detecting fraud
US7577938B2 (en) 2004-02-20 2009-08-18 Microsoft Corporation Data association
US7537153B2 (en) 2004-05-03 2009-05-26 De La Rue International, Limited Method and computer program product for electronically managing payment media
US8885894B2 (en) 2004-06-14 2014-11-11 Michael John Rowen Reduction of transaction fraud through the use of automatic centralized signature/sign verification combined with credit and fraud scoring during real-time payment card authorization processes
WO2006023822A2 (en) 2004-08-18 2006-03-02 Kappa Image Llc Validating negotiable documents using public document validation profiles
US20060041464A1 (en) 2004-08-19 2006-02-23 Transunion Llc. System and method for developing an analytic fraud model
US8914309B2 (en) 2004-08-20 2014-12-16 Ebay Inc. Method and system for tracking fraudulent activity
US20060047725A1 (en) 2004-08-26 2006-03-02 Bramson Steven J Opt-in directory of verified individual profiles
US7480631B1 (en) * 2004-12-15 2009-01-20 Jpmorgan Chase Bank, N.A. System and method for detecting and processing fraud and credit abuse
US20130104251A1 (en) 2005-02-01 2013-04-25 Newsilike Media Group, Inc. Security systems and methods for use with structured and unstructured data
JP4512512B2 (en) 2005-03-29 2010-07-28 大王製紙株式会社 Absorbent article and surface sheet thereof
US20120253805A1 (en) * 2005-04-21 2012-10-04 Anthony Rajakumar Systems, methods, and media for determining fraud risk from audio signals
US8073691B2 (en) 2005-04-21 2011-12-06 Victrio, Inc. Method and system for screening using voice data and metadata
AU2006242555A1 (en) 2005-04-29 2006-11-09 Oracle International Corporation System and method for fraud monitoring, detection, and tiered user authentication
WO2006130874A2 (en) 2005-06-02 2006-12-07 Fair Isaac Corporation Comprehensive identity protection system
WO2007028048A2 (en) 2005-09-02 2007-03-08 Fair Isaac Corporation Systems and methods for detecting fraud
WO2007041709A1 (en) 2005-10-04 2007-04-12 Basepoint Analytics Llc System and method of detecting fraud
US7552865B2 (en) 2005-10-20 2009-06-30 Satyam Computer Services Ltd. System and method for deep interaction modeling for fraud detection
US8346638B2 (en) 2005-10-26 2013-01-01 Capital One Financial Corporation Systems and methods for processing transaction data to perform a merchant chargeback
US20070179840A1 (en) 2005-11-10 2007-08-02 American Express Marketing & Development Corp. Joint redemption account
US20070179849A1 (en) 2006-02-02 2007-08-02 Microsoft Corporation Ad publisher performance and mitigation of click fraud
US20070187491A1 (en) 2006-02-13 2007-08-16 Godwin Bryan W Processing Cashless Transactions of Remote Field Assets
US7693767B2 (en) 2006-03-23 2010-04-06 Oracle International Corporation Method for generating predictive models for a business problem via supervised learning
US7912773B1 (en) 2006-03-24 2011-03-22 Sas Institute Inc. Computer-implemented data storage systems and methods for use with predictive model systems
US7788195B1 (en) 2006-03-24 2010-08-31 Sas Institute Inc. Computer-implemented predictive model generation systems and methods
US7818264B2 (en) 2006-06-19 2010-10-19 Visa U.S.A. Inc. Track data encryption
US8650080B2 (en) 2006-04-10 2014-02-11 International Business Machines Corporation User-browser interaction-based fraud detection system
WO2007120844A2 (en) 2006-04-12 2007-10-25 Chimento Marc A System and method for screening for fraud in commercial transactions
US20070275399A1 (en) 2006-05-04 2007-11-29 Lathrop Richard H Methods for calculating codon pair-based translational kinetics values, and methods for generating polypeptide-encoding nucleotide sequences from such values
US20070100773A1 (en) 2006-08-11 2007-05-03 Regions Asset Company Transaction security system having user defined security parameters
US8688543B2 (en) 2006-08-29 2014-04-01 Visa International Service Association Method and system for processing and authenticating internet purchase transactions
US20080288299A1 (en) 2006-10-31 2008-11-20 Genmobi Technologies, Inc. System and method for user identity validation for online transactions
US8600789B1 (en) * 2006-11-01 2013-12-03 Bank Of America Corporation System and method for processing offending items in a financial system
US20080106726A1 (en) 2006-11-02 2008-05-08 Ellis Park Currency detection & tracking system and method
US20130197998A1 (en) 2012-01-26 2013-08-01 Finsphere Corporation Authenticating entities engaging in automated or electronic transactions or activities
US9922323B2 (en) 2007-03-16 2018-03-20 Visa International Service Association System and method for automated analysis comparing a wireless device location with another geographic location
US20110208601A1 (en) 2010-02-19 2011-08-25 Finshpere Corporation System and method for financial transaction authentication using travel information
US7962418B1 (en) 2007-03-30 2011-06-14 Amazon Technologies, Inc. System and method of fulfilling a transaction
US20090018940A1 (en) 2007-03-30 2009-01-15 Liang Wang Enhanced Fraud Detection With Terminal Transaction-Sequence Processing
US8078515B2 (en) 2007-05-04 2011-12-13 Michael Sasha John Systems and methods for facilitating electronic transactions and deterring fraud
US10769290B2 (en) 2007-05-11 2020-09-08 Fair Isaac Corporation Systems and methods for fraud detection via interactive link analysis
US20090018934A1 (en) 2007-05-15 2009-01-15 Chaorong Peng System and Method for defense ID theft attack security service system in marketing environment
US7575157B2 (en) 2007-05-22 2009-08-18 Bank Of America Corporation Fraud protection
US7739169B2 (en) 2007-06-25 2010-06-15 Visa U.S.A. Inc. Restricting access to compromised account information
US20090083080A1 (en) 2007-09-24 2009-03-26 Tim Brooks Method, apparatus and program product for facilitating transfer of group meeting contracts
GB2467665A (en) * 2007-10-05 2010-08-11 Basepoint Analytics Llc Methods and systems of predicting mortgage payment risk
US8666841B1 (en) 2007-10-09 2014-03-04 Convergys Information Management Group, Inc. Fraud detection engine and method of using the same
US20090099884A1 (en) 2007-10-15 2009-04-16 Mci Communications Services, Inc. Method and system for detecting fraud based on financial records
US8255318B2 (en) 2007-10-18 2012-08-28 First Data Corporation Applicant authentication
US9779403B2 (en) 2007-12-07 2017-10-03 Jpmorgan Chase Bank, N.A. Mobile fraud prevention system and method
US7992746B2 (en) 2008-02-11 2011-08-09 Carefusion 303, Inc. Method and apparatus for removing, inserting and securing receptacles in a receptacle tray
US7857212B1 (en) 2008-02-14 2010-12-28 Capital One Financial Corporation Method and system for authorizing card account transactions by geographic region
US20140200929A1 (en) 2008-04-02 2014-07-17 Yougetitback Limited Systems and methods for dynamically assessing and mitigating risk of an insured entity
US9985978B2 (en) 2008-05-07 2018-05-29 Lookingglass Cyber Solutions Method and system for misuse detection
US20140324677A1 (en) 2008-05-19 2014-10-30 Jpmorgan Chase Bank, N.A. Method and system for detecting, monitoring and investigating first party fraud
US10410220B2 (en) 2008-06-12 2019-09-10 Guardian Analytics, Inc. Fraud detection and analysis system
US20160063501A1 (en) 2008-06-18 2016-03-03 Saraansh Software Solutions Pvt. Ltd. System for detecting banking frauds by examples
US8478692B2 (en) 2008-06-26 2013-07-02 Visa International Service Association Systems and methods for geographic location notifications of payment transactions
US8848924B2 (en) 2008-06-27 2014-09-30 University Of Washington Privacy-preserving location tracking for devices
US8229853B2 (en) 2008-07-24 2012-07-24 International Business Machines Corporation Dynamic itinerary-driven profiling for preventing unauthorized card transactions
JP5206197B2 (en) 2008-07-28 2013-06-12 富士通株式会社 Rule learning method, program and apparatus
US20100106611A1 (en) 2008-10-24 2010-04-29 Uc Group Ltd. Financial transactions systems and methods
US20100114774A1 (en) 2008-11-04 2010-05-06 Moneygram International, Inc. Chargeback decisioning system
CA2742963A1 (en) 2008-11-06 2010-05-14 Visa International Service Association Online challenge-response
US8566235B2 (en) 2008-12-23 2013-10-22 Verifi, Inc. System and method for providing dispute resolution for electronic payment transactions
US8145561B1 (en) 2009-01-05 2012-03-27 Sprint Communications Company L.P. Phone usage pattern as credit card fraud detection trigger
US8140418B1 (en) 2009-01-09 2012-03-20 Apple Inc. Cardholder-not-present authorization
CA2752875A1 (en) 2009-02-20 2010-08-26 Moqom Limited Merchant alert system and method for fraud prevention
US20100274691A1 (en) 2009-04-28 2010-10-28 Ayman Hammad Multi alerts based system
US8712912B2 (en) 2009-04-28 2014-04-29 Visa International Service Association System and method for providing advice to consumer regarding a payment transaction
CA2760769A1 (en) 2009-05-04 2010-11-11 Visa International Service Association Determining targeted incentives based on consumer transaction history
US20100293090A1 (en) 2009-05-14 2010-11-18 Domenikos Steven D Systems, methods, and apparatus for determining fraud probability scores and identity health scores
US8600873B2 (en) 2009-05-28 2013-12-03 Visa International Service Association Managed real-time transaction fraud analysis and decisioning
US8745698B1 (en) 2009-06-09 2014-06-03 Bank Of America Corporation Dynamic authentication engine
US10290053B2 (en) * 2009-06-12 2019-05-14 Guardian Analytics, Inc. Fraud detection and analysis
US8955747B2 (en) * 2009-06-23 2015-02-17 At&T Mobility Ii Llc Devices, systems and methods for wireless point-of-sale
US10242540B2 (en) 2009-09-02 2019-03-26 Fair Isaac Corporation Visualization for payment card transaction fraud analysis
US8620798B2 (en) 2009-09-11 2013-12-31 Visa International Service Association System and method using predicted consumer behavior to reduce use of transaction risk analysis and transaction denials
US8805737B1 (en) 2009-11-02 2014-08-12 Sas Institute Inc. Computer-implemented multiple entity dynamic summarization systems and methods
US10089683B2 (en) 2010-02-08 2018-10-02 Visa International Service Association Fraud reduction system for transactions
US20110238566A1 (en) 2010-02-16 2011-09-29 Digital Risk, Llc System and methods for determining and reporting risk associated with financial instruments
US8413234B1 (en) 2010-02-17 2013-04-02 Sprint Communications Company L.P. Communications-service fraud detection using special social connection
US8626663B2 (en) * 2010-03-23 2014-01-07 Visa International Service Association Merchant fraud risk score
US8600875B2 (en) 2010-04-12 2013-12-03 Visa International Service Association Authentication process using search technology
US8543522B2 (en) 2010-04-21 2013-09-24 Retail Decisions, Inc. Automatic rule discovery from large-scale datasets to detect payment card fraud using classifiers
US20150173733A1 (en) 2010-07-29 2015-06-25 Greatbatch Ltd. Retractor Tool For Minimally Invasive Hip Replacement Surgery
CN102346894B (en) * 2010-08-03 2017-03-01 阿里巴巴集团控股有限公司 The output intent of recommendation information, system and server
US8528054B2 (en) 2010-08-31 2013-09-03 Yahoo! Inc. Multi-step challenge-response test
WO2012051582A2 (en) 2010-10-14 2012-04-19 Visa International Service Association Transaction alerting in a multi-network environment
US8306889B2 (en) 2010-10-20 2012-11-06 Fis Financial Compliance Solutions, Llc System and method for presenting fraud detection information
US20120109821A1 (en) 2010-10-29 2012-05-03 Jesse Barbour System, method and computer program product for real-time online transaction risk and fraud analytics and management
US9552470B2 (en) 2010-11-29 2017-01-24 Biocatch Ltd. Method, device, and system of generating fraud-alerts for cyber-attacks
US8773564B2 (en) 2010-12-14 2014-07-08 Truesense Imaging, Inc. Image sensor with charge multiplication
WO2012088344A1 (en) 2010-12-21 2012-06-28 Visa International Service Association Transaction rate processing apparatuses, methods and systems
US20120173570A1 (en) 2011-01-05 2012-07-05 Bank Of America Corporation Systems and methods for managing fraud ring investigations
US10373160B2 (en) 2011-02-10 2019-08-06 Paypal, Inc. Fraud alerting using mobile phone location
US8346217B2 (en) 2011-02-21 2013-01-01 Knowledge Solutions, LLC Systems, methods and apparatus for controlling access to mobile devices
US8401522B2 (en) 2011-02-21 2013-03-19 Carmela R. Crawford Systems, methods and apparatus for authenticating access to enterprise resources
US8458069B2 (en) 2011-03-04 2013-06-04 Brighterion, Inc. Systems and methods for adaptive identification of sources of fraud
AU2012202173B2 (en) 2011-04-18 2013-09-05 Castle Bookkeeping Wizard Pty Ltd System and method for processing a transaction document including one or more financial transaction entries
US20130013491A1 (en) * 2011-04-19 2013-01-10 Early Warning Sevices, LLC System and method for detecting and mitigating duplicate transaction fraud
US20120278249A1 (en) 2011-04-29 2012-11-01 American Express Travel Related Services Company, Inc. Generating an Identity Theft Score
US20160232465A1 (en) 2011-06-03 2016-08-11 Kenneth Kurtz Subscriber-based system for custom evaluations of business relationship risk
US8620806B2 (en) 2011-07-13 2013-12-31 Mastercard International Incorporated Merchant data cleansing in clearing record
US8606712B2 (en) 2011-07-21 2013-12-10 Bank Of America Corporation Multi-stage filtering for fraud detection with account event data filters
US20130024358A1 (en) 2011-07-21 2013-01-24 Bank Of America Corporation Filtering transactions to prevent false positive fraud alerts
US8447674B2 (en) 2011-07-21 2013-05-21 Bank Of America Corporation Multi-stage filtering for fraud detection with customer history filters
US20130046692A1 (en) 2011-08-19 2013-02-21 Bank Of America Corporation Fraud protection with user location verification
US8930295B2 (en) 2011-09-12 2015-01-06 Stanley Victor CAMPBELL Systems and methods for monitoring and analyzing transactions
US9916538B2 (en) 2012-09-15 2018-03-13 Z Advanced Computing, Inc. Method and system for feature detection
EP2575099A1 (en) 2011-09-30 2013-04-03 Tata Consultancy Services Limited Electronic funds transfer
US20140052621A1 (en) 2011-10-03 2014-02-20 Ap Technology, Llc Merchant electronic check payments
US8478688B1 (en) 2011-12-19 2013-07-02 Emc Corporation Rapid transaction processing
US20130218758A1 (en) 2012-02-16 2013-08-22 Andrew John Bruno Naumann zu Koenigsbrueck Custom scorecard and hybrid fraud model
GB2500212A (en) 2012-03-13 2013-09-18 Validsoft Uk Ltd Method for location based authentication of transaction
US9619852B2 (en) 2012-04-17 2017-04-11 Zighra Inc. Context-dependent authentication system, method and device
US9799020B2 (en) * 2012-04-27 2017-10-24 Mastercard International Incorporated Method for providing payment card security using registrationless telecom geolocation capture
US9953326B2 (en) 2012-05-02 2018-04-24 Jpmorgan Chase Bank, N.A. Alert optimization system and method
WO2013166501A1 (en) 2012-05-04 2013-11-07 Visa International Service Association System and method for local data conversion
US8484132B1 (en) 2012-06-08 2013-07-09 Lexisnexis Risk Solutions Fl Inc. Systems and methods for segmented risk scoring of identity fraud
US9691066B2 (en) 2012-07-03 2017-06-27 Verifone, Inc. Location-based payment system and method
US8762268B2 (en) 2012-07-05 2014-06-24 Index Systems, Inc. Electronic commerce network with transactions analytics
US20140012738A1 (en) * 2012-07-09 2014-01-09 Bennett Woo Methods and systems for measuring accuracy in fraudulent transaction identification
US20140058962A1 (en) 2012-08-22 2014-02-27 Municipality Real Time Payment Systems, Llc Method and system for processing a transaction, effecting payment to a municipality and receiving a certificate of payment
US9519903B2 (en) 2012-08-29 2016-12-13 24/7 Customer, Inc. Method and apparatus for proactive notifications based on the location of a user
US20140067656A1 (en) 2012-09-06 2014-03-06 Shlomo COHEN GANOR Method and system for fraud risk estimation based on social media information
US9208676B2 (en) * 2013-03-14 2015-12-08 Google Inc. Devices, methods, and associated information processing for security in a smart-sensored home
US20140101050A1 (en) 2012-10-04 2014-04-10 Ethoca Technologies, Inc. Do-not-recognize transaction handling
US9940615B2 (en) 2012-10-16 2018-04-10 Fleetcor Technologies Operating Company, Llc Automated pairing of payment products and mobile to mobile devices
US20140114840A1 (en) * 2012-10-19 2014-04-24 Cellco Partnership D/B/A Verizon Wireless Automated fraud detection
US9953321B2 (en) * 2012-10-30 2018-04-24 Fair Isaac Corporation Card fraud detection utilizing real-time identification of merchant test sites
US9836795B2 (en) 2012-11-08 2017-12-05 Hartford Fire Insurance Company Computerized system and method for pre-filling of insurance data using third party sources
US20150286827A1 (en) 2012-12-03 2015-10-08 Nadia Fawaz Method and apparatus for nearly optimal private convolution
US8623321B1 (en) 2012-12-12 2014-01-07 Uop Llc UZM-44 aluminosilicate zeolite
US10504163B2 (en) * 2012-12-14 2019-12-10 Mastercard International Incorporated System for payment, data management, and interchanges for use with global shopping cart
US9043887B2 (en) 2012-12-31 2015-05-26 Apple Inc. Adaptive secondary authentication criteria based on account data
US10262330B2 (en) 2013-01-04 2019-04-16 PlaceIQ, Inc. Location-based analytic platform and methods
US20140207637A1 (en) 2013-01-23 2014-07-24 Mastercard International Incorporated System and Method for Using Credit/Debit Card Transaction Data to Determine Financial Health of a Merchant
US20140207674A1 (en) 2013-01-24 2014-07-24 Mastercard International Incorporated Automated teller machine transaction premium listing to prevent transaction blocking
US20140244503A1 (en) 2013-02-27 2014-08-28 Mastercard International Incorporated System and method for automatic thresholding for payment card spend control
US20140250011A1 (en) 2013-03-01 2014-09-04 Lance Weber Account type detection for fraud risk
US8983868B1 (en) 2013-03-08 2015-03-17 Google Inc. Using location information in electronic commerce
US9286618B2 (en) 2013-03-08 2016-03-15 Mastercard International Incorporated Recognizing and combining redundant merchant designations in a transaction database
US20140279494A1 (en) 2013-03-12 2014-09-18 Mastercard International Incorporated Method and system of detecting and using geofencing for fraud detection and modeling
US20140279503A1 (en) 2013-03-13 2014-09-18 Bank Of America Corporation Providing customer alerts based on geo-thresholds
US9231979B2 (en) 2013-03-14 2016-01-05 Sas Institute Inc. Rule optimization for classification and detection
US9311685B2 (en) 2013-03-14 2016-04-12 Bank Of America Corporation Geolocation check-in system
US20140279312A1 (en) 2013-03-15 2014-09-18 Capital One Financial Corporation System and method for providing automated chargeback operations
US9633322B1 (en) 2013-03-15 2017-04-25 Consumerinfo.Com, Inc. Adjustment of knowledge-based authentication
US11232447B2 (en) 2013-03-15 2022-01-25 Allowify Llc System and method for enhanced transaction authorization
US20140310160A1 (en) 2013-04-11 2014-10-16 Pawan Kumar Alert System with Multiple Transaction Indicators
BR112015025691A2 (en) 2013-04-12 2017-07-18 Mastercard International Inc analytic rules engine for payment processing system
US20140337217A1 (en) 2013-05-09 2014-11-13 Mastercard International Incorporated Card present fraud prevention method using airline passenger detail
US20140358592A1 (en) 2013-05-31 2014-12-04 OneEvent Technologies, LLC Sensors for usage-based property insurance
US9438576B2 (en) 2013-06-12 2016-09-06 Luiz M Franca-Neto Apparatus and method for validation and authorization of device and user by global positioning and non-prompted exchange of information
US9811830B2 (en) 2013-07-03 2017-11-07 Google Inc. Method, medium, and system for online fraud prevention based on user physical location data
US11367073B2 (en) 2013-07-03 2022-06-21 Capital One Services, Llc System and method for fraud control
US20150032604A1 (en) 2013-07-23 2015-01-29 Mastercard International Incorporated Merchant data cleansing in clearing record based on targeted merchants
US20150032621A1 (en) 2013-07-24 2015-01-29 Mastercard International Incorporated Method and system for proximity fraud control
US20150039504A1 (en) 2013-07-31 2015-02-05 Cachet Financial Solutions Inc. Check verification and remote deposit capture
US20150046220A1 (en) 2013-08-12 2015-02-12 Mastercard International Incorporated Predictive model of travel intentions using purchase transaction data method and apparatus
CN105723401A (en) 2013-08-16 2016-06-29 Md保存公司 Network-based marketplace service for facilitating purchases of bundled services and products
US20150058119A1 (en) 2013-08-22 2015-02-26 SocialWire, Inc. Automated Advertisement of Products on Online Sites
CA2860179A1 (en) 2013-08-26 2015-02-26 Verafin, Inc. Fraud detection systems and methods
US20160071104A1 (en) 2013-09-04 2016-03-10 George Gregory Stamatis Securebuy merchant information analytics decision engine
US20150081349A1 (en) 2013-09-13 2015-03-19 Visa International Service Association Systems and Methods to Provide Location Indication in Transaction Data
US9607318B1 (en) 2013-09-19 2017-03-28 Intuit Inc. Method and system for providing relevant sale event notifications using financial transaction data and location data
WO2015048335A1 (en) 2013-09-26 2015-04-02 Dragnet Solutions, Inc. Document authentication based on expected wear
US20150106260A1 (en) 2013-10-11 2015-04-16 G2 Web Services System and methods for global boarding of merchants
US9148869B2 (en) 2013-10-15 2015-09-29 The Toronto-Dominion Bank Location-based account activity alerts
US9727866B2 (en) 2013-10-15 2017-08-08 Intuit Inc. Methods systems and computer program products for verifying consumer identity during transaction
US20150199738A1 (en) 2014-01-14 2015-07-16 Elwha Llc Guaranty investigation
US20150120502A1 (en) 2013-10-29 2015-04-30 Elwha LLC, a limited liability corporation of the State of Delaware Supporting guaranty provisioning via user attribute proffering
US9818105B2 (en) 2013-10-29 2017-11-14 Elwha Llc Guaranty provisioning via wireless service purveyance
US9934498B2 (en) 2013-10-29 2018-04-03 Elwha Llc Facilitating guaranty provisioning for an exchange
US10586234B2 (en) * 2013-11-13 2020-03-10 Mastercard International Incorporated System and method for detecting fraudulent network events
US9936346B2 (en) 2013-11-28 2018-04-03 Microsoft Technology Licensing, Llc Geofences from context and crowd-sourcing
US9483765B2 (en) 2013-12-09 2016-11-01 Mastercard International Incorporated Systems and methods for monitoring payment transactions for fraud using social media
US20150161611A1 (en) 2013-12-10 2015-06-11 Sas Institute Inc. Systems and Methods for Self-Similarity Measure
US20160203490A1 (en) 2013-12-10 2016-07-14 Sas Institute Inc. Systems and Methods for Travel-Related Anomaly Detection
US10115107B2 (en) 2013-12-23 2018-10-30 International Business Machines Corporation Payment card fraud protection
US9330416B1 (en) 2013-12-30 2016-05-03 Emc Corporation Visualization of fraud patterns
WO2015103216A1 (en) 2014-01-02 2015-07-09 Visa International Service Association Location obfuscation for authentication
US20150193768A1 (en) 2014-01-09 2015-07-09 Capital One Financial Corporation Method and system for providing alert messages related to suspicious transactions
US9947055B1 (en) 2014-01-29 2018-04-17 Intuit Inc. System and method for monitoring merchant transactions using aggregated financial data
US20150227934A1 (en) 2014-02-11 2015-08-13 Mastercard International Incorporated Method and system for determining and assessing geolocation proximity
US20150039513A1 (en) 2014-02-14 2015-02-05 Brighterion, Inc. User device profiling in transaction authentications
US20150235221A1 (en) 2014-02-19 2015-08-20 Bank Of America Corporation Proof-of-verification network for third party issuers
US20150242856A1 (en) 2014-02-21 2015-08-27 International Business Machines Corporation System and Method for Identifying Procurement Fraud/Risk
US9786015B1 (en) 2014-02-27 2017-10-10 Intuit Inc. System and method for fraud detection using aggregated financial data
US10032168B2 (en) 2014-03-07 2018-07-24 Fmr Llc Secure validation of financial transactions
US20150262184A1 (en) 2014-03-12 2015-09-17 Microsoft Corporation Two stage risk model building and evaluation
US9552582B2 (en) 2014-03-21 2017-01-24 Ca, Inc. Controlling ecommerce authentication with non-linear analytical models
US9472194B2 (en) 2014-03-21 2016-10-18 Wells Fargo Bank, N.A. Enhanced fraud detection
US9581342B2 (en) 2014-03-28 2017-02-28 Google Inc. Mounting stand for multi-sensing environmental control device
US20180053114A1 (en) 2014-10-23 2018-02-22 Brighterion, Inc. Artificial intelligence for context classifier
US10896421B2 (en) 2014-04-02 2021-01-19 Brighterion, Inc. Smart retail analytics and commercial messaging
US9569767B1 (en) 2014-05-06 2017-02-14 Square, Inc. Fraud protection based on presence indication
EP3036704A4 (en) 2014-05-28 2017-04-26 Emmanuel Gonzalez User profile parameters for financial accounts
US9210156B1 (en) 2014-06-16 2015-12-08 Lexisnexis Risk Solutions Inc. Systems and methods for multi-stage identity authentication
US9934517B2 (en) 2014-06-17 2018-04-03 American Express Travel Related Services Company, Inc. Systems and methods for processing transactions
US11410176B2 (en) 2014-06-27 2022-08-09 Tigergraph, Inc. System and method for enhanced detection of fraudulent electronic transactions
US10963810B2 (en) 2014-06-30 2021-03-30 Amazon Technologies, Inc. Efficient duplicate detection for machine learning data sets
US9509705B2 (en) 2014-08-07 2016-11-29 Wells Fargo Bank, N.A. Automated secondary linking for fraud detection systems
EP2988259A1 (en) 2014-08-22 2016-02-24 Accenture Global Services Limited Intelligent receipt scanning and analysis
JP6630347B2 (en) 2014-09-03 2020-01-15 ナントヘルス,インコーポレーテッド Synthetic genomic variant-based secure transaction devices, systems, and methods
US9418365B2 (en) 2014-09-08 2016-08-16 Mastercard International Incorporated Systems and methods for using social network data to determine payment fraud
US9280774B1 (en) 2014-09-09 2016-03-08 Bank Of America Corporation System and method for investigating fraudulent activity
US20160078444A1 (en) 2014-09-16 2016-03-17 Mastercard International Incorporated Systems and methods for providing fraud indicator data within an authentication protocol
US9460071B2 (en) 2014-09-17 2016-10-04 Sas Institute Inc. Rule development for natural language processing of text
US10140615B2 (en) 2014-09-22 2018-11-27 Visa International Service Association Secure mobile device credential provisioning using risk decision non-overrides
US9298901B1 (en) 2014-10-08 2016-03-29 International Business Machines Corporation Credential validation using multiple computing devices
US10572877B2 (en) 2014-10-14 2020-02-25 Jpmorgan Chase Bank, N.A. Identifying potentially risky transactions
US20160086185A1 (en) 2014-10-15 2016-03-24 Brighterion, Inc. Method of alerting all financial channels about risk in real-time
US20160071017A1 (en) 2014-10-15 2016-03-10 Brighterion, Inc. Method of operating artificial intelligence machines to improve predictive model training and performance
US20160063502A1 (en) 2014-10-15 2016-03-03 Brighterion, Inc. Method for improving operating profits with better automated decision making with artificial intelligence
US10304042B2 (en) 2014-11-06 2019-05-28 Early Warning Services, Llc Location-based authentication of transactions conducted using mobile devices
US10037532B2 (en) 2014-11-13 2018-07-31 Mastercard International Incorporated Systems and methods for detecting transaction card fraud based on geographic patterns of purchases
JP6717198B2 (en) 2014-11-26 2020-07-01 日立化成株式会社 Photosensitive resin composition, photosensitive element, cured product, semiconductor device, resist pattern forming method, and circuit substrate manufacturing method
WO2016090322A1 (en) 2014-12-04 2016-06-09 Cubic Corporation Credit and debit fraud card usage monitoring for transit
US9412108B2 (en) 2014-12-11 2016-08-09 Mastercard International Incorporated Systems and methods for fraud detection by transaction ticket size pattern
US9858575B2 (en) 2014-12-16 2018-01-02 At&T Mobility Ii Llc Fraud detection via mobile device location tracking
US11093845B2 (en) 2015-05-22 2021-08-17 Fair Isaac Corporation Tree pathway analysis for signature inference
US9472098B2 (en) 2015-01-15 2016-10-18 International Business Machines Corporation Vehicle-based abnormal travel event detecting and reporting
IN2015CH00232A (en) 2015-01-15 2015-09-18 Wipro Ltd
US9552578B2 (en) 2015-02-25 2017-01-24 Mastercard International Incorporated Method and system for authentication of payment card transactions
US20150213276A1 (en) 2015-02-28 2015-07-30 Brighterion, Inc. Addrressable smart agent data structures
SG10201502201UA (en) 2015-03-20 2016-10-28 Mastercard Asia Pacific Pte Ltd Method and system for identifying linked cards from authorization records
US20160292666A1 (en) 2015-03-31 2016-10-06 Mastercard International Incorporated Method and system for determining and assessing geolocation proximity
US20160300214A1 (en) 2015-04-08 2016-10-13 Elizabeth Chaffin Methods and systems for automated matter resolution
US9721253B2 (en) 2015-05-06 2017-08-01 Forter Ltd. Gating decision system and methods for determining whether to allow material implications to result from online activities
US20170011382A1 (en) 2015-07-10 2017-01-12 Fair Isaac Corporation Mobile attribute time-series profiling analytics
US9392008B1 (en) 2015-07-23 2016-07-12 Palantir Technologies Inc. Systems and methods for identifying information related to payment card breaches
US10043071B1 (en) 2015-07-30 2018-08-07 Morphotrust Usa, Llc Automated document classification
US9485265B1 (en) 2015-08-28 2016-11-01 Palantir Technologies Inc. Malicious activity detection system capable of efficiently processing data accessed from databases and generating alerts for display in interactive user interfaces
WO2017044836A1 (en) 2015-09-09 2017-03-16 Pay with Privacy, Inc. Systems and methods for automatically securing and validating multi-server electronic communications over a plurality of networks
EP3506613A1 (en) 2015-10-14 2019-07-03 Pindrop Security, Inc. Call detail record analysis to identify fraudulent activity and fraud detection in interactive voice response systems
US10373140B1 (en) * 2015-10-26 2019-08-06 Intuit Inc. Method and system for detecting fraudulent bill payment transactions using dynamic multi-parameter predictive modeling
US10504122B2 (en) 2015-12-17 2019-12-10 Mastercard International Incorporated Systems and methods for predicting chargebacks
US10325436B2 (en) 2015-12-31 2019-06-18 Hand Held Products, Inc. Devices, systems, and methods for optical validation
US10419401B2 (en) 2016-01-08 2019-09-17 Capital One Services, Llc Methods and systems for securing data in the public cloud
US20170221062A1 (en) 2016-01-28 2017-08-03 Verifi, Inc. Order insights system and method
US20170243220A1 (en) 2016-01-29 2017-08-24 Mastercard International Incorporated Systems and methods for blocking ineligible fraud-related chargebacks
ES2917187T3 (en) * 2016-02-04 2022-07-07 Amadeus Sas Monitoring of user authenticity in distributed systems
US10475034B2 (en) 2016-02-12 2019-11-12 Square, Inc. Physical and logical detections for fraud and tampering
US20170270526A1 (en) 2016-03-15 2017-09-21 Hrb Innovations, Inc. Machine learning for fraud detection
US20210264458A1 (en) * 2016-03-25 2021-08-26 State Farm Mutual Automobile Insurance Company Preempting or resolving fraud disputes relating to introductory offer expirations
US11037158B2 (en) * 2016-03-29 2021-06-15 Microsoft Technology Licensing, Llc Bulk dispute challenge system
US10891655B1 (en) 2016-05-05 2021-01-12 State Farm Mutual Automobile Insurance Company Cognitive computing for generating targeted offers to inactive account holders
US10846308B2 (en) 2016-07-27 2020-11-24 Anomalee Inc. Prioritized detection and classification of clusters of anomalous samples on high-dimensional continuous and mixed discrete/continuous feature spaces
US20180060839A1 (en) 2016-08-25 2018-03-01 Mastercard International Incorporated Systems and methods for predicting chargeback stages
US10896422B2 (en) 2016-12-01 2021-01-19 Mastercard International Incorporated Systems and methods for detecting collusion between merchants and cardholders
US11238528B2 (en) 2016-12-22 2022-02-01 American Express Travel Related Services Company, Inc. Systems and methods for custom ranking objectives for machine learning models applicable to fraud and credit risk assessments
US10586152B2 (en) 2017-02-16 2020-03-10 Honeywell International Inc. Determining image forensics using gradient statistics at edges

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11170375B1 (en) 2016-03-25 2021-11-09 State Farm Mutual Automobile Insurance Company Automated fraud classification using machine learning
US11334894B1 (en) 2016-03-25 2022-05-17 State Farm Mutual Automobile Insurance Company Identifying false positive geolocation-based fraud alerts
US11348122B1 (en) * 2016-03-25 2022-05-31 State Farm Mutual Automobile Insurance Company Identifying fraudulent online applications
US11687937B1 (en) 2016-03-25 2023-06-27 State Farm Mutual Automobile Insurance Company Reducing false positives using customer data and machine learning
US11687938B1 (en) 2016-03-25 2023-06-27 State Farm Mutual Automobile Insurance Company Reducing false positives using customer feedback and machine learning
US11699158B1 (en) 2016-03-25 2023-07-11 State Farm Mutual Automobile Insurance Company Reducing false positive fraud alerts for online financial transactions
US11741480B2 (en) 2016-03-25 2023-08-29 State Farm Mutual Automobile Insurance Company Identifying fraudulent online applications
WO2023081617A1 (en) * 2021-11-04 2023-05-11 Capital One Services, Llc Systems and methods for enhancing transaction data
US11886417B2 (en) 2021-11-04 2024-01-30 Capital One Services, Llc Systems and methods for enhancing transaction data

Also Published As

Publication number Publication date
US10949854B1 (en) 2021-03-16
US20220122071A1 (en) 2022-04-21
US10832248B1 (en) 2020-11-10
US11741480B2 (en) 2023-08-29
US11049109B1 (en) 2021-06-29
US20210374764A1 (en) 2021-12-02
US10949852B1 (en) 2021-03-16
US20210158355A1 (en) 2021-05-27
US11348122B1 (en) 2022-05-31
US20230316286A1 (en) 2023-10-05
US11699158B1 (en) 2023-07-11
US20230394500A1 (en) 2023-12-07
US20210264458A1 (en) 2021-08-26
US20210374753A1 (en) 2021-12-02
US11170375B1 (en) 2021-11-09
US20230088436A1 (en) 2023-03-23
US11037159B1 (en) 2021-06-15
US10825028B1 (en) 2020-11-03
US10872339B1 (en) 2020-12-22
US20210065186A1 (en) 2021-03-04
US20220351216A1 (en) 2022-11-03
US11334894B1 (en) 2022-05-17
US20230316285A1 (en) 2023-10-05
US20220366433A1 (en) 2022-11-17
US11004079B1 (en) 2021-05-11
US11687937B1 (en) 2023-06-27
US11687938B1 (en) 2023-06-27

Similar Documents

Publication Publication Date Title
US11699158B1 (en) Reducing false positive fraud alerts for online financial transactions
US11321784B2 (en) Methods and systems for automatically detecting fraud and compliance issues in expense reports and invoices
US11023889B2 (en) Enhanced merchant identification using transaction data
US9552578B2 (en) Method and system for authentication of payment card transactions
US20140258169A1 (en) Method and system for automated verification of customer reviews
US20180330384A1 (en) Systems and methods for processing customer purchase transactions using biometric data
US20160292666A1 (en) Method and system for determining and assessing geolocation proximity
US20170161745A1 (en) Payment account fraud detection using social media heat maps
US11928722B2 (en) Item level data determination device, method, and non-transitory computer-readable media
CN111461859A (en) Loan pairing system and method
US20200160427A1 (en) Systems and methods for aggregating, exchanging, and filtering data over a communications network
US20230316284A1 (en) Reducing false positives using customer data and machine learning
US11954739B2 (en) Methods and systems for automatically detecting fraud and compliance issues in expense reports and invoices
CN114021990A (en) Method, system, apparatus and medium for assessing risk of vehicle accident

Legal Events

Date Code Title Description
AS Assignment

Owner name: STATE FARM MUTUAL AUTOMOBILE INSURANCE COMPANY, ILLINOIS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KRAMME, TIMOTHY;FLOWERS, ELIZABETH;BATRA, REENA;AND OTHERS;SIGNING DATES FROM 20170308 TO 20170320;REEL/FRAME:041707/0040

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION