WO2020205258A1 - System and method for mitigating cyber security threats - Google Patents

System and method for mitigating cyber security threats Download PDF

Info

Publication number
WO2020205258A1
WO2020205258A1 PCT/US2020/023557 US2020023557W WO2020205258A1 WO 2020205258 A1 WO2020205258 A1 WO 2020205258A1 US 2020023557 W US2020023557 W US 2020023557W WO 2020205258 A1 WO2020205258 A1 WO 2020205258A1
Authority
WO
WIPO (PCT)
Prior art keywords
risk
behaviors
factor
risk factor
factors
Prior art date
Application number
PCT/US2020/023557
Other languages
French (fr)
Inventor
Nadir IZRAEL
Shiri Ladelsky Lellouch
Misha Seltzer
Original Assignee
Armis Security Ltd.
Armis Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Armis Security Ltd., Armis Inc. filed Critical Armis Security Ltd.
Priority to CA3135483A priority Critical patent/CA3135483A1/en
Priority to CN202080039806.5A priority patent/CN114270347A/en
Priority to EP20784480.4A priority patent/EP3948600A4/en
Publication of WO2020205258A1 publication Critical patent/WO2020205258A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/50Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
    • G06F21/57Certifying or maintaining trusted computer platforms, e.g. secure boots or power-downs, version controls, system software checks, secure updates or assessing vulnerabilities
    • G06F21/577Assessing vulnerabilities and evaluating computer system security
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/50Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
    • G06F21/55Detecting local intrusion or implementing counter-measures
    • G06F21/552Detecting local intrusion or implementing counter-measures involving long-term monitoring or reporting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/50Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
    • G06F21/55Detecting local intrusion or implementing counter-measures
    • G06F21/554Detecting local intrusion or implementing counter-measures involving event detection and direct action
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/14Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic
    • H04L63/1408Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic by monitoring network traffic
    • H04L63/1425Traffic logging, e.g. anomaly detection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/14Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic
    • H04L63/1433Vulnerability analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2221/00Indexing scheme relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F2221/03Indexing scheme relating to G06F21/50, monitoring users, programs or devices to maintain the integrity of platforms
    • G06F2221/034Test or assess a computer or a system

Definitions

  • the present disclosure relates generally to cyber security, and more particularly to securing networks against threats posed by malicious devices.
  • Some solutions for securing organizations’ computing infrastructures against malicious devices include requiring authentication for new devices logging onto the network and blocking unknown devices from accessing the network.
  • requiring authentication does not necessarily prevent a malicious entity from obtaining the necessary credentials and accessing the network despite the authentication protocols.
  • blocking all unknown devices may result in benign devices being blocked.
  • previously benign devices may become malicious when, for example, infected by malware. Thus, such malware-infected devices may be allowed network access that may be utilized for malicious purposes.
  • Other solutions for securing organizations’ computing infrastructures include monitoring network activity to detect anomalies.
  • existing detection tools may not detect certain types of anomalous behavior, for example a new attack pattern used as part of a zero-day attack that is not yet recognized by the detection tools.
  • devices that do not pose a direct threat to the network e.g., devices that are not directly engaging in malicious behavior but provide network access to other malicious devices
  • Certain embodiments disclosed herein include a method for method for mitigating cyber security threats by devices using risk factors.
  • the method comprises: determining a plurality of risk factors for a device based on a plurality of risk behaviors indicated by network activity and information of the device; determining a risk score for the device based on the plurality of risk factors and a plurality of weights, wherein each of the plurality of weights is applied to one of the plurality of risk factors; and performing at least one mitigation action based on the risk score.
  • Certain embodiments disclosed herein also include a non-transitory computer readable medium having stored thereon causing a processing circuitry to execute a process, the process comprising: determining a plurality of risk factors for a device based on a plurality of risk behaviors indicated by network activity and information of the device; determining a risk score for the device based on the plurality of risk factors and a plurality of weights, wherein each of the plurality of weights is applied to one of the plurality of risk factors; and performing at least one mitigation action based on the risk score.
  • Certain embodiments disclosed herein also include a system for mitigating cyber security threats by devices using risk factors.
  • the system comprises: a processing circuitry; and a memory, the memory containing instructions that, when executed by the processing circuitry, configure the system to: determine a plurality of risk factors for a device based on a plurality of risk behaviors indicated by network activity and information of the device; determine a risk score for the device based on the plurality of risk factors and a plurality of weights, wherein each of the plurality of weights is applied to one of the plurality of risk factors; and perform at least one mitigation action based on the risk score.
  • Figure 1 is a network diagram utilized to describe the various disclosed embodiments.
  • Figure 2 is a flowchart illustrating a method for mitigating cyber security threats by devices using risk factors according to an embodiment.
  • Figure 3 is a schematic diagram illustrating a threat mitigator according to an embodiment.
  • assessment of potential threats should be flexible such that activity which may not present a direct, known cybersecurity threat can be detected.
  • a printer connected to a network may act as a backdoor by broadcasting an unprotected Wi-Fi signal.
  • the printer itself may not present any cyber threats that would be detected by cyber threat detection tools but may allow access to the network by other malicious devices.
  • a smart television that does not have appropriate antivirus software or a smartphone with out-of-date software may present cyber security threats even when the device itself has not yet started acting maliciously.
  • the disclosed embodiments allow for rapid detection and mitigation of potential cyber security threats by devices.
  • Risk factors used according to the disclosed embodiments provide a more flexible approach to detecting potentially malicious devices than existing solutions while maintaining fast reaction times.
  • the risk factors allow for detecting potentially malicious devices in cases such as, but not limited to, devices engaging in cyber-attacks that are not yet known to the detecting system, devices engaged in pre-attack phase (e.g., exploration, infection, or dormant stages) activities, devices that do not directly present risks (e.g., devices providing backdoor network access to other devices or devices that may be susceptible to exploits by other devices and systems), and completely unknown devices for which no activity data is yet available.
  • the various disclosed embodiments include a method and system for mitigating cyber security threats by devices using risk factors.
  • a device to be checked for risk is detected.
  • the detected device may be a device that is connected to a network, a device that is physically proximate to network infrastructure (e.g., to a router), or a device that is otherwise visible to a network.
  • Risk factors are determined for the device based on risk-related behaviors (hereinafter “risk behaviors”). Each risk factor is determined based on one or more observed risk behaviors, one or more assumed risk behaviors, or a combination thereof.
  • the observed risk behaviors are behaviors indicated by data gathered from the device (e.g., device configuration data, protocol data, signal strength data, etc.), indicated by data gathered with respect to activity by the device on the network, or both.
  • the assumed risk behaviors are extrapolated based on contextual data related to the device such as other devices that have accessed the network, public information (e.g., information about a company that manufactures or designs the device, information related to software installed on the device, etc.), entropy of data in traffic to or from the device, or a combination thereof.
  • a risk score is determined for the device.
  • the risk score may be a weighted average of the determined risk factors.
  • the weight applied to each of the risk factors may be predetermined and may further vary based on portions of the network being accessed, a type of device, the specific device, and the like.
  • the mitigation actions include actively interfering with the device connecting to or acting on the network when the risk score is above a threshold and passively monitoring activity by the device when the risk score is below a threshold.
  • Some of the risk factors may be weighted negatively such that the negative risk factors reduce the risk score for the device.
  • the presence of antivirus software installed on the device may result in a negative weight applied to a risk factor for cybersecurity software installed on the device.
  • the negative risk factors allow for a more holistic view of risk as compared to some existing solutions that determine devices to be either safe or not safe when certain conditions are met.
  • the disclosed embodiments include determining risk factors based on risk behaviors.
  • a number may be high for purposes of a risk factor if, for example, the number is above a threshold.
  • the threshold may be changed over time, for example as normal device activity changes.
  • Various of the following examples also mention more, higher, older, or other relative statements.
  • the value of the risk factor may increase as the respective number or degree increases.
  • An attack surface exposure risk factor may be determined based on risk behaviors such as, but not limited to, vulnerabilities, common traffic patterns, threat intelligence, vulnerabilities to traffic, open ports, use of specific protocols, access of cloud domains, wireless protocols, open hotspots, and any other external access provided by a device.
  • Risk behaviors that would result in a higher attack surface exposure risk factor may include, but are not limited to, a high number of open ports, wireless communications, hotspots, or combinations thereof; use of unencrypted protocols; traffic patterns that deviate from the common traffic patterns; threat intelligence for the device indicating that the device likely presents a cyber security threat; and a high number of known vulnerabilities.
  • a cloud synchronization risk factor may be determined based on risk behaviors for the device with respect to cloud services accessed by the device such as, but not limited to, a number of cloud services used by the device, an amount of data transmitted by the device to cloud services, a number of tunnels formed between the device and cloud environments, types of data transmitted by the device, known reputations of domains of cloud environments accessed by the device, and the like.
  • Risk behaviors that would result in a higher cloud synchronization risk factor may include, but are not limited to, a high number of cloud endpoints, unencrypted data or credentials, accessing suspicious cloud domains, and the like.
  • a connection security risk factor may be determined based on risk behaviors for the device with respect to security and potential vulnerabilities of connections by the device such as, but not limited to, vulnerabilities and threat intelligence related to protocols used by the device as well as connections data (e.g., numbers of other distinct devices to which the device is connected, whether connections are encrypted, etc.). Risk behaviors that would result in a higher connection security risk factor may include, but are not limited to, a high number of connections to distinct devices, use of unencrypted connections, use of known potentially vulnerable protocols, and the like.
  • a boundary evasion risk factor may be determined based on risk behaviors for the device with respect to the device connecting to multiple trusted boundaries such as, but not limited to, a number of connections to distinct boundaries, a number of connections to sensitive boundaries (e.g., boundaries of corporate networks), known standard device configurations, and the like.
  • Risk behaviors that would result in a higher boundary evasion risk factor may include, but are not limited to, more connections to distinct boundaries, more connections to sensitive boundaries, a device connecting to more than one network when standard device configurations for similar devices suggest that the device shouldn’t connect to more than one network, and the like.
  • a third party application stores risk factor may be determined based on risk behaviors for the device with respect to the number and risk of third party application stores accessed by the device such as, but not limited to, number of third party application stores accessed, reputations of domains hosting third party application stores accessed, security features (e.g., firewalls, traffic blocks, etc.) of accessed third party application stores, and the like.
  • Risk behaviors that would result in a higher third party application stores risk factor may include, but are not limited to, more third party application stores accessed, higher risk remote domains as indicated by threat intelligence with respect to known attacks, a lack of detected firewalls or traffic blocks, and the like.
  • a malicious domains risk factor may be determined based on risk behaviors for the device with respect to the number and risk of domains accessed by the device such as, but not limited to, a number of known malicious or suspicious domains accessed, higher risk remote domains as indicated by threat intelligence with respect to known attacks, security features (e.g., firewalls, traffic blocks, etc.) of accessed domains, and the like.
  • Risk behaviors that would result in a higher malicious domains risk factor may include, but are not limited to, more connections to known malicious or suspicious domains, higher risk remote domains as indicated by threat intelligence with respect to known attacks, a lack of detected firewalls or traffic blocks, and the like.
  • a vulnerability history risk factor may be determined based on risk behaviors for the device with respect to the number and severity of vulnerabilities detected for the device such as, but not limited to, number of detected vulnerabilities, severity of vulnerabilities, whether vulnerabilities are remotely exploitable, whether mitigations have been performed by the device, and the like. Risk behaviors that would result in a higher vulnerability history risk factor may include, but are not limited to, more vulnerabilities, higher risk vulnerabilities, higher risk remotely exploitable vulnerabilities, lack of mitigations, and the like.
  • a data-at-rest risk factor may be determined based on risk behaviors for the device with respect to hoarding or storing data by the device such as, but not limited to, amounts of data sent to and received by the device, importance or sensitivity of data sent to the device, whether the device has an encrypted disk, and the like.
  • Risk behaviors that would result in a higher data-at-rest risk factor may include, but are not limited to, more data going into the device than coming out, more important or sensitive traffic going to the device, the device lacking an unencrypted disk, and the like.
  • Example important or sensitive data may include, but are not limited to, customer relationship management data, device data, scan data, patient data, data indicating personally identifiable information, and the like.
  • An external connectivity risk factor may be determined based on risk behaviors for the device with respect to opening of external connections by the device such as, but not limited to, number of hotspots, number of open wireless protocols, whether access by the device is encrypted, whether access by the device requires authentication, known vulnerabilities of any external connections, and the like. Risk behaviors that would result in a higher external connectivity risk factor may include, but are not limited to, more hotspots, more open wireless protocols, unencrypted access, unauthenticated access, known vulnerabilities for external connections, and the like.
  • a user authentication risk factor may be determined based on risk behaviors for the device with respect to distinct user authentications on the device and security of credentials for those user authentications such as, but not limited to, number of users using the same device, whether credentials are encrypted, whether authenticating users conform to a known organizational structure, and the like.
  • Risk behaviors that would result in a higher user authentication risk factor may include, but are not limited to, more users using the same device, unencrypted credentials, users that do not conform to the structure of users in the organization, and the like.
  • a software version risk factor may be determined based on risk behaviors for the device with respect to the number and age of operating systems and software applications installed on the device such as, but not limited to, age of applications, age of operating systems, number of applications, number of operating systems, version numbers of applications, version numbers of operating systems, applications or operating systems having known higher risk remotely exploitable vulnerabilities, lack of mitigations of cyber threats by operating systems or software, and the like.
  • Risk behaviors that would result in a higher software version risk factor may include, but are not limited to, older applications or operating systems, more applications or operating systems, applications or operating systems having higher risk remotely exploitable vulnerabilities, lack of detected mitigations, and the like.
  • a certificate reuse risk factor may be determined based on risk behaviors for the device with respect to reuse of certificates by the device such as, but not limited to, a number of devices using the same certificate, whether certificates used by the device are user based or device based, makes and models of devices sharing the certificate, and the like. Risk behaviors that would result in a higher certificate reuse risk factor may include, but are not limited to, more devices using the same certificate, use of user based certificates, sharing of a certificate by devices having different makes and models, and the like.
  • a manufacturer reputation risk factor may be determined based on risk behaviors related to the manufacturer of the device such as, but not limited to, a number of known breaches affecting the manufacturer, known reputations of geographical locations of origin for the manufacturer, a known number of vulnerabilities across devices made by the manufacturer, and the like. Risk behaviors that would result in a higher manufacturer reputation risk factor may include, but are not limited to, higher numbers of known breaches affecting the manufacturer, poor reputation of country of origin, higher number of vulnerabilities of devices made by the manufacturer, and the like.
  • a device model reputation risk factor may be determined based on risk behaviors related to the model of the device such as, but not limited to, a degree of commonality of the device model (e.g., a relative number of users or owners of the devices having the same model), known threat intelligence for the model of the device, number of known vulnerabilities for the device model, and the like.
  • Risk behaviors that would result in a higher device model reputation risk factor may include, but are not limited to, the device model being less common, threat intelligence on the device model suggesting that the device model is likely unsafe, a high number of vulnerabilities for the device model, and the like.
  • Fig. 1 shows an example network diagram 100 utilized to describe the various disclosed embodiments.
  • a threat mitigator 120 is deployed such that it can access a network 1 10.
  • the network 1 10 may be, but is not limited to, a wireless, cellular or wired network, a local area network (LAN), a wide area network (WAN), a metro area network (MAN), similar networks, and any combinations thereof.
  • a device 130 accesses (shown) or is deployed in physical proximity (not shown) to the network 1 10.
  • the device 130 may be, but is not limited to, a personal computer, a laptop, a tablet computer, a smartphone, a wearable computing device, a printer, or any other device connected to the network 1 10 or deployed in physical proximity to network infrastructure (e.g., a router, not shown) of the network 1 10.
  • network infrastructure e.g., a router, not shown
  • the network 1 10 includes a database 1 1 1 and one or more network activity detection tools 1 12.
  • the threat mitigator 120 is configured to access the database 11 1 , the detection tools 1 12, or both, to obtain data related to risk behaviors to be used for determining risk factors.
  • the database 11 1 may store contextual data related to assumed risk behaviors such as, but not limited to, manufacturer reputation information, device model reputation information, number of common vulnerabilities and exposures associated with products by the manufacturer or designer, number of employees of the manufacturer or designer, popular operating systems, and the like.
  • the database 1 11 may also store data collected by the network activity detection tools 1 12 such that the threat mitigator 1 12 may retrieve such information from the database 1 11.
  • the database 1 11 may further store data related to known device behaviors that may be utilized to determine risk factors.
  • the database 1 1 1 may act as a knowledgebase of known device behavior profiles.
  • the data related to known device behaviors may define baseline behaviors for a device representing normal behavior and values (or formulas used for calculating values) of risk factors based on deviations from the baseline behaviors.
  • a baseline behavior for a security camera may be communicating with a single internal server on the network at a fixed geographical location.
  • Risk factors for the security camera may be defined for behaviors including communicating with more than one server, communicating with an external server, ceasing communications with the server, or changing geographic location.
  • the manufacturer reputation information may include information related to previous devices connected to the network, public source information related to the size and notable security breaches for the company, or both.
  • the information related to previous devices may be based on risk scores determined for other devices having the same manufacturer such that high risk scores of such previous devices will increase the resulting risk factor.
  • the size and security breach information may include, for example, a number of CVEs, a severity of each CVE, and a size of the company such that a high number of CVEs and high severity of CVEs relative to the number of employees of the company will result in a higher risk factor.
  • the popular operating systems information may include common operating systems seen installed on other devices, minimum and maximum (i.e., earliest and latest or least secure and most secure) operating system versions available for a device, or both. Devices that do not use common operating systems result in higher risk factors. Devices having operating system versions closer to the minimum version result in higher risk factors than devices having operating system versions closest to the maximum.
  • the detection tools 1 12 are configured to collect data related to the device, network activity by the device 130, or both.
  • data may include data related to observed risk behaviors such as, but is not limited to, data included in traffic to or from the device 130, amounts of traffic sent by the device 130, number of endpoints receiving traffic from the device 130, type of traffic sent by the device 130 (e.g., encrypted or unencrypted, repetitive or non-repeating, etc.) common vulnerabilities and exposure exhibited by the device 130 (e.g., for the device 130, for software running on the device 130, or both), domains and Internet Protocol (IP) addressed accessed by the device 130, types and versions of software installed on the device 130, type and version of an operating system installed on the device 130, amount and type of external communication options (e.g., ports, protocols, amount of service set identifiers broadcasted, amount of different antennae, etc.), geographical location of the device, geographical movements of the device, and the like.
  • IP Internet Protocol
  • the network activity data may be collected with respect to the
  • the threat mitigator 120 is configured to determine a risk score for the device 130 and to perform mitigation actions based on the determined risk score.
  • the risk score is determined based on risk factors.
  • the risk factors are determined based on risk behaviors including observed risk behaviors and assumed risk behaviors.
  • the observed risk behaviors may be indicated in data collected from the device 130, from the network activity detection tools 1 12 with respect to network activity by the device 130, or both.
  • the assumed risk behaviors are extrapolated based on contextual data related to the device such as, but not limited to, activity by other devices (not shown) accessing the network 1 10, public information related to the device (e.g., information about a manufacturer of the device, an assumed operating system used by the device 130, data transmitted to or by the device, etc.), or both.
  • Each risk factor may be determined based on observed risk behaviors, assumed risk behaviors, or a combination thereof.
  • the disclosed embodiments are not limited to the particular layout shown in Fig. 1.
  • the threat mitigator 120 is illustrated in Fig. 1 as being deployed outside the network 1 10, but that the threat mitigator 120 may be equally deployed in the network 1 10 without departing from the scope of the disclosure.
  • the threat mitigator 120 and the detection tools 1 12 are shown separately for simplicity purposes, but that the threat mitigator 120 may be included in one of the detection tools 1 12 or otherwise act as one of the detection tools 1 12 without departing from the scope of the disclosure.
  • Fig. 2 is an example flowchart 200 illustrating a method for mitigating cyber security threats by devices using risk factors according to an embodiment.
  • the method is performed by the threat mitigator 120.
  • a device to be checked for risk is detected.
  • the detection may include detecting a connection of a device to a network, detection of a device turning on in physical proximity to network infrastructure, or detection of a device otherwise becoming visible to the network (e.g., a device that is already turned on moving within a threshold physical distance from any part of the network infrastructure).
  • a device that was already visible to the network may be checked for risk.
  • a device may be continuously analyzed as described with respect to the following steps from connection to disconnection from the network.
  • the data related to the device includes data directly related to the device (e.g., configuration data of the device, identifying information of the device, etc.) and network activity by the device (e.g., data collected from the device or gathered by monitoring activity of the device) and data indicating contextual information related to the device.
  • data directly related to the device e.g., configuration data of the device, identifying information of the device, etc.
  • network activity by the device e.g., data collected from the device or gathered by monitoring activity of the device
  • the collected data may include, but is not limited to, network activity data (e.g., data indicating traffic coming into or out of the device, traffic data (i.e., data transmitted by or to the device), volume of traffic, portions of the network to which the device has connected, etc.), identifying information of the device (e.g., a name, model number, identifier, a manufacturer of the device, etc.), end-of-life or end-of-service-life data, software data related to the device (e.g., programs installed on the device), connection data (e.g., open ports, wireless communications, hotspots, number of networks the device is connected to as an endpoint, whether connections are encrypted, domains and Internet protocol addresses of systems to which the device is connected, etc.), protocol data indicating protocols used by the device, websites visited by the device, geolocation of the device, type of the device (e.g., smartphone, smartwatch, laptop, security camera, personal computer, etc.), data indicating other information related to determining risk factors as described above,
  • risk factors are determined.
  • the risk factors are determined based on risk behaviors such as observed risk behaviors and assumed risk behaviors.
  • each risk factor is determined based on a list of predetermined risk factors and associated risk behaviors, as a function of one or more numerical values representing risk behaviors, or a combination thereof (e.g., a predetermined numerical value may be associated with an associated risk behavior, and the predetermined numerical value may be used as an input to a function of multiple risk behavior values).
  • each risk factor is a number in the range of 1 -10.
  • Each risk factor may be determined based on observed risk behaviors, assumed risk behaviors, or a combination thereof, and may be aggregated based on multiple values representing different risk behaviors. To this end, S230 may further include determining a value for each risk behavior and aggregating the values to determine a risk factor.
  • the risk factors may be determined based on a knowledgebase of known normal device behaviors.
  • a knowledgebase may be stored in a database (e.g., the database 1 1 1 , Fig. 1 ) and includes known normal behaviors of devices.
  • the knowledgebase further defines risk factors for deviations from the known normal behaviors.
  • the knowledgebase may define predetermined values for particular deviating behaviors, formulas for calculating values of risk factors based on values representing deviations from normal behavior, or both.
  • the known normal behaviors may further include different sets of known normal behaviors for different devices, types of devices, users of devices, and the like.
  • the risk factors may be determined based on comparison of behaviors between devices. For example, behaviors of the device may be compared to behaviors of other devices that are connected to the network. More specifically, a risk factor determined for a device based on comparison to other devices on the network may be determined based on whether there is a difference in behavior, a degree of difference in behavior, a number of other devices having the same behavior, a number of other devices having different behavior, numbers of devices having different degrees of difference in behavior, and the like.
  • the comparison may further be between comparable devices, types of devices (e.g., behaviors of laptops may be compared to each other but not to behaviors of servers or security cameras), owners of devices (e.g., behaviors of devices may be compared among entry level employees, between management, between executives or other high- ranking officials, etc.), brands of devices, and the like.
  • types of devices e.g., behaviors of laptops may be compared to each other but not to behaviors of servers or security cameras
  • owners of devices e.g., behaviors of devices may be compared among entry level employees, between management, between executives or other high- ranking officials, etc.
  • brands of devices e.g., etc.
  • a risk factor may be determined based on this deviation.
  • determining risk factors follow.
  • data indicating that the device is moving is associated with a predetermined observed behavior risk factor of 2 and data indicating that the device is immobile is associated with a predetermined observed behavior risk factor of 8.
  • the assumed behavior risk factor for company common vulnerabilities and exposure (CVE) history is determined as a function of the quotient (Number of CVEs)/(Number of Employees) such that a higher number of CVEs relative to the size of the company expressed in number of employees yields a higher risk factor, and the function may further be based on the severity of each of the CVEs.
  • the assumed risk behavior for data predictability may be a function of entropy calculated based on the data such that higher entropy results in a lower risk factor and vice versa.
  • connections to a suspicious cloud domain e.g., from a predetermined list of suspicious cloud domains
  • connections to 20 cloud endpoints may be associated with a value of 9 such that the aggregated risk factor is determined to be 8.
  • S230 further includes determining the observed risk behaviors.
  • the observed risk behaviors are determined directly based on network activity data, configuration data, or both, of the device.
  • network activity data related to traffic may be analyzed to determine an amount of traffic into and out of the device.
  • configuration data for the device may be analyzed to determine whether the device has an encrypted disk.
  • S230 further includes determining the assumed risk behaviors.
  • the assumed risk behaviors are determined indirectly by extrapolating based on contextual information related to the device.
  • the contextual information may include, for example, predetermined associations between certain circumstances indicated by data related to the device and assumed risk behaviors.
  • the circumstances may be determined based on data related to the device such as, but not limited to, configuration data for the device, identifying data for the device (e.g., identifying the device by name, type, model, manufacturer, brand, etc.).
  • configuration data for the device may be analyzed to determine an operating system version of the device (e.g., ABC OS v.5.4), and a known age of the operating system (e.g., an actual age or an age relative to other versions) may be determined based on a list of known versions of the operating system.
  • identifying data for the device may be analyzed to determine a manufacturer of the device (e.g., XYZ Phone Manufacturers, Inc.), and a manufacturer reputation for the device may be determined based on a list of manufacturers and known reputations (e.g., reputations expressed as the number of common vulnerabilities and exposures attributed to the manufacturer relative to the number of employees of the manufacturer).
  • a risk score is determined for the device. In an embodiment, the risk score is determined by applying a weight value to each risk factor. The risk score is the sum of the weighted risk factors.
  • the weights may be predetermined and may further differ depending on the device (e.g., a type of device, a model, a specific device, etc.), activity by the device (e.g., portions of the network to which the device connects), or both.
  • a low weight of 0.1 may be applied to a risk factor for the number of domains accessed when the device is a laptop, while a higher weight of 0.8 may be applied to that risk factor when the device is a security camera.
  • At least some of the weights may be negative such that their corresponding risk factors reduce the overall risk score. This allows for holistic consideration of aggravating and mitigating circumstances related to risk (i.e., circumstances that increase the likelihood that a device is risky and decrease the likelihood that a device is risky, respectively).
  • appropriate mitigation actions are determined.
  • the appropriate mitigation actions may include, but are not limited to, disconnecting the device from the network, preventing the device from connecting to the network (e.g., by reconfiguring access control settings of network infrastructure), limiting activity by the device on the network (e.g., preventing the device from uploading data to the network), limiting portions of the network to which the device can connect (e.g., only allowing the device to connect to a guest network rather than a private network), and the like.
  • Which mitigation actions to perform may vary, for example, based on one or more thresholds for the risk score.
  • the appropriate mitigation actions may include passive monitoring of activity by the device instead of active interference with device activity, for example, when the risk score is below a threshold.
  • the risk factors and risk score for the device may be updated over time (e.g., periodically) based on new information gained through the passive monitoring. This allows for providing network access to devices which have not yet been determined as sufficiently risky to mitigate and reevaluating the risk posed by the device based on network activity. For example, a device that has not yet been connected to the network for a long period of time may have a low risk score initially, but the device may have a higher risk score over time as the device acts within the network. Thus, such a device may be initially permitted to act but may be disconnected or otherwise prevented from accessing the network once it is determined that the device presents a likely cyber security threat.
  • the mitigation actions are performed.
  • the risk factors and risk score for the device may be updated periodically, when suspicious activity is detected (for example, by one or more cybersecurity detection tools such as the detection tools 1 12, Fig. 1 ), or both. In such implementations (not shown), execution continues with S220.
  • Fig. 2 is described with respect to a discrete single assessment of a device with respect to risk factors merely for simplicity and without limitation on the disclosed embodiments.
  • the disclosed embodiments may be equally applied to implementations in which the assessment of the device is performed repeatedly, is performed continuously, or is otherwise updated throughout a session of the device on or in physical proximity to the network.
  • Fig. 3 is an example schematic diagram of a threat mitigator 120 according to an embodiment.
  • the threat mitigator 120 includes a processing circuitry 310 coupled to a memory 320, a storage 330, and a network interface 340.
  • the components of the threat mitigator 120 may be communicatively connected via a bus 350.
  • the processing circuitry 310 may be realized as one or more hardware logic components and circuits.
  • illustrative types of hardware logic components include field programmable gate arrays (FPGAs), application-specific integrated circuits (ASICs), Application-specific standard products (ASSPs), system-on-a-chip systems (SOCs), general-purpose microprocessors, microcontrollers, digital signal processors (DSPs), and the like, or any other hardware logic components that can perform calculations or other manipulations of information.
  • the memory 320 may be volatile (e.g., RAM, etc.), non-volatile (e.g., ROM, flash memory, etc.), or a combination thereof.
  • computer readable instructions to implement one or more embodiments disclosed herein may be stored in the storage 330.
  • the memory 320 is configured to store software.
  • Software shall be construed broadly to mean any type of instructions, whether referred to as software, firmware, middleware, microcode, hardware description language, or otherwise. Instructions may include code (e.g., in source code format, binary code format, executable code format, or any other suitable format of code).
  • the instructions when executed by the processing circuitry 310, cause the processing circuitry 310 to perform the various processes described herein. Specifically, the instructions, when executed, cause the processing circuitry 310 to generate fleet behavior models and detect anomalous behavior in fleets or sub-fleets as described herein.
  • the storage 330 may be magnetic storage, optical storage, and the like, and may be realized, for example, as flash memory or other memory technology, CD-ROM, Digital Versatile Disks (DVDs), or any other medium which can be used to store the desired information.
  • flash memory or other memory technology
  • CD-ROM Compact Discs
  • DVDs Digital Versatile Disks
  • the network interface 340 allows the threat mitigator 120 to communicate with the database 1 1 1 for the purpose of, for example, retrieving assumed behavior data related to the device 130, and the like. Further, the network interface 340 allows the threat mitigator 120 to communicate with the detection tools 1 12 for the purpose of, for example, retrieving network activity data related to the device 130.
  • the various embodiments disclosed herein can be implemented as hardware, firmware, software, or any combination thereof.
  • the software is preferably implemented as an application program tangibly embodied on a program storage unit or computer readable medium consisting of parts, or of certain devices and/or a combination of devices.
  • the application program may be uploaded to, and executed by, a machine comprising any suitable architecture.
  • the machine is implemented on a computer platform having hardware such as one or more central processing units (“CPUs”), a memory, and input/output interfaces.
  • CPUs central processing units
  • the computer platform may also include an operating system and microinstruction code.
  • a non-transitory computer readable medium is any computer readable medium except for a transitory propagating signal.
  • any reference to an element herein using a designation such as“first,”“second,” and so forth does not generally limit the quantity or order of those elements. Rather, these designations are generally used herein as a convenient method of distinguishing between two or more elements or instances of an element. Thus, a reference to first and second elements does not mean that only two elements may be employed there or that the first element must precede the second element in some manner. Also, unless stated otherwise, a set of elements comprises one or more elements.
  • the phrase“at least one of” followed by a listing of items means that any of the listed items can be utilized individually, or any combination of two or more of the listed items can be utilized.
  • the system can include A alone; B alone; C alone; 2A; 2B; 2C; 3A; A and B in combination; B and C in combination; A and C in combination; A, B, and C in combination; 2A and C in combination; A, 3B, and 2C in combination; and the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Security & Cryptography (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Computer And Data Communications (AREA)

Abstract

A system and method for mitigating cyber security threats by devices using risk factors. The method includes determining a plurality of risk factors for a device based on a plurality of risk behaviors indicated by network activity and information of the device; determining a risk score for the device based on the plurality of risk factors and a plurality of weights, wherein each of the plurality of weights is applied to one of the plurality of risk factors; and performing at least one mitigation action based on the risk score.

Description

SYSTEM AND METHOD FOR MITIGATING CYBER SECURITY THREATS
TECHNICAL FIELD
[001] The present disclosure relates generally to cyber security, and more particularly to securing networks against threats posed by malicious devices.
BACKGROUND
[002] Whenever a new device connects to a network of an organization or is activated in a location that is proximate to the organization’s physical location, there is a possibility that the new device could be used by a malicious entity to cause harm to the organization, the network, or both. The number of potential problems that may occur due to network access has increased exponentially due to the rapidly increasing number of new devices that are now capable of network connectivity. Further, many of these new devices are not only capable of connecting to networks but may also be able to create their own networks or hotspots.
[003] Some solutions for securing organizations’ computing infrastructures against malicious devices include requiring authentication for new devices logging onto the network and blocking unknown devices from accessing the network. However, requiring authentication does not necessarily prevent a malicious entity from obtaining the necessary credentials and accessing the network despite the authentication protocols. Also, blocking all unknown devices may result in benign devices being blocked. Moreover, previously benign devices may become malicious when, for example, infected by malware. Thus, such malware-infected devices may be allowed network access that may be utilized for malicious purposes.
[004] Other solutions for securing organizations’ computing infrastructures include monitoring network activity to detect anomalies. However, existing detection tools may not detect certain types of anomalous behavior, for example a new attack pattern used as part of a zero-day attack that is not yet recognized by the detection tools. Additionally, devices that do not pose a direct threat to the network (e.g., devices that are not directly engaging in malicious behavior but provide network access to other malicious devices) may not be detected as malicious or may take longer to detect as malicious than devices engaging in more easily recognizable malicious activity.
[005] It would therefore be advantageous to provide a solution that would overcome the challenges noted above.
SUMMARY
[006] A summary of several example embodiments of the disclosure follows. This summary is provided for the convenience of the reader to provide a basic understanding of such embodiments and does not wholly define the breadth of the disclosure. This summary is not an extensive overview of all contemplated embodiments, and is intended to neither identify key or critical elements of all embodiments nor to delineate the scope of any or all aspects. Its sole purpose is to present some concepts of one or more embodiments in a simplified form as a prelude to the more detailed description that is presented later. For convenience, the term “some embodiments” or“certain embodiments” may be used herein to refer to a single embodiment or multiple embodiments of the disclosure.
[007] Certain embodiments disclosed herein include a method for method for mitigating cyber security threats by devices using risk factors. The method comprises: determining a plurality of risk factors for a device based on a plurality of risk behaviors indicated by network activity and information of the device; determining a risk score for the device based on the plurality of risk factors and a plurality of weights, wherein each of the plurality of weights is applied to one of the plurality of risk factors; and performing at least one mitigation action based on the risk score.
[008] Certain embodiments disclosed herein also include a non-transitory computer readable medium having stored thereon causing a processing circuitry to execute a process, the process comprising: determining a plurality of risk factors for a device based on a plurality of risk behaviors indicated by network activity and information of the device; determining a risk score for the device based on the plurality of risk factors and a plurality of weights, wherein each of the plurality of weights is applied to one of the plurality of risk factors; and performing at least one mitigation action based on the risk score.
[009] Certain embodiments disclosed herein also include a system for mitigating cyber security threats by devices using risk factors. The system comprises: a processing circuitry; and a memory, the memory containing instructions that, when executed by the processing circuitry, configure the system to: determine a plurality of risk factors for a device based on a plurality of risk behaviors indicated by network activity and information of the device; determine a risk score for the device based on the plurality of risk factors and a plurality of weights, wherein each of the plurality of weights is applied to one of the plurality of risk factors; and perform at least one mitigation action based on the risk score.
BRIEF DESCRIPTION OF THE DRAWINGS
[0010] The subject matter disclosed herein is particularly pointed out and distinctly claimed in the claims at the conclusion of the specification. The foregoing and other objects, features, and advantages of the disclosed embodiments will be apparent from the following detailed description taken in conjunction with the accompanying drawings.
[0011 ] Figure 1 is a network diagram utilized to describe the various disclosed embodiments.
[0012] Figure 2 is a flowchart illustrating a method for mitigating cyber security threats by devices using risk factors according to an embodiment.
[0013] Figure 3 is a schematic diagram illustrating a threat mitigator according to an embodiment.
DETAILED DESCRIPTION
[0014] It is important to note that the embodiments disclosed herein are only examples of the many advantageous uses of the innovative teachings herein. In general, statements made in the specification of the present application do not necessarily limit any of the various claimed embodiments. Moreover, some statements may apply to some inventive features but not to others. In general, unless otherwise indicated, singular elements may be in plural and vice versa with no loss of generality. In the drawings, like numerals refer to like parts through several views.
[0015] It has been identified that, whenever a device connects to a network or is turned on in physical proximity to network infrastructure, the new device should be quickly assessed to determine whether and to what extent the new device poses a threat to the organization. Reaction to any malicious entities must be rapid since more damage can be inflicted the longer a malicious device has access to the network. Moreover, devices should be continuously assessed to ensure that changes in activity or combinations of actions do not demonstrate a potential threat.
[0016] However, assessment of potential threats should be flexible such that activity which may not present a direct, known cybersecurity threat can be detected. For example, a printer connected to a network may act as a backdoor by broadcasting an unprotected Wi-Fi signal. The printer itself may not present any cyber threats that would be detected by cyber threat detection tools but may allow access to the network by other malicious devices. As another example, a smart television that does not have appropriate antivirus software or a smartphone with out-of-date software may present cyber security threats even when the device itself has not yet started acting maliciously.
[0017] The disclosed embodiments allow for rapid detection and mitigation of potential cyber security threats by devices. Risk factors used according to the disclosed embodiments provide a more flexible approach to detecting potentially malicious devices than existing solutions while maintaining fast reaction times. In particular, the risk factors allow for detecting potentially malicious devices in cases such as, but not limited to, devices engaging in cyber-attacks that are not yet known to the detecting system, devices engaged in pre-attack phase (e.g., exploration, infection, or dormant stages) activities, devices that do not directly present risks (e.g., devices providing backdoor network access to other devices or devices that may be susceptible to exploits by other devices and systems), and completely unknown devices for which no activity data is yet available.
[0018]To this end, the various disclosed embodiments include a method and system for mitigating cyber security threats by devices using risk factors. A device to be checked for risk is detected. The detected device may be a device that is connected to a network, a device that is physically proximate to network infrastructure (e.g., to a router), or a device that is otherwise visible to a network.
[0019] Risk factors are determined for the device based on risk-related behaviors (hereinafter “risk behaviors”). Each risk factor is determined based on one or more observed risk behaviors, one or more assumed risk behaviors, or a combination thereof. The observed risk behaviors are behaviors indicated by data gathered from the device (e.g., device configuration data, protocol data, signal strength data, etc.), indicated by data gathered with respect to activity by the device on the network, or both. The assumed risk behaviors are extrapolated based on contextual data related to the device such as other devices that have accessed the network, public information (e.g., information about a company that manufactures or designs the device, information related to software installed on the device, etc.), entropy of data in traffic to or from the device, or a combination thereof.
[0020] Based on the risk factors, a risk score is determined for the device. The risk score may be a weighted average of the determined risk factors. The weight applied to each of the risk factors may be predetermined and may further vary based on portions of the network being accessed, a type of device, the specific device, and the like.
[0021] Based on the risk score, one or more mitigation actions are performed. In an example implementation, the mitigation actions include actively interfering with the device connecting to or acting on the network when the risk score is above a threshold and passively monitoring activity by the device when the risk score is below a threshold.
[0022] Some of the risk factors may be weighted negatively such that the negative risk factors reduce the risk score for the device. For example, the presence of antivirus software installed on the device may result in a negative weight applied to a risk factor for cybersecurity software installed on the device. The negative risk factors allow for a more holistic view of risk as compared to some existing solutions that determine devices to be either safe or not safe when certain conditions are met.
[0023] The disclosed embodiments include determining risk factors based on risk behaviors.
Following is an explanation of how certain types of risk behaviors are used to determine various risk factors and examples of specific risk behaviors may affect the resulting risk factors.
[0024] Various of the following examples mention high numbers. A number may be high for purposes of a risk factor if, for example, the number is above a threshold. The threshold may be changed over time, for example as normal device activity changes. Various of the following examples also mention more, higher, older, or other relative statements. For such examples, the value of the risk factor may increase as the respective number or degree increases.
[0025] An attack surface exposure risk factor may be determined based on risk behaviors such as, but not limited to, vulnerabilities, common traffic patterns, threat intelligence, vulnerabilities to traffic, open ports, use of specific protocols, access of cloud domains, wireless protocols, open hotspots, and any other external access provided by a device. Risk behaviors that would result in a higher attack surface exposure risk factor may include, but are not limited to, a high number of open ports, wireless communications, hotspots, or combinations thereof; use of unencrypted protocols; traffic patterns that deviate from the common traffic patterns; threat intelligence for the device indicating that the device likely presents a cyber security threat; and a high number of known vulnerabilities.
[0026] A cloud synchronization risk factor may be determined based on risk behaviors for the device with respect to cloud services accessed by the device such as, but not limited to, a number of cloud services used by the device, an amount of data transmitted by the device to cloud services, a number of tunnels formed between the device and cloud environments, types of data transmitted by the device, known reputations of domains of cloud environments accessed by the device, and the like. Risk behaviors that would result in a higher cloud synchronization risk factor may include, but are not limited to, a high number of cloud endpoints, unencrypted data or credentials, accessing suspicious cloud domains, and the like.
[0027] A connection security risk factor may be determined based on risk behaviors for the device with respect to security and potential vulnerabilities of connections by the device such as, but not limited to, vulnerabilities and threat intelligence related to protocols used by the device as well as connections data (e.g., numbers of other distinct devices to which the device is connected, whether connections are encrypted, etc.). Risk behaviors that would result in a higher connection security risk factor may include, but are not limited to, a high number of connections to distinct devices, use of unencrypted connections, use of known potentially vulnerable protocols, and the like.
[0028] A boundary evasion risk factor may be determined based on risk behaviors for the device with respect to the device connecting to multiple trusted boundaries such as, but not limited to, a number of connections to distinct boundaries, a number of connections to sensitive boundaries (e.g., boundaries of corporate networks), known standard device configurations, and the like. Risk behaviors that would result in a higher boundary evasion risk factor may include, but are not limited to, more connections to distinct boundaries, more connections to sensitive boundaries, a device connecting to more than one network when standard device configurations for similar devices suggest that the device shouldn’t connect to more than one network, and the like.
[0029] A third party application stores risk factor may be determined based on risk behaviors for the device with respect to the number and risk of third party application stores accessed by the device such as, but not limited to, number of third party application stores accessed, reputations of domains hosting third party application stores accessed, security features (e.g., firewalls, traffic blocks, etc.) of accessed third party application stores, and the like. Risk behaviors that would result in a higher third party application stores risk factor may include, but are not limited to, more third party application stores accessed, higher risk remote domains as indicated by threat intelligence with respect to known attacks, a lack of detected firewalls or traffic blocks, and the like.
[0030] A malicious domains risk factor may be determined based on risk behaviors for the device with respect to the number and risk of domains accessed by the device such as, but not limited to, a number of known malicious or suspicious domains accessed, higher risk remote domains as indicated by threat intelligence with respect to known attacks, security features (e.g., firewalls, traffic blocks, etc.) of accessed domains, and the like. Risk behaviors that would result in a higher malicious domains risk factor may include, but are not limited to, more connections to known malicious or suspicious domains, higher risk remote domains as indicated by threat intelligence with respect to known attacks, a lack of detected firewalls or traffic blocks, and the like.
[0031] A vulnerability history risk factor may be determined based on risk behaviors for the device with respect to the number and severity of vulnerabilities detected for the device such as, but not limited to, number of detected vulnerabilities, severity of vulnerabilities, whether vulnerabilities are remotely exploitable, whether mitigations have been performed by the device, and the like. Risk behaviors that would result in a higher vulnerability history risk factor may include, but are not limited to, more vulnerabilities, higher risk vulnerabilities, higher risk remotely exploitable vulnerabilities, lack of mitigations, and the like.
[0032] A data-at-rest risk factor may be determined based on risk behaviors for the device with respect to hoarding or storing data by the device such as, but not limited to, amounts of data sent to and received by the device, importance or sensitivity of data sent to the device, whether the device has an encrypted disk, and the like. Risk behaviors that would result in a higher data-at-rest risk factor may include, but are not limited to, more data going into the device than coming out, more important or sensitive traffic going to the device, the device lacking an unencrypted disk, and the like. Example important or sensitive data may include, but are not limited to, customer relationship management data, device data, scan data, patient data, data indicating personally identifiable information, and the like.
[0033] An external connectivity risk factor may be determined based on risk behaviors for the device with respect to opening of external connections by the device such as, but not limited to, number of hotspots, number of open wireless protocols, whether access by the device is encrypted, whether access by the device requires authentication, known vulnerabilities of any external connections, and the like. Risk behaviors that would result in a higher external connectivity risk factor may include, but are not limited to, more hotspots, more open wireless protocols, unencrypted access, unauthenticated access, known vulnerabilities for external connections, and the like.
[0034] A user authentication risk factor may be determined based on risk behaviors for the device with respect to distinct user authentications on the device and security of credentials for those user authentications such as, but not limited to, number of users using the same device, whether credentials are encrypted, whether authenticating users conform to a known organizational structure, and the like. Risk behaviors that would result in a higher user authentication risk factor may include, but are not limited to, more users using the same device, unencrypted credentials, users that do not conform to the structure of users in the organization, and the like.
[0035] A software version risk factor may be determined based on risk behaviors for the device with respect to the number and age of operating systems and software applications installed on the device such as, but not limited to, age of applications, age of operating systems, number of applications, number of operating systems, version numbers of applications, version numbers of operating systems, applications or operating systems having known higher risk remotely exploitable vulnerabilities, lack of mitigations of cyber threats by operating systems or software, and the like. Risk behaviors that would result in a higher software version risk factor may include, but are not limited to, older applications or operating systems, more applications or operating systems, applications or operating systems having higher risk remotely exploitable vulnerabilities, lack of detected mitigations, and the like.
[0036] A certificate reuse risk factor may be determined based on risk behaviors for the device with respect to reuse of certificates by the device such as, but not limited to, a number of devices using the same certificate, whether certificates used by the device are user based or device based, makes and models of devices sharing the certificate, and the like. Risk behaviors that would result in a higher certificate reuse risk factor may include, but are not limited to, more devices using the same certificate, use of user based certificates, sharing of a certificate by devices having different makes and models, and the like.
[0037] A manufacturer reputation risk factor may be determined based on risk behaviors related to the manufacturer of the device such as, but not limited to, a number of known breaches affecting the manufacturer, known reputations of geographical locations of origin for the manufacturer, a known number of vulnerabilities across devices made by the manufacturer, and the like. Risk behaviors that would result in a higher manufacturer reputation risk factor may include, but are not limited to, higher numbers of known breaches affecting the manufacturer, poor reputation of country of origin, higher number of vulnerabilities of devices made by the manufacturer, and the like.
[0038] A device model reputation risk factor may be determined based on risk behaviors related to the model of the device such as, but not limited to, a degree of commonality of the device model (e.g., a relative number of users or owners of the devices having the same model), known threat intelligence for the model of the device, number of known vulnerabilities for the device model, and the like. Risk behaviors that would result in a higher device model reputation risk factor may include, but are not limited to, the device model being less common, threat intelligence on the device model suggesting that the device model is likely unsafe, a high number of vulnerabilities for the device model, and the like.
[0039] Fig. 1 shows an example network diagram 100 utilized to describe the various disclosed embodiments. In the example network diagram 100, a threat mitigator 120 is deployed such that it can access a network 1 10. The network 1 10 may be, but is not limited to, a wireless, cellular or wired network, a local area network (LAN), a wide area network (WAN), a metro area network (MAN), similar networks, and any combinations thereof.
[0040] A device 130 accesses (shown) or is deployed in physical proximity (not shown) to the network 1 10. The device 130 may be, but is not limited to, a personal computer, a laptop, a tablet computer, a smartphone, a wearable computing device, a printer, or any other device connected to the network 1 10 or deployed in physical proximity to network infrastructure (e.g., a router, not shown) of the network 1 10.
[0041] In an example implementation, the network 1 10 includes a database 1 1 1 and one or more network activity detection tools 1 12. The threat mitigator 120 is configured to access the database 11 1 , the detection tools 1 12, or both, to obtain data related to risk behaviors to be used for determining risk factors.
[0042] The database 11 1 may store contextual data related to assumed risk behaviors such as, but not limited to, manufacturer reputation information, device model reputation information, number of common vulnerabilities and exposures associated with products by the manufacturer or designer, number of employees of the manufacturer or designer, popular operating systems, and the like. The database 1 11 may also store data collected by the network activity detection tools 1 12 such that the threat mitigator 1 12 may retrieve such information from the database 1 11.
[0043] The database 1 11 may further store data related to known device behaviors that may be utilized to determine risk factors. Thus, the database 1 1 1 may act as a knowledgebase of known device behavior profiles. The data related to known device behaviors may define baseline behaviors for a device representing normal behavior and values (or formulas used for calculating values) of risk factors based on deviations from the baseline behaviors. As a non-limiting example, a baseline behavior for a security camera may be communicating with a single internal server on the network at a fixed geographical location. Risk factors for the security camera may be defined for behaviors including communicating with more than one server, communicating with an external server, ceasing communications with the server, or changing geographic location.
[0044] The manufacturer reputation information may include information related to previous devices connected to the network, public source information related to the size and notable security breaches for the company, or both. The information related to previous devices may be based on risk scores determined for other devices having the same manufacturer such that high risk scores of such previous devices will increase the resulting risk factor. The size and security breach information may include, for example, a number of CVEs, a severity of each CVE, and a size of the company such that a high number of CVEs and high severity of CVEs relative to the number of employees of the company will result in a higher risk factor.
[0045] The popular operating systems information may include common operating systems seen installed on other devices, minimum and maximum (i.e., earliest and latest or least secure and most secure) operating system versions available for a device, or both. Devices that do not use common operating systems result in higher risk factors. Devices having operating system versions closer to the minimum version result in higher risk factors than devices having operating system versions closest to the maximum.
[0046] The detection tools 1 12 are configured to collect data related to the device, network activity by the device 130, or both. Such data may include data related to observed risk behaviors such as, but is not limited to, data included in traffic to or from the device 130, amounts of traffic sent by the device 130, number of endpoints receiving traffic from the device 130, type of traffic sent by the device 130 (e.g., encrypted or unencrypted, repetitive or non-repeating, etc.) common vulnerabilities and exposure exhibited by the device 130 (e.g., for the device 130, for software running on the device 130, or both), domains and Internet Protocol (IP) addressed accessed by the device 130, types and versions of software installed on the device 130, type and version of an operating system installed on the device 130, amount and type of external communication options (e.g., ports, protocols, amount of service set identifiers broadcasted, amount of different antennae, etc.), geographical location of the device, geographical movements of the device, and the like. The network activity data may be collected with respect to the device 130, an operating system running on the device 130, for each application running on the device 130, or a combination thereof.
[0047] The threat mitigator 120 is configured to determine a risk score for the device 130 and to perform mitigation actions based on the determined risk score. The risk score is determined based on risk factors. The risk factors are determined based on risk behaviors including observed risk behaviors and assumed risk behaviors. The observed risk behaviors may be indicated in data collected from the device 130, from the network activity detection tools 1 12 with respect to network activity by the device 130, or both. The assumed risk behaviors are extrapolated based on contextual data related to the device such as, but not limited to, activity by other devices (not shown) accessing the network 1 10, public information related to the device (e.g., information about a manufacturer of the device, an assumed operating system used by the device 130, data transmitted to or by the device, etc.), or both. Each risk factor may be determined based on observed risk behaviors, assumed risk behaviors, or a combination thereof.
[0048] It should be noted that the disclosed embodiments are not limited to the particular layout shown in Fig. 1. For example, the threat mitigator 120 is illustrated in Fig. 1 as being deployed outside the network 1 10, but that the threat mitigator 120 may be equally deployed in the network 1 10 without departing from the scope of the disclosure. Additionally, the threat mitigator 120 and the detection tools 1 12 are shown separately for simplicity purposes, but that the threat mitigator 120 may be included in one of the detection tools 1 12 or otherwise act as one of the detection tools 1 12 without departing from the scope of the disclosure.
[0049] Fig. 2 is an example flowchart 200 illustrating a method for mitigating cyber security threats by devices using risk factors according to an embodiment. In an embodiment, the method is performed by the threat mitigator 120.
[0050] At optional S210, a device to be checked for risk is detected. The detection may include detecting a connection of a device to a network, detection of a device turning on in physical proximity to network infrastructure, or detection of a device otherwise becoming visible to the network (e.g., a device that is already turned on moving within a threshold physical distance from any part of the network infrastructure). In some implementations, a device that was already visible to the network may be checked for risk. In particular, a device may be continuously analyzed as described with respect to the following steps from connection to disconnection from the network.
[0051] At S220, data related to the device is collected. The data related to the device includes data directly related to the device (e.g., configuration data of the device, identifying information of the device, etc.) and network activity by the device (e.g., data collected from the device or gathered by monitoring activity of the device) and data indicating contextual information related to the device.
[0052]The collected data may include, but is not limited to, network activity data (e.g., data indicating traffic coming into or out of the device, traffic data (i.e., data transmitted by or to the device), volume of traffic, portions of the network to which the device has connected, etc.), identifying information of the device (e.g., a name, model number, identifier, a manufacturer of the device, etc.), end-of-life or end-of-service-life data, software data related to the device (e.g., programs installed on the device), connection data (e.g., open ports, wireless communications, hotspots, number of networks the device is connected to as an endpoint, whether connections are encrypted, domains and Internet protocol addresses of systems to which the device is connected, etc.), protocol data indicating protocols used by the device, websites visited by the device, geolocation of the device, type of the device (e.g., smartphone, smartwatch, laptop, security camera, personal computer, etc.), data indicating other information related to determining risk factors as described above, and the like.
[0053] At S230, based on the collected data, risk factors are determined. The risk factors are determined based on risk behaviors such as observed risk behaviors and assumed risk behaviors. In an embodiment, each risk factor is determined based on a list of predetermined risk factors and associated risk behaviors, as a function of one or more numerical values representing risk behaviors, or a combination thereof (e.g., a predetermined numerical value may be associated with an associated risk behavior, and the predetermined numerical value may be used as an input to a function of multiple risk behavior values). In an example implementation, each risk factor is a number in the range of 1 -10.
[0054] Each risk factor may be determined based on observed risk behaviors, assumed risk behaviors, or a combination thereof, and may be aggregated based on multiple values representing different risk behaviors. To this end, S230 may further include determining a value for each risk behavior and aggregating the values to determine a risk factor.
[0055] In an embodiment, the risk factors may be determined based on a knowledgebase of known normal device behaviors. Such a knowledgebase may be stored in a database (e.g., the database 1 1 1 , Fig. 1 ) and includes known normal behaviors of devices. The knowledgebase further defines risk factors for deviations from the known normal behaviors. To this end, the knowledgebase may define predetermined values for particular deviating behaviors, formulas for calculating values of risk factors based on values representing deviations from normal behavior, or both. The known normal behaviors may further include different sets of known normal behaviors for different devices, types of devices, users of devices, and the like.
[0056] In another embodiment, the risk factors may be determined based on comparison of behaviors between devices. For example, behaviors of the device may be compared to behaviors of other devices that are connected to the network. More specifically, a risk factor determined for a device based on comparison to other devices on the network may be determined based on whether there is a difference in behavior, a degree of difference in behavior, a number of other devices having the same behavior, a number of other devices having different behavior, numbers of devices having different degrees of difference in behavior, and the like.
[0057]The comparison may further be between comparable devices, types of devices (e.g., behaviors of laptops may be compared to each other but not to behaviors of servers or security cameras), owners of devices (e.g., behaviors of devices may be compared among entry level employees, between management, between executives or other high- ranking officials, etc.), brands of devices, and the like. As a non-limiting example, if other ABC brand laptop computers on the network have an operating system version 10.1 and the device is a laptop on the network having an operating system version 9.0, a risk factor may be determined based on this deviation.
[0058] Some examples for determining risk factors follow. First, data indicating that the device is moving is associated with a predetermined observed behavior risk factor of 2 and data indicating that the device is immobile is associated with a predetermined observed behavior risk factor of 8. Second, the assumed behavior risk factor for company common vulnerabilities and exposure (CVE) history is determined as a function of the quotient (Number of CVEs)/(Number of Employees) such that a higher number of CVEs relative to the size of the company expressed in number of employees yields a higher risk factor, and the function may further be based on the severity of each of the CVEs. Third, the assumed risk behavior for data predictability may be a function of entropy calculated based on the data such that higher entropy results in a lower risk factor and vice versa. Fourth, connections to a suspicious cloud domain (e.g., from a predetermined list of suspicious cloud domains) may be associated with a value of 7 and connections to 20 cloud endpoints may be associated with a value of 9 such that the aggregated risk factor is determined to be 8.
[0059] Additional example risk factors and the risk behaviors they are determined based on are described above.
[0060] In an embodiment, S230 further includes determining the observed risk behaviors.
The observed risk behaviors are determined directly based on network activity data, configuration data, or both, of the device. As a non-limiting example, network activity data related to traffic may be analyzed to determine an amount of traffic into and out of the device. As another non-limiting example, configuration data for the device may be analyzed to determine whether the device has an encrypted disk.
[0061] In an embodiment, S230 further includes determining the assumed risk behaviors.
The assumed risk behaviors are determined indirectly by extrapolating based on contextual information related to the device. To this end, the contextual information may include, for example, predetermined associations between certain circumstances indicated by data related to the device and assumed risk behaviors. The circumstances may be determined based on data related to the device such as, but not limited to, configuration data for the device, identifying data for the device (e.g., identifying the device by name, type, model, manufacturer, brand, etc.). As a non-limiting example, configuration data for the device may be analyzed to determine an operating system version of the device (e.g., ABC OS v.5.4), and a known age of the operating system (e.g., an actual age or an age relative to other versions) may be determined based on a list of known versions of the operating system. As another non-limiting example, identifying data for the device may be analyzed to determine a manufacturer of the device (e.g., XYZ Phone Manufacturers, Inc.), and a manufacturer reputation for the device may be determined based on a list of manufacturers and known reputations (e.g., reputations expressed as the number of common vulnerabilities and exposures attributed to the manufacturer relative to the number of employees of the manufacturer). [0062] At S240, based on the risk factors, a risk score is determined for the device. In an embodiment, the risk score is determined by applying a weight value to each risk factor. The risk score is the sum of the weighted risk factors. The weights may be predetermined and may further differ depending on the device (e.g., a type of device, a model, a specific device, etc.), activity by the device (e.g., portions of the network to which the device connects), or both. As a non-limiting example, a low weight of 0.1 may be applied to a risk factor for the number of domains accessed when the device is a laptop, while a higher weight of 0.8 may be applied to that risk factor when the device is a security camera.
[0063] In some implementations, at least some of the weights may be negative such that their corresponding risk factors reduce the overall risk score. This allows for holistic consideration of aggravating and mitigating circumstances related to risk (i.e., circumstances that increase the likelihood that a device is risky and decrease the likelihood that a device is risky, respectively).
[0064] At S250, based on the risk score, appropriate mitigation actions are determined. The appropriate mitigation actions may include, but are not limited to, disconnecting the device from the network, preventing the device from connecting to the network (e.g., by reconfiguring access control settings of network infrastructure), limiting activity by the device on the network (e.g., preventing the device from uploading data to the network), limiting portions of the network to which the device can connect (e.g., only allowing the device to connect to a guest network rather than a private network), and the like. Which mitigation actions to perform may vary, for example, based on one or more thresholds for the risk score.
[0065] In some implementations, the appropriate mitigation actions may include passive monitoring of activity by the device instead of active interference with device activity, for example, when the risk score is below a threshold. In such implementations, the risk factors and risk score for the device may be updated over time (e.g., periodically) based on new information gained through the passive monitoring. This allows for providing network access to devices which have not yet been determined as sufficiently risky to mitigate and reevaluating the risk posed by the device based on network activity. For example, a device that has not yet been connected to the network for a long period of time may have a low risk score initially, but the device may have a higher risk score over time as the device acts within the network. Thus, such a device may be initially permitted to act but may be disconnected or otherwise prevented from accessing the network once it is determined that the device presents a likely cyber security threat.
[0066] At S260, the mitigation actions are performed. In some implementations, the risk factors and risk score for the device may be updated periodically, when suspicious activity is detected (for example, by one or more cybersecurity detection tools such as the detection tools 1 12, Fig. 1 ), or both. In such implementations (not shown), execution continues with S220.
[0067] It should be noted that Fig. 2 is described with respect to a discrete single assessment of a device with respect to risk factors merely for simplicity and without limitation on the disclosed embodiments. The disclosed embodiments may be equally applied to implementations in which the assessment of the device is performed repeatedly, is performed continuously, or is otherwise updated throughout a session of the device on or in physical proximity to the network.
[0068] Fig. 3 is an example schematic diagram of a threat mitigator 120 according to an embodiment. The threat mitigator 120 includes a processing circuitry 310 coupled to a memory 320, a storage 330, and a network interface 340. In an embodiment, the components of the threat mitigator 120 may be communicatively connected via a bus 350.
[0069]The processing circuitry 310 may be realized as one or more hardware logic components and circuits. For example, and without limitation, illustrative types of hardware logic components that can be used include field programmable gate arrays (FPGAs), application-specific integrated circuits (ASICs), Application-specific standard products (ASSPs), system-on-a-chip systems (SOCs), general-purpose microprocessors, microcontrollers, digital signal processors (DSPs), and the like, or any other hardware logic components that can perform calculations or other manipulations of information.
[0070]The memory 320 may be volatile (e.g., RAM, etc.), non-volatile (e.g., ROM, flash memory, etc.), or a combination thereof. In one configuration, computer readable instructions to implement one or more embodiments disclosed herein may be stored in the storage 330. [0071] In another embodiment, the memory 320 is configured to store software. Software shall be construed broadly to mean any type of instructions, whether referred to as software, firmware, middleware, microcode, hardware description language, or otherwise. Instructions may include code (e.g., in source code format, binary code format, executable code format, or any other suitable format of code). The instructions, when executed by the processing circuitry 310, cause the processing circuitry 310 to perform the various processes described herein. Specifically, the instructions, when executed, cause the processing circuitry 310 to generate fleet behavior models and detect anomalous behavior in fleets or sub-fleets as described herein.
[0072] The storage 330 may be magnetic storage, optical storage, and the like, and may be realized, for example, as flash memory or other memory technology, CD-ROM, Digital Versatile Disks (DVDs), or any other medium which can be used to store the desired information.
[0073] The network interface 340 allows the threat mitigator 120 to communicate with the database 1 1 1 for the purpose of, for example, retrieving assumed behavior data related to the device 130, and the like. Further, the network interface 340 allows the threat mitigator 120 to communicate with the detection tools 1 12 for the purpose of, for example, retrieving network activity data related to the device 130.
[0074] It should be understood that the embodiments described herein are not limited to the specific architecture illustrated in Fig. 3, and other architectures may be equally used without departing from the scope of the disclosed embodiments.
[0075] The various embodiments disclosed herein can be implemented as hardware, firmware, software, or any combination thereof. Moreover, the software is preferably implemented as an application program tangibly embodied on a program storage unit or computer readable medium consisting of parts, or of certain devices and/or a combination of devices. The application program may be uploaded to, and executed by, a machine comprising any suitable architecture. Preferably, the machine is implemented on a computer platform having hardware such as one or more central processing units (“CPUs”), a memory, and input/output interfaces. The computer platform may also include an operating system and microinstruction code. The various processes and functions described herein may be either part of the microinstruction code or part of the application program, or any combination thereof, which may be executed by a CPU, whether or not such a computer or processor is explicitly shown. In addition, various other peripheral units may be connected to the computer platform such as an additional data storage unit and a printing unit. Furthermore, a non-transitory computer readable medium is any computer readable medium except for a transitory propagating signal.
[0076] All examples and conditional language recited herein are intended for pedagogical purposes to aid the reader in understanding the principles of the disclosed embodiment and the concepts contributed by the inventor to furthering the art, and are to be construed as being without limitation to such specifically recited examples and conditions. Moreover, all statements herein reciting principles, aspects, and embodiments of the disclosed embodiments, as well as specific examples thereof, are intended to encompass both structural and functional equivalents thereof. Additionally, it is intended that such equivalents include both currently known equivalents as well as equivalents developed in the future, i.e., any elements developed that perform the same function, regardless of structure.
[0077] It should be understood that any reference to an element herein using a designation such as“first,”“second,” and so forth does not generally limit the quantity or order of those elements. Rather, these designations are generally used herein as a convenient method of distinguishing between two or more elements or instances of an element. Thus, a reference to first and second elements does not mean that only two elements may be employed there or that the first element must precede the second element in some manner. Also, unless stated otherwise, a set of elements comprises one or more elements.
[0078] As used herein, the phrase“at least one of” followed by a listing of items means that any of the listed items can be utilized individually, or any combination of two or more of the listed items can be utilized. For example, if a system is described as including“at least one of A, B, and C,” the system can include A alone; B alone; C alone; 2A; 2B; 2C; 3A; A and B in combination; B and C in combination; A and C in combination; A, B, and C in combination; 2A and C in combination; A, 3B, and 2C in combination; and the like.

Claims

CLAIMS What is claimed is:
1 . A method for mitigating cyber security threats by devices using risk factors, comprising:
determining a plurality of risk factors for a device based on a plurality of risk behaviors indicated by network activity and information of the device;
determining a risk score for the device based on the plurality of risk factors and a plurality of weights, wherein each of the plurality of weights is applied to one of the plurality of risk factors; and
performing at least one mitigation action based on the risk score.
2. The method of claim 1 , wherein the plurality of risk behaviors includes observed risk behaviors and assumed risk behaviors, wherein the observed risk behaviors are indicated by data related to network activity by the device, wherein the assumed risk behaviors are extrapolated based on known contextual information related to the device.
3. The method of claim 1 , wherein the plurality of risk factors is determined for the device when at least one of: the device connects to a network, the device is turned on in physical proximity to a network, and the device becomes physically proximate to network infrastructure.
4. The method of claim 1 , wherein the plurality of risk behaviors includes observed risk behaviors, wherein determining the plurality of risk factors further comprises:
determining the observed risk behaviors based on data related to at least one of: configuration of the device, network activity by the device, geographic movement of the device, signal strength of the device, and a protocol used by the device.
5. The method of claim 1 , wherein the plurality of risk behaviors includes observed risk behaviors, wherein determining the plurality of risk factors further comprises: determining the assumed risk behaviors based on at least one of: manufacturer reputation information, device model reputation information, known software vulnerabilities, and known operating system vulnerabilities.
6. The method of claim 1 , wherein the at least one mitigation action includes monitoring network activity by the device when the risk score is below a threshold, further comprising:
updating the risk score based on the monitored network activity; and
performing at least one subsequent mitigation action based on the updated risk score.
7. The method of claim 1 , wherein the plurality of weights includes at least one negative weight.
8. The method of claim 1 , wherein the plurality of risk factors includes a manufacturer reputation risk factor, wherein the manufacturer reputation risk factor is determined based on the quotient of a number of common vulnerabilities and exposures attributed to a manufacturer of the device over a number of employees of the manufacturer of the device.
9. The method of claim 1 , wherein the plurality of risk factors includes a data entropy risk factor, wherein the data entropy risk factor is determined based on entropy of at least one of: data received by the device, and data sent by the device.
10. The method of claim 1 , wherein the plurality of risk factors includes at least one of: an attack surface exposure risk factor, a cloud synchronization risk factor, a connection security risk factor, a boundary evasion risk factor, a third party application stores risk factor, a malicious domains risk factor, a vulnerability history risk factor, a data-at-rest risk factor, an external connectivity risk factor, a user authentication risk factor, a software version risk factor, a certificate reuse risk factor, a manufacturer reputation risk factor, and a device model reputation risk factor.
1 1 . The method of claim 1 , wherein the plurality of risk factors is determined based further on a plurality of known device behaviors, wherein each of the plurality of known device behaviors is associated with a plurality of known risk factors, wherein each of the plurality of known risk factors is associated with at least one risk behavior.
12. The method of claim 1 , wherein the plurality of risk factors is determined based further on a plurality of risk behaviors of at least one other device.
13. A non-transitory computer readable medium having stored thereon instructions for causing a processing circuitry to execute a process, the process comprising:
determining a plurality of risk factors for a device based on a plurality of risk behaviors indicated by network activity and information of the device;
determining a risk score for the device based on the plurality of risk factors and a plurality of weights, wherein each of the plurality of weights is applied to one of the plurality of risk factors; and
performing at least one mitigation action based on the risk score.
14. A system for mitigating cyber security threats by devices using risk factors, comprising:
a processing circuitry; and
a memory, the memory containing instructions that, when executed by the processing circuitry, configure the system to:
determine a plurality of risk factors for a device based on a plurality of risk behaviors indicated by network activity and information of the device;
determine a risk score for the device based on the plurality of risk factors and a plurality of weights, wherein each of the plurality of weights is applied to one of the plurality of risk factors; and
perform at least one mitigation action based on the risk score.
15. The system of claim 14, wherein the plurality of risk behaviors includes observed risk behaviors and assumed risk behaviors, wherein the observed risk behaviors are indicated by data related to network activity by the device, wherein the assumed risk behaviors are extrapolated based on known contextual information related to the device.
16. The system of claim 14, wherein the plurality of risk factors is determined for the device when at least one of: the device connects to a network, the device is turned on in physical proximity to a network, and the device becomes physically proximate to network infrastructure.
17. The system of claim 14, wherein the plurality of risk behaviors includes observed risk behaviors, wherein the system is further configured to:
determine the observed risk behaviors based on data related to at least one of: configuration of the device, network activity by the device, geographic movement of the device, signal strength of the device, and a protocol used by the device.
18. The system of claim 14, wherein the plurality of risk behaviors includes observed risk behaviors, wherein the system is further configured to:
determine the assumed risk behaviors based on at least one of: manufacturer reputation information, device model reputation information, known software vulnerabilities, and known operating system vulnerabilities.
19. The system of claim 14, wherein the at least one mitigation action includes monitoring network activity by the device when the risk score is below a threshold, f wherein the system is further configured to:
update the risk score based on the monitored network activity; and
perform at least one subsequent mitigation action based on the updated risk score.
20. The system of claim 14, wherein the plurality of weights includes at least one negative weight.
21. The system of claim 14, wherein the plurality of risk factors includes a manufacturer reputation risk factor, wherein the manufacturer reputation risk factor is determined based on the quotient of a number of common vulnerabilities and exposures attributed to a manufacturer of the device over a number of employees of the manufacturer of the device.
22. The system of claim 14, wherein the plurality of risk factors includes a data entropy risk factor, wherein the data entropy risk factor is determined based on entropy of at least one of: data received by the device, and data sent by the device.
23. The system of claim 14, wherein the plurality of risk factors includes at least one of: an attack surface exposure risk factor, a cloud synchronization risk factor, a connection security risk factor, a boundary evasion risk factor, a third party application stores risk factor, a malicious domains risk factor, a vulnerability history risk factor, a data-at-rest risk factor, an external connectivity risk factor, a user authentication risk factor, a software version risk factor, a certificate reuse risk factor, a manufacturer reputation risk factor, and a device model reputation risk factor.
24. The system of claim 14, wherein the plurality of risk factors is determined based further on a plurality of known device behaviors, wherein each of the plurality of known device behaviors is associated with a plurality of known risk factors, wherein each of the plurality of known risk factors is associated with at least one risk behavior.
25. The system of claim 14, wherein the plurality of risk factors is determined based further on a plurality of risk behaviors of at least one other device.
PCT/US2020/023557 2019-04-01 2020-03-19 System and method for mitigating cyber security threats WO2020205258A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CA3135483A CA3135483A1 (en) 2019-04-01 2020-03-19 System and method for mitigating cyber security threats
CN202080039806.5A CN114270347A (en) 2019-04-01 2020-03-19 System and method for mitigating network security threats
EP20784480.4A EP3948600A4 (en) 2019-04-01 2020-03-19 System and method for mitigating cyber security threats

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US16/371,794 2019-04-01
US16/371,794 US11363051B2 (en) 2019-04-01 2019-04-01 System and method for mitigating cyber security threats by devices using risk factors

Publications (1)

Publication Number Publication Date
WO2020205258A1 true WO2020205258A1 (en) 2020-10-08

Family

ID=72607447

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2020/023557 WO2020205258A1 (en) 2019-04-01 2020-03-19 System and method for mitigating cyber security threats

Country Status (5)

Country Link
US (2) US11363051B2 (en)
EP (1) EP3948600A4 (en)
CN (1) CN114270347A (en)
CA (1) CA3135483A1 (en)
WO (1) WO2020205258A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US12015634B2 (en) 2019-04-01 2024-06-18 Armis Security Ltd. System and method for mitigating cyber security threats by devices using risk factors

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11949700B2 (en) 2017-05-15 2024-04-02 Forcepoint Llc Using content stored in an entity behavior catalog in combination with an entity risk score
US11632382B2 (en) 2017-05-15 2023-04-18 Forcepoint Llc Anomaly detection using endpoint counters
US10999296B2 (en) * 2017-05-15 2021-05-04 Forcepoint, LLC Generating adaptive trust profiles using information derived from similarly situated organizations
US11995593B2 (en) * 2018-11-28 2024-05-28 Merck Sharp & Dohme Llc Adaptive enterprise risk evaluation
US11283826B2 (en) * 2019-05-31 2022-03-22 Servicenow, Inc. Dynamic preview of security vulnerability levels in a managed network
US11102222B1 (en) * 2019-06-17 2021-08-24 Rapid7, Inc. Multi-stage network scanning
US11316886B2 (en) * 2020-01-31 2022-04-26 International Business Machines Corporation Preventing vulnerable configurations in sensor-based devices
US11610020B2 (en) * 2020-04-07 2023-03-21 Mcafee, Llc Securing sensitive user data stored locally by an application
US20210336947A1 (en) * 2020-04-27 2021-10-28 Microsoft Technology Licensing, Llc Rogue certificate detection
US20220159029A1 (en) * 2020-11-13 2022-05-19 Cyberark Software Ltd. Detection of security risks based on secretless connection data
CN112597499B (en) * 2020-12-30 2024-02-20 北京启明星辰信息安全技术有限公司 Nondestructive security inspection method and system for video monitoring equipment
US11637852B2 (en) * 2021-01-04 2023-04-25 Microsoft Technology Licensing, Llc Internet-facing device identification
US12069082B2 (en) * 2021-06-11 2024-08-20 Cisco Technology, Inc. Interpreting and remediating network risk using machine learning
US20230254334A1 (en) * 2022-02-04 2023-08-10 Kyndryl, Inc. Intelligent workflow for protecting servers from outside threats

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130246088A1 (en) 2012-03-14 2013-09-19 Hill-Rom Services, Inc. Algorithm for predicting and mitigating adverse events
US20150163242A1 (en) * 2013-12-06 2015-06-11 Cyberlytic Limited Profiling cyber threats detected in a target environment and automatically generating one or more rule bases for an expert system usable to profile cyber threats detected in a target environment
US20160173521A1 (en) 2014-12-13 2016-06-16 Security Scorecard Calculating and benchmarking an entity's cybersecurity risk score
US20170214701A1 (en) * 2016-01-24 2017-07-27 Syed Kamran Hasan Computer security based on artificial intelligence
EP3111614B1 (en) 2014-02-28 2018-04-18 British Telecommunications public limited company Malicious encrypted network traffic identification
US20180124091A1 (en) * 2016-10-27 2018-05-03 Src, Inc. Method for the Continuous Calculation of a Cyber Security Risk Index
US20180144139A1 (en) 2016-11-21 2018-05-24 Zingbox, Ltd. Iot device risk assessment

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9894093B2 (en) * 2009-04-21 2018-02-13 Bandura, Llc Structuring data and pre-compiled exception list engines and internet protocol threat prevention
WO2014128253A1 (en) * 2013-02-22 2014-08-28 Adaptive Mobile Security Limited System and method for embedded mobile (em)/machine to machine (m2m) security, pattern detection, mitigation
US9104865B2 (en) * 2013-08-29 2015-08-11 International Business Machines Corporation Threat condition management
US10095866B2 (en) * 2014-02-24 2018-10-09 Cyphort Inc. System and method for threat risk scoring of security threats
US9413786B1 (en) * 2015-02-04 2016-08-09 International Business Machines Corporation Dynamic enterprise security control based on user risk factors
US9600320B2 (en) * 2015-02-11 2017-03-21 International Business Machines Corporation Mitigation of virtual machine security breaches
US10699018B2 (en) * 2015-02-16 2020-06-30 G-Software, Inc. Automated and continuous risk assessment related to a cyber liability insurance transaction
US10735456B2 (en) * 2015-10-28 2020-08-04 Qomplx, Inc. Advanced cybersecurity threat mitigation using behavioral and deep analytics
US10230745B2 (en) * 2016-01-29 2019-03-12 Acalvio Technologies, Inc. Using high-interaction networks for targeted threat intelligence
CA2960531C (en) * 2016-03-11 2019-06-25 The Toronto-Dominion Bank Application platform security enforcement in cross device and ownership structures
CA2968710A1 (en) * 2016-05-31 2017-11-30 Valarie Ann Findlay Security threat information gathering and incident reporting systems and methods
US10333965B2 (en) * 2016-09-12 2019-06-25 Qualcomm Incorporated Methods and systems for on-device real-time adaptive security based on external threat intelligence inputs
US11210670B2 (en) * 2017-02-28 2021-12-28 Early Warning Services, Llc Authentication and security for mobile-device transactions
US11363051B2 (en) 2019-04-01 2022-06-14 Armis Security Ltd. System and method for mitigating cyber security threats by devices using risk factors

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130246088A1 (en) 2012-03-14 2013-09-19 Hill-Rom Services, Inc. Algorithm for predicting and mitigating adverse events
US20150163242A1 (en) * 2013-12-06 2015-06-11 Cyberlytic Limited Profiling cyber threats detected in a target environment and automatically generating one or more rule bases for an expert system usable to profile cyber threats detected in a target environment
EP3111614B1 (en) 2014-02-28 2018-04-18 British Telecommunications public limited company Malicious encrypted network traffic identification
US20160173521A1 (en) 2014-12-13 2016-06-16 Security Scorecard Calculating and benchmarking an entity's cybersecurity risk score
US20170214701A1 (en) * 2016-01-24 2017-07-27 Syed Kamran Hasan Computer security based on artificial intelligence
US20180124091A1 (en) * 2016-10-27 2018-05-03 Src, Inc. Method for the Continuous Calculation of a Cyber Security Risk Index
US20180144139A1 (en) 2016-11-21 2018-05-24 Zingbox, Ltd. Iot device risk assessment

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3948600A4

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US12015634B2 (en) 2019-04-01 2024-06-18 Armis Security Ltd. System and method for mitigating cyber security threats by devices using risk factors

Also Published As

Publication number Publication date
US11363051B2 (en) 2022-06-14
US12015634B2 (en) 2024-06-18
US20200314134A1 (en) 2020-10-01
CA3135483A1 (en) 2020-10-08
US20220263853A1 (en) 2022-08-18
EP3948600A4 (en) 2022-12-07
CN114270347A (en) 2022-04-01
EP3948600A1 (en) 2022-02-09

Similar Documents

Publication Publication Date Title
US12015634B2 (en) System and method for mitigating cyber security threats by devices using risk factors
US11290478B2 (en) Methods, systems, and devices for dynamically modeling and grouping endpoints for edge networking
EP3298527B1 (en) Secured access control to cloud-based applications
US9622081B1 (en) Systems and methods for evaluating reputations of wireless networks
US20240154981A1 (en) Logging configuration system and method
US20240089271A1 (en) System and method for providing security to iot devices

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20784480

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 3135483

Country of ref document: CA

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2020784480

Country of ref document: EP

Effective date: 20211102