US20150324586A1 - Methods and apparatus for control and detection of malicious content using a sandbox environment - Google Patents
Methods and apparatus for control and detection of malicious content using a sandbox environment Download PDFInfo
- Publication number
- US20150324586A1 US20150324586A1 US14/797,847 US201514797847A US2015324586A1 US 20150324586 A1 US20150324586 A1 US 20150324586A1 US 201514797847 A US201514797847 A US 201514797847A US 2015324586 A1 US2015324586 A1 US 2015324586A1
- Authority
- US
- United States
- Prior art keywords
- application
- instance
- behavior
- indication
- sandbox environment
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 199
- 244000035744 Hura crepitans Species 0.000 title claims abstract description 166
- 238000001514 detection method Methods 0.000 title description 25
- 230000002547 anomalous effect Effects 0.000 claims abstract description 64
- 230000000977 initiatory effect Effects 0.000 claims abstract description 19
- 230000004044 response Effects 0.000 claims abstract description 16
- 230000006399 behavior Effects 0.000 claims description 209
- 238000011156 evaluation Methods 0.000 claims description 25
- 230000015654 memory Effects 0.000 claims description 17
- 238000012545 processing Methods 0.000 claims description 7
- 238000012544 monitoring process Methods 0.000 claims description 5
- 238000002347 injection Methods 0.000 claims description 4
- 239000007924 injection Substances 0.000 claims description 4
- 230000008569 process Effects 0.000 abstract description 182
- 208000015181 infectious disease Diseases 0.000 description 57
- 238000004891 communication Methods 0.000 description 35
- 230000000694 effects Effects 0.000 description 10
- 230000001413 cellular effect Effects 0.000 description 9
- 230000002155 anti-virotic effect Effects 0.000 description 8
- 230000004048 modification Effects 0.000 description 7
- 238000012986 modification Methods 0.000 description 7
- 238000012549 training Methods 0.000 description 7
- 230000009471 action Effects 0.000 description 5
- 238000003860 storage Methods 0.000 description 5
- 101100264195 Caenorhabditis elegans app-1 gene Proteins 0.000 description 4
- 241000700605 Viruses Species 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000003542 behavioural effect Effects 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- VYZAMTAEIAYCRO-UHFFFAOYSA-N Chromium Chemical compound [Cr] VYZAMTAEIAYCRO-UHFFFAOYSA-N 0.000 description 1
- 230000003213 activating effect Effects 0.000 description 1
- 238000013475 authorization Methods 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 239000010979 ruby Substances 0.000 description 1
- 229910001750 ruby Inorganic materials 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/50—Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
- G06F21/52—Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems during program execution, e.g. stack integrity ; Preventing unwanted data erasure; Buffer overflow
- G06F21/53—Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems during program execution, e.g. stack integrity ; Preventing unwanted data erasure; Buffer overflow by executing in a restricted environment, e.g. sandbox or secure virtual machine
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/50—Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
- G06F21/55—Detecting local intrusion or implementing counter-measures
- G06F21/56—Computer malware detection or handling, e.g. anti-virus arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/50—Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
- G06F21/55—Detecting local intrusion or implementing counter-measures
- G06F21/56—Computer malware detection or handling, e.g. anti-virus arrangements
- G06F21/566—Dynamic detection, i.e. detection performed at run-time, e.g. emulation, suspicious activities
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2221/00—Indexing scheme relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F2221/03—Indexing scheme relating to G06F21/50, monitoring users, programs or devices to maintain the integrity of platforms
- G06F2221/034—Test or assess a computer or a system
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2221/00—Indexing scheme relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F2221/21—Indexing scheme relating to G06F21/00 and subgroups addressing additional information or applications relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F2221/2101—Auditing as a secondary aspect
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2221/00—Indexing scheme relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F2221/21—Indexing scheme relating to G06F21/00 and subgroups addressing additional information or applications relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F2221/2141—Access rights, e.g. capability lists, access control lists, access tables, access matrices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2221/00—Indexing scheme relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F2221/21—Indexing scheme relating to G06F21/00 and subgroups addressing additional information or applications relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F2221/2149—Restricted operating environment
Definitions
- Some embodiments described herein relate generally to a sandboxed protection system that protects a computing device from malicious content.
- a non-transitory processor-readable medium storing code represents instructions to cause a processor to receive a set of indications of allowed behavior associated with an application.
- the processor is also caused to initiate an instance of the application within a sandbox environment.
- the processor is further caused to receive, from a monitor module associated with the sandbox environment, a set of indications of actual behavior of the instance of the application in response to initiating the instance of the application within the sandbox environment.
- the processor is also caused to send an indication associated with an anomalous behavior if at least one indication from the set of indications of actual behavior does not correspond to an indication from the set of indications of allowed behavior.
- FIG. 1A is a schematic illustration of a sandboxed protection system, according to an embodiment.
- FIG. 1B is a schematic illustrations of a sandbox environment within a user device, according to an embodiment.
- FIG. 2 is a schematic block diagram of a computer network system to control and detect malicious content, according to an embodiment.
- FIG. 3 is a flowchart of a process for implementing a sandboxed protection system, according to an embodiment.
- FIG. 4 is a flowchart of a process for control and detection of malicious behavior of an application based on allowed behavior, according to an embodiment.
- FIG. 5 is a flowchart of a process for control and detection of malicious behavior of an application based on application type, according to an embodiment.
- FIG. 6 is a flowchart of a process for control and detection of malicious behavior of an application based on another application, according to an embodiment.
- Known anti-virus software is used to prevent, detect and remove various types of malware from computer devices.
- Some known anti-virus clients use signature-based detection methods, in which a computer is scanned for traces of known threats. Signatures can be developed based on examining known malware. Thus, threats previously unknown to anti-virus clients can go unnoticed and/or undetected.
- a sandbox can be used to separate and/or isolate running programs from each other.
- a sandbox can provide a controlled set of resources for use by suspicious programs.
- a sandboxed protection system can detect malware in a way that overcomes shortcomings of the traditional anti-virus clients and network gateway products.
- the sandboxed protection system can use behavioral-based malware detection.
- behavior-based malware detection includes recognizing behaviors that software applications are allowed to perform. Behaviors not recognized can be considered suspect. For a typical user's computer, this can be an extraordinarily complex problem because the configuration is a dynamic operation. For example, over time, many applications can be installed and run on the computer. Many of these applications can access unsafe content on a network such as, for example, the Internet.
- the sandboxed protection system can be designed to monitor processes and/or applications such as web browsers, document readers, etc. that are running in a sandbox environment. Such a sandbox environment can include applications installed by and/or recognized by the sandboxed protection system.
- applications not installed by and/or recognized by the sandboxed protection system can be excluded from the sandbox environment.
- the sandboxed protection system can understand the behaviors of the applications running in the sandbox environment.
- any behavior that falls outside of “normal behavior” i.e., anomalous behavior
- the sandbox environment is no longer clean.
- the event that produced the anomalous behavior can be labeled as an infection.
- the user can be informed of the infection and/or can be prompted to restore the sandbox to a clean state.
- the behaviors originating from the infection can be collected and/or organized into a forensic report. This forensic report can be provided to a separate data server when the user restores the sandbox to a clean state. In other embodiments, the forensic report can be stored locally instead of or in addition to providing the report to the separate data server. A system administrator can, based on the forensic report, view the forensic data.
- the forensic data can describe and/or show in detail what happened during the infection session, where the infection came from, which portions of the sandbox environment were infected and/or which devices within a network were infected.
- the forensic data can include a summary of the infections.
- system administrators can view the full scope of infections that could have occurred on their network.
- a non-transitory processor-readable medium storing code representing instructions to cause a processor to perform a process includes code to cause the processor to receive a set of indications of allowed behavior associated with an application.
- the processor is also caused to initiate an instance of the application within a sandbox environment.
- the processor is further caused to receive, from a monitor module associated with the sandbox environment, a set of indications of actual behavior of the instance of the application in response to initiating the instance of the application within the sandbox environment.
- the processor is also caused to send an indication associated with an anomalous behavior if at least one indication from the set of indications of actual behavior does not correspond to an indication from the set of indications of allowed behavior.
- an apparatus includes a control module implemented in at least one of a memory or a processing device.
- the control module is configured to initiate an instance of a first application and an instance of a second application within a sandbox environment.
- the control module is also configured to receive, from a monitor module associated with the sandbox environment, a set of indications of actual behavior of the instance of the first application and a set of indications of actual behavior of the instance of the second application.
- An indication of a behavior is within both the set of indications of actual behavior of the instance of the first application and the set of indications of actual behavior of the instance of the second application.
- the control module is also configured to classify the behavior as an anomalous behavior for the first application based on a baseline behavior set for the first application.
- the control module is configured to not classify the behavior as an anomalous behavior for the second application based on a baseline behavior set for the second application.
- the control module is further configured to send a signal in response to classifying the behavior as an anomalous behavior for the
- an apparatus includes a control module implemented in at least one of a memory or a processing device.
- the control module is configured to receive a set of indications of allowed behavior associated with a first application.
- the control module is configured to initiate an instance of the first application within a sandbox environment such that the instance of the first application initiates an instance of a second application within the sandbox environment.
- the control module is also configured to receive, from a monitor module associated with the sandbox environment, a set of indications of actual behavior of the instance of the second application in response to the instance of the first application initiating the instance of the second application.
- the control module is also configured to send an indication associated with an anomalous behavior if at least one indication from the set of indications of actual behavior of the instance of the second application does not correspond to an indication from the set of indications of allowed behavior associated with the first application.
- sandbox environment is intended to mean a single sandbox environment or a combination of sandbox environments (e.g., sandbox environments with a series of configurations for classes of applications, for applications with different levels of trust, etc.).
- FIG. 1A is a schematic illustration of a sandboxed protection system, according to an embodiment.
- a sandboxed protection system 121 a can be operatively coupled to at least one User Equipment (UE) 101 a .
- UE User Equipment
- a UE 101 a can be a personal computer, a tablet computer, a mobile telephone, a smart telephone, a personal data assistant (PDA), etc.
- the UE 101 a can include at least a sandbox environment 105 a that operates under a host operating system 103 of the UE 101 a .
- the UE 101 a can further include a data store 109 a .
- a data store can be, for example, a memory, a data storage device such as a hard drive, or any component or recording media used to retain digital data.
- the sandbox environment 105 a can include an application helper module 107 a .
- the application helper module 107 a is an application extension or add-on that monitors activity within an application running on UE 101 a .
- an application helper module 107 a can be used, under the supervision of the sandboxed protection system 121 a , to monitor an identifier of a source of an anomalous behavior.
- the application helper module 107 a can monitor Uniform Resource Locators (URLs) accessed by a user of UE 101 a using a browser application.
- URLs Uniform Resource Locators
- a module can be, for example, any assembly and/or set of operatively-coupled electrical components, and can include, for example, a memory, a processor, electrical traces, optical connectors, software (executing or to be executed in hardware) and/or the like. Furthermore, a module can be capable of performing one or more specific functions associated with the module, as discussed further below.
- the sandboxed protection system 121 a can provide a security mechanism for separating running programs associated with UE 101 a .
- a sandbox environment 105 a can be used by UE 101 a under control of the sandboxed protection system 121 a to execute untested code, or untrusted programs from unverified third-parties, suppliers, untrusted users and untrusted websites.
- the sandboxed protection system 121 a can provide a tightly-controlled set of resources in which guest programs can run, such as scratch space on a data store 109 a .
- the sandboxed protection system 121 a can limit, disallow, or restrict capabilities of a UE 101 a such as, for example, network access, ability to inspect a host system, read from input devices, etc.
- a sandboxed protection system 121 a can include an application control module 123 a , a sandbox monitor module 125 a , an event engine module 127 a , a guest monitor module 129 a , a kernel mode module 131 , a rule generation module 133 , a data store 135 , and a trust classification module 137 . Furthermore, the sandboxed protection system 121 a communicates with one or more User Equipments (UEs) 101 a or other devices, computer network systems, etc., via input signal 139 and output signal 141 .
- FIG. 1A is merely an example illustrating the types of components that can be included within a sandboxed protection system 121 a.
- the sandboxed protection system 121 a and its components may be located anywhere within a communication network system (not shown in FIG. 1A ) including, but not limited to, within the UE 101 a , within one or more service provider devices (not shown), or in separate locations within the communication network system.
- the application control module 123 a is configured to control the sandbox environment 105 a by sending an output signal to the sandbox environment 105 a via the output signal 141 .
- the application control module 123 a can initiate a module within the UE 101 a that is executed by the host operating system 103 a using the sandbox environment 105 a of UE 101 a .
- the application control module 123 a can initiate the sandbox environment 105 a , for example by activating one or more applications and/or processes within the sandbox environment 105 a .
- the application control module 123 a can also terminate a sandbox environment 105 a by terminating one or more applications and/or processes running within the sandbox environment 105 a .
- the application control module 123 a can also initialize or clean the sandbox environment 105 a by, for example, restarting the sandbox environment 105 a or restoring a standard or predefined set of configurations to the sandbox environment 105 .
- the sandbox control module 123 a is configured to initiate a module to run within the sandbox environment 105 a to monitor application activity within the sandbox environment 105 a .
- the application helper module 107 a can monitor activity within the sandbox environment 105 a that the host operating system 103 a is unable to monitor outside the sandbox environment 105 a.
- the kernel mode module 131 can control a kernel mode of operation of software applications within the host operating system 103 a .
- a process in kernel mode operation, a process can have complete and unrestricted access to the underlying hardware of the UE 101 a .
- the kernel mode module 131 enables a running process to execute any Central Processing Unit (CPU) (not shown) instruction and/or reference any memory (not shown) address.
- the kernel mode module 131 can reserve the kernel mode operation for lowest-level, most trusted functions of the host operating system 103 a.
- the event engine module 127 a can control a user mode of operation of software applications within the host operating system 103 a . In some instances, in user mode operation, a process is unable to directly access hardware and/or reference memory of the UE 101 a . In some instances, the event engine module 127 a enables a process running in user mode to delegate requests to system Application Programming Interfaces (APIs) to access hardware or memory of the UE 101 a . The protection provided by the user mode isolation, enables the host operating system 103 a to recover crashes in user mode. In some instances, most of the processes and/or code associated with applications running on a UE 101 a can be executed in user mode controlled by the event engine module 127 a.
- APIs Application Programming Interfaces
- the application control module 123 a can activate the event engine module 127 a .
- the event engine module 127 a can be included within the application control module 123 a .
- the event engine module 127 a is a rule-based event processing engine that can determine if an event (e.g., a process, a service, etc.) is allowed to be executed on UE 101 a or is an infection (e.g., a malware).
- the event engine module 127 a can be responsible for receiving event data and analyzing the received data to determine if an anomalous behavior has occurred. In some instances, the event engine module 127 a can receive data associated with events in substantially real-time.
- the sandboxed protection system 121 a can substantially constantly and/or periodically detect anomalous behavior and/or collect data associated with the anomalous behavior from applications and/or processes in the sandbox environment 105 a .
- the collected data can be evaluated by the event engine module 127 a using a script (e.g., a rule-based filter) and the associated processes can be added to a tree of processes and/or process behaviors referred to herein as an evaluation tree, stored in data store 109 a or in data store 135 .
- the evaluation tree can be used by the event engine module 127 a to keep track of and/or monitor the current state of the sandbox environment 105 a .
- App 1 launches application App 2
- App 1 is defined as a node in the evaluation tree with App 2 as its child.
- behaviors associated with a potential infection are added to the evaluation tree and/or behaviors not associated to a potential infection are not added to the evaluation tree.
- new information can be associated with the node as an attribute.
- the evaluation tree can be serialized to a forensic report.
- the forensic report can be stored in a data store 109 a or a data store 135 as a record of an infected session.
- features of a product including the application and/or the process can be defined ahead of time, for example by the provider, or at run time, for example by the sandboxed protection system 121 a , via regulated authorization.
- a rule can be defined to identify behaviors of a sandboxed application, process and/or service that may trigger an infection.
- the rule can be used by the event engine module 127 a to determine a behavior as allowed or not-allowed for the sandboxed application, process and/or service.
- some system administrators can have the authority to pre-configure the sandboxed applications with their own software applications, processes and/or browser add-ons.
- the system administrators can build and/or define a reliable set of rules that covers the allowed behaviors of the applications and/or processes and enters the rules into the sandboxed protection system 121 a via input signal 139 .
- the rules may be stored by the sandboxed protection system 121 a in rules database 136 a and accessed by the application control module 123 a , event engine module 127 a or other components of the sandboxed protection system 121 a .
- the sandboxed protection system 121 a can associate levels of trust to the applications and/or processes being monitored in the sandbox environment 105 a.
- the event engine module 127 a can receive a notification from the sandbox monitor module 125 a associated with a behavior of an instance of an application running within the sandbox environment 105 a .
- the event engine module 127 a may receive a notification from the sandbox monitor module 125 a indicating that an instance of a Web browser running within the sandbox environment 105 a is trying to modify a system file.
- the event engine module 127 a can receive a set of rules from the rules database 136 a describing allowed behavior associated with the application.
- the set of rules may include rules describing file modification rights of the Web browser application.
- the event engine module 131 can analyze the set of rules and determine whether the system file modification by the Web browser is allowed.
- the event engine module 127 a can send a signal to the application control module 123 a that the modification attempt by the Web browser is an allowed behavior.
- the application control module 123 a can, in response, allow the Web browser instance to continue execution and modify the system file. Otherwise, if the analysis results by the event engine module 127 a indicate that the attempt by the Web browser to modify the file system is not allowed, the event engine module 127 a indicates the system file modification as an anomalous behavior and sends the indication to the application control module 123 a .
- the application control module 123 a can, in response, terminate the file system modification attempt, terminate the Web browser, terminate the sandbox environment 105 a , or take other predetermined actions associated with such anomalous behaviors.
- the event engine module 127 a can store the anomalous behavior in the rules database 136 a associated with the Web browser application.
- the trust classification module 137 can associate a trust level to each application, process and/or service that is executed within the sandbox environment 105 a .
- the trust classification module 137 can also associate a set of trusted processes with each application, process and/or service.
- the trust level and/or the set of trusted processes may be defined at least based on past behavior of the application, process and/or service, user input, application, process and/or service type, type of data associated with UE 101 a that the application, process and/or service attempts to access, etc.
- the trust classification module 137 can store the trust level and/or the set of trusted processes associated with each application, process and/or service at the rules database 136 a.
- the sandbox monitor module 125 a can monitor activities of an instance of an application within the sandbox environment 105 a by tracing network activities of the running instance of the application within the sandbox environment 105 a .
- the sandbox monitor module 125 a can also monitor one or more executable files of the instance of the application.
- different levels of trust can be associated with applications and/or processes based on monitoring of the applications and/or processes by the sandbox monitor module 125 a .
- an application, process and/or service can be classified as “trusted”, “untrusted”, “suspect”, or “unknown”.
- such a classification can be performed by a trust classification module 137 of the sandboxed protection system 121 a .
- the rule author e.g., a system administrator
- the sandboxed protection system 121 a can then automatically assign trust levels suspect and unknown after further evaluation of the application, process and/or service.
- the rule author can also make the determination of whether the application, process and/or service is suspect or unknown.
- the sandboxed protection system 121 a can automatically assign a trust level to each application, process and/or service based on a type of the application, process and/or service (e.g., game, browser, word processor, etc.), observed behavior of the application, process and/or service, etc.
- a system administrator can request the sandboxed protection system 121 a to override a trust level for an application, process and/or service.
- a system administrator can request the sandboxed protection system 121 a to modify the trust levels of applications and/or processes. Classifying the trustworthiness of each process in a guest application can provide flexibility for customization of the scope of malware detection.
- declaring a process as trusted in the rules can essentially imply that any possible behavior exhibited and/or actions performed by the process and/or the process's child processes are considered as allowed behavior.
- trusted processes can be allowed to perform any action.
- the sandboxed protection system 121 a can effectively disable detection for that process. In such instances, for example, if a trusted process launches a child process, the child process can inherit the trust and become trusted. As another example, if a file is written by a trusted process and launched by a different process, the new process can be trusted because the file inherits the trust from the trusted process. This behavior can be common for programs that update themselves. In other instances, a system administrator can request the sandboxed protection system 121 a for manual or rule-based reclassification of each child process launched and/or written by a trusted process.
- Trusting a process is, typically, a less secure option than not trusting the process.
- the sandboxed protection system 121 a can warn the system administrators about the risk of adding new software into the sandbox environment 105 a .
- a system administrator that would like to install a WebEx application, process and/or service in the sandbox environment 105 a but would also like to avoid defining allowed behaviors for the WebEx process and all child processes can declare WebEx.exe as a trusted process.
- a system administrator can declare a process as trusted by providing the path to the root process to a trusted process section of the rules stored in rules database 136 a and processed by the event engine module 127 a.
- processes that are monitored by the sandbox monitor module 125 a as potential propagation methods for infection can be declared and/or classified as untrusted by the trust classification module 137 .
- the event engine module 127 a detects that an untrusted application, process and/or service performs a suspicious behavior that is not explicitly defined as allowed, this is a trigger for an infection.
- the rules database 136 a can include whitelisted rules for untrusted applications and/or processes that define normal trusted operating behaviors of the applications and/or processes.
- a main account in the sandbox environment 105 a can run and/or execute as a limited user. Therefore, some processes associated with the main account may not have permission to modify registry keys inside HKEY_LOCAL_MACHINE. If a process without having the permission attempts to modify a portion of the registry key, the sandboxed protection system 121 a can recognize the attempted modification as a trigger behavior for an infection. A process P, however, which runs in the system context, can have permission to modify registry keys inside HKEY_LOCAL_MACHINE. In order to prevent process P from triggering an infection, a rule that allows P to access the registry key can be defined (as seen in the Heuristic Table shown below).
- the process that caused the infection can be classified, by the sandboxed protection system 121 a , as suspect.
- the sandboxed protection system 121 a can record subsequent behaviors being performed by the process and/or its child processes. In some instances, similar to trusted processes, behavior originating from a suspect process can inherit the suspect trust level.
- a forensic report e.g., a report file in data store 135 . The forensic report can be analyzed by the event engine module 127 a and can provide details of the suspect events.
- additional information on suspect behaviors and/or events associated with processes and applications such as, for example, cryptographic hash functions (e.g., Message Digest (MD5) hashes, Secure Hash Algorithm (SHA) hashes, etc.) providing unique signatures for files, can be captured by the sandboxed protection system 121 a .
- cryptographic hash functions e.g., Message Digest (MD5) hashes, Secure Hash Algorithm (SHA) hashes, etc.
- MD5 Message Digest
- SHA Secure Hash Algorithm
- the hash value can be used to lookup and/or compare the triggering application, process and/or service against a third-party malware (e.g., virus) database. This lookup can provide information about number of anti-virus clients that scanned the process as being infected.
- third-party malware e.g., virus
- the sandboxed protection system 121 a also includes a user-configurable feature which can terminate and/or kill any suspect processes as they are detected. In some instances, the process termination can be accomplished by terminating and/or killing a process as soon it is classified as suspect. In other instances, the sandboxed protection system 121 a can provide information about the suspect processes to a system administrator and allow the administrator to terminate and/or kill a process.
- a process can be classified as unknown if its path is not explicitly listed as trusted or untrusted in the rules database 136 a .
- this process can perform actions and/or behaviors without triggering an infection.
- the unknown trust of this process is not inherited across child processes.
- An unknown process can, for example, launch an application App 1 , and application App 1 can be tracked as untrusted.
- the behaviors of an unknown process can be added to the evaluation tree and can show up in an infection report (e.g., a forensic report) if the session becomes tainted. While four trust levels are shown and described herein, in other instances any number of trust levels can be used.
- designing the allowable behavior rules can depend on the types of behaviors that can trigger an infection.
- infections can be determined based on the following heuristics table.
- other infection triggers can be added with additional sensors and/or types of events.
- an attempt can be made by the sandboxed protection system 121 a to determine the originating source of the infection. For example, if the infection is triggered by a web browser (e.g., Internet Explorer, Firefox, etc.), or a child process of a web browser, then the source can be determined based on a URL history of a user of UE 101 a .
- the URL trace events can be captured from an application helper module 107 a attached to the web browser.
- the URL trace events can be fed to the event engine module 127 a , which then correlates the main URL and any URLs redirected from that main URL with an infection.
- the event engine module 127 a can correlate the URLs by tracing backwards in time from the most recent URL matching a Process Identifier (PID) associated with an infection root.
- PID Process Identifier
- a process, application, service and/or the like can be identified as the originating source of the infection.
- the sandboxed protection system 121 a can allow a user of UE 101 a to launch a portable document format (PDF) file on the UE 101 a for viewing inside the sandbox environment 105 a .
- PDF portable document format
- the sandboxed protection system 121 a can trace the source to the PDF file if an infection is originated from the PDF reader or its child processes. The sandboxed protection system 121 a can do this by keeping and/or maintaining a map of processes in the sandbox environment 105 a that were launched when the PDF file was redirected from the host operating system 103 a .
- the sandboxed protection system 121 a can check the map to see if the PDF file belongs to an infected process. This technique can be used to track an infection if other redirected documents from the host operating system 103 a are determined to be the source of an infection.
- Heuristics Table 1 If an untrusted process launches another process that is not explicitly allowed, 2. If an untrusted process terminates another process that is not explicitly allowed, 3. If an untrusted process writes a file and the file is launched by a process with unknown trust, 4. If an untrusted process modifies a sensitive registry value that is not explicitly allowed, for example: a. HKEY_LOCAL_MACHINE b. HKEY_CLASSES_ROOT c. Any windows startup key d. Any policy key, 5. If an untrusted process modifies a sensitive file path that is not explicitly allowed, Examples of restricted paths are: a. C: ⁇ Windows ⁇ System32 b. C: ⁇ Program Files 6.
- the source can be added as a node in the evaluation tree as a child of the infected process and/or application.
- This source can be included in the summary of infection report (e.g., the forensic report) provided to a system administrator by the sandboxed protection system 121 a .
- the name of the infected process can be used.
- rules can be defined and stored in the rules database 136 a by users, by system administrators, by application providers, etc.
- a set of rules can be associated with each application, process and/or service.
- the set of rules associated with an application, process and/or service define allowed behavior of the application, process and/or service.
- the rules associated with an application, process and/or service can be used by the event engine module 127 a to determine whether a behavior triggers and/or is associated with an infection.
- the rules can be divided up into a plurality of rule sets, for example a main rule set (e.g., rules associated with the host operating system 103 a ), a custom rule set (e.g., rules associated with applications and/or processes), etc.
- the main rule set can be provided with and/or standard to the sandboxed protection system 121 a .
- system administrators can introduce their own custom rule sets by entering rules into the sandboxed protection system 121 a via input 139 .
- the applications executed in sandbox environment 105 a e.g., browsers
- a custom rule set file can enable the event engine module 127 a to define allowable behaviors for new processes that would otherwise trigger an infection.
- the main rule set can be merged with the custom rule sets and, for example, compiled into a byte code.
- the sets of rules may be stored in the rules database 136 a.
- the sandboxed protection system 121 a includes a rule generation module 133 that automatically generates rules for an application, process and/or service based on previously declared infections associated with same or similar applications and/or processes (e.g., applications with the same type). For example, the rules generation module 133 can generate rules based on infections that have been declared as false positives. The rule generation module 133 can also generate rules based on sets of rules existing in the rules database 136 a for the application, process and/or service. The rule generation module 133 can address various issues associated with application, process and/or service behavioral detection. For example, the rule generation module 133 can define allowed behaviors for new applications.
- the rule generation module 133 can also provide a false positive infection report (e.g., forensic report) and convert the report into a new set of rules.
- the rule generation module 133 can further enable a system administrator to develop rules for their own set of plug-ins that have not been already declared and/or included in the main rule set.
- the sandboxed protection system 121 a can include functionality to automatically generate allowed rules from an infection that may be a false positive.
- the system administrators can view the forensic report provided by the sandboxed protection system 121 a and provide an indication to allow a process and/or behavior of a process.
- the system administrator can click a button labeled “Allow . . . ” next to the entry for a process and/or behavior of a process.
- this button can enable the system administrator to add the process and/or behavior of the process to the main rule set (e.g., windows.rul) and/or a custom rule set (e.g., custom.rul).
- the main rule set e.g., windows.rul
- a custom rule set e.g., custom.rul
- a system administrator can provide an input to the sandboxed protection system 121 a to trust new processes, for example by selecting a checkbox for “Trust newly encountered processes”, the system administrator can automatically classify new processes as trusted, which, while less secure, can reduce the chance of false positives.
- the forensic report, the main rule set, and/or the custom rule set (if exists) can be sent from UE 101 a to the rule generation module 133 via input signal 139 .
- the rule generation module 133 can receive the inputs and generate a new custom rule set, for example to be saved in rules database 136 a .
- the sandboxed protection system 121 a can send a signal to the UE 101 a via output signal 141 .
- the UE 101 a can provide an option to the user (e.g., a button) to display a save dialog that allows the user to retrieve the newly defined custom rule set from the rules database 136 a.
- the sandboxed protection system 121 a can be set to a training mode.
- the training mode can be, for example, a command line parameter.
- infections sent to the UE 101 a can be flagged by the event engine module 127 a as training samples and distinguished visually (e.g., in a report and/or in the rule generation module 133 ) from real infections with an indicator such as a different color or special icon. This can help distinguish between actual infection reports and training samples that are used to generate rules by the rule generation module 133 .
- training samples can then be converted into additional rules automatically or via interaction with the system administrator.
- the events collected in the training sample can be filtered, reduced and/or output as new detection rules.
- the new rules can be downloaded and used by the event engine module 127 a for future detection and/or analysis.
- multiple sensors can be defined and/or set up on the UE 101 a to capture events (e.g. anomalous behavior) and collect data about events.
- the sensors can include module detection (e.g., process detection).
- module detection e.g., process detection
- the sandboxed protection system 121 a can classify a new process as trusted, false positives can still occur. For example, when writing rules for a WebEx plug-in, although WebEx is installed and webex.exe has been declared and/or classified as a trusted process, webex.dll is still loaded into a browser (e.g., Internet Explorer).
- the sandboxed protection system 121 a can whitelist suspicious behaviors for the browser that could otherwise be compromising. In order to solve this problem, the sandboxed protection system can understand what modules are doing irrespective of the process into which they are loaded.
- the sandboxed protection system can separate behaviors from modules and processes.
- the sandboxed protection system can declare, classify and/or recognize webex.dll as a trusted module. Anything webex.dll does can be allowed, however, the browser (e.g., iexplore.exe) can trigger an infection if it performs the same suspicious behavior.
- the sandboxed protection system can declare, classify and/or recognize webex.dll as untrusted but whitelist the behaviors originating from that module for added security.
- the guest monitor module 129 a provides an operating system device driver that can monitor and/or report activity associated with a network, a file system, a registry and/or initiation and/or termination of processes within UE 101 a.
- the data collected by the sensors can be stored in data store 109 a and sent to the event engine module 127 a to be processed and analyzed based on a previously defined list of allowed behavior for the processes.
- the events can be captured in substantially real-time.
- events can be captured using a kernel driver that hooks onto and/or monitors process events, file events, registry events, and/or network events as the events are being performed.
- the application helper module 107 a can be loaded in a web browser (e.g., Internet Explorer or Firefox) to capture Uniform Resource Locator (URL) events.
- URL Uniform Resource Locator
- These captured events can be sent to the event engine module 127 a for analysis.
- the event engine module 127 a can process each event by running it through a rule-based filter.
- the rules for the rule-based filter can be defined using a scripting language optimized to filter events efficiently.
- the host operating system 103 a executes a sandbox environment 105 a controlled by a sandboxed protection system 121 a.
- the sandboxed protection system 121 a can be activated by instructions stored in a memory (e.g., a non-transitory processor-readable medium 113 ) on the UE 101 a .
- the sandboxed protection system 121 a can include additional modules and/or engines such as, for example, a trust classification module, a rule engine, a detection engine, and/or the like (each not shown).
- Such modules and/or engines can be hardware modules and/or engines, or software modules and/or engines executing in hardware.
- FIG. 1B is a schematic illustrations of a sandbox environment 105 b within a UE 101 b .
- components 101 b , 103 b , 105 b , 107 b , 123 ab , 125 b , 127 b , and 129 b are structurally and/or functionally similar to components 101 a , 103 a , 105 a , 107 a , 123 a , 125 a , 127 a , and 129 a of FIG. 1A , respectively.
- FIG. 1B components 101 b , 103 b , 105 b , 107 b , 123 ab , 125 b , 127 b , and 129 b are structurally and/or functionally similar to components 101 a , 103 a , 105 a , 107 a , 123 a , 125 a , 127 a , and 129 a
- the application control module 123 b can execute within the host operating system 103 b
- the event engine module 127 b can execute within the application control module 123 b
- the sandbox monitor module 125 b can execute within the sandbox environment 105 b .
- the sandbox monitor module 125 b can execute on the host operating system 103 b but outside the sandbox environment 105 b.
- applications 153 such as web browsers, documents viewers, etc. are executed within the sandbox environment 105 b .
- the sandbox monitor module 125 b and the application helper module 107 b report application events (e.g., application behaviors) 155 associated with applications 153 to the application control module 123 b .
- the application control module 123 b can control the sandbox environment 105 b (shown as arrow 157 ), for example by initiating, terminating, or modifying applications and/or processes executed within the sandbox environment 105 b using the application events 155 .
- the guest monitor module 129 b reports kernel mode events such as events associated with the host operating system 103 b to the application control module 123 b (shown as arrow 159 ).
- the application control module 123 b and/or the event engine module 127 b can use the operating system events 159 for controlling the applications 153 within the sandbox environment 105 b.
- FIG. 2 is a schematic block diagram of a computer network system to control and detect malicious content, according to an embodiment.
- a computer network system 200 can include one or more user devices or user equipments (UEs) 201 a - 201 n .
- UEs 201 a - 201 n can be structurally and/or functionally similar to UE 101 a of FIG. 1A .
- the computer network system 200 further includes a sandboxed protection system 203 (structurally and/or functionally similar to the sandboxed protection system 121 a of FIG.
- FIGS. 1A and 1B are merely examples illustrating the types of devices that can be included within a computer network system 200 .
- the sandbox protection system 203 can control the sandbox environments 207 a - 207 n via the communication network 205 .
- control signals can be sent between the sandbox protection system 203 and the sandbox environments 207 a - 207 n via the communication network 205 .
- the sandbox environments 207 a - 207 n can be running on an execution server 211 that a user of UE 201 a - 201 n can access remotely by a sandbox access module 221 a - 221 n via the communication network 205 .
- Each UE 201 a - 201 n has access to a sandbox environment 207 a - 207 n located within one or more execution server(s) 211 .
- the sandbox environments 207 a - 207 n are structurally and/or functionally similar to the sandbox environment 105 a and 105 b of FIGS. 1A and 1B .
- the sandboxed protection system 203 can be operatively coupled to the UEs 201 a - 201 n via, for example, a communication network 205 .
- control signals can be sent between the sandboxed protection system 203 and the UEs 201 a - 201 n via the communication network 205 .
- the sandboxed protection system 203 and its components may be located anywhere within a communication network system 205 including, but not limited to, within the UEs 201 a - 201 n , or in separate locations within the communication network system 205 .
- the execution server(s) 211 are equipped with one or more processors and one or more data storages (e.g., memories).
- a storage location on the execution server(s) 211 can be associated with the UE 201 a as a sandbox environment 201 .
- the UE 201 a may have access to a sandbox environment 207 a installed on the execution server(s) 211 such that the UE 201 a can access the sandbox environment 207 a via the communication network 205 and can execute applications and/or processes within the sandbox environment 207 a on the execution server(s) 211 .
- the UE 201 a can initiate downloading an application from the application provider(s) 209 into the sandbox environment 207 a (on the execution server(s) 211 ). The UE 201 a can then initiate execution of the application on the execution server(s) 211 . The sandbox protection system 203 can control the execution of the downloaded application on the execution server(s) 211 via the communication network 205 .
- Communication network 205 can be any communication network, such as the Internet, an Intranet, a Local Area Network (LAN), a Wide Area Network (WAN), a telephone network, an Ethernet network, a fiber-optic network, a wireless network, a cellular network, etc., configurable to allow the one or more UEs 201 a - 201 n , the sandboxed protection system 203 , the one or more application provider(s) 209 , and the one or more execution server(s) 211 to communicate with communication network 205 and/or to each other through communication network 205 .
- LAN Local Area Network
- WAN Wide Area Network
- communication network 205 can include multiple networks operatively coupled to one another by, for example, network bridges, routers, switches and/or gateways.
- the UEs 201 a - 201 n can be operatively coupled to a cellular network and the application provider(s) 209 , the sandboxed protection system 203 , and/or the one or more execution server(s) 211 can be operatively coupled to a fiber-optic network.
- the cellular network and the fiber-optic network can each be operatively coupled to one another via one or more network bridges, routers, switches, and/or gateways such that the cellular network, the Ethernet network and the fiber-optic network are operatively coupled to form a communication network.
- the cellular network and fiber-optic network can each be operatively coupled to one another via one or more additional networks.
- the cellular network and the fiber-optic network can each be operatively coupled to the Internet such that the cellular network, the fiber-optic network and the Internet are operatively coupled to form a communication network.
- UEs 201 a - 201 n are operatively coupled to communication network 205 via network connection 213 ; application provider(s) 209 are operatively coupled to communication network 205 via network connection 215 ; execution server(s) 211 are operatively coupled to communication network 205 via network connection 217 ; and the sandboxed protection system 203 is operatively coupled to communication network 205 via network connection 219 .
- Network connections 213 , 215 , 217 , and 219 can be any appropriate network connection for operatively coupling UEs 201 a - 201 n , application provider(s) 209 , execution server(s) 211 , and the sandboxed protection system 203 .
- a network connection can be a wireless network connection such as, for example, a wireless fidelity (“Wi-Fi”) or wireless local area network (“WLAN”) connection, a wireless wide area network (“WWAN”) connection, and/or a cellular connection.
- Wi-Fi wireless fidelity
- WLAN wireless local area network
- WWAN wireless wide area network
- a network connection can be a wired connection such as, for example, an Ethernet connection, a digital subscription line (“DSL”) connection, a broadband coaxial connection, and/or a fiber-optic connection.
- DSL digital subscription line
- a computer network system 200 can include more than one UE 201 a - 201 n , more than one sandboxed protection system 203 , more than one application provider(s) 209 , and more than one execution server(s) 211 .
- a UE 201 a - 201 n , a sandboxed protection system 203 , an application provider 209 , and/or an execution server(s) 211 can be operatively coupled to the communication network 205 by heterogeneous network connections.
- a first UE 201 a - 201 n can be operatively coupled to the communication network 205 by a WWAN network connection
- a second UE 201 a - 201 n can be operatively coupled to the communication network 205 by a DSL network connection
- a sandboxed protection system 203 can be operatively coupled to the communication network 205 by a fiber-optic network connection.
- the application provider(s) 209 can be, for example, a web server configured to provide various applications to electronic devices, such as UEs 201 a - 201 n .
- the UE 201 a - 201 n can be in communication with the application provider(s) 209 via the communication network 205 under the supervision of the sandboxed protection system 203 .
- the UEs 201 a - 201 n can be any of a variety of electronic devices that can be operatively coupled to communication network 205 .
- a UE 201 a - 201 n can be a personal computer, a laptop computer, a personal digital assistant (PDA), a cellular telephone, a portable/mobile internet device and/or some other electronic communication device.
- the UEs 201 a - 201 n can include a web browser configured to access a webpage or website hosted on or accessible via the application provider(s) 209 over communication network 205 .
- the UEs 201 a - 201 n can be configured to support, for example, HTML using JavaScript.
- the UEs 201 a - 201 n can include a web browser, such as, Firefox, Safari, Opera and Chrome.
- An Internet page or website can be accessed by a user of a web browser at a UE 201 a - 201 n by providing the web browser with a reference such as a uniform resource locator (URL), for example, of a webpage.
- a user of a UE 201 a - 201 n can access an application provider 209 via a URL designated for the application provider 209 .
- UEs 201 a - 201 n can include specialized software for accessing a web server other than a browser, such as, for example, a specialized network-enabled application or program.
- portions of a website accessible via a web server can be located in an execution server(s) 211 accessible to the application provider 209 .
- a memory, within the execution server(s) 211 can be at least one of a database, a data warehouse, and/or the like.
- a UE 201 a - 201 n can also include a display, monitor or user interface (UI), a keyboard, various ports (e.g., a USB port), and other user interface features, such as, for example, touch screen controls, audio components, and/or video components (each not shown).
- UI monitor or user interface
- keyboard e.g., a USB port
- other user interface features such as, for example, touch screen controls, audio components, and/or video components (each not shown).
- a UE 201 a - 201 n can be operatively coupled to communication network 205 via the UI and network connection 213 .
- FIG. 3 is a flowchart of a process for implementing a sandboxed protection system, according to an embodiment.
- the sandboxed protection system 121 a of FIG. 1A executes the process 300 .
- processor-readable instructions to execute the process can be stored in data store 109 a and/or in data store 135 .
- the sandbox control module 123 a initiates a sandbox environment 105 a .
- the sandbox environment 105 a can, for example, be executed within the host operating system 103 that runs on a processor of the UE 101 a .
- the sandbox environment 105 a can be configured to execute applications and/or processes such as, for example, a web browser or any other application, process and/or service at risk of receiving malicious content.
- a set of rules for processes within the sandbox environment 105 a is defined by the rule generation module 133 .
- the rules can include, for example, classifying processes and/or applications using trust classifications provided by the trust classification module 137 .
- the rules can include predefined rules for the system and/or custom defined rules by a system administrator.
- the rules are stored in the rules database 136 a.
- the sandbox monitor module 125 a monitors events and activities associated with running applications and/or processes within the sandbox environment 105 a , based on the rules associated with the applications and/or processes.
- the sandbox processes can be monitored based on the rules in the rules database 136 a .
- the sandbox monitor module 125 a can report any anomalous event or activity to the event engine module 127 a.
- the event engine module 127 a analyzes the report received from the sandbox monitor module 125 a with reference to the associated rules in the rules database 136 a to determine whether the anomalous behavior is malicious.
- the event engine module 127 a can declares an infection based on the detected malicious behavior and the rules of the rules database 136 a.
- the event engine module 127 a adds the detected infection to an evaluation tree within the data store 135 .
- the evaluation tree can be used for further analysis of the infection by the sandboxed protection system 121 a or by a system administrator.
- the analysis results can be used by the rule generation module 133 for updating the rules database 136 a.
- the sandboxed protection system 121 a can automatically and/or dynamically disable and/or ignore an infection detection. Disabling and/or ignoring detection events can be used, for example, in cases where some events are generated intentionally that otherwise would trigger the event engine module 127 a to report an infection. For example, if a user of UE 101 a intentionally runs an unknown process such as an application or an installer in the sandbox environment 105 a without previously defining rules to trust the process.
- the rule generation module 133 revises the rules in the rules database 136 a based on the detected infections reported as the evaluation tree.
- the rule generation module 133 can automatically revise the rules in the rules database 136 a based on the monitored behavior of the sandbox environment 105 a by the sandbox monitor module 125 a and the analysis of the monitored behavior, including the infections detected by the event engine module 127 a .
- a system administrator can manually revise the rules based on the detected infections, evaluation tree content, etc.
- the rules may remain unchanged based on the infection.
- the sandboxed control module 123 a restarts the sandbox environment 105 a , such that the infected environment can be terminated and a new clean sandbox environment 105 a can be restarted.
- the process of monitoring the sandbox environment 105 a by the sandbox monitor module 125 a based on the rules of the rules database 136 a can then continue with the clean sandbox environment 105 a .
- the sandbox control module 123 a can terminate and/or restart the application, process and/or service with malicious behavior without restarting the sandbox environment 105 a.
- FIG. 4 is a flowchart of a process for control and detection of malicious behavior of an application based on allowed behavior, according to an embodiment.
- the application control module 123 a can execute a code to receive a set of indications of allowed behavior associated with an application, based on a set of rules in the rules database 136 a .
- the application control module 123 a can, for example, retrieve the indications from data store 135 , receive the indications from a system administrator via input signal 139 , and/or the like.
- the application control module 123 a can store the received indications in data store 135 .
- the application control module 123 a sends a signal to the host operating system 103 via output signal 141 to initiate an instance of the application in the sandbox environment 105 a.
- the event engine module 127 a receives, from the sandbox monitor module 125 a , a set of indications of actual behavior of the instance of the application within the sandbox environment 105 a .
- the event engine module 127 a can store the actual behavior indications in the data store 135 .
- the event engine module 127 a analyzes the actual behavior of the instance of the application in comparison with the allowed behavior associated with the instance of the application in the data store 135 , to detect correspondences and discrepancies. If the allowed behavior and the actual behavior correspond, the event engine module 127 a can conclude that no infection has occurred and malicious behavior has not been detected.
- the event engine module 127 a sends an indication associated with an anomalous behavior to the application control module 123 a .
- the indication associated with the anomalous behavior can be used by the application control module 123 a to terminate the application and/or to terminate sandbox environment's process, to send an indication of the anomalous behavior to a system administrator or to a user, and/or the like.
- the application control module 123 a can send the indication associated with the anomalous behavior to the event engine module 127 a , for example to define an evaluation tree.
- the evaluation tree including the indication associated with the anomalous behavior can be stored in data store 135 associated with the application, for example added to the rules database 136 a associated with the application.
- the sandbox monitor module 125 a can be configured to monitor at least one event of the instance of the application within the sandbox environment 105 a such as, for example, process events (e.g., executing unauthorized processes) of the instance of the application, file events (e.g., access of unauthorized files) of the instance of the application, registry events (e.g., registry updates) of the instance of the application, network events (e.g., network connection events) of the instance of the application, or thread injection events (e.g., behavior associated with inserting and run executable code within the address space of another process) of the instance of the application, etc.
- process events e.g., executing unauthorized processes
- file events e.g., access of unauthorized files
- registry events e.g., registry updates
- network events e.g., network connection events
- thread injection events e.g., behavior associated with inserting and run executable code within the address space of another process
- the rule generation module 133 uses the anomalous behavior detected by the event engine module 127 a and revises the set of indications (rules) of allowed behavior associated with the application.
- the revised set of rules can replace the set of rules in the rules database 136 a.
- the indication associated with the anomalous behavior can include a trace associated with a source of the anomalous behavior.
- the application can be a web browser application and the indication associated with the anomalous behavior can include a uniform resource locator (URL) trace associated with the web browser application.
- the source trace can include an identifier of an application, process, service and/or the like. The source trace enables the rule generation module 133 to revise the rules to enable the sandbox control module 123 a to control application access to the sources that cause anomalous behavior of the application.
- the set of rules indicating allowed behavior associated with the application is based at least in part on a trust level associated with the application.
- the trust level can be defined by the trust classification module 137 based on, for example, user input, previous behavior of the application (e.g., history of actual behavior in data store 135 ), application type, etc.
- the set of indications of allowed behavior associated with the application can include an identifier of a trusted process associated with the application.
- the sandbox monitor module 125 a can monitor activities of an instance of an application within the sandbox environment 105 a by tracing network activities of the running instance of the application within the sandbox environment 105 a .
- the sandbox monitor module 125 a can also monitor one or more executable files of the instance of the application.
- FIG. 5 is a flowchart of a process for control and detection of malicious behavior of an application based on application type, according to an embodiment.
- the sandbox environment 105 a can have multiple instances or different applications or instances of the same application running simultaneously.
- the sandbox monitor 125 a and the application control module 123 a can monitor and control the running applications.
- the application control module 123 a initiates an instance of a first application and an instance of a second application within a sandbox environment 105 a.
- the event engine module 127 a receives, from a sandbox monitor module 125 a associated with the sandbox environment 105 a , a set of indications of actual behavior of the instance of the first application and a set of indications of actual behavior of the instance of the second application.
- the first application and the second application can be from the same application type (two instances of the same application running simultaneously on sandbox environment 105 a ) or from different application types (two different applications running simultaneously on sandbox environment 105 a ).
- the set of indications of actual behavior of the instance of the first application and the set of indications of actual behavior of the instance of the second application can overlap such that the instance of the first application and the instance of the second application may show the same actual behavior.
- the same actual behavior that can be declared as anomalous behavior for an instance of the first application can be declared as an allowed behavior for an instance of the second application, or vise versa.
- the event engine module 127 a checks the rules database 136 a to find a baseline behavior set for the first application (e.g., a set of allowed behavior for the first application). If the actual behavior is an anomalous behavior for the first application based on the baseline behavior set for the first application in the rules database 136 a , the event engine module 127 a classifies the actual behavior as anomalous behavior for the first application.
- a baseline behavior set for the first application e.g., a set of allowed behavior for the first application.
- the event engine module 127 a checks the rules database 136 a to find a baseline behavior set for the second application. If the actual behavior is not an anomalous behavior for the second application based on the baseline behavior set for the second application (e.g., a set of allowed behavior for the second application) in the rules database 136 a , event engine module 127 a does not classify the actual behavior as anomalous behavior for the second application.
- the baseline behavior set for the second application e.g., a set of allowed behavior for the second application
- the event engine module 127 a can classify a behavior B as an anomalous behavior for the first application based on a baseline behavior set for the first application, but not classify the behavior B as an anomalous behavior for the second application based on a baseline behavior set for the second application.
- the sandboxed protection system 203 can be located within the UEs 201 a - 201 n , or in a separate location coupled with the UEs 201 a - 201 n via the communication network 205 . Therefore, in some instances, as shown in FIG. 2 , the sandboxed protection module 203 can include the sandbox environments 207 a - 207 n.
- the sandbox control module 123 a can be operatively coupled to the sandbox environment 105 a via a network.
- the sandboxed protection system 203 that includes the sandbox control module 123 a is coupled to the UEs 201 a - 201 n , that include sandbox environments 207 a - 207 n , via communication network 205 by network connections 213 and 219 .
- the set of indications of actual behavior of the instance of the first (or second) application includes at least one of a process event identifier of the instance of the first (or second) application, a file event identifier of the instance of the first (second) application, a registry event identifier of the instance of the first (or second) application or a network event identifier of the instance of the first (second) application.
- the event engine module 127 a can send a signal to other components of the sandboxed protection system 121 a such as the application control module 123 a to indicate classifying of the actual behavior as anomalous behavior for the first application.
- the application control module 123 a can send a signal to initiate actions such as first application termination, sandbox environment 105 a termination, notifying a system administrator or a user about the anomalous behavior, etc.
- FIG. 6 is a flowchart of a process for control and detection of malicious behavior of an application based on another application, according to an embodiment.
- an application running within the sandbox environment 105 a can initiate one or more other applications within the sandbox 135 .
- the sandbox control module 123 a receives, from a rules database 136 a , a set of indications of allowed behavior associated with a first application.
- the sandbox control module 123 a initiates an instance of the first application within a sandbox environment 105 a such that the instance of the first application initiates an instance of a second application within the sandbox environment 105 a .
- execution of the first application can start execution of a second application without an intervention by the application control module 123 a .
- the sandbox monitor module 125 a can detect initiation of the second application and report the initiation to the application control module 123 a and/or to the event engine module 127 a.
- event engine module 127 a receives, from the sandbox monitor module 125 a associated with the sandbox environment 105 a , a set of indications of actual behavior of the instance of the second application, as previously discussed with regards to FIG. 5 , in response to the instance of the first application initiating the instance of the second application.
- the first application and the second application can be from the same application type or from different application types.
- the event engine module 127 a sends an indication associated with an anomalous behavior associated with the second application to the application control module 123 a .
- the indication associated with the anomalous behavior can be used by the application control module 123 a to terminate the second application and/or the sandbox environment process, etc.
- the event engine module 127 a can add the anomalous behavior to an evaluation tree associated with the second application.
- the rule generation module 133 uses the anomalous behavior of the second application detected by the event engine module 127 a and revises the set of indications (rules) of allowed behavior associated with the first application and/or the set of indications (rules) of allowed behavior associated with the second application.
- the revised set of indications can replace the set of indications in rules database 136 a.
- the second application can be, for example, a web browser application and the indication associated with the anomalous behavior can include a uniform resource locator (URL) trace associated with the web browser application.
- the source trace enables the rule generation module 133 to revise the set of indications (rules) of allowed behavior associated with the first and/or second application such that access of second application to the sources that caused anomalous behavior can be controlled.
- the set of indications of allowed behavior associated with the first application is based at least in part on a trust level associated with the first application.
- the trust level can be defined by the trust classification module 137 based on, for example, user input, previous behavior of the first application (e.g., history of actual behavior in data store 135 ), first application type, previous behavior of the second application (e.g., history of actual behavior in data store 135 ), second application type, etc.
- the set of indications of allowed behavior associated with the first application can includes an identifier of a trusted process associated with the first application.
- Hardware modules may include, for example, a general-purpose processor, a field programmable gate array (FPGA), and/or an application specific integrated circuit (ASIC).
- Software modules (executed on hardware) can be expressed in a variety of software languages (e.g., computer code), including C, C++, JavaTM, Ruby, Visual BasicTM, and other object-oriented, procedural, or other programming language and development tools.
- Examples of computer code include, but are not limited to, micro-code or micro-instructions, machine instructions, such as produced by a compiler, code used to produce a web service, and files containing higher-level instructions that are executed by a computer using an interpreter. Additional examples of computer code include, but are not limited to, control signals, encrypted code, and compressed code.
- Some embodiments described herein relate to a computer storage product with a non-transitory computer-readable medium (also can be referred to as a non-transitory processor-readable medium) having instructions or computer code thereon for performing various computer-implemented operations.
- the computer-readable medium or processor-readable medium
- the media and computer code may be those designed and constructed for the specific purpose or purposes.
- non-transitory computer-readable media include, but are not limited to: magnetic storage media such as hard disks, floppy disks, and magnetic tape; optical storage media such as Compact Disc/Digital Video Discs (CD/DVDs), Compact Disc-Read Only Memories (CD-ROMs), and holographic devices; magneto-optical storage media such as optical disks; carrier wave signal processing modules; and hardware devices that are specially configured to store and execute program code, such as Application-Specific Integrated Circuits (ASICs), Programmable Logic Devices (PLDs), Read-Only Memory (ROM) and Random-Access Memory (RAM) devices.
- ASICs Application-Specific Integrated Circuits
- PLDs Programmable Logic Devices
- ROM Read-Only Memory
- RAM Random-Access Memory
Landscapes
- Engineering & Computer Science (AREA)
- Computer Security & Cryptography (AREA)
- Software Systems (AREA)
- Theoretical Computer Science (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Virology (AREA)
- Information Transfer Between Computers (AREA)
Abstract
Description
- This application is a continuation of U.S. patent application Ser. No. 13/690,452, entitled “Methods and Apparatus for Control and Detection of Malicious Content Using a Sandbox Environment,” filed Nov. 30, 2012, which claims priority to and the benefit of U.S. Provisional Patent Application No. 61/566,162, filed Dec. 2, 2011 and entitled “Methods and Apparatus for Control and Detection of Malicious Content Using a Sandbox,” each of which is incorporated herein by reference in its entirety.
- Some embodiments described herein relate generally to a sandboxed protection system that protects a computing device from malicious content.
- Information and application providers are continually challenged to deliver value and convenience to consumers by, for example, providing compelling information protection mechanisms. As the amount of the digitally available information increases, the amount of viruses and other malicious information or software (malware) that can harm computing devices and endanger information safety and security is also on the rise. Some known anti-virus software is used to prevent, detect and remove various types of malware. Some known anti-virus clients, however, for example, use signature-based detection methods, in which a computer is scanned for traces of known threats. Signatures can be developed based on examining known malware. Thus, threats previously unknown to anti-virus clients can go unnoticed and/or undetected.
- Thus, it would be desirable to provide an alternative protection system, such as a sandbox that instead of using a signature-based detection paradigms, can detect malicious software based on, for example, software behavior.
- In some embodiments, a non-transitory processor-readable medium storing code represents instructions to cause a processor to receive a set of indications of allowed behavior associated with an application. The processor is also caused to initiate an instance of the application within a sandbox environment. The processor is further caused to receive, from a monitor module associated with the sandbox environment, a set of indications of actual behavior of the instance of the application in response to initiating the instance of the application within the sandbox environment. The processor is also caused to send an indication associated with an anomalous behavior if at least one indication from the set of indications of actual behavior does not correspond to an indication from the set of indications of allowed behavior.
-
FIG. 1A is a schematic illustration of a sandboxed protection system, according to an embodiment. -
FIG. 1B is a schematic illustrations of a sandbox environment within a user device, according to an embodiment. -
FIG. 2 is a schematic block diagram of a computer network system to control and detect malicious content, according to an embodiment. -
FIG. 3 is a flowchart of a process for implementing a sandboxed protection system, according to an embodiment. -
FIG. 4 is a flowchart of a process for control and detection of malicious behavior of an application based on allowed behavior, according to an embodiment. -
FIG. 5 is a flowchart of a process for control and detection of malicious behavior of an application based on application type, according to an embodiment. -
FIG. 6 is a flowchart of a process for control and detection of malicious behavior of an application based on another application, according to an embodiment. - Known anti-virus software is used to prevent, detect and remove various types of malware from computer devices. Some known anti-virus clients, however, use signature-based detection methods, in which a computer is scanned for traces of known threats. Signatures can be developed based on examining known malware. Thus, threats previously unknown to anti-virus clients can go unnoticed and/or undetected.
- A sandbox can be used to separate and/or isolate running programs from each other. A sandbox can provide a controlled set of resources for use by suspicious programs. In some embodiments, a sandboxed protection system can detect malware in a way that overcomes shortcomings of the traditional anti-virus clients and network gateway products. In some embodiments, for example, the sandboxed protection system can use behavioral-based malware detection.
- In some embodiments, behavior-based malware detection includes recognizing behaviors that software applications are allowed to perform. Behaviors not recognized can be considered suspect. For a typical user's computer, this can be an extraordinarily complex problem because the configuration is a dynamic operation. For example, over time, many applications can be installed and run on the computer. Many of these applications can access unsafe content on a network such as, for example, the Internet. In some embodiments, the sandboxed protection system can be designed to monitor processes and/or applications such as web browsers, document readers, etc. that are running in a sandbox environment. Such a sandbox environment can include applications installed by and/or recognized by the sandboxed protection system. In some embodiments, applications not installed by and/or recognized by the sandboxed protection system can be excluded from the sandbox environment. Thus, the sandboxed protection system can understand the behaviors of the applications running in the sandbox environment. In some embodiments, any behavior that falls outside of “normal behavior” (i.e., anomalous behavior) can be classified as a potential infection.
- In some embodiments, after the sandboxed protection system detects anomalous behavior, the sandbox environment is no longer clean. The event that produced the anomalous behavior can be labeled as an infection. In some embodiments, the user can be informed of the infection and/or can be prompted to restore the sandbox to a clean state. Additionally, in some embodiments, the behaviors originating from the infection can be collected and/or organized into a forensic report. This forensic report can be provided to a separate data server when the user restores the sandbox to a clean state. In other embodiments, the forensic report can be stored locally instead of or in addition to providing the report to the separate data server. A system administrator can, based on the forensic report, view the forensic data. In some embodiments, the forensic data can describe and/or show in detail what happened during the infection session, where the infection came from, which portions of the sandbox environment were infected and/or which devices within a network were infected. In other embodiments, the forensic data can include a summary of the infections. In some embodiments, system administrators can view the full scope of infections that could have occurred on their network.
- Apparatus, systems and methods are described herein to provide control and detection of malicious content using a sandbox environment. In some embodiments, a non-transitory processor-readable medium storing code representing instructions to cause a processor to perform a process includes code to cause the processor to receive a set of indications of allowed behavior associated with an application. The processor is also caused to initiate an instance of the application within a sandbox environment. The processor is further caused to receive, from a monitor module associated with the sandbox environment, a set of indications of actual behavior of the instance of the application in response to initiating the instance of the application within the sandbox environment. The processor is also caused to send an indication associated with an anomalous behavior if at least one indication from the set of indications of actual behavior does not correspond to an indication from the set of indications of allowed behavior.
- In some embodiments, an apparatus includes a control module implemented in at least one of a memory or a processing device. The control module is configured to initiate an instance of a first application and an instance of a second application within a sandbox environment. The control module is also configured to receive, from a monitor module associated with the sandbox environment, a set of indications of actual behavior of the instance of the first application and a set of indications of actual behavior of the instance of the second application. An indication of a behavior is within both the set of indications of actual behavior of the instance of the first application and the set of indications of actual behavior of the instance of the second application. The control module is also configured to classify the behavior as an anomalous behavior for the first application based on a baseline behavior set for the first application. The control module is configured to not classify the behavior as an anomalous behavior for the second application based on a baseline behavior set for the second application. The control module is further configured to send a signal in response to classifying the behavior as an anomalous behavior for the first application.
- In some embodiments, an apparatus includes a control module implemented in at least one of a memory or a processing device. The control module is configured to receive a set of indications of allowed behavior associated with a first application. The control module is configured to initiate an instance of the first application within a sandbox environment such that the instance of the first application initiates an instance of a second application within the sandbox environment. The control module is also configured to receive, from a monitor module associated with the sandbox environment, a set of indications of actual behavior of the instance of the second application in response to the instance of the first application initiating the instance of the second application. The control module is also configured to send an indication associated with an anomalous behavior if at least one indication from the set of indications of actual behavior of the instance of the second application does not correspond to an indication from the set of indications of allowed behavior associated with the first application.
- As used herein, the singular forms “a,” “an” and “the” include plural referents unless the context clearly dictates otherwise. Thus, for example, the term “a “sandbox environment” is intended to mean a single sandbox environment or a combination of sandbox environments (e.g., sandbox environments with a series of configurations for classes of applications, for applications with different levels of trust, etc.).
-
FIG. 1A is a schematic illustration of a sandboxed protection system, according to an embodiment. As shown inFIG. 1A , asandboxed protection system 121 a can be operatively coupled to at least one User Equipment (UE) 101 a. AUE 101 a can be a personal computer, a tablet computer, a mobile telephone, a smart telephone, a personal data assistant (PDA), etc. TheUE 101 a can include at least asandbox environment 105 a that operates under a host operating system 103 of theUE 101 a. TheUE 101 a can further include adata store 109 a. A data store can be, for example, a memory, a data storage device such as a hard drive, or any component or recording media used to retain digital data. - In some instances, the
sandbox environment 105 a can include anapplication helper module 107 a. Theapplication helper module 107 a is an application extension or add-on that monitors activity within an application running onUE 101 a. For example, anapplication helper module 107 a can be used, under the supervision of thesandboxed protection system 121 a, to monitor an identifier of a source of an anomalous behavior. For example, theapplication helper module 107 a can monitor Uniform Resource Locators (URLs) accessed by a user ofUE 101 a using a browser application. - As used herein, a module can be, for example, any assembly and/or set of operatively-coupled electrical components, and can include, for example, a memory, a processor, electrical traces, optical connectors, software (executing or to be executed in hardware) and/or the like. Furthermore, a module can be capable of performing one or more specific functions associated with the module, as discussed further below.
- The
sandboxed protection system 121 a can provide a security mechanism for separating running programs associated withUE 101 a. In some instances, asandbox environment 105 a can be used byUE 101 a under control of thesandboxed protection system 121 a to execute untested code, or untrusted programs from unverified third-parties, suppliers, untrusted users and untrusted websites. Thesandboxed protection system 121 a can provide a tightly-controlled set of resources in which guest programs can run, such as scratch space on adata store 109 a. In some instances, thesandboxed protection system 121 a can limit, disallow, or restrict capabilities of aUE 101 a such as, for example, network access, ability to inspect a host system, read from input devices, etc. - As shown in
FIG. 1A , asandboxed protection system 121 a can include anapplication control module 123 a, asandbox monitor module 125 a, anevent engine module 127 a, aguest monitor module 129 a, a kernel mode module 131, a rule generation module 133, a data store 135, and a trust classification module 137. Furthermore, thesandboxed protection system 121 a communicates with one or more User Equipments (UEs) 101 a or other devices, computer network systems, etc., viainput signal 139 andoutput signal 141. Thus,FIG. 1A is merely an example illustrating the types of components that can be included within asandboxed protection system 121 a. - In various instances, the
sandboxed protection system 121 a and its components may be located anywhere within a communication network system (not shown inFIG. 1A ) including, but not limited to, within theUE 101 a, within one or more service provider devices (not shown), or in separate locations within the communication network system. - In some instances, the
application control module 123 a is configured to control thesandbox environment 105 a by sending an output signal to thesandbox environment 105 a via theoutput signal 141. For example, theapplication control module 123 a can initiate a module within theUE 101 a that is executed by thehost operating system 103 a using thesandbox environment 105 a ofUE 101 a. Theapplication control module 123 a can initiate thesandbox environment 105 a, for example by activating one or more applications and/or processes within thesandbox environment 105 a. Theapplication control module 123 a can also terminate asandbox environment 105 a by terminating one or more applications and/or processes running within thesandbox environment 105 a. Theapplication control module 123 a can also initialize or clean thesandbox environment 105 a by, for example, restarting thesandbox environment 105 a or restoring a standard or predefined set of configurations to the sandbox environment 105. - In some instances, the
sandbox control module 123 a is configured to initiate a module to run within thesandbox environment 105 a to monitor application activity within thesandbox environment 105 a. In some embodiments, theapplication helper module 107 a can monitor activity within thesandbox environment 105 a that thehost operating system 103 a is unable to monitor outside thesandbox environment 105 a. - In some instances, the kernel mode module 131 can control a kernel mode of operation of software applications within the
host operating system 103 a. In some instances, in kernel mode operation, a process can have complete and unrestricted access to the underlying hardware of theUE 101 a. In such instances, the kernel mode module 131 enables a running process to execute any Central Processing Unit (CPU) (not shown) instruction and/or reference any memory (not shown) address. In some instances, the kernel mode module 131 can reserve the kernel mode operation for lowest-level, most trusted functions of thehost operating system 103 a. - In some instances, the
event engine module 127 a can control a user mode of operation of software applications within thehost operating system 103 a. In some instances, in user mode operation, a process is unable to directly access hardware and/or reference memory of theUE 101 a. In some instances, theevent engine module 127 a enables a process running in user mode to delegate requests to system Application Programming Interfaces (APIs) to access hardware or memory of theUE 101 a. The protection provided by the user mode isolation, enables thehost operating system 103 a to recover crashes in user mode. In some instances, most of the processes and/or code associated with applications running on aUE 101 a can be executed in user mode controlled by theevent engine module 127 a. - In some instances the
application control module 123 a can activate theevent engine module 127 a. Theevent engine module 127 a can be included within theapplication control module 123 a. Theevent engine module 127 a is a rule-based event processing engine that can determine if an event (e.g., a process, a service, etc.) is allowed to be executed onUE 101 a or is an infection (e.g., a malware). - In some instances, the
event engine module 127 a can be responsible for receiving event data and analyzing the received data to determine if an anomalous behavior has occurred. In some instances, theevent engine module 127 a can receive data associated with events in substantially real-time. - In some instances, the
sandboxed protection system 121 a can substantially constantly and/or periodically detect anomalous behavior and/or collect data associated with the anomalous behavior from applications and/or processes in thesandbox environment 105 a. In some instances, the collected data can be evaluated by theevent engine module 127 a using a script (e.g., a rule-based filter) and the associated processes can be added to a tree of processes and/or process behaviors referred to herein as an evaluation tree, stored indata store 109 a or in data store 135. In some instances, the evaluation tree can be used by theevent engine module 127 a to keep track of and/or monitor the current state of thesandbox environment 105 a. For example, if application App1 launches application App2, then App1 is defined as a node in the evaluation tree with App2 as its child. In some instances, behaviors associated with a potential infection are added to the evaluation tree and/or behaviors not associated to a potential infection are not added to the evaluation tree. In some instances, as more information is gathered about a node (e.g., an application) in the evaluation tree, new information can be associated with the node as an attribute. Upon occurrence of an anomalous behavior, the evaluation tree can be serialized to a forensic report. The forensic report can be stored in adata store 109 a or a data store 135 as a record of an infected session. - In some instances, in order to effectively define what a sandboxed application, process and/or service is allowed to do, features of a product including the application and/or the process can be defined ahead of time, for example by the provider, or at run time, for example by the
sandboxed protection system 121 a, via regulated authorization. For example, a rule can be defined to identify behaviors of a sandboxed application, process and/or service that may trigger an infection. The rule can be used by theevent engine module 127 a to determine a behavior as allowed or not-allowed for the sandboxed application, process and/or service. In some instances, when defining allowable rules for file and registry events, techniques such as, for example, wild cards can be used as a shortcut to match a single rule to multiple files and/or folders. Thus, in such instances, a balance can be reached regarding how strict each rule should be. If too many events associated with a process are allowed, then a potential infection can be missed. However, if too many events are restricted, then a false positive can alert the user ofUE 101 a, in which an allowed behavior is detected as anomalous. - Additionally, in some instances, some system administrators can have the authority to pre-configure the sandboxed applications with their own software applications, processes and/or browser add-ons. In such instances, the system administrators can build and/or define a reliable set of rules that covers the allowed behaviors of the applications and/or processes and enters the rules into the
sandboxed protection system 121 a viainput signal 139. The rules may be stored by thesandboxed protection system 121 a inrules database 136 a and accessed by theapplication control module 123 a,event engine module 127 a or other components of thesandboxed protection system 121 a. In some instances, to handle these challenges, thesandboxed protection system 121 a can associate levels of trust to the applications and/or processes being monitored in thesandbox environment 105 a. - In some instances, the
event engine module 127 a can receive a notification from thesandbox monitor module 125 a associated with a behavior of an instance of an application running within thesandbox environment 105 a. For example, theevent engine module 127 a may receive a notification from thesandbox monitor module 125 a indicating that an instance of a Web browser running within thesandbox environment 105 a is trying to modify a system file. Theevent engine module 127 a can receive a set of rules from therules database 136 a describing allowed behavior associated with the application. The set of rules may include rules describing file modification rights of the Web browser application. The event engine module 131 can analyze the set of rules and determine whether the system file modification by the Web browser is allowed. If the modification is allowed by the Web browser theevent engine module 127 a can send a signal to theapplication control module 123 a that the modification attempt by the Web browser is an allowed behavior. Theapplication control module 123 a can, in response, allow the Web browser instance to continue execution and modify the system file. Otherwise, if the analysis results by theevent engine module 127 a indicate that the attempt by the Web browser to modify the file system is not allowed, theevent engine module 127 a indicates the system file modification as an anomalous behavior and sends the indication to theapplication control module 123 a. Theapplication control module 123 a can, in response, terminate the file system modification attempt, terminate the Web browser, terminate thesandbox environment 105 a, or take other predetermined actions associated with such anomalous behaviors. Furthermore, theevent engine module 127 a can store the anomalous behavior in therules database 136 a associated with the Web browser application. - In some instances, the trust classification module 137 can associate a trust level to each application, process and/or service that is executed within the
sandbox environment 105 a. The trust classification module 137 can also associate a set of trusted processes with each application, process and/or service. The trust level and/or the set of trusted processes may be defined at least based on past behavior of the application, process and/or service, user input, application, process and/or service type, type of data associated withUE 101 a that the application, process and/or service attempts to access, etc. The trust classification module 137 can store the trust level and/or the set of trusted processes associated with each application, process and/or service at therules database 136 a. - In some instances, the
sandbox monitor module 125 a can monitor activities of an instance of an application within thesandbox environment 105 a by tracing network activities of the running instance of the application within thesandbox environment 105 a. Thesandbox monitor module 125 a can also monitor one or more executable files of the instance of the application. - In some instances, different levels of trust can be associated with applications and/or processes based on monitoring of the applications and/or processes by the
sandbox monitor module 125 a. For example, in some instances, an application, process and/or service can be classified as “trusted”, “untrusted”, “suspect”, or “unknown”. In some instances, such a classification can be performed by a trust classification module 137 of thesandboxed protection system 121 a. In some instances, the rule author (e.g., a system administrator) can make a determination of whether an application, process and/or service in thesandbox environment 105 a is trusted or untrusted. Thesandboxed protection system 121 a can then automatically assign trust levels suspect and unknown after further evaluation of the application, process and/or service. In other instances, the rule author can also make the determination of whether the application, process and/or service is suspect or unknown. In still other instances, thesandboxed protection system 121 a can automatically assign a trust level to each application, process and/or service based on a type of the application, process and/or service (e.g., game, browser, word processor, etc.), observed behavior of the application, process and/or service, etc. In some instances, a system administrator can request thesandboxed protection system 121 a to override a trust level for an application, process and/or service. Similarly, in some instances, a system administrator can request thesandboxed protection system 121 a to modify the trust levels of applications and/or processes. Classifying the trustworthiness of each process in a guest application can provide flexibility for customization of the scope of malware detection. - In some instances, declaring a process as trusted in the rules, can essentially imply that any possible behavior exhibited and/or actions performed by the process and/or the process's child processes are considered as allowed behavior. In some instances, trusted processes can be allowed to perform any action. In such instances, by classifying a process as trusted, the
sandboxed protection system 121 a can effectively disable detection for that process. In such instances, for example, if a trusted process launches a child process, the child process can inherit the trust and become trusted. As another example, if a file is written by a trusted process and launched by a different process, the new process can be trusted because the file inherits the trust from the trusted process. This behavior can be common for programs that update themselves. In other instances, a system administrator can request thesandboxed protection system 121 a for manual or rule-based reclassification of each child process launched and/or written by a trusted process. - Trusting a process is, typically, a less secure option than not trusting the process. Thus, in some instances the
sandboxed protection system 121 a can warn the system administrators about the risk of adding new software into thesandbox environment 105 a. For example, a system administrator that would like to install a WebEx application, process and/or service in thesandbox environment 105 a but would also like to avoid defining allowed behaviors for the WebEx process and all child processes can declare WebEx.exe as a trusted process. In some instances, a system administrator can declare a process as trusted by providing the path to the root process to a trusted process section of the rules stored inrules database 136 a and processed by theevent engine module 127 a. - In some instances, processes that are monitored by the
sandbox monitor module 125 a as potential propagation methods for infection can be declared and/or classified as untrusted by the trust classification module 137. In some instances, if theevent engine module 127 a detects that an untrusted application, process and/or service performs a suspicious behavior that is not explicitly defined as allowed, this is a trigger for an infection. In some instances, therules database 136 a can include whitelisted rules for untrusted applications and/or processes that define normal trusted operating behaviors of the applications and/or processes. - For example, in a Microsoft Windows environment, a main account in the
sandbox environment 105 a can run and/or execute as a limited user. Therefore, some processes associated with the main account may not have permission to modify registry keys inside HKEY_LOCAL_MACHINE. If a process without having the permission attempts to modify a portion of the registry key, thesandboxed protection system 121 a can recognize the attempted modification as a trigger behavior for an infection. A process P, however, which runs in the system context, can have permission to modify registry keys inside HKEY_LOCAL_MACHINE. In order to prevent process P from triggering an infection, a rule that allows P to access the registry key can be defined (as seen in the Heuristic Table shown below). - In some instances, after an infection has occurred, the process that caused the infection can be classified, by the
sandboxed protection system 121 a, as suspect. In some instances, after the process is classified as suspect, thesandboxed protection system 121 a can record subsequent behaviors being performed by the process and/or its child processes. In some instances, similar to trusted processes, behavior originating from a suspect process can inherit the suspect trust level. After an event is evaluated as being suspect by theevent engine module 127 a, it can be added to the evaluation tree so that it can be serialized into a forensic report (e.g., a report file in data store 135). The forensic report can be analyzed by theevent engine module 127 a and can provide details of the suspect events. - In some instances, additional information on suspect behaviors and/or events associated with processes and applications such as, for example, cryptographic hash functions (e.g., Message Digest (MD5) hashes, Secure Hash Algorithm (SHA) hashes, etc.) providing unique signatures for files, can be captured by the
sandboxed protection system 121 a. After a hash value is calculated, it can be sent to theevent engine module 127 a to be attached as an attribute to the application and/or the process in the evaluation tree. The hash value can be used to lookup and/or compare the triggering application, process and/or service against a third-party malware (e.g., virus) database. This lookup can provide information about number of anti-virus clients that scanned the process as being infected. - In some instances, the
sandboxed protection system 121 a also includes a user-configurable feature which can terminate and/or kill any suspect processes as they are detected. In some instances, the process termination can be accomplished by terminating and/or killing a process as soon it is classified as suspect. In other instances, thesandboxed protection system 121 a can provide information about the suspect processes to a system administrator and allow the administrator to terminate and/or kill a process. - In some instances, a process can be classified as unknown if its path is not explicitly listed as trusted or untrusted in the
rules database 136 a. In some instances, because this process has not yet been configured and/or classified, it can perform actions and/or behaviors without triggering an infection. In some instances, however, unlike trusted processes, the unknown trust of this process is not inherited across child processes. An unknown process can, for example, launch an application App1, and application App1 can be tracked as untrusted. In some instances, because unknown process behaviors are important to know about and evaluate, the behaviors of an unknown process can be added to the evaluation tree and can show up in an infection report (e.g., a forensic report) if the session becomes tainted. While four trust levels are shown and described herein, in other instances any number of trust levels can be used. - In some instances, designing the allowable behavior rules can depend on the types of behaviors that can trigger an infection. In some instances, for example, infections can be determined based on the following heuristics table. In various instances, other infection triggers can be added with additional sensors and/or types of events.
- In some instances, when an infection is detected, an attempt can be made by the
sandboxed protection system 121 a to determine the originating source of the infection. For example, if the infection is triggered by a web browser (e.g., Internet Explorer, Firefox, etc.), or a child process of a web browser, then the source can be determined based on a URL history of a user ofUE 101 a. The URL trace events can be captured from anapplication helper module 107 a attached to the web browser. The URL trace events can be fed to theevent engine module 127 a, which then correlates the main URL and any URLs redirected from that main URL with an infection. Theevent engine module 127 a can correlate the URLs by tracing backwards in time from the most recent URL matching a Process Identifier (PID) associated with an infection root. For another example, a process, application, service and/or the like can be identified as the originating source of the infection. - In some instances, for example, the
sandboxed protection system 121 a can allow a user ofUE 101 a to launch a portable document format (PDF) file on theUE 101 a for viewing inside thesandbox environment 105 a. Since PDF readers are a prime attack propagation method for viruses, thesandboxed protection system 121 a can trace the source to the PDF file if an infection is originated from the PDF reader or its child processes. Thesandboxed protection system 121 a can do this by keeping and/or maintaining a map of processes in thesandbox environment 105 a that were launched when the PDF file was redirected from thehost operating system 103 a. When an infection occurs, thesandboxed protection system 121 a can check the map to see if the PDF file belongs to an infected process. This technique can be used to track an infection if other redirected documents from thehost operating system 103 a are determined to be the source of an infection. -
Heuristics Table 1. If an untrusted process launches another process that is not explicitly allowed, 2. If an untrusted process terminates another process that is not explicitly allowed, 3. If an untrusted process writes a file and the file is launched by a process with unknown trust, 4. If an untrusted process modifies a sensitive registry value that is not explicitly allowed, for example: a. HKEY_LOCAL_MACHINE b. HKEY_CLASSES_ROOT c. Any windows startup key d. Any policy key, 5. If an untrusted process modifies a sensitive file path that is not explicitly allowed, Examples of restricted paths are: a. C:\Windows\System32 b. C:\Program Files 6. If an untrusted process deletes, renames, or overwrites an executable file, 7. If an untrusted process makes a TCP/UCP connection that is not explicitly allowed, 8. If an untrusted process set up a TCP/UDP listener that is not explicitly allowed. - After the source of an infection is determined, the source can be added as a node in the evaluation tree as a child of the infected process and/or application. This source can be included in the summary of infection report (e.g., the forensic report) provided to a system administrator by the
sandboxed protection system 121 a. In some instances, if no source for the infection is found, the name of the infected process can be used. - As previously discussed, rules can be defined and stored in the
rules database 136 a by users, by system administrators, by application providers, etc. A set of rules can be associated with each application, process and/or service. The set of rules associated with an application, process and/or service define allowed behavior of the application, process and/or service. In some instances, the rules associated with an application, process and/or service can be used by theevent engine module 127 a to determine whether a behavior triggers and/or is associated with an infection. The rules can be divided up into a plurality of rule sets, for example a main rule set (e.g., rules associated with thehost operating system 103 a), a custom rule set (e.g., rules associated with applications and/or processes), etc. In some instances, the main rule set can be provided with and/or standard to thesandboxed protection system 121 a. In some instances, system administrators can introduce their own custom rule sets by entering rules into thesandboxed protection system 121 a viainput 139. In some instances, the applications executed insandbox environment 105 a (e.g., browsers) can be customizable for installing software. A custom rule set file can enable theevent engine module 127 a to define allowable behaviors for new processes that would otherwise trigger an infection. When theevent engine module 127 a is initialized, the main rule set can be merged with the custom rule sets and, for example, compiled into a byte code. The sets of rules may be stored in therules database 136 a. - In some instances, the
sandboxed protection system 121 a includes a rule generation module 133 that automatically generates rules for an application, process and/or service based on previously declared infections associated with same or similar applications and/or processes (e.g., applications with the same type). For example, the rules generation module 133 can generate rules based on infections that have been declared as false positives. The rule generation module 133 can also generate rules based on sets of rules existing in therules database 136 a for the application, process and/or service. The rule generation module 133 can address various issues associated with application, process and/or service behavioral detection. For example, the rule generation module 133 can define allowed behaviors for new applications. The rule generation module 133 can also provide a false positive infection report (e.g., forensic report) and convert the report into a new set of rules. The rule generation module 133 can further enable a system administrator to develop rules for their own set of plug-ins that have not been already declared and/or included in the main rule set. - In some instances, the
sandboxed protection system 121 a can include functionality to automatically generate allowed rules from an infection that may be a false positive. In some instances, for example, the system administrators can view the forensic report provided by thesandboxed protection system 121 a and provide an indication to allow a process and/or behavior of a process. For example, the system administrator can click a button labeled “Allow . . . ” next to the entry for a process and/or behavior of a process. In some instances, this button can enable the system administrator to add the process and/or behavior of the process to the main rule set (e.g., windows.rul) and/or a custom rule set (e.g., custom.rul). In some instances, a system administrator can provide an input to thesandboxed protection system 121 a to trust new processes, for example by selecting a checkbox for “Trust newly encountered processes”, the system administrator can automatically classify new processes as trusted, which, while less secure, can reduce the chance of false positives. - In some instances, when a user of a
UE 101 a generates custom rules, the forensic report, the main rule set, and/or the custom rule set (if exists) can be sent fromUE 101 a to the rule generation module 133 viainput signal 139. The rule generation module 133 can receive the inputs and generate a new custom rule set, for example to be saved inrules database 136 a. After the new custom rule set is generated, thesandboxed protection system 121 a can send a signal to theUE 101 a viaoutput signal 141. In response to the signal, theUE 101 a can provide an option to the user (e.g., a button) to display a save dialog that allows the user to retrieve the newly defined custom rule set from therules database 136 a. - Additionally, in some instances, the
sandboxed protection system 121 a can be set to a training mode. The training mode can be, for example, a command line parameter. In some instances, while thesandboxed protection system 121 a is in training mode, infections sent to theUE 101 a can be flagged by theevent engine module 127 a as training samples and distinguished visually (e.g., in a report and/or in the rule generation module 133) from real infections with an indicator such as a different color or special icon. This can help distinguish between actual infection reports and training samples that are used to generate rules by the rule generation module 133. In some instances, training samples can then be converted into additional rules automatically or via interaction with the system administrator. The events collected in the training sample can be filtered, reduced and/or output as new detection rules. The new rules can be downloaded and used by theevent engine module 127 a for future detection and/or analysis. - In some instances, for determining behavior of the processes executed in the
sandbox environment 105 a, multiple sensors (not shown) can be defined and/or set up on theUE 101 a to capture events (e.g. anomalous behavior) and collect data about events. In some instances, the sensors can include module detection (e.g., process detection). In some instances, even though thesandboxed protection system 121 a can classify a new process as trusted, false positives can still occur. For example, when writing rules for a WebEx plug-in, although WebEx is installed and webex.exe has been declared and/or classified as a trusted process, webex.dll is still loaded into a browser (e.g., Internet Explorer). Anything webex.dll does can show up as a behavior originating from the browser (e.g., iexplore.exe). Therefore in order to fully whitelist WebEx, thesandboxed protection system 121 a can whitelist suspicious behaviors for the browser that could otherwise be compromising. In order to solve this problem, the sandboxed protection system can understand what modules are doing irrespective of the process into which they are loaded. - Similarly stated, the sandboxed protection system can separate behaviors from modules and processes. Referring to the above example, the sandboxed protection system can declare, classify and/or recognize webex.dll as a trusted module. Anything webex.dll does can be allowed, however, the browser (e.g., iexplore.exe) can trigger an infection if it performs the same suspicious behavior. Similarly, in some embodiments, the sandboxed protection system can declare, classify and/or recognize webex.dll as untrusted but whitelist the behaviors originating from that module for added security.
- In some instances, the
guest monitor module 129 a provides an operating system device driver that can monitor and/or report activity associated with a network, a file system, a registry and/or initiation and/or termination of processes withinUE 101 a. - The data collected by the sensors can be stored in
data store 109 a and sent to theevent engine module 127 a to be processed and analyzed based on a previously defined list of allowed behavior for the processes. - In some instances, the events can be captured in substantially real-time. For example, events can be captured using a kernel driver that hooks onto and/or monitors process events, file events, registry events, and/or network events as the events are being performed. In some instances, for example, the
application helper module 107 a can be loaded in a web browser (e.g., Internet Explorer or Firefox) to capture Uniform Resource Locator (URL) events. These captured events can be sent to theevent engine module 127 a for analysis. Theevent engine module 127 a can process each event by running it through a rule-based filter. In some instances, the rules for the rule-based filter can be defined using a scripting language optimized to filter events efficiently. As shown inFIG. 1A , thehost operating system 103 a executes asandbox environment 105 a controlled by asandboxed protection system 121 a. - In some instances, the
sandboxed protection system 121 a can be activated by instructions stored in a memory (e.g., a non-transitory processor-readable medium 113) on theUE 101 a. Additionally, in some instances, thesandboxed protection system 121 a can include additional modules and/or engines such as, for example, a trust classification module, a rule engine, a detection engine, and/or the like (each not shown). Such modules and/or engines can be hardware modules and/or engines, or software modules and/or engines executing in hardware. -
FIG. 1B is a schematic illustrations of asandbox environment 105 b within aUE 101 b. InFIG. 1B components components FIG. 1A , respectively. In the instance shown inFIG. 1B , theapplication control module 123 b can execute within thehost operating system 103 b, theevent engine module 127 b can execute within theapplication control module 123 b, and thesandbox monitor module 125 b can execute within thesandbox environment 105 b. In other embodiments, thesandbox monitor module 125 b can execute on thehost operating system 103 b but outside thesandbox environment 105 b. - In some instances,
applications 153 such as web browsers, documents viewers, etc. are executed within thesandbox environment 105 b. Thesandbox monitor module 125 b and theapplication helper module 107 b report application events (e.g., application behaviors) 155 associated withapplications 153 to theapplication control module 123 b. Theapplication control module 123 b can control thesandbox environment 105 b (shown as arrow 157), for example by initiating, terminating, or modifying applications and/or processes executed within thesandbox environment 105 b using theapplication events 155. Furthermore, theguest monitor module 129 b reports kernel mode events such as events associated with thehost operating system 103 b to theapplication control module 123 b (shown as arrow 159). In some instances, theapplication control module 123 b and/or theevent engine module 127 b can use theoperating system events 159 for controlling theapplications 153 within thesandbox environment 105 b. - Examples of the processes by
sandboxed protection system 121 a are further described below in connection with flowcharts ofFIG. 3 ,FIG. 4 ,FIG. 5 , andFIG. 6 . -
FIG. 2 is a schematic block diagram of a computer network system to control and detect malicious content, according to an embodiment. In some instances, acomputer network system 200 can include one or more user devices or user equipments (UEs) 201 a-201 n. UEs 201 a-201 n can be structurally and/or functionally similar toUE 101 a ofFIG. 1A . Thecomputer network system 200 further includes a sandboxed protection system 203 (structurally and/or functionally similar to thesandboxed protection system 121 a ofFIG. 1A ), one or more application provider(s) 209, one or more execution server(s) 211, each of which can be operatively coupled to each other via acommunication network 205. Thus,FIGS. 1A and 1B are merely examples illustrating the types of devices that can be included within acomputer network system 200. - In some instances, the
sandbox protection system 203 can control the sandbox environments 207 a-207 n via thecommunication network 205. In such instances, control signals can be sent between thesandbox protection system 203 and the sandbox environments 207 a-207 n via thecommunication network 205. For example, the sandbox environments 207 a-207 n can be running on anexecution server 211 that a user of UE 201 a-201 n can access remotely by a sandbox access module 221 a-221 n via thecommunication network 205. Each UE 201 a-201 n has access to a sandbox environment 207 a-207 n located within one or more execution server(s) 211. The sandbox environments 207 a-207 n are structurally and/or functionally similar to thesandbox environment FIGS. 1A and 1B . - In some embodiments, the
sandboxed protection system 203 can be operatively coupled to the UEs 201 a-201 n via, for example, acommunication network 205. In such embodiments, control signals can be sent between thesandboxed protection system 203 and the UEs 201 a-201 n via thecommunication network 205. As previously discussed, thesandboxed protection system 203 and its components may be located anywhere within acommunication network system 205 including, but not limited to, within the UEs 201 a-201 n, or in separate locations within thecommunication network system 205. - The execution server(s) 211 are equipped with one or more processors and one or more data storages (e.g., memories). A storage location on the execution server(s) 211 can be associated with the
UE 201 a as a sandbox environment 201. For example, theUE 201 a may have access to asandbox environment 207 a installed on the execution server(s) 211 such that theUE 201 a can access thesandbox environment 207 a via thecommunication network 205 and can execute applications and/or processes within thesandbox environment 207 a on the execution server(s) 211. In this example, theUE 201 a can initiate downloading an application from the application provider(s) 209 into thesandbox environment 207 a (on the execution server(s) 211). TheUE 201 a can then initiate execution of the application on the execution server(s) 211. Thesandbox protection system 203 can control the execution of the downloaded application on the execution server(s) 211 via thecommunication network 205. -
Communication network 205 can be any communication network, such as the Internet, an Intranet, a Local Area Network (LAN), a Wide Area Network (WAN), a telephone network, an Ethernet network, a fiber-optic network, a wireless network, a cellular network, etc., configurable to allow the one or more UEs 201 a-201 n, thesandboxed protection system 203, the one or more application provider(s) 209, and the one or more execution server(s) 211 to communicate withcommunication network 205 and/or to each other throughcommunication network 205. - In some instances,
communication network 205 can include multiple networks operatively coupled to one another by, for example, network bridges, routers, switches and/or gateways. For example, the UEs 201 a-201 n can be operatively coupled to a cellular network and the application provider(s) 209, thesandboxed protection system 203, and/or the one or more execution server(s) 211 can be operatively coupled to a fiber-optic network. The cellular network and the fiber-optic network can each be operatively coupled to one another via one or more network bridges, routers, switches, and/or gateways such that the cellular network, the Ethernet network and the fiber-optic network are operatively coupled to form a communication network. Alternatively, the cellular network and fiber-optic network can each be operatively coupled to one another via one or more additional networks. For example, the cellular network and the fiber-optic network can each be operatively coupled to the Internet such that the cellular network, the fiber-optic network and the Internet are operatively coupled to form a communication network. - As illustrated in
FIG. 2 , UEs 201 a-201 n are operatively coupled tocommunication network 205 vianetwork connection 213; application provider(s) 209 are operatively coupled tocommunication network 205 vianetwork connection 215; execution server(s) 211 are operatively coupled tocommunication network 205 vianetwork connection 217; and thesandboxed protection system 203 is operatively coupled tocommunication network 205 vianetwork connection 219.Network connections sandboxed protection system 203. - A network connection can be a wireless network connection such as, for example, a wireless fidelity (“Wi-Fi”) or wireless local area network (“WLAN”) connection, a wireless wide area network (“WWAN”) connection, and/or a cellular connection. A network connection can be a wired connection such as, for example, an Ethernet connection, a digital subscription line (“DSL”) connection, a broadband coaxial connection, and/or a fiber-optic connection.
- As mentioned above, in some instances, a
computer network system 200 can include more than one UE 201 a-201 n, more than onesandboxed protection system 203, more than one application provider(s) 209, and more than one execution server(s) 211. A UE 201 a-201 n, asandboxed protection system 203, anapplication provider 209, and/or an execution server(s) 211 can be operatively coupled to thecommunication network 205 by heterogeneous network connections. For example, a first UE 201 a-201 n can be operatively coupled to thecommunication network 205 by a WWAN network connection, a second UE 201 a-201 n can be operatively coupled to thecommunication network 205 by a DSL network connection, and asandboxed protection system 203 can be operatively coupled to thecommunication network 205 by a fiber-optic network connection. - The application provider(s) 209 can be, for example, a web server configured to provide various applications to electronic devices, such as UEs 201 a-201 n. For example, the UE 201 a-201 n can be in communication with the application provider(s) 209 via the
communication network 205 under the supervision of thesandboxed protection system 203. - The UEs 201 a-201 n can be any of a variety of electronic devices that can be operatively coupled to
communication network 205. A UE 201 a-201 n can be a personal computer, a laptop computer, a personal digital assistant (PDA), a cellular telephone, a portable/mobile internet device and/or some other electronic communication device. The UEs 201 a-201 n can include a web browser configured to access a webpage or website hosted on or accessible via the application provider(s) 209 overcommunication network 205. The UEs 201 a-201 n can be configured to support, for example, HTML using JavaScript. For example, the UEs 201 a-201 n can include a web browser, such as, Firefox, Safari, Opera and Chrome. An Internet page or website can be accessed by a user of a web browser at a UE 201 a-201 n by providing the web browser with a reference such as a uniform resource locator (URL), for example, of a webpage. For example, a user of a UE 201 a-201 n can access anapplication provider 209 via a URL designated for theapplication provider 209. In some instances, UEs 201 a-201 n can include specialized software for accessing a web server other than a browser, such as, for example, a specialized network-enabled application or program. - In some instances, portions of a website accessible via a web server, for example an
application provider 209, can be located in an execution server(s) 211 accessible to theapplication provider 209. A memory, within the execution server(s) 211, can be at least one of a database, a data warehouse, and/or the like. A UE 201 a-201 n can also include a display, monitor or user interface (UI), a keyboard, various ports (e.g., a USB port), and other user interface features, such as, for example, touch screen controls, audio components, and/or video components (each not shown). A UE 201 a-201 n can be operatively coupled tocommunication network 205 via the UI andnetwork connection 213. -
FIG. 3 is a flowchart of a process for implementing a sandboxed protection system, according to an embodiment. In one embodiment, thesandboxed protection system 121 a ofFIG. 1A executes theprocess 300. Additionally, processor-readable instructions to execute the process can be stored indata store 109 a and/or in data store 135. Atstep 301, thesandbox control module 123 a initiates asandbox environment 105 a. Thesandbox environment 105 a can, for example, be executed within the host operating system 103 that runs on a processor of theUE 101 a. Furthermore, thesandbox environment 105 a can be configured to execute applications and/or processes such as, for example, a web browser or any other application, process and/or service at risk of receiving malicious content. - At 303, a set of rules for processes within the
sandbox environment 105 a is defined by the rule generation module 133. The rules can include, for example, classifying processes and/or applications using trust classifications provided by the trust classification module 137. In some instances, the rules can include predefined rules for the system and/or custom defined rules by a system administrator. The rules are stored in therules database 136 a. - At 305, the
sandbox monitor module 125 a monitors events and activities associated with running applications and/or processes within thesandbox environment 105 a, based on the rules associated with the applications and/or processes. The sandbox processes can be monitored based on the rules in therules database 136 a. Thesandbox monitor module 125 a can report any anomalous event or activity to theevent engine module 127 a. - At 307 the
event engine module 127 a analyzes the report received from thesandbox monitor module 125 a with reference to the associated rules in therules database 136 a to determine whether the anomalous behavior is malicious. Theevent engine module 127 a can declares an infection based on the detected malicious behavior and the rules of therules database 136 a. - At 309, the
event engine module 127 a adds the detected infection to an evaluation tree within the data store 135. The evaluation tree can be used for further analysis of the infection by thesandboxed protection system 121 a or by a system administrator. The analysis results can be used by the rule generation module 133 for updating therules database 136 a. - In some instances (not shown in
FIG. 3 ), thesandboxed protection system 121 a can automatically and/or dynamically disable and/or ignore an infection detection. Disabling and/or ignoring detection events can be used, for example, in cases where some events are generated intentionally that otherwise would trigger theevent engine module 127 a to report an infection. For example, if a user ofUE 101 a intentionally runs an unknown process such as an application or an installer in thesandbox environment 105 a without previously defining rules to trust the process. - At 311, the rule generation module 133 revises the rules in the
rules database 136 a based on the detected infections reported as the evaluation tree. In some instances, the rule generation module 133 can automatically revise the rules in therules database 136 a based on the monitored behavior of thesandbox environment 105 a by thesandbox monitor module 125 a and the analysis of the monitored behavior, including the infections detected by theevent engine module 127 a. In other instances, a system administrator can manually revise the rules based on the detected infections, evaluation tree content, etc. In still other instances, the rules may remain unchanged based on the infection. - At 313, the
sandboxed control module 123 a restarts thesandbox environment 105 a, such that the infected environment can be terminated and a newclean sandbox environment 105 a can be restarted. The process of monitoring thesandbox environment 105 a by thesandbox monitor module 125 a based on the rules of therules database 136 a can then continue with theclean sandbox environment 105 a. Alternatively, thesandbox control module 123 a can terminate and/or restart the application, process and/or service with malicious behavior without restarting thesandbox environment 105 a. -
FIG. 4 is a flowchart of a process for control and detection of malicious behavior of an application based on allowed behavior, according to an embodiment. In some instances, at 401 theapplication control module 123 a can execute a code to receive a set of indications of allowed behavior associated with an application, based on a set of rules in therules database 136 a. Theapplication control module 123 a can, for example, retrieve the indications from data store 135, receive the indications from a system administrator viainput signal 139, and/or the like. Theapplication control module 123 a can store the received indications in data store 135. At 403, theapplication control module 123 a sends a signal to the host operating system 103 viaoutput signal 141 to initiate an instance of the application in thesandbox environment 105 a. - At 405, the
event engine module 127 a, receives, from thesandbox monitor module 125 a, a set of indications of actual behavior of the instance of the application within thesandbox environment 105 a. Theevent engine module 127 a can store the actual behavior indications in the data store 135. - At 407, the
event engine module 127 a analyzes the actual behavior of the instance of the application in comparison with the allowed behavior associated with the instance of the application in the data store 135, to detect correspondences and discrepancies. If the allowed behavior and the actual behavior correspond, theevent engine module 127 a can conclude that no infection has occurred and malicious behavior has not been detected. - If at least one indication from the set of indications of actual behavior does not correspond to an indication from the set of indications of allowed behavior, at 409 the
event engine module 127 a sends an indication associated with an anomalous behavior to theapplication control module 123 a. The indication associated with the anomalous behavior can be used by theapplication control module 123 a to terminate the application and/or to terminate sandbox environment's process, to send an indication of the anomalous behavior to a system administrator or to a user, and/or the like. - In some instances, the
application control module 123 a can send the indication associated with the anomalous behavior to theevent engine module 127 a, for example to define an evaluation tree. The evaluation tree including the indication associated with the anomalous behavior can be stored in data store 135 associated with the application, for example added to therules database 136 a associated with the application. - In some instances, the
sandbox monitor module 125 a can be configured to monitor at least one event of the instance of the application within thesandbox environment 105 a such as, for example, process events (e.g., executing unauthorized processes) of the instance of the application, file events (e.g., access of unauthorized files) of the instance of the application, registry events (e.g., registry updates) of the instance of the application, network events (e.g., network connection events) of the instance of the application, or thread injection events (e.g., behavior associated with inserting and run executable code within the address space of another process) of the instance of the application, etc. - In some instances, the rule generation module 133 uses the anomalous behavior detected by the
event engine module 127 a and revises the set of indications (rules) of allowed behavior associated with the application. The revised set of rules can replace the set of rules in therules database 136 a. - In some instances the indication associated with the anomalous behavior can include a trace associated with a source of the anomalous behavior. For example, the application can be a web browser application and the indication associated with the anomalous behavior can include a uniform resource locator (URL) trace associated with the web browser application. For another example, the source trace can include an identifier of an application, process, service and/or the like. The source trace enables the rule generation module 133 to revise the rules to enable the
sandbox control module 123 a to control application access to the sources that cause anomalous behavior of the application. - In some instances, the set of rules indicating allowed behavior associated with the application is based at least in part on a trust level associated with the application. The trust level can be defined by the trust classification module 137 based on, for example, user input, previous behavior of the application (e.g., history of actual behavior in data store 135), application type, etc. Furthermore, the set of indications of allowed behavior associated with the application can include an identifier of a trusted process associated with the application.
- In some instances, the
sandbox monitor module 125 a can monitor activities of an instance of an application within thesandbox environment 105 a by tracing network activities of the running instance of the application within thesandbox environment 105 a. Thesandbox monitor module 125 a can also monitor one or more executable files of the instance of the application. -
FIG. 5 is a flowchart of a process for control and detection of malicious behavior of an application based on application type, according to an embodiment. Thesandbox environment 105 a can have multiple instances or different applications or instances of the same application running simultaneously. The sandbox monitor 125 a and theapplication control module 123 a can monitor and control the running applications. As shown inFIG. 5 , at 501 theapplication control module 123 a initiates an instance of a first application and an instance of a second application within asandbox environment 105 a. - At 503 the
event engine module 127 a receives, from asandbox monitor module 125 a associated with thesandbox environment 105 a, a set of indications of actual behavior of the instance of the first application and a set of indications of actual behavior of the instance of the second application. The first application and the second application can be from the same application type (two instances of the same application running simultaneously onsandbox environment 105 a) or from different application types (two different applications running simultaneously onsandbox environment 105 a). In some instances the set of indications of actual behavior of the instance of the first application and the set of indications of actual behavior of the instance of the second application can overlap such that the instance of the first application and the instance of the second application may show the same actual behavior. However, the same actual behavior that can be declared as anomalous behavior for an instance of the first application can be declared as an allowed behavior for an instance of the second application, or vise versa. - At 505, the
event engine module 127 a checks therules database 136 a to find a baseline behavior set for the first application (e.g., a set of allowed behavior for the first application). If the actual behavior is an anomalous behavior for the first application based on the baseline behavior set for the first application in therules database 136 a, theevent engine module 127 a classifies the actual behavior as anomalous behavior for the first application. - At 507 the
event engine module 127 a checks therules database 136 a to find a baseline behavior set for the second application. If the actual behavior is not an anomalous behavior for the second application based on the baseline behavior set for the second application (e.g., a set of allowed behavior for the second application) in therules database 136 a,event engine module 127 a does not classify the actual behavior as anomalous behavior for the second application. - For example, the
event engine module 127 a can classify a behavior B as an anomalous behavior for the first application based on a baseline behavior set for the first application, but not classify the behavior B as an anomalous behavior for the second application based on a baseline behavior set for the second application. - Note that, as shown in
FIG. 2 , thesandboxed protection system 203 can be located within the UEs 201 a-201 n, or in a separate location coupled with the UEs 201 a-201 n via thecommunication network 205. Therefore, in some instances, as shown inFIG. 2 , thesandboxed protection module 203 can include the sandbox environments 207 a-207 n. - In some instances, the
sandbox control module 123 a can be operatively coupled to thesandbox environment 105 a via a network. For example, as shown inFIG. 2 , thesandboxed protection system 203 that includes thesandbox control module 123 a is coupled to the UEs 201 a-201 n, that include sandbox environments 207 a-207 n, viacommunication network 205 bynetwork connections - In some instances, the set of indications of actual behavior of the instance of the first (or second) application includes at least one of a process event identifier of the instance of the first (or second) application, a file event identifier of the instance of the first (second) application, a registry event identifier of the instance of the first (or second) application or a network event identifier of the instance of the first (second) application.
- At 509, the
event engine module 127 a can send a signal to other components of thesandboxed protection system 121 a such as theapplication control module 123 a to indicate classifying of the actual behavior as anomalous behavior for the first application. Theapplication control module 123 a can send a signal to initiate actions such as first application termination,sandbox environment 105 a termination, notifying a system administrator or a user about the anomalous behavior, etc. -
FIG. 6 is a flowchart of a process for control and detection of malicious behavior of an application based on another application, according to an embodiment. In some instances, an application running within thesandbox environment 105 a can initiate one or more other applications within the sandbox 135. At 601 thesandbox control module 123 a receives, from arules database 136 a, a set of indications of allowed behavior associated with a first application. - At 603, the
sandbox control module 123 a initiates an instance of the first application within asandbox environment 105 a such that the instance of the first application initiates an instance of a second application within thesandbox environment 105 a. For example, execution of the first application can start execution of a second application without an intervention by theapplication control module 123 a. Thesandbox monitor module 125 a can detect initiation of the second application and report the initiation to theapplication control module 123 a and/or to theevent engine module 127 a. - At 605,
event engine module 127 a receives, from thesandbox monitor module 125 a associated with thesandbox environment 105 a, a set of indications of actual behavior of the instance of the second application, as previously discussed with regards toFIG. 5 , in response to the instance of the first application initiating the instance of the second application. The first application and the second application can be from the same application type or from different application types. - If at least one indication from the set of indications of actual behavior of the instance of the second application does not correspond to an indication from the set of indications of allowed behavior associated with the first application, at 607, the
event engine module 127 a sends an indication associated with an anomalous behavior associated with the second application to theapplication control module 123 a. The indication associated with the anomalous behavior can be used by theapplication control module 123 a to terminate the second application and/or the sandbox environment process, etc. Furthermore, theevent engine module 127 a can add the anomalous behavior to an evaluation tree associated with the second application. - In some instances, the rule generation module 133 uses the anomalous behavior of the second application detected by the
event engine module 127 a and revises the set of indications (rules) of allowed behavior associated with the first application and/or the set of indications (rules) of allowed behavior associated with the second application. The revised set of indications can replace the set of indications inrules database 136 a. - In some instances the second application can be, for example, a web browser application and the indication associated with the anomalous behavior can include a uniform resource locator (URL) trace associated with the web browser application. The source trace enables the rule generation module 133 to revise the set of indications (rules) of allowed behavior associated with the first and/or second application such that access of second application to the sources that caused anomalous behavior can be controlled.
- In some instances, the set of indications of allowed behavior associated with the first application is based at least in part on a trust level associated with the first application. The trust level can be defined by the trust classification module 137 based on, for example, user input, previous behavior of the first application (e.g., history of actual behavior in data store 135), first application type, previous behavior of the second application (e.g., history of actual behavior in data store 135), second application type, etc. Furthermore, the set of indications of allowed behavior associated with the first application can includes an identifier of a trusted process associated with the first application.
- It is intended that the systems and methods described herein can be performed by software (executed on hardware), hardware, or a combination thereof. Hardware modules may include, for example, a general-purpose processor, a field programmable gate array (FPGA), and/or an application specific integrated circuit (ASIC). Software modules (executed on hardware) can be expressed in a variety of software languages (e.g., computer code), including C, C++, Java™, Ruby, Visual Basic™, and other object-oriented, procedural, or other programming language and development tools. Examples of computer code include, but are not limited to, micro-code or micro-instructions, machine instructions, such as produced by a compiler, code used to produce a web service, and files containing higher-level instructions that are executed by a computer using an interpreter. Additional examples of computer code include, but are not limited to, control signals, encrypted code, and compressed code.
- Some embodiments described herein relate to a computer storage product with a non-transitory computer-readable medium (also can be referred to as a non-transitory processor-readable medium) having instructions or computer code thereon for performing various computer-implemented operations. The computer-readable medium (or processor-readable medium) is non-transitory in the sense that it does not include transitory propagating signals per se (e.g., a propagating electromagnetic wave carrying information on a transmission medium such as space or a cable). The media and computer code (also can be referred to as code) may be those designed and constructed for the specific purpose or purposes. Examples of non-transitory computer-readable media include, but are not limited to: magnetic storage media such as hard disks, floppy disks, and magnetic tape; optical storage media such as Compact Disc/Digital Video Discs (CD/DVDs), Compact Disc-Read Only Memories (CD-ROMs), and holographic devices; magneto-optical storage media such as optical disks; carrier wave signal processing modules; and hardware devices that are specially configured to store and execute program code, such as Application-Specific Integrated Circuits (ASICs), Programmable Logic Devices (PLDs), Read-Only Memory (ROM) and Random-Access Memory (RAM) devices.
- While various embodiments have been described above, it should be understood that they have been presented by way of example only, and not limitation. Where methods and steps described above indicate certain events occurring in certain order, the ordering of certain steps may be modified. Additionally, certain of the steps may be performed concurrently in a parallel process when possible, as well as performed sequentially as described above. Although various embodiments have been described as having particular features and/or combinations of components, other embodiments are possible having any combination or sub-combination of any features and/or components from any of the embodiments described herein.
Claims (23)
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/797,847 US9519779B2 (en) | 2011-12-02 | 2015-07-13 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US15/359,004 US10043001B2 (en) | 2011-12-02 | 2016-11-22 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US16/018,720 US10467406B2 (en) | 2011-12-02 | 2018-06-26 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US16/671,664 US10984097B2 (en) | 2011-12-02 | 2019-11-01 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US17/211,412 US12019734B2 (en) | 2011-12-02 | 2021-03-24 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US18/672,750 US20240320323A1 (en) | 2011-12-02 | 2024-05-23 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201161566162P | 2011-12-02 | 2011-12-02 | |
US13/690,452 US9081959B2 (en) | 2011-12-02 | 2012-11-30 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US14/797,847 US9519779B2 (en) | 2011-12-02 | 2015-07-13 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/690,452 Continuation US9081959B2 (en) | 2011-12-02 | 2012-11-30 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/359,004 Continuation US10043001B2 (en) | 2011-12-02 | 2016-11-22 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150324586A1 true US20150324586A1 (en) | 2015-11-12 |
US9519779B2 US9519779B2 (en) | 2016-12-13 |
Family
ID=48525005
Family Applications (7)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/690,452 Active US9081959B2 (en) | 2011-12-02 | 2012-11-30 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US14/797,847 Active US9519779B2 (en) | 2011-12-02 | 2015-07-13 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US15/359,004 Active US10043001B2 (en) | 2011-12-02 | 2016-11-22 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US16/018,720 Active US10467406B2 (en) | 2011-12-02 | 2018-06-26 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US16/671,664 Active US10984097B2 (en) | 2011-12-02 | 2019-11-01 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US17/211,412 Active 2034-06-14 US12019734B2 (en) | 2011-12-02 | 2021-03-24 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US18/672,750 Pending US20240320323A1 (en) | 2011-12-02 | 2024-05-23 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/690,452 Active US9081959B2 (en) | 2011-12-02 | 2012-11-30 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
Family Applications After (5)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/359,004 Active US10043001B2 (en) | 2011-12-02 | 2016-11-22 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US16/018,720 Active US10467406B2 (en) | 2011-12-02 | 2018-06-26 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US16/671,664 Active US10984097B2 (en) | 2011-12-02 | 2019-11-01 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US17/211,412 Active 2034-06-14 US12019734B2 (en) | 2011-12-02 | 2021-03-24 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US18/672,750 Pending US20240320323A1 (en) | 2011-12-02 | 2024-05-23 | Methods and apparatus for control and detection of malicious content using a sandbox environment |
Country Status (2)
Country | Link |
---|---|
US (7) | US9081959B2 (en) |
WO (1) | WO2013082437A1 (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9436822B2 (en) | 2009-06-30 | 2016-09-06 | George Mason Research Foundation, Inc. | Virtual browsing environment |
US9602524B2 (en) | 2008-09-12 | 2017-03-21 | George Mason Research Foundation, Inc. | Methods and apparatus for application isolation |
CN106919837A (en) * | 2016-10-20 | 2017-07-04 | 深圳市安之天信息技术有限公司 | A kind of unknown self-starting recognition methods of malicious code and system |
US9846588B2 (en) | 2007-03-01 | 2017-12-19 | George Mason Research Foundation, Inc. | On-demand disposable virtual work system |
US20180083996A1 (en) * | 2016-09-21 | 2018-03-22 | Sentient Technologies (Barbados) Limited | Detecting behavioral anomaly in machine learned rule sets |
US10043001B2 (en) | 2011-12-02 | 2018-08-07 | Invincea, Inc. | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US20180300474A1 (en) * | 2016-10-31 | 2018-10-18 | Soliton Systems K.K. | Program behavior monitoring control apparatus, distributed object creation management apparatus, recording medium, and program behavior monitoring system |
US20180336351A1 (en) * | 2017-05-22 | 2018-11-22 | Microsoft Technology Licensing, Llc | Isolated Container Event Monitoring |
US12093414B1 (en) * | 2019-12-09 | 2024-09-17 | Amazon Technologies, Inc. | Efficient detection of in-memory data accesses and context information |
Families Citing this family (92)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9390263B2 (en) | 2010-03-31 | 2016-07-12 | Sophos Limited | Use of an application controller to monitor and control software file and application environments |
US9300760B2 (en) * | 2013-01-28 | 2016-03-29 | Google Inc. | Machine-specific instruction set translation |
US9660919B2 (en) * | 2013-01-28 | 2017-05-23 | Apple Inc. | Adaptive data connection retry by a wireless communication device |
US9208317B2 (en) * | 2013-02-17 | 2015-12-08 | Check Point Software Technologies Ltd. | Simultaneous screening of untrusted digital files |
US9355247B1 (en) * | 2013-03-13 | 2016-05-31 | Fireeye, Inc. | File extraction from memory dump for malicious content analysis |
US9104867B1 (en) * | 2013-03-13 | 2015-08-11 | Fireeye, Inc. | Malicious content analysis using simulated user interaction without user involvement |
US9424421B2 (en) | 2013-05-03 | 2016-08-23 | Visa International Service Association | Security engine for a secure operating environment |
US9536091B2 (en) * | 2013-06-24 | 2017-01-03 | Fireeye, Inc. | System and method for detecting time-bomb malware |
US9208310B2 (en) * | 2013-06-26 | 2015-12-08 | Cognizant Technology Solutions India Pvt. Ltd. | System and method for securely managing enterprise related applications and data on portable communication devices |
US9230106B2 (en) | 2013-06-28 | 2016-01-05 | Kaspersky Lab Ao | System and method for detecting malicious software using malware trigger scenarios in a modified computer environment |
RU2653985C2 (en) * | 2013-06-28 | 2018-05-15 | Закрытое акционерное общество "Лаборатория Касперского" | Method and system for detecting malicious software by control of software implementation running under script |
US9591003B2 (en) * | 2013-08-28 | 2017-03-07 | Amazon Technologies, Inc. | Dynamic application security verification |
US9292684B2 (en) | 2013-09-06 | 2016-03-22 | Michael Guidry | Systems and methods for security in computer systems |
US9323931B2 (en) * | 2013-10-04 | 2016-04-26 | Bitdefender IPR Management Ltd. | Complex scoring for malware detection |
US10223530B2 (en) * | 2013-11-13 | 2019-03-05 | Proofpoint, Inc. | System and method of protecting client computers |
US9223964B2 (en) * | 2013-12-05 | 2015-12-29 | Mcafee, Inc. | Detecting JAVA sandbox escaping attacks based on JAVA bytecode instrumentation and JAVA method hooking |
RU2571723C2 (en) * | 2013-12-05 | 2015-12-20 | Закрытое акционерное общество "Лаборатория Касперского" | System and method of reducing load on operating system when executing antivirus application |
US9202046B2 (en) | 2014-03-03 | 2015-12-01 | Bitdefender IPR Management Ltd. | Systems and methods for executing arbitrary applications in secure environments |
US9734332B2 (en) * | 2014-03-17 | 2017-08-15 | Proofpoint, Inc. | Behavior profiling for malware detection |
US9881157B1 (en) * | 2014-03-18 | 2018-01-30 | Bitdefender IPR Management Ltd. | Anti-malware systems and methods using hardware-assisted code injection |
CN103955645B (en) * | 2014-04-28 | 2017-03-08 | 百度在线网络技术(北京)有限公司 | The detection method of malicious process behavior, apparatus and system |
US9760712B2 (en) * | 2014-05-23 | 2017-09-12 | Vmware, Inc. | Application whitelisting using user identification |
CN106663170B (en) | 2014-06-17 | 2019-06-25 | 日本电信电话株式会社 | Information processing system, control method |
US9866582B2 (en) | 2014-06-30 | 2018-01-09 | Paypal, Inc. | Detection of scripted activity |
US9967282B2 (en) * | 2014-09-14 | 2018-05-08 | Sophos Limited | Labeling computing objects for improved threat detection |
US10965711B2 (en) | 2014-09-14 | 2021-03-30 | Sophos Limited | Data behavioral tracking |
US10122687B2 (en) | 2014-09-14 | 2018-11-06 | Sophos Limited | Firewall techniques for colored objects on endpoints |
CN104267977B (en) * | 2014-09-16 | 2017-12-29 | 小米科技有限责任公司 | application program operation method and device |
US9870239B2 (en) * | 2014-09-16 | 2018-01-16 | Xiaomi Inc. | Method and device for running application program |
US10560353B1 (en) * | 2014-09-16 | 2020-02-11 | Amazon Technologies, Inc. | Deployment monitoring for an application |
US9411959B2 (en) * | 2014-09-30 | 2016-08-09 | Juniper Networks, Inc. | Identifying an evasive malicious object based on a behavior delta |
US9692773B1 (en) * | 2014-12-11 | 2017-06-27 | Symantec Corporation | Systems and methods for identifying detection-evasion behaviors of files undergoing malware analyses |
US9575740B2 (en) * | 2015-01-21 | 2017-02-21 | Samsung Electronics Co., Ltd. | Apparatus and method for running multiple instances of a same application in mobile devices |
US9774654B2 (en) * | 2015-02-02 | 2017-09-26 | Linkedin Corporation | Service call graphs for website performance |
TWI616770B (en) | 2015-02-03 | 2018-03-01 | 緯創資通股份有限公司 | Cloud data management method, electronic apparatus and cloud server |
US9600682B2 (en) * | 2015-06-08 | 2017-03-21 | Accenture Global Services Limited | Mapping process changes |
US10185480B1 (en) * | 2015-06-15 | 2019-01-22 | Symantec Corporation | Systems and methods for automatically making selections in user interfaces |
US9977896B2 (en) | 2015-10-08 | 2018-05-22 | Digital Guardian, Inc. | Systems and methods for generating policies for an application using a virtualized environment |
US10963565B1 (en) * | 2015-10-29 | 2021-03-30 | Palo Alto Networks, Inc. | Integrated application analysis and endpoint protection |
US20170149828A1 (en) | 2015-11-24 | 2017-05-25 | International Business Machines Corporation | Trust level modifier |
CN106921608B (en) * | 2015-12-24 | 2019-11-22 | 华为技术有限公司 | A kind of detection terminal security situation method, apparatus and system |
US10055579B2 (en) * | 2015-12-31 | 2018-08-21 | Cybereason, Inc. | System resources for sandboxing |
US10152596B2 (en) * | 2016-01-19 | 2018-12-11 | International Business Machines Corporation | Detecting anomalous events through runtime verification of software execution using a behavioral model |
US10893059B1 (en) | 2016-03-31 | 2021-01-12 | Fireeye, Inc. | Verification and enhancement using detection systems located at the network periphery and endpoint devices |
US10826933B1 (en) | 2016-03-31 | 2020-11-03 | Fireeye, Inc. | Technique for verifying exploit/malware at malware detection appliance through correlation with endpoints |
US11102238B2 (en) | 2016-04-22 | 2021-08-24 | Sophos Limited | Detecting triggering events for distributed denial of service attacks |
US10986109B2 (en) | 2016-04-22 | 2021-04-20 | Sophos Limited | Local proxy detection |
US11165797B2 (en) | 2016-04-22 | 2021-11-02 | Sophos Limited | Detecting endpoint compromise based on network usage history |
US11277416B2 (en) | 2016-04-22 | 2022-03-15 | Sophos Limited | Labeling network flows according to source applications |
US10938781B2 (en) | 2016-04-22 | 2021-03-02 | Sophos Limited | Secure labeling of network flows |
CN107315954B (en) * | 2016-04-27 | 2020-06-12 | 腾讯科技(深圳)有限公司 | File type identification method and server |
US12021831B2 (en) | 2016-06-10 | 2024-06-25 | Sophos Limited | Network security |
US10331883B2 (en) * | 2016-09-28 | 2019-06-25 | International Business Machines Corporation | Malicious code avoidance using transparent containers |
US10331885B2 (en) | 2016-12-02 | 2019-06-25 | Microsoft Technology Licensing, Llc | Identification of entity performing operation on local file(s) and notification to reduce misuse risk |
US10275596B1 (en) * | 2016-12-15 | 2019-04-30 | Symantec Corporation | Activating malicious actions within electronic documents |
US11436328B1 (en) * | 2017-02-24 | 2022-09-06 | Acronis International Gmbh | Systems and methods of safeguarding user data |
JP2018200642A (en) * | 2017-05-29 | 2018-12-20 | 富士通株式会社 | Threat detection program, threat detection method, and information processing apparatus |
US10331535B1 (en) * | 2017-06-05 | 2019-06-25 | AppiSocial Co., Ltd. | Detecting discrepancy in mobile event tracking network |
US11240207B2 (en) * | 2017-08-11 | 2022-02-01 | L3 Technologies, Inc. | Network isolation |
IL253987B (en) * | 2017-08-14 | 2019-05-30 | Cyberbit Ltd | Cyber threat detection system and method |
US11601467B2 (en) | 2017-08-24 | 2023-03-07 | L3 Technologies, Inc. | Service provider advanced threat protection |
US11062021B2 (en) | 2017-08-29 | 2021-07-13 | NortonLifeLock Inc. | Systems and methods for preventing malicious applications from exploiting application services |
US11178104B2 (en) | 2017-09-26 | 2021-11-16 | L3 Technologies, Inc. | Network isolation with cloud networks |
US11184323B2 (en) | 2017-09-28 | 2021-11-23 | L3 Technologies, Inc | Threat isolation using a plurality of containers |
US11374906B2 (en) | 2017-09-28 | 2022-06-28 | L3 Technologies, Inc. | Data exfiltration system and methods |
US11223601B2 (en) | 2017-09-28 | 2022-01-11 | L3 Technologies, Inc. | Network isolation for collaboration software |
US11336619B2 (en) | 2017-09-28 | 2022-05-17 | L3 Technologies, Inc. | Host process and memory separation |
US11552987B2 (en) | 2017-09-28 | 2023-01-10 | L3 Technologies, Inc. | Systems and methods for command and control protection |
US11550898B2 (en) | 2017-10-23 | 2023-01-10 | L3 Technologies, Inc. | Browser application implementing sandbox based internet isolation |
US11170096B2 (en) | 2017-10-23 | 2021-11-09 | L3 Technologies, Inc. | Configurable internet isolation and security for mobile devices |
US11120125B2 (en) | 2017-10-23 | 2021-09-14 | L3 Technologies, Inc. | Configurable internet isolation and security for laptops and similar devices |
KR102558741B1 (en) * | 2017-12-12 | 2023-07-24 | 삼성전자주식회사 | Device and method to register user |
US11855971B2 (en) * | 2018-01-11 | 2023-12-26 | Visa International Service Association | Offline authorization of interactions and controlled tasks |
US11500665B2 (en) | 2018-08-30 | 2022-11-15 | Micron Technology, Inc. | Dynamic configuration of a computer processor based on the presence of a hypervisor |
US10915465B2 (en) | 2018-08-30 | 2021-02-09 | Micron Technology, Inc. | Memory configured to store predefined set of domain registers for instructions being executed in computer processors |
US11182507B2 (en) | 2018-08-30 | 2021-11-23 | Micron Technology, Inc. | Domain crossing in executing instructions in computer processors |
US11914726B2 (en) | 2018-08-30 | 2024-02-27 | Micron Technology, Inc. | Access control for processor registers based on execution domains |
US10942863B2 (en) * | 2018-08-30 | 2021-03-09 | Micron Technology, Inc. | Security configurations in page table entries for execution domains using a sandbox application operation |
US11481241B2 (en) | 2018-08-30 | 2022-10-25 | Micron Technology, Inc. | Virtual machine register in a computer processor |
US10915457B2 (en) | 2018-08-30 | 2021-02-09 | Micron Technology, Inc. | Memory access control through permissions specified in page table entries for execution domains |
US11295011B2 (en) * | 2019-01-08 | 2022-04-05 | Vmware, Inc. | Event-triggered behavior analysis |
US11321453B2 (en) * | 2019-04-18 | 2022-05-03 | Check Point Software Technologies Ltd. | Method and system for detecting and classifying malware based on families |
WO2020246011A1 (en) * | 2019-06-06 | 2020-12-10 | 日本電気株式会社 | Rule generation device, rule generation method, and computer readable storge medium |
US11321481B1 (en) * | 2019-06-26 | 2022-05-03 | Norton LifeLock, Inc. | Method for determining to grant or deny a permission request based on empirical data aggregation |
US10789354B1 (en) | 2019-10-01 | 2020-09-29 | Capital One Services, Llc | Computer-based systems configured for detecting and sandboxing external resources and methods of use thereof |
FR3104761A1 (en) * | 2019-12-12 | 2021-06-18 | Orange | Method for monitoring data passing through user equipment |
RU2755252C2 (en) * | 2020-02-26 | 2021-09-14 | Акционерное общество "Лаборатория Касперского" | Method and system for assessing impact of software under study on availability of industrial automation systems |
US11716310B2 (en) * | 2020-12-31 | 2023-08-01 | Proofpoint, Inc. | Systems and methods for in-process URL condemnation |
US20220207134A1 (en) * | 2020-12-31 | 2022-06-30 | Acronis International Gmbh | Systems and methods for securing cloud meetings using containers |
CN113177205B (en) * | 2021-04-27 | 2023-09-15 | 国家计算机网络与信息安全管理中心 | Malicious application detection system and method |
TWI801856B (en) * | 2021-04-30 | 2023-05-11 | 精品科技股份有限公司 | Method of application control for child-program execution |
CN117521087B (en) * | 2024-01-04 | 2024-03-15 | 江苏通付盾科技有限公司 | Equipment risk behavior detection method, system and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080016339A1 (en) * | 2006-06-29 | 2008-01-17 | Jayant Shukla | Application Sandbox to Detect, Remove, and Prevent Malware |
US20080235764A1 (en) * | 2007-03-22 | 2008-09-25 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | Resource authorizations dependent on emulation environment isolation policies |
US20120317645A1 (en) * | 2011-06-13 | 2012-12-13 | Microsoft Corporation | Threat level assessment of applications |
US8719924B1 (en) * | 2005-03-04 | 2014-05-06 | AVG Technologies N.V. | Method and apparatus for detecting harmful software |
US9111089B1 (en) * | 2011-02-08 | 2015-08-18 | Symantec Corporation | Systems and methods for safely executing programs |
Family Cites Families (165)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5297269A (en) | 1990-04-26 | 1994-03-22 | Digital Equipment Company | Cache coherency protocol for multi processor computer system |
US5394555A (en) | 1992-12-23 | 1995-02-28 | Bull Hn Information Systems Inc. | Multi-node cluster computer system incorporating an external coherency unit at each node to insure integrity of information stored in a shared, distributed memory |
JP2708386B2 (en) | 1994-03-18 | 1998-02-04 | インターナショナル・ビジネス・マシーンズ・コーポレイション | Method and apparatus for recovering duplicate database through simultaneous update and copy procedure |
WO1995033237A1 (en) | 1994-06-01 | 1995-12-07 | Quantum Leap Innovations Inc. | Computer virus trap |
US6157721A (en) | 1996-08-12 | 2000-12-05 | Intertrust Technologies Corp. | Systems and methods using cryptography to protect secure computing environments |
US5854916A (en) | 1995-09-28 | 1998-12-29 | Symantec Corporation | State-based cache for antivirus software |
US5696822A (en) | 1995-09-28 | 1997-12-09 | Symantec Corporation | Polymorphic virus detection module |
US6067410A (en) | 1996-02-09 | 2000-05-23 | Symantec Corporation | Emulation repair system |
US7058822B2 (en) | 2000-03-30 | 2006-06-06 | Finjan Software, Ltd. | Malicious mobile code runtime monitoring system and methods |
US5974549A (en) | 1997-03-27 | 1999-10-26 | Soliton Ltd. | Security monitor |
US5895467A (en) | 1997-04-11 | 1999-04-20 | Informix Software, Inc. | Selectively switching memory access permission for manipulating data in a database |
US6718482B2 (en) | 1997-09-12 | 2004-04-06 | Hitachi, Ltd. | Fault monitoring system |
US6037934A (en) | 1997-11-21 | 2000-03-14 | International Business Machines Corporation | Named bookmark sets |
US6584495B1 (en) | 1998-01-30 | 2003-06-24 | Microsoft Corporation | Unshared scratch space |
US6832120B1 (en) | 1998-05-15 | 2004-12-14 | Tridium, Inc. | System and methods for object-oriented control of diverse electromechanical systems using a computer network |
US6496847B1 (en) | 1998-05-15 | 2002-12-17 | Vmware, Inc. | System and method for virtualizing computer systems |
US6397242B1 (en) | 1998-05-15 | 2002-05-28 | Vmware, Inc. | Virtualization system including a virtual machine monitor for a computer with a segmented architecture |
US7363264B1 (en) | 1999-10-29 | 2008-04-22 | Computer Sciences Corporation | Processing business transactions using dynamic database packageset switching |
EP1275061A1 (en) | 2000-02-11 | 2003-01-15 | Acta Technologies, Inc. | High availability database system using live/load database copies |
JP3768775B2 (en) | 2000-04-27 | 2006-04-19 | 三菱電機株式会社 | Backup apparatus and backup method |
US7111201B2 (en) | 2000-05-19 | 2006-09-19 | Self Repairing Computers, Inc. | Self repairing computer detecting need for repair and having switched protected storage |
US7100075B2 (en) | 2000-05-19 | 2006-08-29 | Sel Repairing Computers, Inc. | Computer system having data store protected from internet contamination by virus or malicious code and method for protecting |
US7137034B2 (en) | 2000-05-19 | 2006-11-14 | Vir2Us, Inc. | Self repairing computer having user accessible switch for modifying bootable storage device configuration to initiate repair |
US20060277433A1 (en) | 2000-05-19 | 2006-12-07 | Self Repairing Computers, Inc. | Computer having special purpose subsystems and cyber-terror and virus immunity and protection features |
US7096381B2 (en) | 2001-05-21 | 2006-08-22 | Self Repairing Computer, Inc. | On-the-fly repair of a computer |
US6880110B2 (en) | 2000-05-19 | 2005-04-12 | Self Repairing Computers, Inc. | Self-repairing computer having protected software template and isolated trusted computing environment for automated recovery from virus and hacker attack |
JP2002041305A (en) | 2000-07-26 | 2002-02-08 | Hitachi Ltd | Allocating method of computer resource in virtual computer system, and virtual computer system |
US6883098B1 (en) | 2000-09-20 | 2005-04-19 | International Business Machines Corporation | Method and computer system for controlling access by applications to this and other computer systems |
EP1195679A1 (en) | 2000-10-06 | 2002-04-10 | Hewlett-Packard Company, A Delaware Corporation | Performing operating system recovery from external back-up media in a headless computer entity |
US7086090B1 (en) | 2000-10-20 | 2006-08-01 | International Business Machines Corporation | Method and system for protecting pervasive devices and servers from exchanging viruses |
US6526491B2 (en) | 2001-03-22 | 2003-02-25 | Sony Corporation Entertainment Inc. | Memory protection system and method for computer architecture for broadband networks |
US6922774B2 (en) | 2001-05-14 | 2005-07-26 | The United States Of America As Represented By The National Security Agency | Device for and method of secure computing using virtual machines |
US7392541B2 (en) | 2001-05-17 | 2008-06-24 | Vir2Us, Inc. | Computer system architecture and method providing operating-system independent virus-, hacker-, and cyber-terror-immune processing environments |
US7849360B2 (en) | 2001-05-21 | 2010-12-07 | Vir2Us, Inc. | Computer system and method of controlling communication port to prevent computer contamination by virus or malicious code |
US6944785B2 (en) | 2001-07-23 | 2005-09-13 | Network Appliance, Inc. | High-availability cluster virtual server system |
US20030023895A1 (en) | 2001-07-27 | 2003-01-30 | Manish Sinha | Peripheral failover system |
US7536598B2 (en) | 2001-11-19 | 2009-05-19 | Vir2Us, Inc. | Computer system capable of supporting a plurality of independent computing environments |
WO2003067435A2 (en) | 2001-11-19 | 2003-08-14 | Self Repairing Computers, Inc. | Computer system capable of supporting a plurality of independent computing environments |
US20030105882A1 (en) | 2001-11-30 | 2003-06-05 | Ali Syed M. | Transparent injection of intelligent proxies into existing distributed applications |
US7788699B2 (en) | 2002-03-06 | 2010-08-31 | Vir2Us, Inc. | Computer and method for safe usage of documents, email attachments and other content that may contain virus, spy-ware, or malicious code |
US7277404B2 (en) | 2002-05-20 | 2007-10-02 | Airdefense, Inc. | System and method for sensing wireless LAN activity |
US7899867B1 (en) | 2002-07-31 | 2011-03-01 | FaceTime Communications, Inc, | SpIM blocking and user approval techniques for real-time messaging networks |
US7210144B2 (en) | 2002-08-02 | 2007-04-24 | Microsoft Corporation | Method for monitoring and emulating privileged instructions of programs in a virtual machine |
AU2003278706A1 (en) | 2002-08-09 | 2004-02-25 | Alphatech, Inc. | Control systems and methods using a partially-observable markov decision process (po-mdp) |
US7287052B2 (en) | 2002-11-09 | 2007-10-23 | Microsoft Corporation | Challenge and response interaction between client and server computing devices |
US20040123117A1 (en) | 2002-12-18 | 2004-06-24 | Symantec Corporation | Validation for behavior-blocking system |
US7904797B2 (en) | 2003-01-21 | 2011-03-08 | Microsoft Corporation | Rapid media group annotation |
US7246156B2 (en) | 2003-06-09 | 2007-07-17 | Industrial Defender, Inc. | Method and computer program product for monitoring an industrial network |
US7464158B2 (en) * | 2003-10-15 | 2008-12-09 | International Business Machines Corporation | Secure initialization of intrusion detection system |
US7222062B2 (en) | 2003-12-23 | 2007-05-22 | Intel Corporation | Method and system to support a trusted set of operational environments using emulated trusted hardware |
US7360249B1 (en) * | 2004-01-13 | 2008-04-15 | Symantec Corporation | Refining behavioral detections for early blocking of malicious code |
EP1709536A2 (en) | 2004-01-15 | 2006-10-11 | Vir2us, Inc. | Isolated multiplexed multi-dimensional processing in a virtual processing space having virus, spyware, and hacker protection features |
US20100005531A1 (en) | 2004-12-23 | 2010-01-07 | Kenneth Largman | Isolated multiplexed multi-dimensional processing in a virtual processing space having virus, spyware, and hacker protection features |
US7693991B2 (en) | 2004-01-16 | 2010-04-06 | International Business Machines Corporation | Virtual clustering and load balancing servers |
US8881282B1 (en) | 2004-04-01 | 2014-11-04 | Fireeye, Inc. | Systems and methods for malware attack detection and identification |
US8793787B2 (en) | 2004-04-01 | 2014-07-29 | Fireeye, Inc. | Detecting malicious network content using virtual environment components |
US8375444B2 (en) * | 2006-04-20 | 2013-02-12 | Fireeye, Inc. | Dynamic signature creation and enforcement |
US7707572B2 (en) | 2004-05-19 | 2010-04-27 | Bea Systems, Inc. | System and method for application container architecture |
US7908653B2 (en) | 2004-06-29 | 2011-03-15 | Intel Corporation | Method of improving computer security through sandboxing |
US7606821B2 (en) | 2004-06-30 | 2009-10-20 | Ebay Inc. | Method and system for preventing fraudulent activities |
WO2006101549A2 (en) | 2004-12-03 | 2006-09-28 | Whitecell Software, Inc. | Secure system for allowing the execution of authorized computer program code |
KR100649713B1 (en) | 2004-12-06 | 2006-11-28 | 한국전자통신연구원 | Method for hierarchical system configuration and integrated scheduling to provide multimedia streaming service on a two-level double cluster system |
US7882317B2 (en) | 2004-12-06 | 2011-02-01 | Microsoft Corporation | Process isolation using protection domains |
US8621458B2 (en) | 2004-12-21 | 2013-12-31 | Microsoft Corporation | Systems and methods for exposing processor topology for virtual machines |
US7409719B2 (en) * | 2004-12-21 | 2008-08-05 | Microsoft Corporation | Computer security management, such as in a virtual machine or hardened operating system |
US7979889B2 (en) | 2005-01-07 | 2011-07-12 | Cisco Technology, Inc. | Methods and apparatus providing security to computer systems and networks |
US7584503B1 (en) | 2005-02-22 | 2009-09-01 | Juniper Networks, Inc. | Federating trust in a heterogeneous network |
US7685633B2 (en) | 2005-02-25 | 2010-03-23 | Microsoft Corporation | Providing consistent application aware firewall traversal |
US7581229B2 (en) | 2005-03-11 | 2009-08-25 | Microsoft Corporation | Systems and methods for supporting device access from multiple operating systems |
US7552479B1 (en) | 2005-03-22 | 2009-06-23 | Symantec Corporation | Detecting shellcode that modifies IAT entries |
US20060271661A1 (en) | 2005-05-27 | 2006-11-30 | International Business Machines Corporation | Method for adaptively modifying the observed collective behavior of individual sensor nodes based on broadcasting of parameters |
US8078740B2 (en) | 2005-06-03 | 2011-12-13 | Microsoft Corporation | Running internet applications with low rights |
JP2007004661A (en) | 2005-06-27 | 2007-01-11 | Hitachi Ltd | Control method and program for virtual machine |
US20070044151A1 (en) | 2005-08-22 | 2007-02-22 | International Business Machines Corporation | System integrity manager |
US20070079307A1 (en) | 2005-09-30 | 2007-04-05 | Puneet Dhawan | Virtual machine based network carriers |
WO2007048062A2 (en) | 2005-10-21 | 2007-04-26 | Vir2Us, Inc. | Computer security method having operating system virtualization allowing multiple operating system instances to securely share single machine resources |
US9055093B2 (en) * | 2005-10-21 | 2015-06-09 | Kevin R. Borders | Method, system and computer program product for detecting at least one of security threats and undesirable computer files |
US8028336B2 (en) * | 2005-11-08 | 2011-09-27 | Oracle America, Inc. | Intrusion detection using dynamic tracing |
US7836303B2 (en) | 2005-12-09 | 2010-11-16 | University Of Washington | Web browser operating system |
US7685627B2 (en) | 2005-12-30 | 2010-03-23 | Microsoft Corporation | Unified networking diagnostics |
US8196205B2 (en) | 2006-01-23 | 2012-06-05 | University Of Washington Through Its Center For Commercialization | Detection of spyware threats within virtual machine |
KR100791290B1 (en) | 2006-02-10 | 2008-01-04 | 삼성전자주식회사 | Apparatus and method for using information of malicious application's behavior across devices |
US8694797B2 (en) | 2006-02-14 | 2014-04-08 | Lenovo (Sinapore) Pte Ltd | Method for preventing malicious software from execution within a computer system |
US7941813B1 (en) | 2006-02-17 | 2011-05-10 | Parallels Holdings, Ltd. | System and method for using virtual machine for driver installation sandbox |
US7774459B2 (en) | 2006-03-01 | 2010-08-10 | Microsoft Corporation | Honey monkey network exploration |
US7725922B2 (en) | 2006-03-21 | 2010-05-25 | Novell, Inc. | System and method for using sandboxes in a managed shell |
US20070240212A1 (en) * | 2006-03-30 | 2007-10-11 | Check Point Software Technologies, Inc. | System and Methodology Protecting Against Key Logger Spyware |
US8015563B2 (en) | 2006-04-14 | 2011-09-06 | Microsoft Corporation | Managing virtual machines with system-wide policies |
US8181244B2 (en) | 2006-04-20 | 2012-05-15 | Webroot Inc. | Backward researching time stamped events to find an origin of pestware |
WO2008039241A1 (en) | 2006-04-21 | 2008-04-03 | Av Tech, Inc | Methodology, system and computer readable medium for detecting and managing malware threats |
US7845009B2 (en) | 2006-05-16 | 2010-11-30 | Intel Corporation | Method and apparatus to detect kernel mode rootkit events through virtualization traps |
US8615800B2 (en) | 2006-07-10 | 2013-12-24 | Websense, Inc. | System and method for analyzing web content |
US8272048B2 (en) | 2006-08-04 | 2012-09-18 | Apple Inc. | Restriction of program process capabilities |
US20080059556A1 (en) | 2006-08-31 | 2008-03-06 | Egenera, Inc. | Providing virtual machine technology as an embedded layer within a processing platform |
US20080215852A1 (en) | 2006-08-31 | 2008-09-04 | Kenneth Largman | System and Device Architecture For Single-Chip Multi-Core Processor Having On-Board Display Aggregator and I/O Device Selector Control |
US20080127348A1 (en) | 2006-08-31 | 2008-05-29 | Kenneth Largman | Network computer system and method using thin user client and virtual machine to provide immunity to hacking, viruses and spy ware |
US20080082976A1 (en) | 2006-09-29 | 2008-04-03 | Matthias Steinwagner | Usage of virtualization software for shipment of software products |
US8234640B1 (en) | 2006-10-17 | 2012-07-31 | Manageiq, Inc. | Compliance-based adaptations in managed virtual systems |
US7900243B2 (en) | 2006-10-19 | 2011-03-01 | Oracle America, Inc. | Method and system for managing execution of an application module |
US7633864B2 (en) | 2006-12-20 | 2009-12-15 | Sun Microsystems, Inc. | Method and system for creating a demilitarized zone using network stack instances |
US8447880B2 (en) | 2006-12-20 | 2013-05-21 | Oracle America, Inc. | Network stack instance architecture with selection of transport layers |
US8468244B2 (en) | 2007-01-05 | 2013-06-18 | Digital Doors, Inc. | Digital information infrastructure and method for security designated data and with granular data stores |
US7840801B2 (en) | 2007-01-19 | 2010-11-23 | International Business Machines Corporation | Architecture for supporting attestation of a virtual machine in a single step |
US8184631B2 (en) | 2007-01-22 | 2012-05-22 | Oracle America, Inc. | Method for specifying a MAC identifier for a network-interface-device |
WO2008092031A2 (en) | 2007-01-24 | 2008-07-31 | Vir2Us, Inc. | Computer system architecture having isolated file system management for secure and reliable data processing |
US8856782B2 (en) | 2007-03-01 | 2014-10-07 | George Mason Research Foundation, Inc. | On-demand disposable virtual work system |
US20080320594A1 (en) | 2007-03-19 | 2008-12-25 | Xuxian Jiang | Malware Detector |
US8584236B2 (en) | 2007-03-28 | 2013-11-12 | British Telecommunications Public Limited Company | Method and apparatus for detecting abnormal traffic in a network |
US8572735B2 (en) | 2007-03-29 | 2013-10-29 | George Mason Research Foundation, Inc. | Attack resistant continuous network service trustworthiness controller |
US8127412B2 (en) | 2007-03-30 | 2012-03-06 | Cisco Technology, Inc. | Network context triggers for activating virtualized computer applications |
US20080271019A1 (en) | 2007-04-24 | 2008-10-30 | Stratton Robert J | System and Method for Creating a Virtual Assurance System |
US20080271025A1 (en) | 2007-04-24 | 2008-10-30 | Stacksafe, Inc. | System and method for creating an assurance system in a production environment |
US7873635B2 (en) | 2007-05-31 | 2011-01-18 | Microsoft Corporation | Search ranger system and double-funnel model for search spam analyses and browser protection |
US8074229B2 (en) | 2007-07-18 | 2011-12-06 | Sap Ag | Co-execution of objects from divergent runtime environments |
US20090034423A1 (en) * | 2007-07-30 | 2009-02-05 | Anthony Terrance Coon | Automated detection of TCP anomalies |
US8763115B2 (en) | 2007-08-08 | 2014-06-24 | Vmware, Inc. | Impeding progress of malicious guest software |
US8677352B2 (en) | 2007-10-31 | 2014-03-18 | Vmware, Inc. | Interchangeable guest and host execution environments |
US8220029B2 (en) | 2007-11-13 | 2012-07-10 | Samsung Electronics Co., Ltd. | Method and system for enforcing trusted computing policies in a hypervisor security module architecture |
CN101855606B (en) * | 2007-11-13 | 2012-08-29 | 惠普开发有限公司 | Launching an application from a power management state |
US7797748B2 (en) | 2007-12-12 | 2010-09-14 | Vmware, Inc. | On-access anti-virus mechanism for virtual machine architecture |
US8181174B2 (en) | 2007-12-28 | 2012-05-15 | Accenture Global Services Limited | Virtual machine configuration system |
US9264441B2 (en) | 2008-03-24 | 2016-02-16 | Hewlett Packard Enterprise Development Lp | System and method for securing a network from zero-day vulnerability exploits |
US9058483B2 (en) | 2008-05-08 | 2015-06-16 | Google Inc. | Method for validating an untrusted native code module |
US8910255B2 (en) | 2008-05-27 | 2014-12-09 | Microsoft Corporation | Authentication for distributed secure content management system |
US20090300599A1 (en) | 2008-05-30 | 2009-12-03 | Matthew Thomas Piotrowski | Systems and methods of utilizing virtual machines to protect computer systems |
US8356352B1 (en) | 2008-06-16 | 2013-01-15 | Symantec Corporation | Security scanner for user-generated web content |
US8881284B1 (en) | 2008-06-16 | 2014-11-04 | Symantec Operating Corporation | Method and system for secure network access using a virtual machine |
US8332825B2 (en) | 2008-06-26 | 2012-12-11 | Microsoft Corporation | Dynamically monitoring application behavior |
US8856937B1 (en) | 2008-06-27 | 2014-10-07 | Symantec Corporation | Methods and systems for identifying fraudulent websites |
US8434093B2 (en) | 2008-08-07 | 2013-04-30 | Code Systems Corporation | Method and system for virtualization of software applications |
US8776038B2 (en) | 2008-08-07 | 2014-07-08 | Code Systems Corporation | Method and system for configuration of virtualized software applications |
US8290763B1 (en) | 2008-09-04 | 2012-10-16 | Mcafee, Inc. | Emulation system, method, and computer program product for passing system calls to an operating system for direct execution |
US9098698B2 (en) | 2008-09-12 | 2015-08-04 | George Mason Research Foundation, Inc. | Methods and apparatus for application isolation |
US8370931B1 (en) * | 2008-09-17 | 2013-02-05 | Trend Micro Incorporated | Multi-behavior policy matching for malware detection |
US8347386B2 (en) * | 2008-10-21 | 2013-01-01 | Lookout, Inc. | System and method for server-coupled malware prevention |
US8850571B2 (en) | 2008-11-03 | 2014-09-30 | Fireeye, Inc. | Systems and methods for detecting malicious network content |
US8010085B2 (en) | 2008-11-19 | 2011-08-30 | Zscaler, Inc. | Traffic redirection in cloud based security services |
US9213566B2 (en) | 2008-11-26 | 2015-12-15 | Red Hat, Inc. | Implementing security in process-based virtualization |
US8745361B2 (en) | 2008-12-02 | 2014-06-03 | Microsoft Corporation | Sandboxed execution of plug-ins |
US20100146523A1 (en) | 2008-12-05 | 2010-06-10 | Tripod Ventures Inc./ Entreprises Tripod Inc. | Browser environment application and local file server application system |
DK2384375T3 (en) | 2009-01-16 | 2017-10-16 | Zeus Ind Products Inc | ELECTROSPINING PTFE WITH HIGH-VISUAL MATERIALS |
US8464252B2 (en) | 2009-02-27 | 2013-06-11 | Red Hat, Inc. | Per process virtual machines |
US8413139B2 (en) | 2009-03-11 | 2013-04-02 | Microsoft Corporation | Programming model for application and data access and synchronization within virtual environments |
US8935773B2 (en) | 2009-04-09 | 2015-01-13 | George Mason Research Foundation, Inc. | Malware detector |
US8752180B2 (en) * | 2009-05-26 | 2014-06-10 | Symantec Corporation | Behavioral engine for identifying patterns of confidential data use |
US8205035B2 (en) | 2009-06-22 | 2012-06-19 | Citrix Systems, Inc. | Systems and methods for integration between application firewall and caching |
US8001606B1 (en) * | 2009-06-30 | 2011-08-16 | Symantec Corporation | Malware detection using a white list |
US8839422B2 (en) | 2009-06-30 | 2014-09-16 | George Mason Research Foundation, Inc. | Virtual browsing environment |
US8627451B2 (en) * | 2009-08-21 | 2014-01-07 | Red Hat, Inc. | Systems and methods for providing an isolated execution environment for accessing untrusted content |
US8572740B2 (en) * | 2009-10-01 | 2013-10-29 | Kaspersky Lab, Zao | Method and system for detection of previously unknown malware |
US8479286B2 (en) * | 2009-12-15 | 2013-07-02 | Mcafee, Inc. | Systems and methods for behavioral sandboxing |
US9684785B2 (en) * | 2009-12-17 | 2017-06-20 | Red Hat, Inc. | Providing multiple isolated execution environments for securely accessing untrusted content |
US8401982B1 (en) * | 2010-01-14 | 2013-03-19 | Symantec Corporation | Using sequencing and timing information of behavior events in machine learning to detect malware |
US8578345B1 (en) * | 2010-04-15 | 2013-11-05 | Symantec Corporation | Malware detection efficacy by identifying installation and uninstallation scenarios |
US20120297457A1 (en) | 2010-11-15 | 2012-11-22 | Brian Schulte | Interactive Malware Detector |
US9117075B1 (en) * | 2010-11-22 | 2015-08-25 | Trend Micro Inc. | Early malware detection by cross-referencing host data |
US8826436B2 (en) * | 2010-12-08 | 2014-09-02 | At&T Intellectual Property I, L.P. | Systems, methods and apparatus to apply permissions to applications |
US8468600B1 (en) * | 2011-03-04 | 2013-06-18 | Adobe Systems Incorporated | Handling instruction received from a sandboxed thread of execution |
DK2707003T3 (en) | 2011-05-12 | 2019-08-05 | Paratek Pharm Innc | Crystalline Salts of (4S, 4AS, 5AR, 12AS) -4-DIMETHYLAMINO-3,10,12,12A-TETRAHYDROXY-7 - [(METHOXY (METHYL) AMINO) -METHYL] -1,11-DIOXO-1,4 , 4A, 5.5A, 6,11,12A-OCTAHYDRO-NAPHTHACEN-2-CARBOXYLIC ACID AMIDE AND PROCEDURES FOR USING IT |
US9047441B2 (en) * | 2011-05-24 | 2015-06-02 | Palo Alto Networks, Inc. | Malware analysis system |
US8799862B2 (en) * | 2011-06-24 | 2014-08-05 | Alcatel Lucent | Application testing using sandboxes |
US9065826B2 (en) * | 2011-08-08 | 2015-06-23 | Microsoft Technology Licensing, Llc | Identifying application reputation based on resource accesses |
US20130117006A1 (en) | 2011-11-07 | 2013-05-09 | Microsoft Corporation | Simulated boot process to detect introduction of unauthorized information |
US9081959B2 (en) | 2011-12-02 | 2015-07-14 | Invincea, Inc. | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US11411992B2 (en) | 2019-11-07 | 2022-08-09 | Mcafee, Llc | Visual detection of phishing websites via headless browser |
-
2012
- 2012-11-30 US US13/690,452 patent/US9081959B2/en active Active
- 2012-11-30 WO PCT/US2012/067311 patent/WO2013082437A1/en active Application Filing
-
2015
- 2015-07-13 US US14/797,847 patent/US9519779B2/en active Active
-
2016
- 2016-11-22 US US15/359,004 patent/US10043001B2/en active Active
-
2018
- 2018-06-26 US US16/018,720 patent/US10467406B2/en active Active
-
2019
- 2019-11-01 US US16/671,664 patent/US10984097B2/en active Active
-
2021
- 2021-03-24 US US17/211,412 patent/US12019734B2/en active Active
-
2024
- 2024-05-23 US US18/672,750 patent/US20240320323A1/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8719924B1 (en) * | 2005-03-04 | 2014-05-06 | AVG Technologies N.V. | Method and apparatus for detecting harmful software |
US20080016339A1 (en) * | 2006-06-29 | 2008-01-17 | Jayant Shukla | Application Sandbox to Detect, Remove, and Prevent Malware |
US20080235764A1 (en) * | 2007-03-22 | 2008-09-25 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | Resource authorizations dependent on emulation environment isolation policies |
US9111089B1 (en) * | 2011-02-08 | 2015-08-18 | Symantec Corporation | Systems and methods for safely executing programs |
US20120317645A1 (en) * | 2011-06-13 | 2012-12-13 | Microsoft Corporation | Threat level assessment of applications |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9846588B2 (en) | 2007-03-01 | 2017-12-19 | George Mason Research Foundation, Inc. | On-demand disposable virtual work system |
US10956184B2 (en) | 2007-03-01 | 2021-03-23 | George Mason Research Foundation, Inc. | On-demand disposable virtual work system |
US10187417B2 (en) | 2008-09-12 | 2019-01-22 | George Mason Research Foundation, Inc. | Methods and apparatus for application isolation |
US9602524B2 (en) | 2008-09-12 | 2017-03-21 | George Mason Research Foundation, Inc. | Methods and apparatus for application isolation |
US11310252B2 (en) | 2008-09-12 | 2022-04-19 | George Mason Research Foundation, Inc. | Methods and apparatus for application isolation |
US9871812B2 (en) | 2008-09-12 | 2018-01-16 | George Mason Research Foundation, Inc. | Methods and apparatus for application isolation |
US10567414B2 (en) | 2008-09-12 | 2020-02-18 | George Mason Research Foundation, Inc. | Methods and apparatus for application isolation |
US9436822B2 (en) | 2009-06-30 | 2016-09-06 | George Mason Research Foundation, Inc. | Virtual browsing environment |
US10120998B2 (en) | 2009-06-30 | 2018-11-06 | George Mason Research Foundation, Inc. | Virtual browsing environment |
US10043001B2 (en) | 2011-12-02 | 2018-08-07 | Invincea, Inc. | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US12019734B2 (en) | 2011-12-02 | 2024-06-25 | Invincea, Inc. | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US10467406B2 (en) | 2011-12-02 | 2019-11-05 | Invincea, Inc. | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US10984097B2 (en) | 2011-12-02 | 2021-04-20 | Invincea, Inc. | Methods and apparatus for control and detection of malicious content using a sandbox environment |
US20180083996A1 (en) * | 2016-09-21 | 2018-03-22 | Sentient Technologies (Barbados) Limited | Detecting behavioral anomaly in machine learned rule sets |
US11336672B2 (en) | 2016-09-21 | 2022-05-17 | Cognizant Technology Solutions U.S. Corporation | Detecting behavioral anomaly in machine learned rule sets |
US10735445B2 (en) * | 2016-09-21 | 2020-08-04 | Cognizant Technology Solutions U.S. Corporation | Detecting behavioral anomaly in machine learned rule sets |
CN106919837A (en) * | 2016-10-20 | 2017-07-04 | 深圳市安之天信息技术有限公司 | A kind of unknown self-starting recognition methods of malicious code and system |
US10831885B2 (en) * | 2016-10-31 | 2020-11-10 | Soliton Systems K.K. | Program behavior monitoring control apparatus, distributed object creation management apparatus, recording medium, and program behavior monitoring system |
US20180300474A1 (en) * | 2016-10-31 | 2018-10-18 | Soliton Systems K.K. | Program behavior monitoring control apparatus, distributed object creation management apparatus, recording medium, and program behavior monitoring system |
US10885189B2 (en) * | 2017-05-22 | 2021-01-05 | Microsoft Technology Licensing, Llc | Isolated container event monitoring |
WO2018217368A1 (en) * | 2017-05-22 | 2018-11-29 | Microsoft Technology Licensing, Llc | Isolated container event monitoring |
CN110651269A (en) * | 2017-05-22 | 2020-01-03 | 微软技术许可有限责任公司 | Isolated container event monitoring |
US20180336351A1 (en) * | 2017-05-22 | 2018-11-22 | Microsoft Technology Licensing, Llc | Isolated Container Event Monitoring |
US12093414B1 (en) * | 2019-12-09 | 2024-09-17 | Amazon Technologies, Inc. | Efficient detection of in-memory data accesses and context information |
Also Published As
Publication number | Publication date |
---|---|
US20180314823A1 (en) | 2018-11-01 |
US10043001B2 (en) | 2018-08-07 |
US20130145463A1 (en) | 2013-06-06 |
US20240320323A1 (en) | 2024-09-26 |
US20170200004A1 (en) | 2017-07-13 |
WO2013082437A1 (en) | 2013-06-06 |
US9081959B2 (en) | 2015-07-14 |
US20210209225A1 (en) | 2021-07-08 |
US20200242236A1 (en) | 2020-07-30 |
US9519779B2 (en) | 2016-12-13 |
US10984097B2 (en) | 2021-04-20 |
US10467406B2 (en) | 2019-11-05 |
US12019734B2 (en) | 2024-06-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US12019734B2 (en) | Methods and apparatus for control and detection of malicious content using a sandbox environment | |
US11636206B2 (en) | Deferred malware scanning | |
US11562068B2 (en) | Performing threat detection by synergistically combining results of static file analysis and behavior analysis | |
US9251343B1 (en) | Detecting bootkits resident on compromised computers | |
KR101265173B1 (en) | Apparatus and method for inspecting non-portable executable files | |
US11973780B2 (en) | Deobfuscating and decloaking web-based malware with abstract execution | |
US20130232576A1 (en) | Systems and methods for cyber-threat detection | |
CN110119619B (en) | System and method for creating anti-virus records | |
JP6134395B2 (en) | System and method for risk-based rules for application control | |
KR101086203B1 (en) | A proactive system against malicious processes by investigating the process behaviors and the method thereof | |
US12041070B2 (en) | Detecting malicious activity on an endpoint based on real-time system events | |
Čeponis et al. | Towards a robust method of dataset generation of malicious activity for anomaly-based HIDS training and presentation of AWSCTD dataset | |
CN102984134A (en) | Safe defense system | |
US10880316B2 (en) | Method and system for determining initial execution of an attack | |
Supriya et al. | Malware detection techniques: a survey | |
CN105844161A (en) | Security defense method, device and system | |
Patel et al. | Malware Detection Using Yara Rules in SIEM | |
EP3522058B1 (en) | System and method of creating antivirus records | |
US20240211597A1 (en) | Apparatus and methods for an application programming interface to detect and locate malware in memory | |
Major | A Taxonomic Evaluation of Rootkit Deployment, Behavior and Detection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INVINCEA, INC., VIRGINIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GHOSH, ANUP;COSBY, SCOTT;KEISTER, ALAN;AND OTHERS;SIGNING DATES FROM 20130221 TO 20130305;REEL/FRAME:036160/0358 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: COMERICA BANK, MICHIGAN Free format text: SECURITY INTEREST;ASSIGNOR:INVINCEA, INC.;REEL/FRAME:040766/0220 Effective date: 20150325 |
|
CC | Certificate of correction | ||
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.) |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
AS | Assignment |
Owner name: INVINCEA, INC., VIRGINIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:COMERICA BANK;REEL/FRAME:053118/0809 Effective date: 20200702 |
|
AS | Assignment |
Owner name: GOLDMAN SACHS BANK USA, AS COLLATERAL AGENT, NEW YORK Free format text: PATENT SECURITY AGREEMENT FIRST LIEN;ASSIGNOR:INVINCEA, INC.;REEL/FRAME:053124/0314 Effective date: 20200703 Owner name: OWL ROCK CAPITAL CORPORATION, AS COLLATERAL AGENT, NEW YORK Free format text: PATENT SECURITY AGREEMENT SECOND LIEN;ASSIGNOR:INVINCEA, INC.;REEL/FRAME:053124/0380 Effective date: 20200703 |
|
AS | Assignment |
Owner name: INVINCEA, INC., MASSACHUSETTS Free format text: RELEASE OF SECURITY INTEREST IN PATENTS AT R/F 053124/0380;ASSIGNOR:OWL ROCK CAPITAL CORPORATION, AS COLLATERAL AGENT;REEL/FRAME:055555/0308 Effective date: 20210308 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |