added colors to indicate advml techniques

This commit is contained in:
Keith Manville 2020-10-16 13:05:43 -04:00
parent 12bc650e83
commit 2af432253c
3 changed files with 44 additions and 44 deletions

BIN
images/color_advml.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 108 B

BIN
images/color_cyber.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 108 B

View File

@ -33,49 +33,49 @@ However, there are two main differences:
### Reconnaissance
#### Acquire OSINT Information
#### ![AdvML](/images/color_advml.png) Acquire OSINT Information
Adversaries may leverage publicly available information about an organization that could identify where or how machine learning is being used in a system, and help tailor an attack to make it more effective. These sources of information include technical publications, blog posts, press releases, software repositories, public data repositories, and social media postings.
#### ML Model Discovery
#### ![AdvML](/images/color_advml.png) ML Model Discovery
Adversaries may attempt to identify machine learning pipelines that exist on the system and gather information about them, including the software stack used to train and deploy models, training and testing data repositories, model repositories, and software repositories containing algorithms. This information can be used to identify targets for further collection, exfiltration, or disruption, or to tailor and improve attacks.
> ##### Reveal ML Ontology
> ##### ![AdvML](/images/color_advml.png) Reveal ML Ontology
>
> stub
>
> ##### Reveal ML Model Family
> ##### ![AdvML](/images/color_advml.png) Reveal ML Model Family
>
> stub
#### Gathering Datasets
#### ![AdvML](/images/color_advml.png) Gathering Datasets
Adversaries may collect datasets similar to those used by a particular organization or in a specific approach. Datasets may be identified when [Acquiring OSINT Information](#Acquire-OSINT-Information). This may allow the adversary to replicate a private model's funcionality, constituting [Intellectual Property Theft](#Stolen-Intellectual-Property), or enable the adversary to carry out other attacks such as an [Evasion Attack](#Evasion-Attack).
#### Exploit Physical Environment
#### ![AdvML](/images/color_advml.png) Exploit Physical Environment
Stub
#### Model Replication
#### ![AdvML](/images/color_advml.png) Model Replication
Stub
> ##### Exploit API
> ##### ![AdvML](/images/color_advml.png) Exploit API
>
> An adversary my replicate a machine learning model's functionality by exploiting its inference API.
>
> ##### Pre-Trained Weights
> ##### ![AdvML](/images/color_advml.png) Pre-Trained Weights
>
> Once made publicly available, an adversary may use pre-trained weights to replicate a model's functionality.
### ML Model Stealing
#### ![AdvML](/images/color_advml.png) ML Model Stealing
Machine learning models' functionality can be stolen exploiting an inference API. This constitutes [Stolen Intellectual Property](#Stolen-Intellectual-Property).
### Initial Access
#### Exploit Public-Facing Application
#### ![Cyber](/images/color_cyber.png) Exploit Public-Facing Application
Adversaries may attempt to take advantage of a weakness in an Internet-facing computer or program using software, data, or commands in order to cause unintended or unanticipated behavior. The weakness in the system can be a bug, a glitch, or a design vulnerability. These applications are often websites, but can include databases (like SQL)(Citation: NVD CVE-2016-6662), standard services (like SMB(Citation: CIS Multiple SMB Vulnerabilities) or SSH), and any other applications with Internet accessible open sockets, such as web servers and related services.(Citation: NVD CVE-2014-7169) Depending on the flaw being exploited this may include [Exploitation for Defense Evasion](https://attack.mitre.org/techniques/T1211).
@ -83,29 +83,29 @@ If an application is hosted on cloud-based infrastructure, then exploiting it ma
For websites and databases, the OWASP top 10 and CWE top 25 highlight the most common web-based vulnerabilities.(Citation: OWASP Top 10)(Citation: CWE top 25)
#### Valid Accounts
#### ![Cyber](/images/color_cyber.png) Valid Accounts
Adversaries may obtain and abuse credentials of existing accounts as a means of gaining Initial Access, Persistence, Privilege Escalation, or Defense Evasion. Compromised credentials may be used to bypass access controls placed on various resources on systems within the network and may even be used for persistent access to remote systems and externally available services, such as VPNs, Outlook Web Access and remote desktop. Compromised credentials may also grant an adversary increased privilege to specific systems or access to restricted areas of the network. Adversaries may choose not to use malware or tools in conjunction with the legitimate access those credentials provide to make it harder to detect their presence.
The overlap of permissions for local, domain, and cloud accounts across a network of systems is of concern because the adversary may be able to pivot across accounts and systems to reach a high level of access (i.e., domain or enterprise administrator) to bypass access controls set within the enterprise. (Citation: TechNet Credential Theft)
#### Phishing
#### ![Cyber](/images/color_cyber.png) Phishing
Adversaries may send phishing messages to elicit sensitive information and/or gain access to victim systems. All forms of phishing are electronically delivered social engineering. Phishing can be targeted, known as spearphishing. In spearphishing, a specific individual, company, or industry will be targeted by the adversary. More generally, adversaries can conduct non-targeted phishing, such as in mass malware spam campaigns.
Adversaries may send victims emails containing malicious attachments or links, typically to execute malicious code on victim systems or to gather credentials for use of [Valid Accounts](https://attack.mitre.org/techniques/T1078). Phishing may also be conducted via third-party services, like social media platforms.
#### External Remote Services
#### ![Cyber](/images/color_cyber.png) External Remote Services
Adversaries may leverage external-facing remote services to initially access and/or persist within a network. Remote services such as VPNs, Citrix, and other access mechanisms allow users to connect to internal enterprise network resources from external locations. There are often remote service gateways that manage connections and credential authentication for these services. Services such as [Windows Remote Management](https://attack.mitre.org/techniques/T1021/006/) can also be used externally.
Access to [Valid Accounts](https://attack.mitre.org/techniques/T1078) to use the service is often a requirement, which could be obtained through credential pharming or by obtaining the credentials from users after compromising the enterprise network. Access to remote services may be used as a redundant or persistent access mechanism during an operation.
#### Pre-Trained ML Model with Backdoor
#### ![AdvML](/images/color_advml.png) Pre-Trained ML Model with Backdoor
Adversaries may gain initial access to a system by compromising portions of the ML supply chain. This could include GPU hardware, data and its annotations, parts of the ML software stack, or the model itself. In some instances the attacker will need secondary access to fully carry out an attack using compromised components of the supply chain.
#### Trusted Relationship
#### ![Cyber](/images/color_cyber.png) Trusted Relationship
Adversaries may breach or otherwise leverage organizations who have access to intended victims. Access through trusted third party relationship exploits an existing connection that may not be protected or receives less scrutiny than standard mechanisms of gaining access to a network.
@ -113,19 +113,19 @@ Organizations often grant elevated access to second or third-party external prov
### Execution
#### Execute Unsafe ML Models
#### ![AdvML](/images/color_advml.png) Execute Unsafe ML Models
An Adversary may utilize unsafe ML Models that when executed have an unintended effect. The adversary can use this technique to establish persistent access to systems. These models may be introduced via a [Pre-Trained Model with Backdoor](#Pre-Trained-Model-with-Backdoor).
> ##### ML Models from Compromised Sources
> ##### ![AdvML](/images/color_advml.png) ML Models from Compromised Sources
>
> An adversary may introduce an unsafe model from a compromised source.
>
> ##### Pickle Embedding
> ##### ![AdvML](/images/color_advml.png) Pickle Embedding
>
> An adversary may use pickle embedding to introduce malicious data payloads.
#### Native API
#### ![Cyber](/images/color_cyber.png) Native API
Adversaries may directly interact with the native OS application programming interface (API) to execute behaviors. Native APIs provide a controlled means of calling low-level OS services within the kernel, such as those involving hardware/devices, memory, and processes. These native APIs are leveraged by the OS during system boot (when other system components are not yet initialized) as well as carrying out tasks and requests during routine operations.
@ -135,21 +135,21 @@ Higher level software frameworks, such as Microsoft .NET and macOS Cocoa, are al
Adversaries may abuse these native API functions as a means of executing behaviors. Similar to [Command and Scripting Interpreter](https://attack.mitre.org/techniques/T1059/), the native API and its hierarchy of interfaces, provide mechanisms to interact with and utilize various components of a victimized system.
#### Traditional Software Attacks
#### ![Cyber](/images/color_cyber.png) Traditional Software Attacks
Stubs
### Persistence
#### Unsafe ML Model Execution
#### ![AdvML](/images/color_advml.png) Unsafe ML Model Execution
An Adversary may utilize unsafe ML Models that when executed have an unintended effect. The adversary can use this technique to establish persistent access to systems. These models may be introduced via a [Pre-trained Model with Backdoor](#Pre-Trained-Model-with-Backdoor). An example of this technique is to use pickle embedding to introduce malicious data payloads.
#### Account Manipulation
#### ![Cyber](/images/color_cyber.png) Account Manipulation
Adversaries may manipulate accounts to maintain access to victim systems. Account manipulation may consist of any action that preserves adversary access to a compromised account, such as modifying credentials or permission groups. These actions could also include account activity designed to subvert security policies, such as performing iterative password updates to bypass password duration policies and preserve the life of compromised credentials. In order to create or manipulate accounts, the adversary must already have sufficient permissions on systems or the domain.
#### Implant Container Image
#### ![Cyber](/images/color_cyber.png) Implant Container Image
Adversaries may implant cloud container images with malicious code to establish persistence. Amazon Web Service (AWS) Amazon Machine Images (AMI), Google Cloud Platform (GCP) Images, and Azure Images as well as popular container runtimes such as Docker can be implanted or backdoored. Depending on how the infrastructure is provisioned, this could provide persistent access if the infrastructure provisioning tool is instructed to always use the latest image.(Citation: Rhino Labs Cloud Image Backdoor Technique Sept 2019)
@ -158,73 +158,73 @@ A tool has been developed to facilitate planting backdoors in cloud container im
### Defense Evasion
#### Evasion Attack
#### ![AdvML](/images/color_advml.png) Evasion Attack
Adversaries can create data inputs that prevent a machine learning model from positively identifying the data sample. This technique can be used to evade detection on the network, or to evade a downstream task where machine learning is utilized.
Example evasion attacks include Simple Transformation, Common Corruption, Adversarial Examples, Happy String
> ##### Offline Evasion
> ##### ![AdvML](/images/color_advml.png) Offline Evasion
>
> Simple Transformation, Common Corruption, Adversarial Examples, Happy Strings
>
> ##### Online Evasion
> ##### ![AdvML](/images/color_advml.png) Online Evasion
>
> Simple Transformation, Common Corruption, Adversarial Examples, Happy Strings
#### Model Poisoning
#### ![AdvML](/images/color_advml.png) Model Poisoning
Adversaries can train machine learning that are performant, but contain backdoors that produce inference errors when presented with input containing a trigger defined by the adversary. A model with a backdoor can be introduced by an innocent user via a [pre-trained model with backdoor](#Pre-Trained-Model-with-Backdoor) or can be a result of [Data Poisoning](#Data-Poisoning). This backdoored model can be exploited at inference time with an [Evasion Attack](#Evasion-Attack)
#### Data Poisoning
#### ![AdvML](/images/color_advml.png) Data Poisoning
Adversaries may attempt to poison datasets used by a ML system by modifying the underlying data or its labels. This allows the adversary to embed vulnerabilities in ML models trained on the data that may not be easily detectable. The embedded vulnerability can be activated at a later time by providing the model with data containing the trigger. Data Poisoning can help enable attacks such as [ML Model Evasion](#Evasion-Attack).
> ###### Tainting Data from Acquisition - Label Corruption
> ###### ![AdvML](/images/color_advml.png) Tainting Data from Acquisition - Label Corruption
>
> Adversaries may attempt to alter labels in a training set. This would cause a model to misclassify an input
>
> ###### Tainting Data from Open Source Supply Chains
> ###### ![AdvML](/images/color_advml.png) Tainting Data from Open Source Supply Chains
>
> Adversaries may attempt to add their own data to an open source dataset which could create a classfication backdoor.
>
> ###### Tainting Data from Acquisition - Chaff Data
> ###### ![AdvML](/images/color_advml.png) Tainting Data from Acquisition - Chaff Data
>
> Adding noise to a dataset would lower the accuracy of the model, potentially making the model more vulnerable to misclassifications
>
> ###### Tainting Data in Training - Label Corruption
> ###### ![AdvML](/images/color_advml.png) Tainting Data in Training - Label Corruption
>
> Changing training labels could create a backdoor in the model, such that a malicious input would always be classified to the benefit of the adversary
### Exfiltration
#### Exfiltrate Training Data
#### ![AdvML](/images/color_advml.png) Exfiltrate Training Data
Adversaries may exfiltrate private information related to machine learning models via their inference APIs. Additionally, adversaries can use these APIs to create copy-cat or proxy models.
> ##### Membership Inference Attack
> ##### ![AdvML](/images/color_advml.png) Membership Inference Attack
>
> The membership of a data sample in a training set may be infered by an adversary with access to an inference API.
>
> ##### ML Model Inversion
> ##### ![AdvML](/images/color_advml.png) ML Model Inversion
>
> Machine learning models' training data could be reconstructed by exploiting an inference API.
#### ML Model Stealing
#### ![AdvML](/images/color_advml.png) ML Model Stealing
Machine learning models' functionality can be stolen exploiting an inference API. This constitutes [Stolen Intellectual Property](#Stolen-Intellectual-Property).
#### Insecure Storage
#### ![Cyber](/images/color_cyber.png) Insecure Storage
Adversaries may exfiltrate proprietary machine learning models or private training and testing data by exploiting insecure storage mechanisms. Adversaries may [discover](#ML-Model-Discovery), and exfiltrate components of a ML pipeline, resulting in [Stolen Intellectual Property](#Stolen-Intellectual-Property)
### Impact
#### ML Model Integrity Attacks
#### ![AdvML](/images/color_advml.png) ML Model Integrity Attacks
Adversaries may attack the integrity of machine learning models by crafting adversarial examples that appear to be normal inputs, but cause the model to make errors at inference time. These attacks can reduce prediction confidence and cause false positives or false negatives, eroding confidence in the system over time.
#### Defacement
#### ![AdvML](/images/color_advml.png) Defacement
Adversaries can create data inputs that prevent a machine learning model from positively identifying the data sample. This technique can be used to evade detection on the network, or to evade a downstream task where machine learning is utilized.
@ -232,23 +232,23 @@ Example evasion attacks include Simple Transformation, Common Corruption, Advers
Exploit Open CVEs in Package
#### Denial of Service
#### ![AdvML](/images/color_advml.png) Denial of Service
Adversaries may target different Machine Learning services to conduct a DoS.
One example of this type of attack is Sponge attack.
#### Stolen Intellectual Property
#### ![Cyber](/images/color_cyber.png) Stolen Intellectual Property
Adversaries may steal intellectual property by [Model Replication](#ML-Model-Replication) or [Model Stealing](#ML-Model-Stealing).
#### Data Encrypted for Impact
#### ![AdvML](/images/color_advml.png) Data Encrypted for Impact
Adversaries may encrypt data on target systems or on large numbers of systems in a network to interrupt availability to system and network resources. They can attempt to render stored data inaccessible by encrypting files or data on local and remote drives and withholding access to a decryption key. This may be done in order to extract monetary compensation from a victim in exchange for decryption or a decryption key (ransomware) or to render data permanently inaccessible in cases where the key is not saved or transmitted.(Citation: US-CERT Ransomware 2016)(Citation: FireEye WannaCry 2017)(Citation: US-CERT NotPetya 2017)(Citation: US-CERT SamSam 2018) In the case of ransomware, it is typical that common user files like Office documents, PDFs, images, videos, audio, text, and source code files will be encrypted. In some cases, adversaries may encrypt critical system files, disk partitions, and the MBR.(Citation: US-CERT NotPetya 2017)
To maximize impact on the target organization, malware designed for encrypting data may have worm-like features to propagate across a network by leveraging other attack techniques like [Valid Accounts](https://attack.mitre.org/techniques/T1078), [OS Credential Dumping](https://attack.mitre.org/techniques/T1003), and [SMB/Windows Admin Shares](https://attack.mitre.org/techniques/T1021/002).(Citation: FireEye WannaCry 2017)(Citation: US-CERT NotPetya 2017)
#### Stop System Shutdown/Reboot
#### ![AdvML](/images/color_advml.png) Stop System Shutdown/Reboot
Stub