GenAI Security Risk and Mitigation
[ 1 ] Sensitive Information Disclosure (Data Engineering - Source Data)
Risk: Sensitive data, such as personal information or confidential business data, could be exposed during model training or usage.
Mitigation Strategies:
[ 2 ] Training Data Poisoning (Data Engineering - Source Data)
Risk: Malicious actors can inject corrupted or misleading data into the training set, causing the model to make inaccurate predictions or behave undesirably.
Mitigation Strategies:
[ 3 ] Supply Chain Vulnerability (Data Engineering)
Risk: Supply chain vulnerabilities arise from third-party tools, libraries, or data sources, which may introduce backdoors, malware, or vulnerabilities.
Mitigation Strategies:
[ 4 ] Model Theft (Data Engineering)
Risk: Models are valuable intellectual property, and adversaries might steal or reverse-engineer them for unauthorized usage or exploitation.
Mitigation Strategies:
[ 5 ] Insecure Plugin Design (LLM Usage - LLM Plugins)
Risk: Insecurely designed third-party plugins may introduce vulnerabilities into the LLM ecosystem, leading to data leaks or system exploits.
Mitigation Strategies:
Recommended by LinkedIn
[ 6 ] Prompt Injection (LLM Usage - Prompt Engineering)
Risk: Prompt injection involves maliciously crafting inputs that manipulate the model to generate harmful or unauthorized outputs.
Mitigation Strategies:
[ 7 ] Insecure Output Handling (LLM Usage - Prompt Engineering)
Risk: Generated outputs may unintentionally contain sensitive or harmful information, leading to breaches, misinformation, or inappropriate behavior.
Mitigation Strategies:
[ 8 ] Overreliance (LLM Usage - Prompt Engineering)
Risk: Blind reliance on LLM outputs without proper validation or oversight can lead to poor decision-making or propagation of biases.
Mitigation Strategies:
[ 9 ] Model Theft (LLM Application Development & Model Engineering)
Risk: Attackers may attempt to steal or reverse-engineer models from the engineering phase, leading to unauthorized use or replication.
Mitigation Strategies:
[ 10 ] Supply Chain Vulnerability (LLM Application Development)
Risk: Use of third-party libraries, models, or tools can introduce supply chain vulnerabilities that compromise the security of the model and its ecosystem.
Mitigation Strategies:
Conclusion
Each layer of the GenAI system introduces unique risks that can be mitigated through a combination of technical solutions, governance, and operational controls. From the security of the data pipeline to the development and use of LLMs, addressing these risks through encryption, access controls, regular audits, and human oversight helps maintain the integrity, confidentiality, and availability of AI systems.