Risk Assessment Framework
The Hidden
Threats of
Generative AI
A structured examination of three critical risk domains—data privacy, copyright exposure, and cybersecurity—that every organization deploying GenAI must confront.
privacy as their top GenAI concern
litigation exposure industry-wide
vs. traditional software
GenAI systems ingest, memorize, and reproduce personal data in ways that fundamentally challenge existing privacy frameworks. The very capability that makes them powerful—learning from vast datasets—creates profound exposure.
Models trained on scraped web data may memorize PII, medical records, or private communications and reproduce them verbatim when prompted correctly.
↳ Researchers have extracted real names, addresses, and SSNs from GPT-2 and GPT-3 via targeted prompting.
Adversaries can infer sensitive attributes—health status, sexual orientation, political beliefs—from model responses without direct data access.
↳ Membership inference attacks can confirm whether specific individuals appeared in training data.
In shared GenAI deployments, context windows may inadvertently surface another user’s queries, documents, or personally identifiable information.
↳ Samsung engineers accidentally uploaded proprietary semiconductor source code to ChatGPT in 2023.
Data originally collected for one purpose (e.g., customer support) is repurposed for model training without explicit user consent, violating GDPR and similar laws.
↳ CNIL (France) fined a company for using customer chat data in LLM fine-tuning without lawful basis.
“GenAI doesn’t just process data—it absorbs it. Every input becomes a potential memory, and every output a potential disclosure.”
Mitigation Strategies
- Differential privacy in model training
- Strict data minimization policies
- PII scrubbing before ingestion
- Right-to-erasure mechanisms
- Federated learning architectures
- Regular model audits for memorization
- Explicit consent frameworks
- Data residency controls
Regulatory Landscape
EU General Data Protection Regulation
Requires lawful basis for processing; AI outputs that include personal data are subject to full GDPR obligations including Article 22 on automated decisions.
EU Artificial Intelligence Act (2024)
Classifies certain AI systems as high-risk with mandatory data governance requirements; GPAI model providers must maintain detailed training data documentation.
California Consumer Privacy Act
Grants consumers rights to know what data is used in AI training; opt-out rights apply to automated profiling of personal information.
The intellectual property implications of GenAI are tectonic. Models trained on copyrighted works produce outputs that may constitute infringement—yet existing legal frameworks were never designed for this paradigm.
Using copyrighted books, code, images, and music to train models without licensing may constitute direct infringement at massive scale.
↳ The New York Times v. OpenAI alleges billions of copyrighted articles were used without authorization or compensation.
Models can reproduce near-verbatim excerpts from training data—generating full song lyrics, code functions, or book passages on demand.
↳ GitHub Copilot has been documented reproducing entire MIT/GPL-licensed code blocks without attribution.
Who owns AI-generated content? Current law in most jurisdictions requires human authorship for copyright protection, creating a dangerous ownership vacuum.
↳ The US Copyright Office has repeatedly denied registration for purely AI-generated works.
GenAI outputs that are “substantially similar” to copyrighted source material—even when transformed—may qualify as infringing derivative works.
↳ Image generators producing work in the “style of” a living artist have faced multiple infringement suits.
“Every GenAI deployment carries a latent copyright liability that crystallizes the moment an output goes public.”
Mitigation Strategies
- License training data explicitly
- Implement output filtering for verbatim reproduction
- Use permissively-licensed model alternatives
- Maintain provenance records for training data
- Implement “opt-out” registries for creators
- Watermark AI-generated content
- IP indemnification clauses in vendor contracts
- Human review for high-stakes outputs
Key Legal Battlegrounds
NYT v. OpenAI & Microsoft
The New York Times alleges copyright infringement via training on millions of articles; seeks billions in damages and model destruction.
Getty Images v. Stability AI
Getty claims Stable Diffusion was trained on 12M+ watermarked images without license; seeks injunction and damages in UK and US courts.
Authors Guild v. OpenAI
Class action by thousands of authors alleging their books were ingested for training without compensation or consent.
GenAI has simultaneously lowered the barrier to cyberattacks while introducing entirely new vulnerability classes. It is both weapon and target—organizations face threats from both directions.
Attackers embed malicious instructions in content the AI processes—emails, documents, web pages—hijacking the AI’s actions or extracting sensitive system prompts.
↳ “Indirect prompt injection” attacks on AI assistants have exfiltrated user data via crafted web pages the AI browses.
Threat actors use GenAI to craft hyper-personalized phishing, generate malware variants, automate vulnerability research, and create deepfake social engineering.
↳ WormGPT and FraudGPT—jailbroken LLMs sold on dark web—generate convincing BEC phishing at scale.
Adversaries inject malicious data into model training pipelines, causing the model to behave incorrectly or maliciously in specific, attacker-controlled scenarios.
↳ Backdoor attacks can cause models to misclassify inputs containing specific triggers with near-100% reliability.
Open-source model weights and datasets from repositories like Hugging Face may contain embedded malware, trojans, or data poisoning, infecting downstream deployments.
↳ JFrog researchers discovered PyTorch models on Hugging Face containing malicious pickle payloads.
“Prompt injection is the SQL injection of the AI era—except the attack surface is every piece of text the model touches.”
Mitigation Strategies
- Input/output sanitization pipelines
- Principle of least privilege for AI agents
- Red-team testing for prompt injection
- Model scanning before deployment
- Robust system prompt separation
- AI activity monitoring and logging
- Human-in-the-loop for sensitive actions
- Zero-trust architecture for AI integrations
Emerging Attack Taxonomy
Direct Prompt Injection
User directly inputs adversarial instructions to override system prompts or extract confidential model context and instructions.
Indirect / Environmental Injection
Malicious instructions embedded in external content (emails, files, web pages) that the AI agent reads during an agentic task.
Safety Bypass & Jailbreaking
Techniques to circumvent model safety guardrails—roleplaying, token manipulation, multi-turn social engineering—to produce harmful outputs.

