Samsung's ChatGPT Trade Secret Leak: Why Consumer AI Destroys IP Protection for Law Firms
In April 2023, Samsung Electronics discovered that engineers had uploaded proprietary semiconductor source code, yield optimization processes, and notes from a confidential internal meeting to ChatGPT — three separate incidents within a span of twenty days. The incidents occurred immediately after Samsung had lifted an internal ChatGPT ban and before it had established any AI use policy. Under the Defend Trade Secrets Act and the Uniform Trade Secrets Act, uploading trade secrets to a consumer AI platform with third-party data access constitutes disclosure that may permanently eliminate trade secret protection. The implications for law firms handling client IP — source code, M&A strategy, litigation strategy, regulatory submissions — are direct and severe.
🔒 Samsung ChatGPT Disclosure Incidents — April 2023
| Company | Samsung Electronics Co., Ltd. |
| Discovery Date | April 2023 (three incidents within 20 days) |
| AI Platform | ChatGPT (OpenAI public consumer version) |
| Incident 1 | Engineer uploaded source code from semiconductor equipment measurement tools to ask ChatGPT to identify and fix a bug |
| Incident 2 | Engineer uploaded source code related to Samsung's proprietary yield optimization process |
| Incident 3 | Employee uploaded notes from a confidential internal meeting and asked ChatGPT to generate a meeting summary |
| Legal Framework | Defend Trade Secrets Act (DTSA), 18 U.S.C. § 1836; Uniform Trade Secrets Act (UTSA) |
| Samsung Response | Company-wide ChatGPT ban reinstated; Samsung began developing internal proprietary AI |
The Samsung incidents were not the result of malicious actors or corporate espionage. They were the entirely predictable consequence of allowing employees to use a consumer productivity tool — one that happened to process user inputs on third-party servers under terms of service that permitted use of conversations for model training — without any policy framework governing what information could be submitted. The engineers were doing their jobs. They were using a tool that appeared useful for those jobs. Nobody had told them that doing so might permanently compromise Samsung's intellectual property rights in its most valuable technological assets.
The Three Incidents in Detail
Incident 1 — Semiconductor Equipment Measurement Source Code
An engineer working on Samsung's semiconductor equipment measurement tools uploaded source code from that system to ChatGPT and asked the model to identify and fix a bug. The source code represented a proprietary implementation of measurement algorithms used in Samsung's chip fabrication process — information with substantial independent economic value derived from its secrecy from Samsung's competitors in the semiconductor industry.
Incident 2 — Yield Optimization Process Source Code
A separate engineer uploaded source code related to Samsung's yield optimization process — the proprietary methodology by which Samsung maximizes the percentage of functional chips produced per silicon wafer. Yield optimization is one of the most commercially valuable and closely guarded technical processes in semiconductor manufacturing. The source code uploaded to ChatGPT represented the concrete implementation of that methodology.
Incident 3 — Confidential Internal Meeting Notes
An employee uploaded notes taken during a confidential internal meeting to ChatGPT and asked the model to generate a polished meeting summary. The notes contained information discussed by Samsung executives in the context of confidential business planning — information that, once uploaded to ChatGPT and processed under OpenAI's consumer terms of service, was no longer under Samsung's exclusive control.
How ChatGPT Training Works and Why Upload Equals Potential Permanent Disclosure
The legal significance of the Samsung incidents cannot be understood without understanding the technical mechanism by which consumer AI tools process and potentially retain user data. This is the technical foundation on which trade secret claims — and their elimination — rest.
The OpenAI Consumer Terms of Service at the Time
At the time of the Samsung incidents in April 2023, OpenAI's Terms of Service for the consumer ChatGPT product explicitly stated that conversations with the model could be used to train and improve future AI models. This was not buried in a footnote — it was a stated use in the primary terms. When a Samsung engineer uploaded proprietary source code and asked ChatGPT to debug it, that source code was transmitted to OpenAI's servers, processed by the model, and — under the applicable terms — potentially retained and used as training data for future model iterations.
Training data incorporation is not merely a theoretical risk. A model trained on a specific proprietary algorithm may, when queried by a subsequent user about the same problem domain, generate output that reflects patterns learned from that algorithm — effectively making Samsung's proprietary implementation accessible to anyone who asks the right question. The trade secret has not been "published" in the traditional sense, but it has been disclosed to a party under no obligation of confidentiality and has potentially been diffused into a system accessible to competitors.
Once proprietary information has been incorporated into a model's training data, it cannot be "removed." Model unlearning techniques exist but are imperfect and cannot guarantee complete elimination of a specific piece of training data's influence on model outputs. This means that a trade secret disclosed to a consumer AI training corpus may be permanently compromised — not reversible through legal action against the AI vendor.
The Architecture of Consumer AI Data Flows
Understanding consumer AI data flows requires distinguishing between several distinct stages at which proprietary information can be exposed, retained, or incorporated:
- Transmission: User input is transmitted over the internet to the vendor's servers. This transmission crosses multiple network infrastructure points and is processed by the vendor's infrastructure, which may involve sub-processors (cloud providers, GPU infrastructure, monitoring services) who are not parties to any confidentiality agreement with the user.
- Processing: The model processes the input to generate a response. During this stage, the input is in memory on third-party hardware under third-party operational control.
- Retention: Consumer AI products typically retain conversation history for some period — both to enable context-aware responses within a session and for quality review, safety monitoring, and potentially training.
- Training incorporation: Under the terms in effect at the time of the Samsung incidents, conversations could be used for model training, embedding the pattern of the input into the model's weight updates.
- Staff access: Consumer AI vendors typically reserve the right to access conversations for safety review, quality assurance, and moderation — meaning vendor employees may read uploaded proprietary information.
UTSA and DTSA: How Consumer AI Destroys the "Reasonable Measures" Element
Trade secret protection under both the Uniform Trade Secrets Act (which most states have adopted) and the federal Defend Trade Secrets Act (18 U.S.C. § 1836) requires two elements: (1) information that derives independent economic value from its secrecy, and (2) the owner's use of reasonable measures to maintain that secrecy. Consumer AI use attacks the second element directly and fundamentally.
The DTSA Definition of Trade Secret
Under 18 U.S.C. § 1839(3), a trade secret is information that: (A) the owner has taken reasonable measures to keep secret; and (B) derives independent economic value, actual or potential, from not being generally known to, and not being readily ascertainable through proper means by, another person who can obtain economic value from the disclosure or use of the information. Source code for proprietary semiconductor processes and yield optimization algorithms easily satisfies the economic value prong — they represent Samsung's competitive advantage in a trillion-dollar global industry. The reasonable measures prong is where consumer AI use creates catastrophic legal risk.
UTSA § 1(a) and the "Disclosure" Problem
The Uniform Trade Secrets Act defines misappropriation to include disclosure to a person who has no duty to maintain secrecy. UTSA § 1(a). Uploading trade secret information to ChatGPT constitutes exactly this: disclosure to a person (OpenAI and its infrastructure) under no contractual duty to maintain the secrecy of the specific information disclosed. Consumer terms of service do not create a duty of secrecy for specific proprietary content — they create a broad license to use the data for the vendor's purposes.
The critical legal consequence: An intellectual property owner who uploads trade secret information to a consumer AI platform with third-party data access rights may have voluntarily disclosed that information to a party under no obligation of secrecy. This disclosure can eliminate the "reasonable measures" element required for trade secret protection — potentially destroying protection not just for the disclosed information, but for related information whose secrecy was dependent on the same protections.
The "Reasonable Measures" Failure
Samsung's situation at the time of the incidents was particularly acute: the company had lifted its ChatGPT ban without establishing any AI use policy. This meant that Samsung's own internal policies provided no protection against the upload — there was nothing for an employee to violate, and no policy that Samsung could point to as evidence of "reasonable measures" to maintain the secrecy of the uploaded information. Courts examining whether an employer took reasonable measures to protect trade secrets look at internal policies, access controls, confidentiality agreements, and employee training. An employer who permits unlimited consumer AI use without any policy framework has, in effect, failed to take reasonable measures with respect to any information an employee might choose to upload.
The Corporate Espionage Risk
Even setting aside direct training data incorporation, the corporate espionage risk from consumer AI disclosure is not hypothetical. Cyberhaven's 2023 research on enterprise AI use found that employees regularly upload sensitive business information to ChatGPT — including source code, business strategies, M&A analysis, customer personally identifiable information, and internal financial data. Once that information is on a third-party platform under permissive terms of service, it is accessible to the vendor's staff, potentially incorporated into training data, and potentially accessible through carefully crafted queries by sophisticated actors who know what to look for.
The Direct Implications for Law Firms
Law firms handling client intellectual property face a version of the Samsung problem that is legally more severe, not less. Attorneys owe confidentiality obligations to clients under ABA Model Rule 1.6 and analogous state rules — obligations that predate and are independent of any trade secret analysis. But when a law firm attorney uploads a client's proprietary source code, patent application, product specification, or M&A due diligence materials to a consumer AI tool, the attorney has potentially committed two violations simultaneously: a breach of client confidentiality under Rule 1.6, and a contribution to the destruction of the client's trade secret protection.
Consider the typical scenarios in which law firm attorneys might upload client IP to a consumer AI tool:
- Patent prosecution: Attorney uploads client's invention disclosure or draft patent application to ask ChatGPT to help improve claim language — disclosing pre-publication invention information to a third-party platform with training data rights
- IP litigation: Attorney uploads client's source code to ask for an analysis of the similarities to accused infringing code — potentially disclosing trade secret implementation details
- M&A due diligence: Attorney uploads target company's technical specification or financial model to ask ChatGPT for a summary — disclosing confidential M&A information that may itself constitute a trade secret
- Contract drafting: Attorney uploads proprietary algorithm description to ask ChatGPT to draft a software license — disclosing the technical architecture the license is designed to protect
- Regulatory filings: Attorney uploads draft FDA submission, SEC filing, or FTC consent decree to ask for editing help — disclosing regulatory strategy that may be confidential
Consumer AI Terms of Service for Trade Secrets
- Data transmitted to and processed on third-party servers
- Conversations may be used for model training (consumer tier)
- Vendor staff may access for safety/quality review
- No specific confidentiality obligation for uploaded content
- No data isolation — shared infrastructure with all users
- Sub-processors (cloud/GPU) have access to conversation data
- Retention period often not clearly specified
- No contractual trade secret designation mechanism
- UTSA/DTSA "reasonable measures" element eliminated
- No incident notification if data is accessed or breached
Claire Enterprise Terms of Service for Trade Secrets
- Data processed within firm's own isolated tenant environment
- Zero use of client data for any model training, ever
- No vendor staff access to client data for any purpose
- Contractual confidentiality obligation for all processed data
- Isolated tenant — no shared infrastructure with other customers
- Sub-processors disclosed, contracted, and bound by same terms
- Data retention controlled by firm, not vendor
- DPA with trade secret designation and treatment provisions
- UTSA/DTSA "reasonable measures" element preserved
- Contractual breach notification within defined SLA
IP Audit Methodology for Law Firms
Law firms handling client intellectual property should conduct a systematic IP audit to identify the current risk exposure from consumer AI use by attorneys and staff. The following methodology provides a structured framework for that assessment:
Phase 1: Inventory Current AI Use
Survey all attorneys, paralegals, and staff about AI tools currently in use. Many law firms discover that unauthorized consumer AI use is far more widespread than management believed — staff use tools like ChatGPT for drafting, research, summarization, and formatting without understanding that client information uploaded to these tools creates IP and confidentiality risk.
Phase 2: Categorize Client Information at Risk
Identify the categories of client information that firm personnel are most likely to upload to AI tools in the course of normal work. For IP-intensive practices, this includes: patent applications and invention disclosures, source code and technical specifications, product roadmaps, competitive analyses, M&A transaction documents, and regulatory submissions. Prioritize the highest-risk categories for immediate policy intervention.
Phase 3: Review AI Vendor Terms
For each AI tool currently in use, obtain and review the applicable terms of service, privacy policy, and enterprise agreements. Identify: whether conversations are used for training, who has access to conversation data, what the data retention period is, whether data is processed by sub-processors, and what geographic jurisdiction governs data storage. Consumer products will typically fail all of these criteria for trade secret protection purposes.
Phase 4: Assess Trade Secret Exposure
For matters where consumer AI tools were used and client proprietary information may have been uploaded, assess the severity of trade secret exposure. Consult with the client about disclosure. In some cases — particularly where the information is highly valuable and the upload occurred under OpenAI's training-permissive consumer terms — notification and remediation discussions with the client may be required under both Rule 1.4 (communication) and Rule 1.6.
12-Item Enterprise AI Trade Secret Protection Checklist
AI Trade Secret Protection Checklist for Law Firms
Any AI tool used for client matters must provide a contractual guarantee — not a default opt-out, not a policy statement — that client data will never be used to train, fine-tune, or otherwise improve any AI model. This is the single most important trade secret protection requirement.
Before using any AI tool for matters involving client proprietary information, execute a Data Processing Agreement (DPA) that explicitly addresses trade secret treatment, confidentiality obligations for uploaded content, and vendor obligations that mirror your client confidentiality duties under Rule 1.6.
Confirm that your AI deployment operates in an isolated tenant environment — meaning your firm's data is not processed on shared infrastructure with other customers. Shared infrastructure creates the theoretical possibility of cross-customer data exposure that consumer platforms accept as an inherent risk.
Adopt a written firm policy explicitly prohibiting the use of consumer AI tools (ChatGPT free/Plus, Claude.ai consumer, Gemini consumer) for any task involving client trade secrets, patent applications, source code, technical specifications, M&A materials, or regulatory submissions.
Under DTSA and UTSA, the owner's reasonable measures to protect secrecy are an element of trade secret status. Your firm's AI use policy, vendor agreements, and technical controls constitute part of the "reasonable measures" documentation for client IP matters. Maintain and organize this documentation by matter.
Obtain a complete list of sub-processors used by your AI vendor — cloud infrastructure providers, GPU operators, monitoring services — who may have access to data you submit. Verify that each sub-processor is contractually bound by the same confidentiality and no-training-data-use obligations as the primary vendor.
Disclose AI use to clients in engagement letters. Specify the AI tools or architecture used, describe the trade secret and confidentiality protections in place, and obtain client consent before using AI to process their proprietary information. This satisfies both Rule 1.4 and demonstrates reasonable measures under DTSA.
Like Samsung's engineers, law firm employees may use consumer AI tools for legitimate work tasks without understanding the trade secret implications. Mandatory training on AI data practices and trade secret law is required — not optional CLE credit. The Samsung incidents demonstrate that even sophisticated technical employees make these errors without explicit guidance.
For matters already underway, assess whether consumer AI tools may have been used to process client proprietary information before firm policies were established. Where exposure is identified, evaluate disclosure obligations to clients and assess whether remediation or notification is required.
For clients with international operations, establish data residency requirements for AI processing. GDPR and other data protection frameworks may restrict processing of certain data outside specific jurisdictions. Enterprise AI deployments that support data residency controls protect both trade secrets and regulatory compliance.
Your vendor agreement must include an obligation to notify you promptly — within a defined timeframe, typically 24-72 hours — if your data is accessed by unauthorized parties or if a security incident occurs that may affect your client data. Consumer AI products typically provide no such notification.
For matters involving the most sensitive client IP — pending patents, core competitive algorithms, M&A targets — consider deploying AI in a fully air-gapped or MCP-isolated configuration where the model operates entirely within your network perimeter without any external data transmission.
How Claire Maintains Trade Secret Protection
Claire's Trade Secret-Safe Architecture
The Samsung incidents and the DTSA analysis they illustrate are architectural problems: consumer AI was not designed for environments where the "reasonable measures" element of trade secret protection must be preserved. Claire's enterprise deployment was designed from the start to operate within that requirement.
No Training Data Use — Contractually and Architecturally
Client data submitted to Claire during a session is processed in-session using ephemeral memory and is discarded when the session terminates. It does not persist to any database, is not used for fine-tuning or model improvement, and cannot influence the model's behavior in any future session. This is both a contractual commitment in Claire's enterprise agreement and an architectural characteristic of the deployment — verifiable through technical review, not merely a policy assertion.
Isolated Tenant — No Shared Infrastructure
Each Claire enterprise deployment operates in a fully isolated tenant environment. Your firm's AI instance processes no data from any other firm. There is no shared compute, no shared storage, and no shared model state between tenants. Cross-client data exposure — the theoretical vector that the New Jersey Advisory Committee identified as a concern requiring architectural elimination — is impossible by design, not merely contractually prohibited.
No Vendor Staff Access to Client Data
Claire's operational model does not include vendor staff access to client data for quality assurance, safety review, or any other purpose. The DPA reflects this as a contractual prohibition with defined consequences for breach. This eliminates the human-access vector through which trade secret information could be accessed by individuals at the AI vendor who are under no obligation of confidentiality to your clients.
MCP-Based System Isolation for Highest-Risk Matters
For matters involving the most sensitive client IP, Claire supports deployment in MCP-isolated configurations where the model operates within your own network perimeter. In this configuration, no client data leaves your infrastructure at any point during processing. This provides the strongest possible architectural support for the "reasonable measures" element of trade secret protection.
Complete Sub-Processor Disclosure and Contracting
Claire provides a complete list of sub-processors who may interact with your firm's data and contractual commitments binding each sub-processor to the same confidentiality and no-training obligations as the primary agreement. This sub-processor chain is auditable and forms part of the "reasonable measures" documentation your firm maintains for client IP matters.
Trade Secret-Aware Engagement Letter Templates
Claire provides engagement letter language that discloses AI use, describes the architectural protections in place for client IP, and obtains informed consent for AI-assisted work on matters involving trade secrets. This documentation satisfies both ABA Rule 1.4 disclosure requirements and DTSA/UTSA "reasonable measures" documentation requirements.
Consumer AI vs. Enterprise AI: Trade Secret Terms Comparison
The Samsung incidents should serve as a permanent reference point for any organization that handles valuable proprietary information and is considering AI tool deployment. The engineers who uploaded that code were not reckless actors — they were using a publicly available tool to do their jobs more efficiently, exactly as the tool was designed to be used. The problem was not the employees. The problem was an organizational failure to understand that consumer AI tools, by their architecture and their terms of service, are incompatible with the preservation of trade secret rights in information submitted to them.
For law firms, this incompatibility is not merely an internal corporate governance problem. It implicates direct professional obligations under ABA Model Rule 1.6, which requires attorneys to use reasonable measures to prevent unauthorized disclosure of client confidential information — including trade secrets. The Samsung incidents are the clearest possible illustration of what "unauthorized disclosure through consumer AI" looks like in practice, and they happened at one of the most sophisticated technology companies in the world.
For the regulatory compliance dimension of AI use in legal practice, see ABA Model Rules 1.1, 1.6, 5.3 and AI: The Legal Ethics Framework Every Law Firm Needs. For the discovery implications of AI-generated documents, see AI-Generated Documents as Discoverable ESI: FRCP Rule 26 and the New Litigation Hold Obligations.