The AI training data risk is real. In early 2023, Samsung engineers were given access to ChatGPT.
The worst part was not the initial incident. It was the realisation that followed: there was no way to get any of it back. Samsung banned the use of generative AI tools on company devices shortly after. But the data that had already been submitted was already gone — not in the sense of being deleted, but in the sense of potentially being used to train the very models that the rest of the world, including Samsung's competitors, would go on to use.
Samsung is a global enterprise with dedicated security teams and the resources to detect and respond quickly. Most UK SMEs will not discover this has happened to them at all.
Three data leaks in twenty days. Source code, internal meeting recordings, and trade secrets — all submitted via ChatGPT consumer accounts. None of it retrievable. Samsung subsequently banned generative AI tools on company networks entirely.
Consumer accounts versus enterprise accounts — and why the difference matters enormously
Not all AI accounts are created equal. The distinction between a consumer account and an enterprise account is not just about features or price. It is about what happens to your data after you submit it.
On a consumer account — the free or standard paid tier that most individuals sign up for — AI providers typically reserve the right to use your conversation data to train and improve their models. This is usually disclosed in the terms of service, buried in clauses that most users never read. It is the standard trade: access to a powerful AI tool in exchange for your inputs becoming training material.
On an enterprise account — the business tier designed for commercial use — providers typically commit not to use your data for model training. Training is disabled by default. There may also be additional commitments around data retention, residency, and deletion.
The gap between these two configurations is enormous. And here is the uncomfortable reality: the majority of SMEs in the UK that are using AI tools right now are doing so on consumer accounts. They have not made a deliberate choice to accept training data risk. They simply signed up for the free tool, found it useful, and started using it — without knowing that the enterprise tier exists or why it matters.
What "in the model weights" actually means
When people hear that their data might be used for training, they often imagine it is stored somewhere as a retrievable file — like a document sitting in a server that could, in theory, be found and deleted. That is not how it works.
Training data is used to adjust the numerical parameters — the "weights" — of a neural network. The model learns from patterns across enormous volumes of text. Once that learning has occurred, the specific inputs that contributed to it are not separately stored or identifiable. The data is not in the model in any recoverable sense. It has become part of the model's behaviour.
This has a critical implication for GDPR's Right to Erasure. Under UK data protection law, individuals can request that their personal data be deleted. But if that data has already been incorporated into model weights, there is no mechanism by which it can be erased. The legal obligation exists — the technical means to fulfil it does not. AI providers have acknowledged this tension publicly. There is no clean resolution.
What kinds of data are actually at risk?
The instinct is to think of dramatic examples — nuclear launch codes, classified government documents. But the data that businesses are routinely submitting to AI tools on consumer accounts is far more mundane, and in many cases equally valuable:
- Source code — Pasted in to check for bugs, optimise performance, or get explanations. This is often the most commercially sensitive IP a technology business owns.
- Financial projections — Submitted to help write board reports, investor updates, or internal summaries. Revenue figures, margin data, growth plans.
- Strategy documents — Uploaded to be summarised or refined. Competitive analysis, product roadmaps, acquisition plans.
- Client lists and CRM data — Pasted in to help draft communications or segment audiences. Names, companies, contact details, deal values.
- Internal communications — Meeting notes, performance reviews, HR correspondence, internal policy drafts — submitted to be cleaned up or turned into action items.
None of these feel like dramatic security incidents in the moment. They feel like using a useful tool to do your job more efficiently. That is precisely why the risk is so underestimated.
Enterprise AI accounts disable training by default. Consumer accounts often don't — and most SMEs are on consumer accounts.
Detect, Assess, Defend — getting this under control
Training data risk is entirely manageable once you know it exists and understand the configuration differences between account types. The challenge is that most businesses have never audited their AI tool usage with this lens. A structured approach covers three phases.
The practical fix for most businesses is more straightforward than it sounds. Identify which AI tools are in use. Check whether those tools are being accessed via consumer or enterprise accounts. Where consumer accounts are in use for business purposes, either upgrade to enterprise configurations or restrict use until that upgrade is in place. Then document it — because your clients, your insurers, and potentially the ICO may one day ask.
The harder task is cultural: helping staff understand why this matters. Most people who pasted company data into a consumer AI account were not being careless. They were being efficient. The policy change is straightforward. The behavioural change requires explanation, context, and training — ideally with real examples, like Samsung, that make the stakes concrete.
How BBS helps with this
- Shadow AI Discovery — Audit which tools your staff are using and whether they are on consumer or enterprise accounts with training disabled. Most businesses find tools in use that IT has no visibility over.
- Vendor Configuration Review — Verify training opt-out settings across your entire AI tool stack and fix them. We check not just whether the option exists, but whether it has actually been enabled on your accounts.
- AI Acceptable Use Policy — Mandate enterprise accounts and correct data handling as a condition of AI tool use. Staff get clear, practical rules — not a policy document that lives in a shared drive.
- Staff Awareness Training — Explain what training data risk means in practice, using real examples like Samsung, so that staff understand the why behind the policy — not just the what.