What is AI encryption (at rest, in transit, in use)? Why it matters for your business

A person at a desk reading a printed document with an open laptop next to them
TL;DR

Encryption operates across three states: at rest, in transit, and in use. The first two are mature and almost every AI vendor offers them. The third is the AI-specific gap: when you send a prompt to a hosted AI service, the prompt is decrypted at the inference server so the model can read it. Confidential computing on AWS, Azure, Google Cloud, and Apple's Private Cloud Compute closes that gap, but you have to ask for it by name.

Key takeaways

- Encryption operates across three states. At rest is data on a disk. In transit is data on a network. In use is data being actively processed by a model. - "Encrypted at rest and in transit" is true of almost every AI vendor and is no longer a meaningful security claim on its own. - The AI-specific gap is encryption in use. During inference, the prompt is decrypted on the vendor's infrastructure so the model can run on it. - Confidential computing closes the gap. AWS Nitro Enclaves, Azure Confidential Computing, Google Cloud Confidential VMs, and Apple's Private Cloud Compute are all production-ready in 2026, at roughly a 20-30% compute premium. - For NHS, FCA-regulated, or any sensitive data, ask vendors specifically about encryption in use and customer-managed keys, not the at-rest and in-transit headline.

A client sent through a security questionnaire ahead of a contract renewal. One clause asked whether any third-party processor handling the firm’s data uses encryption at rest and in transit. He ticked the box, because every AI tool in the stack does. Then he read the next clause, which asked whether prompts and responses are exposed in plaintext during inference. He did not know. He went to the AI vendor’s documentation. Encryption at rest was explained in detail. Encryption in transit was explained in detail. Encryption in use was not mentioned, because in that product it does not exist.

That gap is the reason a clause that read as complete five years ago for a CRM is no longer the same promise for an AI service today. Encryption operates across three states, and AI workloads sit differently from any other software you procure on the third.

What is AI encryption?

Encryption operates across three states for any AI service. At rest is data sitting on a disk or in a backup. In transit is data moving across a network between your laptop and a vendor’s server. In use is data being actively processed, which for AI means a model running on it. The first two are mature engineering. The third is where AI workloads break the assumptions every other software contract was written under.

The mechanics are simple. When you send a prompt to a hosted AI service, the prompt travels over TLS 1.3, the standard in-transit protocol. The vendor’s edge server decrypts the prompt, logs it, and forwards it to a GPU inference server in plaintext, because the model has to read the words to generate a response. The response is logged, encrypted again, and returned to you. From the edge to the log, your data is plaintext on the vendor’s systems. That window is structural, not a configuration mistake.

Why does it matter for your business?

UK regulators treat encryption as the baseline of data protection, and the AI workload pattern leaves a hole in the middle of that baseline. Article 32 of UK GDPR requires appropriate technical measures including encryption. NHS DSPT requires customer-managed keys. The FCA expects financial services firms to control the encryption of regulated data. Cyber insurance underwriters now ask about AI tool encryption at renewal. The clauses you tick are read against this floor.

The practical consequence for an owner-managed firm is that the security questionnaires from your clients have grown a new section. A legal practice sending case material into a generative AI tool, a healthcare consultancy sending de-identified patient information, a financial firm sending portfolio data, a recruiter sending CVs, all of these are exposing that data to a vendor’s plaintext logs. This is the documented behaviour of the tool, sitting behind a clause that reads as if the data were protected throughout.

The third reason it matters is that it shapes which products you can defensibly use. For high-classification data, the answer is often a different tier of the same product. Microsoft’s standard ChatGPT API and Azure OpenAI Service can serve the same underlying model, but Azure offers confidential computing as an enterprise option that the standard API does not. The variable that matters is the encryption posture of the contract, not the model behind it.

Where will you actually meet it?

You will meet it in vendor security pages first. A page that lists “AES-256 encryption at rest” and “TLS 1.3 in transit” in detail, then says nothing about inference processing, is the standard pattern in 2026. The omission is rarely deceptive. The vendor is honestly describing what it does. Encryption in use is a separate technical capability that consumer and standard commercial AI APIs do not yet offer by default.

You will meet it in your own client questionnaires next. A growing share of mid-market and enterprise procurement forms ask explicitly whether prompts and responses are exposed in plaintext during inference, whether the customer can manage the encryption key, and whether the vendor publishes a cryptographic attestation report. If you cannot answer those questions for the AI tools you use, the box is unticked and the contract stalls.

Then you will meet it in the cloud-provider product pages, where the answer lives. AWS Nitro Enclaves, Azure Confidential Computing, and Google Cloud Confidential VMs are all production-ready in 2026. NVIDIA H100 and H200 GPUs support confidential computing, which is the layer that matters for large model inference. The cost is roughly a 20 to 30 percent compute premium. Apple’s Private Cloud Compute, announced in June 2024, is the cleanest verifiable example: Apple publishes the architecture, third-party researchers can audit it, and a cryptographic attestation lets a customer check the enclave has not been tampered with. Few commercial AI providers publish anything equivalent, which means the security claim cannot be independently checked.

When to ask about it, when to ignore it

Ask about it when the data going into the prompt would matter if it leaked. Client identifiers, health information, financial details, legal case material, employee records, internal commercial documents, anything covered by a confidentiality agreement or a regulatory regime. For these, encryption posture is part of the procurement decision, and the questions you send the vendor need to be specific enough that they cannot be answered with marketing copy.

Five questions earn their place on the form. Is data encrypted at rest with AES-256, and can we manage the keys ourselves. Is TLS 1.3 enforced on every connection. During inference, are prompts and responses exposed in plaintext or protected by confidential computing. Can you provide a cryptographic attestation report or third-party security audit. What is the retention period for prompt and response logs, and where do they sit. A vendor who can answer these directly is in a different operational class from one who replies with “enterprise-grade security”.

Hold the question lightly when the prompt content would not matter if it were public. A general writing assistant drafting a press release, a meeting summariser running over a recording you would happily publish, a code completion tool reading a non-confidential repository. The encryption-in-use gap exists in those services too, but the practical exposure is low. Spending procurement effort on closing it is a misallocation. The discriminator is the data classification of what you are about to put into the prompt.

For Article 32 risk assessments, NHS DSPT submissions, FCA outsourcing notifications, EU AI Act Article 15 compliance, or any specific legal question, route through the NCSC, the ICO, your sector regulator, and a qualified cyber consultancy. This piece is the question framing, not the legal answer.

Confidential computing is the production-ready technology that closes the encryption-in-use gap. Code and data run inside a hardware-isolated execution environment on the CPU or GPU, where neither the operating system, the hypervisor, nor an administrator with root access can read the contents. Intel calls its version Software Guard Extensions, AMD calls it Secure Encrypted Virtualisation, and ARM calls it Confidential Computing Architecture. For AI workloads in 2026, this is the answer.

Hardware security modules, HSMs, are tamper-resistant devices that store cryptographic keys and perform encryption operations without ever exposing the key in plaintext. When you see BYOK in a contract, the key is being held in an HSM the vendor staff cannot extract from.

Attestation is the cryptographic proof that a piece of code is running inside a confidential enclave and has not been modified. The enclave generates a signed report, and a customer or auditor can verify it independently. Apple Private Cloud Compute publishes attestation reports for its inference servers, which is why it works as the reference example.

Homomorphic encryption and secure multi-party computation are research-frontier techniques that allow computation on encrypted data without decryption at all. They appear in vendor literature and procurement documents. For mainstream AI workloads in 2026, confidential computing is the practical path; the other two are worth knowing as terms, not deploying as solutions.

The Storm-0558 incident in July 2023, when a nation-state actor obtained a Microsoft consumer signing key and used it to forge access tokens for several US government Outlook accounts, is the standing reminder that key custody is not theoretical. The cryptography held; the key did not. Every encryption claim in your AI vendor stack rests on the same assumption, that the key is held somewhere your data is not. Worth checking before you tick the box.

Sources

Information Commissioner's Office (2024). Article 32 Security of Processing under UK GDPR. The regulatory anchor for encryption obligations on personal data. https://ico.org.uk/for-organisations/uk-gdpr-guidance-and-resources/security/a-guide-to-data-security/ National Cyber Security Centre (2024). Cryptography guidance and approved algorithms. The UK government baseline for AES-256 at rest and TLS 1.3 in transit. https://www.ncsc.gov.uk/collection/cryptography National Cyber Security Centre (2024). Cloud Security Principles, including customer key management. The framework underpinning UK cloud procurement decisions. https://www.ncsc.gov.uk/collection/cloud Apple (2024). Private Cloud Compute architecture and security overview. The cleanest verifiable example of confidential AI inference with public attestation. https://www.apple.com/privacy/private-cloud-compute/ Microsoft (2025). Microsoft Mitigates China-Based Threat Actor Storm-0558 Targeting of Customer Email. Canonical 2023 incident showing why key custody matters even when the cryptography is sound. https://msrc.microsoft.com/blog/2023/07/microsoft-mitigates-china-based-threat-actor-storm-0558-targeting-of-customer-email/ AWS (2025). AWS Nitro Enclaves: isolated compute environments for confidential workloads. The production-ready confidential computing offering in AWS. https://aws.amazon.com/enclaves/ Microsoft (2025). Azure Confidential Computing documentation. Production-ready confidential computing on Intel SGX, AMD SEV-SNP, and ARM CCA. https://learn.microsoft.com/en-us/azure/confidential-computing/ Google Cloud (2025). Confidential Computing on Google Cloud. Confidential VMs and Confidential GKE clusters using AMD SEV-SNP. https://cloud.google.com/confidential-computing NHS England (2024). Data Security and Protection Toolkit. The mandatory certification scheme for organisations processing NHS data, including encryption requirements. https://www.dsptoolkit.nhs.uk/ EUR-Lex (2024). Regulation (EU) 2024/1689 on artificial intelligence (EU AI Act), Article 15 cybersecurity requirements. Relevant to UK firms supplying AI services into EU markets. https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:32024R1689

Frequently asked questions

Is "end-to-end encryption" the same as encryption in use?

No. End-to-end encryption protects data on the wire and at rest, where only the sender and receiver hold the keys. A messaging app like Signal works this way. A hosted AI API does not. The vendor's server has to read your prompt to run the model on it, which means decrypting it at the edge. Encryption in use is the separate, harder problem of keeping data encrypted even while the model is processing it.

Do I actually need encryption in use, or is at rest and in transit enough?

For a marketing newsletter draft, at rest and in transit is fine. For client data covered by UK GDPR, NHS DSPT, FCA outsourcing rules, or professional indemnity confidentiality, the answer changes. Encryption in use is increasingly expected for sensitive data, and cyber insurance underwriters now ask about it directly. The right test is what data you are putting into the prompt, not the AI tool's general security posture.

What does "BYOK" mean and why does it matter?

BYOK is Bring Your Own Key. It means you generate the encryption key for your data and the vendor stores it in a hardware security module they cannot extract it from. The alternative, vendor-managed keys, leaves the vendor in control of the key and able to decrypt your data if asked. For NHS data, regulated financial data, or anything covered by a strict data processing agreement, BYOK is now the expected baseline rather than a premium feature.

This post is general information and education only, not legal, regulatory, financial, or other professional advice. Regulations evolve, fee benchmarks shift, and every situation is different, so please take qualified professional advice before acting on anything you read here. See the Terms of Use for the full position.

Ready to talk it through?

Book a free 30 minute conversation. No pitch, no pressure, just a useful chat about where AI fits in your business.

Book a conversation

Related reading

If any of this sounds familiar, let's talk.

The next step is a conversation. No pitch, no pressure. Just an honest discussion about where you are and whether I can help.

Book a conversation