How to use AI whilst keeping your Data Private and Safe

AI can pay off quickly—copilots that accelerate knowledge work, smarter customer operations, and faster software delivery. The risk is not AI itself; it is how you handle data. Look at privacy (what you expose), security (who can access), compliance (what you can prove), and sovereignty (where processing happens) as separate lenses. The playbook is simple: classify the data you’ll touch; choose one of four deployment models; apply a few guardrails—identity, logging, and simple rules people understand; then measure value and incidents. Start “as open as safely possible” with the less sensitive cases for speed, and move to tighter control as sensitivity increases.


What “Private & Safe” actually means

Private and safe AI means using the least amount of sensitive information, tightly controlling who and what AI can access, proving that your handling meets legal and industry obligations, and ensuring processing happens in approved locations. In practice you minimise exposure, authenticate users, encrypt and log activity, and keep a clear record of decisions and data flows so auditors and customers can trust the outcome.

To make this work across the enterprise, bring the right people together around each use case. The CIO and CISO own the platform choices and controls; the CDO curates which data sources are approved; Legal sets lawful use and documentation; business owners define value and success; HR and Works Council get involved where employee data or work patterns change. Run a short, repeatable intake: describe the use case, identify the data, select the deployment model, confirm the controls, and agree how quality and incidents will be monitored.


How to classify “Sensitive Data” – a simple four-tier guide

Not all data is equal. Classifying it upfront tells you how careful you need to be and which setup to use.

Tier 1 – Low sensitivity. Think public information or generic content such as first drafts of marketing copy. Treat this as the training ground for speed: use packaged tools, keep records of usage, and avoid connecting unnecessary internal sources.

Decision check: “Could this appear on our website tomorrow?”Yes = Tier 1

Tier 2 – Internal. Everyday company knowledge—policy summaries, project notes, internal wikis. Allow AI to read from approved internal sources, but restrict access to teams who need it and retain basic logs so you can review what was asked and answered.

Decision check: “Would sharing this externally require approval?”Yes = Tier 2+

Tier 3 – Confidential. Material that would harm you or your customers if leaked—client lists, pricing models, source code. Use controlled company services that you manage, limit which repositories can be searched, keep detailed activity records, and review results for quality and leakage before scaling.

Decision check: “Would leakage breach a contract or NDA?”Yes = Tier 3+

Tier 4 – Restricted or regulated. Legally protected or mission-critical information—patient or financial records, trade secrets, M&A. Run in tightly controlled environments you operate, separate this work from general productivity tools, test thoroughly before go-live, and document decisions for auditors and boards.

Decision check: “Is this regulated or business-critical?”Yes = Tier 4


Common mistakes – and how to fix them

Using personal AI accounts with company data.
This bypasses your protections and creates invisible risk. Make it company accounts only, block personal tools on the network, and provide approved alternatives that people actually want to use.

Assuming “enterprise tier” means safe by default.
Labels vary and settings differ by vendor. Ask for clear terms: your questions and documents are not used to improve public systems, processing locations are under your control, and retention of queries and answers is off unless you choose otherwise.

Building clever assistants without seeing what actually flows.
Teams connect documents and systems, then no one reviews which questions, files, or outputs move through the pipeline. Turn on logging, review usage, and allow only a short list of approved data connections.

Skipping basic training and a simple policy.
People guess what’s allowed, leading to inconsistent—and risky—behaviour. Publish a one-page “how we use AI here,” include it in onboarding, and name owners who check usage and costs.


AI Deployment Models

Model 1 — Secure packaged tools (fastest path to value).
Ready-made apps with business controls—ideal for broad productivity on low-to-moderate sensitivity work such as drafting, summarising, meeting notes, and internal Q&A. Examples: Microsoft Copilot for Microsoft 365, Google Workspace Gemini, Notion AI, Salesforce Einstein Copilot, ServiceNow Now Assist. Use this when speed matters and the content is not highly sensitive; step up to other models for regulated data or deeper system connections.

Model 2 — Enterprise AI services from major providers.
You access powerful models through your company’s account; your inputs aren’t used to train public systems and you can choose where processing happens. Well-suited to building your own assistants and workflows that read approved internal data. Examples: Azure OpenAI, AWS Bedrock, Google Vertex AI, OpenAI Enterprise, Anthropic for Business. Choose this for flexibility without running the underlying software yourself; consider Model 3 if you need stronger control and detailed records.

Model 3 — Managed models running inside your cloud.
The models and search components run within your own cloud environment, giving you stronger control and visibility while the vendor still manages the runtime. A good fit for confidential or regulated work where oversight and location matter. Examples: Bedrock in your AWS account, Vertex AI in your Google Cloud Platform, Azure OpenAI in your subscription, Databricks Mosaic AI, Snowflake Cortex. Use this when you need enterprise-grade control with fewer operational burdens than full self-hosting.

Model 4 — Self-hosted and open-source models.
You operate the models yourself—on-premises or in your cloud. This gives maximum control and sovereignty, at the cost of more engineering, monitoring, and testing. Suits the most sensitive use cases or IP-heavy R&D. Examples: Llama, Mistral, DBRX—supported by platforms such as Databricks, Nvidia NIM, VMware Private AI, Hugging Face, and Red Hat OpenShift AI. Use this when the business case and risk profile justify the investment and you have the talent to run it safely.


Building Blocks and How to Implement (by company size)

Essential Building blocks

A few building blocks change outcomes more than anything else. Connect AI to approved data sources through a standard “search-then-answer” approach—often called Retrieval-Augmented Generation (RAG), where the AI first looks up facts in your trusted sources and only then drafts a response.

This reduces the need to copy data into the AI system and keeps authority with your original records. Add a simple filter to remove personal or secret information before questions are sent. Control access with single sign-on and clear roles. Record questions and answers so you can review quality, fix issues, and evidence compliance. Choose processing regions deliberately and, where possible, manage your own encryption keys. Keep costs in check with team budgets and a monthly review of usage and benefits.

Large enterprises

Move fastest with a dual approach. Enable packaged tools for day-to-day productivity, and create a central runway based on enterprise AI services for most custom assistants. For sensitive domains, provide managed environments inside your cloud with the standard connection pattern, built-in filtering, and ready-made quality tests. Reserve full self-hosting for the few cases that genuinely need it. Success looks like rapid adoption, measurable improvements in time or quality, and no data-handling incidents.

Mid-market organisations

Get leverage by standardising on one enterprise AI service from their primary cloud, while selectively enabling packaged tools where they clearly save time. Offer a single reusable pattern for connecting to internal data, with logging and simple redaction built in. Keep governance light: a short policy, a quarterly review of model quality and costs, and a named owner for each assistant.

Small-Mid sized companies

Should keep it simple. Use packaged tools for daily work and a single enterprise AI service for tasks that need internal data. Turn off retention of questions and answers where available, restrict connections to a small list of approved sources, and keep work inside the company account—no personal tools or copying content out. A one-page “how we use AI here,” plus a monthly check of usage and spend, is usually enough.


What success looks like

Within 90 days, 20–40% of knowledge workers are using AI for routine tasks. Teams report time saved or quality improved on specific workflows. You have zero data-handling incidents and can show auditors your data flows, access controls, and review process. Usage and costs are tracked monthly, and you’ve refined your approved-tools list based on what actually gets adopted.

You don’t need a bespoke platform or a 200-page policy to use AI safely. You need clear choices, a short playbook, and the discipline to apply it.

What I Learned from Google & Kaggle’s Generative AI Intensive Course

Last week, I joined over 100,000 participants in a 5-day Generative AI Intensive Course hosted by Google and Kaggle—a free and fast-paced program designed to equip professionals with practical knowledge on how to harness the power of GenAI in real-world settings.

Why did I join? Because GenAI is no longer a concept—it’s here, and it’s evolving faster than most organizations can absorb. As leaders in digital transformation, we can’t afford to wait. We need to understand not just the what, but also the how of these technologies.

This course offered an excellent foundation of the current status of GenAI technologies, how they can be applied today, and even provided glimpses into where they are likely to evolve next.

Below is a summary of the course—structured for executives and transformation leaders seeking clarity on how GenAI will impact their businesses.


Day 1: Foundational Large Language Models & Text Generation

Why it matters: Understanding the fundamentals is critical before scaling GenAI use cases. Day one unpacked the Transformer architecture, the core engine behind tools like ChatGPT and Gemini.

Key Takeaways:

  • LLMs are the brains behind GenAI—they interpret and generate human-like language at scale.
  • Transformer models help these systems understand context and nuance.
  • Fine-tuning allows you to adapt general models to business-specific tasks, such as customer service or marketing.

Google whitepaper: “Foundational Large Language Models & Text Generation”


Day 2: Embeddings and Vector Stores

Why it matters: Without intelligent data structuring, GenAI becomes just another flashy tool. This session focused on how to make AI actually useful inside your organization.

Key Takeaways:

  • Embeddings turn complex data into searchable formats.
  • Vector stores make this information retrievable at speed and scale.
  • Retrieval-Augmented Generation (RAG) combines LLMs with your proprietary data for smarter, context-rich answers.

Google whitepaper: “Embeddings & Vector Stores”


Day 3: Generative AI Agents

Why it matters: GenAI is moving beyond chatbots—into agents that can autonomously perform tasks, interact with systems, and even make decisions.

Key Takeaways:

  • AI agents integrate tools, logic, and memory to act independently.
  • Platforms like LangChain and Vertex AI Agents provide orchestration layers for real-world applications.
  • Think of these as junior digital employees—capable of assisting operations, support, or analysis at scale.

Google whitepapers: “Agents” and “Agents Companion”


Day 4: Solving Domain-Specific Problems Using LLMs

Why it matters: Generic models only take you so far. Tailoring AI to your industry delivers far more strategic value.

Key Takeaways:

  • Domain-specific LLMs adapt to unique challenges in sectors like healthcare and cybersecurity.
  • SecLM enhances threat detection and response capabilities in cybersecurity.
  • MedLM supports clinical workflows and patient information retrieval in healthcare.

Google whitepaper: “Solving Domain-Specific Problems Using LLMs”


Day 5: Operationalizing GenAI on Vertex AI with MLOps

Why it matters: Scaling GenAI requires more than a good prompt—it demands structured deployment, governance, and monitoring.

Key Takeaways:

  • MLOps for GenAI adapts best practices from machine learning to this new frontier of GenAI applications.
  • Understanding the GenAI lifecycle—from experimentation to production—is key to long-term success.
  • Platforms like Vertex AI help organizations deploy and manage GenAI responsibly and at scale.

Google whitepaper: “Operationalizing Generative AI on Vertex AI using MLOps”


My Reflections

This course reinforced a simple truth: GenAI is becoming more capable rapidly. And like any capability, it needs strategy, structure, and experimentation to create real business value.

If you’re in a leadership role, here are three questions to reflect on:

  1. Where can GenAI complement or augment your current operations?
  2. Do you have the data foundation to make it effective?
  3. Are you equipping your teams to experiment safely and learn quickly?

What’s Next

I’ll be diving deeper into some of these topics in future articles —especially GenAI agents and domain-specific applications.

Let’s continue learning and leading—together