Key Takeaways
- Estimates suggest 50-70% of AI usage in organizations occurs through unsanctioned tools.
- The goal is not to eliminate shadow AI but to channel it into governed, secure pathways that let teams innovate.
- A four-phase framework — Discovery, Policy, Monitoring, Protection — gives IT leaders a practical roadmap for bringing unsanctioned AI under control.
Shadow AI is multiplying across enterprises as employees experiment with unsanctioned chatbots, automation apps and model integrations. Instead of fighting this trend, IT leaders can bring it under control by setting clear governance policies.
Industry experts point to a phased approach — starting with discovery and ending with real-time protection — as the most practical path forward.
Table of Contents
- What Is Shadow AI?
- How Widespread Is the Problem?
- Why Shadow AI Proliferates
- A Governance Framework: From Discovery to Protection
- The Phased Approach
What Is Shadow AI?
Shadow AI refers to the use of unsanctioned AI tools by employees within an organization — everything from chatbots to custom model integrations and code assistants adopted without IT approval. It's especially prevalent in organizations with decentralized innovation cultures or limited AI governance.
This unsanctioned AI use is typically well-meaning but unwise: employees using AI tools like ChatGPT, Claude or equivalents to summarize organizational documents, analyze them for common themes, flag inconsistencies or get quick guidance on a topic.
“The biggest risk is actually in this very benign, though misguided, usage,” said Grace Trinidad, research director in IDC's Security & Trust research practice.
How Widespread Is the Problem?
The scale of shadow AI is significant. According to IDC estimates, this issue within organizations ranges from 50% of employees using unsanctioned tools to 70% of all AI usage occurring in unsanctioned tools. "But without an AI governance solution in place, the true rate of unsanctioned AI usage, or shadow AI, is unknown," noted Trinidad.
The core risk is that employees may be uploading sensitive or private information to GenAI platforms, inadvertently or carelessly sharing privileged information with external services.
“It's pretty much guaranteed that shadow AI is happening in your organization,” said Trinidad.
Why Shadow AI Proliferates
Shadow AI thrives in environments where governance structures are absent or insufficient. Jayesh Chaurasia, senior analyst at Forrester, identified the root causes as a lack of:
- AI usage policies
- Inventory tracking
- Simple risk assessment workflows
Chaurasia also pointed to a critical blind spot: even when technical controls are in place, they're ineffective when employees use AI tools without the company's knowledge or use personal devices or tools to do office work when the office doesn't provide them.
In other words, when employees feel blocked from using AI at work, they often find workarounds, making the problem harder to detect and manage.
Related Article: Best Free AI Tools: Top Picks for Writing, Design, Productivity and More
A Governance Framework: From Discovery to Protection
Rather than fighting the trend, IT leaders can bring shadow AI under control through a structured governance approach. The goal is not to shut down AI experimentation — it's to channel it safely into the enterprise where it can drive productivity without compromising security.
Phase 1: Discovery
The first step is understanding what AI tools are already in use across the organization. AI governance and discovery tools enable organizations to see which GenAI resources employees are accessing.
“For many organizations, the first step is discovering which AI platforms are in use without enterprise restriction, also a useful way to guide organizational investment in AI products,” Trinidad said.
Once discovered, organizations can evaluate and whitelist AI products that meet governance standards. Many AI governance platforms will:
- Help identify tools that are transparent in model versioning and lineage
- Have drift monitoring in place
- Include data privacy and handling controls
- Prevent harmful or policy-violating outputs
“This helps to build confidence in the AI products that are brought into the organization,” Trinidad noted.
Phase 2: Policy and Governance Structure
With a clear picture of existing AI usage, organizations should establish formal governance. Chaurasia recommended five core components:
- Define clear usage policies: Establish what is and is not acceptable AI use, including which data types can be shared with AI tools.
- Establish an AI intake and review process: Create a formal pathway for employees to request and get approval for new AI tools.
- Create a federated governance model: Distribute governance responsibility across business units rather than centralizing all decisions in IT.
- Promote transparency through documentation: Maintain clear records of AI tools in use, their purposes and their risk profiles.
- Enable responsible experimentation: Provide approved channels for employees to explore AI, reducing the incentive to go rogue.
“A balanced AI governance framework should define clear usage policies, establish an AI intake and review process, create a federated governance model, and promote transparency through proper documentation,” said Chaurasia.
Phase 3: Monitoring and Measurement
Ongoing governance requires continuous measurement and real-time monitoring. Trinidad identified four key dimensions organizations must track:
| Dimension | What to Track |
|---|---|
| Volume of Usage | How much AI is being used across the organization and by whom |
| Data Types Shared | What categories of data (sensitive, public, internal) are being sent to AI tools |
| Models in Use | Which AI models and platforms employees are accessing |
| Tool Access Patterns | What external tools and integrations the AI platforms are connecting to |
“Measurement is ongoing quantification and characterization of all usage occurring in the organization,” Trinidad explained. “Monitoring or tracking all behavior in real time against organizational policies is also required.”
Phase 4: Protection — Securing Inputs and Outputs
The final layer of the framework focuses on actively protecting organizational data as it flows through AI tools. According to Trinidad, this means enforcing guardrails on what data can be shared, blocking risky actions before they happen and flagging potential issues for human follow-up.
However, none of these protections work without a strong Identity and Access Management (IAM) foundation.
Key protective measures include:
- Regular audits of AI tool usage and data flows
- Secure APIs to control how AI tools connect to enterprise systems
- Role-based permissions to limit access based on employee function and clearance
- Guardrails on inputs to prevent sensitive data from leaving the organization
- Output monitoring to flag or block harmful, non-compliant or policy-violating responses
Related Article: The Quiet Growth Engine: How Modern Data & AI Governance Unlocks Value Potential
The Phased Approach
The path from unmanaged shadow AI to responsible enterprise AI adoption follows a clear progression:
| Phase | Focus | Key Actions |
|---|---|---|
| 1 | Discovery | Identify all AI tools in use; assess which meet governance standards |
| 2 | Policy | Define usage policies, intake process and federated governance model |
| 3 | Monitoring | Track usage volume, data types, models and tool access in real time |
| 4 | Protection | Enforce IAM, guardrails, audits, secure APIs and role-based permissions |
Remember, the goal is not to eliminate shadow AI, but to channel it into governed, secure pathways that allow teams to innovate while protecting organizational data and compliance.