Back to blog

Article

Which AI Tools Are Your Employees Using? How to Find Out

A practical discovery framework for SMB teams to identify real AI tool usage, reduce shadow AI blind spots, and build a defensible governance baseline.

3 min readBy Varentus Team

If you cannot answer which AI tools your employees are using, you do not have AI governance.

You have assumptions.

Most SMB leaders underestimate how quickly AI adoption spreads inside their organization.

Marketing experiments with copy tools.
Sales tests summarization assistants.
Developers use AI copilots.
Operations explores workflow automation.

None of this requires formal approval.

And that is the problem.


You cannot govern what you cannot see.


Why AI tool discovery is the first control

Before you write policy updates.
Before you build an approved tools list.
Before you evaluate vendors.

You need visibility.

AI tool discovery reduces:

  • Shadow AI risk
  • Data exposure blind spots
  • Vendor review surprises
  • Incident investigation scope

It is the foundation of enforceable governance.


Step 1: Ask directly — but structure it

Employee declarations are the simplest starting point.

Send a structured request:

  • What AI tools are you using?
  • For what workflows?
  • Are you using personal or company accounts?
  • What types of data are entered?

Make it clear that the objective is governance, not punishment.

If employees believe discovery equals enforcement crackdowns, they will underreport.

Transparency drives accuracy.


Step 2: Review expense and SaaS records

Shadow AI frequently appears in:

  • Corporate card statements
  • Reimbursement reports
  • SaaS management platforms
  • Browser extension marketplaces

Look for:

  • AI subscription charges
  • API usage fees
  • Unknown vendor domains

This step often reveals tools leadership did not realize were in use.


Step 3: Analyze workspace and SSO logs

If your company uses:

  • Google Workspace
  • Microsoft 365
  • Okta
  • Azure AD

Review:

  • Connected third-party apps
  • OAuth integrations
  • External application logins

SSO data is often the most reliable visibility source for AI tool discovery.


Step 4: Review developer environments

Engineering teams frequently adopt AI tools independently.

Audit:

  • IDE plugins
  • Code assistant integrations
  • API keys embedded in repositories
  • External AI service calls

Proprietary code exposure is one of the highest-risk shadow AI patterns.

If you want to understand the downstream cost of invisible AI usage, review the breakdown in Shadow AI breach costs and prevention.


Step 5: Cross-reference against data sensitivity

Once you identify tools, classify them:

  • Approved
  • Restricted
  • Unapproved

Then map each tool against:

  • Customer data exposure
  • Financial model access
  • PHI handling
  • Source code interaction

Discovery without classification does not reduce risk.

Classification enables control.

If you have not yet built a structured approval framework, use the methodology outlined in How to create an AI-approved tools list.


The common mistake: stopping at visibility

Many teams perform one discovery exercise and assume the problem is solved.

AI adoption evolves monthly.

New tools appear.
Terms change.
Integrations expand.

Discovery must become recurring.

Quarterly review is a reasonable baseline for SMB teams.


How to turn discovery into governance

Discovery is only the first step.

After identifying tools:

  1. Publish or update your AI usage policy.
  2. Formalize an approved tools list.
  3. Require enterprise accounts.
  4. Define restricted data categories.
  5. Track employee acknowledgement.

If you need a fast starting baseline, generate one using the free AI policy generator, then pressure-test it using the Shadow AI risk guide.

Governance becomes credible when visibility connects to enforcement.


Why this matters commercially

When customers ask how you govern AI usage, you should not respond with:

“We think we know what tools are in use.”

You should respond with:

  • Documented discovery process
  • Approved tools list
  • Vendor review criteria
  • Attestation tracking
  • Review cadence

Visibility signals maturity.

Guesswork signals exposure.


Bottom line

Which AI tools are your employees using?

If the answer is uncertain, governance is incomplete.

Discovery is not optional.

It is the first control.

Build visibility. Classify tools. Enforce boundaries.

That is how informal experimentation becomes governed adoption.