Your AI Privacy Guide

What you need to know before sharing sensitive data with ChatGPT, Gemini, Claude, or any AI tool

AI Privacy Guide for Professionals

Why This Guide Matters

Artificial intelligence tools like ChatGPT, Claude, Gemini, and Perplexity have become everyday productivity tools. But behind their convenience is a critical consideration: privacy. Whether you're using AI to draft emails, summarize documents, or generate ideas, it's important to understand how your data is being collected, used, and stored.

This guide offers a straightforward breakdown of what major AI platforms do with your data, how long they keep it, who may access it, and what settings you can control. It also includes recent legal developments (like the New York Times v. OpenAI lawsuit) and provides direct links and step-by-step instructions for updating your privacy settings.

If you're using AI tools in a professional context, especially if you're working with client information or sensitive content (e.g., health records, financials, client records, legal documents, or internal strategy materials), this guide is essential reading. Always check with your employer before using AI tools on the job as many organizations have internal policies on acceptable use.

What to Look For When Evaluating AI Privacy

When evaluating any AI tool, particularly for professional or organizational use, consider the following questions:

  • What types of data are collected? Are prompts, files, metadata, or connected app information included?

  • Is the data used for model training? Is this enabled by default, or only with consent?

  • What is the data retention policy? How long is your data stored, and can you delete it?

  • Who has access? Are internal teams, third-party vendors, or reviewers involved?

  • What control do you have? Can you disable training, delete history, or restrict integrations?

Start by reviewing the tool’s Privacy Policy and Terms of Service, typically linked in the website footer. Then navigate to your account or workspace settings under "Privacy," "Data Usage," or "Security" to assess your options.

Pay special attention to tools that request access to third-party services such as Gmail, Google Calendar, Outlook, Slack, Dropbox, or Notion. Connecting these services may give the AI access to:

  • Email content, meeting schedules, attachments, or shared documents

  • Historical communications and metadata

  • Internal data tied to your professional accounts

Before granting access:

  • Carefully review what permissions are being requested during integration

  • Confirm whether your connected data could be used to improve the model

  • Disconnect services that aren’t essential to your workflow

  • Ensure that you’re complying with your organization’s data use policies

When in doubt, opt for the most privacy-conscious settings, and favor enterprise-level services for handling any sensitive or regulated information

As part of ongoing litigation with The New York Times, OpenAI has been ordered to retain user conversations, including deleted chats, for potential legal discovery. This impacts all Free, Plus, and Team users of ChatGPT. Users on Enterprise plans or using API endpoints with "zero data retention" are not affected.

This development underscores the importance of understanding what happens to your data, even after you think you've deleted it.

AI Privacy Matrix (with Step-by-Step Opt-Out Instructions)

Below is a detailed comparison of how leading AI platforms handle your data, including what they collect, whether your input is used to train their models, how long it's stored, and how you can manage or opt out of data sharing. Use this as a quick reference to understand your level of control and risk across each platform.

Platform

What Data They Collect

Used to Train the AI?

How Long It's Kept

Who Can See It

Privacy Controls (Step-by-Step)

OpenAI (ChatGPT)

Prompts, uploads, metadata

Yes, unless you opt out. Enterprise/API with zero data retention: No. Source

Deleted chats normally held ~30 days. Due to NYT lawsuit, OpenAI is currently required to retain all data. Source

OpenAI staff, legal, support teams

To turn off training use:1. Go to ChatGPT settings2. Click your name → Settings → Data Controls3. Toggle off “Improve the model with your data”To use Temporary Chats:1. Same menu → Toggle on “Temporary Chat”

Google Gemini

Prompts, files, metadata, Gmail/Calendar data (if connected)

Yes for free users. No for Cloud customers. Source

Chat history stored for 72 hours; Activity data may persist up to 18 months unless turned off. Source

Google staff, reviewers (for training)

To turn off Gemini activity:1. Visit Gemini Activity Settings2. Click “Turn off” under Gemini ActivityTo avoid training entirely:Use Vertex AI (Cloud) instead of free Gemini

Anthropic Claude

Prompts, usage data, metadata

No, unless you opt in by submitting feedback. Source

History cleared immediately, full deletion within 30 days unless content is flagged. Source

Minimal staff with encrypted access

Nothing to toggle unless you submit feedback1. Avoid clicking thumbs-up/down icons2. Delete history via sidebar (bin/trash icon)3. Read Claude’s Privacy Policy for updates

Meta AI / Llama

Public posts, chat inputs via Meta apps

Yes for public posts unless you opt out (EU only). Private chats not used. Source

Retained per Meta policy; not end-to-end encrypted. Source

Meta engineers, product teams

To opt out (EU users only):1. Go to Meta AI Settings2. Log in and scroll to “Your Rights” section3. Toggle “Don’t use my content for training”

Perplexity AI

Prompts, file uploads, usage data

Yes by default on free plans. No on Enterprise plans. Source

Uploaded files auto-delete after 7 days. Threads retained until you delete. Source

Staff under strict internal controls; GDPR-compliant

To turn off training use:1. Visit Perplexity Settings2. Scroll to “AI Data Usage”3. Toggle OFF the switchEnterprise accounts never use data for training

Final Recommendations

  • Avoid pasting sensitive or confidential data into AI tools

  • Disable training and activity tracking when possible

  • Prefer enterprise-grade plans for stronger data protections

  • Regularly review and delete your chat histories (though with ongoing litigation, this may become irrelevant)

  • Confirm your organization’s policy before using AI tools professionally