Data Privacy With Generative AI

— Written By
en Español / em Português
Español

El inglés es el idioma de control de esta página. En la medida en que haya algún conflicto entre la traducción al inglés y la traducción, el inglés prevalece.

Al hacer clic en el enlace de traducción se activa un servicio de traducción gratuito para convertir la página al español. Al igual que con cualquier traducción por Internet, la conversión no es sensible al contexto y puede que no traduzca el texto en su significado original. NC State Extension no garantiza la exactitud del texto traducido. Por favor, tenga en cuenta que algunas aplicaciones y/o servicios pueden no funcionar como se espera cuando se traducen.


Português

Inglês é o idioma de controle desta página. Na medida que haja algum conflito entre o texto original em Inglês e a tradução, o Inglês prevalece.

Ao clicar no link de tradução, um serviço gratuito de tradução será ativado para converter a página para o Português. Como em qualquer tradução pela internet, a conversão não é sensivel ao contexto e pode não ocorrer a tradução para o significado orginal. O serviço de Extensão da Carolina do Norte (NC State Extension) não garante a exatidão do texto traduzido. Por favor, observe que algumas funções ou serviços podem não funcionar como esperado após a tradução.


English

English is the controlling language of this page. To the extent there is any conflict between the English text and the translation, English controls.

Clicking on the translation link activates a free translation service to convert the page to Spanish. As with any Internet translation, the conversion is not context-sensitive and may not translate the text to its original meaning. NC State Extension does not guarantee the accuracy of the translated text. Please note that some applications and/or services may not function as expected when translated.

Collapse ▲

Why Worry

Depending on the generative AI you are using, you may be inadvertently providing training data for their training model. The Large Language Models that power services like ChatGPT, Claude and Gemini need as much information as possible to increase the accuracy of their AI. The more data these models have access to, the more powerful they become. Some of these companies are running into the issue of running out of readily available data to use in training, so they are relying on user input data. This data is often from the free tiers of their user base.

Ignoring using the data for training, any data you have given them permission to use can be stored by the service you are using. In the event of a data breach, sensitive data may be compromised.

Best Practices

While you should always check the data privacy policy for any service you are using, there are a few best practices that you should follow to protect your data regardless of the policy. If you follow these, your data should be relatively safe, regardless of the company’s policies.

  •             With few exceptions, don’t enter information that is not publicly available or that you would not feel comfortable posting publicly on social media.
  •             Never enter demographic data. Do not include specific information about yourself or any other individual in prompts. There is never a reason to include this information.

Disabling Chat History and Training

ChatGPT

In the bottom left of your window click on your username. From the Context Menu that appears choose settings. In the settings window choose Data Control then disable Chat history & training. Keep in mind this will prevent you from revisiting old chats.

Open AI Data Controls Policy

Claude

By default, Claude Pro does not train using user inputs or prompts. Data will only be used “if you give us explicit permission by submitting feedback through the thumbs up/down feature or by reaching out to us with a request”.

ClaudePro FAQ on Data Privacy