5 Things You Should Never Share with ChatGPT and Other AI Chatbots
Google recently made changes to its privacy policy, allowing it to use all public data and information from the internet to train its ChatGPT rivals. To oppose this change, you can only delete your account, but even then, the data you have previously shared online may still be used to train Google’s Bard and other ChatGPT alternatives.
This change in Google’s privacy policy serves as a reminder that oversharing with AI chatbots should be avoided. Until these programs can be trusted with your privacy, it’s essential to keep certain information from them.
Currently, generative AI innovation lacks proper regulation. However, governments worldwide will eventually implement best practices to protect user privacy and copyrighted content.
There will come a time when generative AI operates independently on devices without relying on external servers. Products like Humane’s Ai Pin and Apple’s Vision Pro could be examples of this. Until then, it is crucial to treat ChatGPT, Google Bard, and Bing Chat as strangers in your home or office – refraining from sharing personal information or work-related secrets with them.
While it has been mentioned before, it is worth reiterating that personal details like your full name, address, birthday, and social security number should not be shared with ChatGPT and other AI bots. OpenAI introduced privacy features after the release of ChatGPT, but even with these settings enabled, it is not enough to ensure the confidentiality of your information. Bugs or other issues may compromise the effectiveness of these privacy settings.
The concern is not that ChatGPT will profit from your personal information, nor that OpenAI will misuse it. However, this data is used to train the AI model. Moreover, OpenAI has previously experienced a data breach due to hacking, which may result in your data falling into the wrong hands. Although finding specific data may be challenging, it is not impossible, leading to potential misuse such as identity theft.
Hackers primarily seek login information during data breaches. Having access to usernames and passwords can be detrimental, especially if you reuse credentials across multiple platforms. To safeguard your passwords effectively, it is recommended to use secure password management apps like Proton Pass and 1Password.
Avoid sharing your login credentials with generative AI, even though it might be convenient in the future to use AI to log in to various applications. Sharing personal banking information with ChatGPT is also unnecessary as OpenAI will never require credit card numbers or bank account details. Disclosing this highly sensitive data can result in severe financial repercussions if it falls into the wrong hands.
Beware of any mobile or computer apps claiming to be ChatGPT clients that request financial information. Such requests are red flags indicating potential ChatGPT malware. Do not provide any financial data under any circumstances. Instead, delete the app and only use official generative AI apps from reputable sources like OpenAI, Google, or Microsoft.
Previously, confidential information from Samsung reached OpenAI’s servers. This incident led to Samsung banning generative AI bots, and other companies, including Apple, followed suit. Therefore, it is crucial to keep work-related secrets confidential and find alternative, secure methods to seek assistance from ChatGPT.
Lastly, sharing extensive health data with chatbots is not advisable. Although prompts containing what if scenarios of specific symptoms might seem harmless, it is important to avoid disclosing all your health data to ChatGPT-like services unless they are personal AI products operating on your device. While generative AI products may have the potential to provide insight into health matters in the future, it is important to approach these platforms with caution.
Additionally, sharing your most personal thoughts with chatbots, which some individuals may rely on for therapy, poses privacy concerns. While utilizing chatbots for therapeutic purposes is a personal decision, it is crucial to remember that ChatGPT and other chatbots do not offer privacy that can be completely trusted. Your personal thoughts will be stored on the servers of OpenAI, Google, or Microsoft and used to train the AI models.
It is vital to note that ChatGPT and other chatbots may provide incorrect information, even regarding health issues. When interacting with these bots, always request sources for their responses. However, resist the temptation to provide additional personal information in hopes of receiving more tailored answers.
Lastly, beware of malicious apps posing as generative AI programs that may request personal data. Falling victim to such apps might leave you unaware until it’s too late, with hackers potentially exploiting your personal information against you.
In conclusion, it is crucial to exercise caution and avoid sharing certain information with AI chatbots like ChatGPT. By prioritizing your privacy and refraining from oversharing personal and sensitive data, you can protect yourself from potential misuse and maintain control over your information.