BETA
This is a BETA experience. You may opt-out by clicking here

More From Forbes

Edit Story

Is Your Data Safe With GenAI Practices? Here's How To Stay Secure

Forbes Technology Council

Anshu is the founder/CEO of CloudDefense.AI—a CNAPP that secures both applications and cloud infrastructure.

Since the dawn of the internet, the ability to leverage vast amounts of data has fueled innovation across every industry. This data revolution has given rise to powerful new tools such as generative AI (GenAI) services like ChatGPT and Gemini. These innovative tools can craft realistic text formats, translate languages and even generate creative content, all at the user's command.

However, as CEO of a cloud security company, I'm also keenly aware of its potential dark side: data leaks. Unlike traditional software, GenAI services learn and adapt by processing massive amounts of data. This raises a critical question: Could our very interactions with these tools inadvertently expose sensitive information?

The answer, unfortunately, is yes. Data leaks involving GenAI are a real and growing concern. But the good news is that we can proactively mitigate these risks. Let's discuss real-world practices to keep your sensitive information safe while leveraging the power of GenAI.

Data Security Concerns With GenAI

At its core, GenAI services leverage complex algorithms trained on voluminous datasets encompassing text, code and diverse information formats. This training allows the AI to identify patterns and relationships within the data and use them accordingly.

But here's where the data security concern arises. To function effectively, GenAI services often require access to this training data, which can include a mix of publicly available and potentially sensitive information. The risk lies in how this data is secured by the GenAI provider and how users interact with the service.

Let's delve into a few scenarios where data leaks can occur:

Unintentional Exposure

When interacting with GenAI services, users might unknowingly include sensitive information in their prompts or instructions. Imagine a marketing team using a GenAI tool to generate ad copy. If they accidentally include snippets of customer data in their prompts, this information could be inadvertently incorporated into the generated content.

Insufficient Training Data Security

The security practices of the GenAI providers themselves play a crucial role. If the provider doesn't have robust safeguards in place to protect the training data, there's a chance it could be accessed by unauthorized individuals or accidentally leaked.

User Errors And Malicious Intent

Even with strong provider safeguards, user actions can introduce vulnerabilities. Accidental mistakes during data prompting, where users instruct the GenAI service what to do, could lead to sensitive information being unintentionally revealed. Additionally, malicious actors within an organization could misuse GenAI services to deliberately exfiltrate data.

Data Sharing During Development

The training process for GenAI models often involves vast amounts of data from various sources. While anonymization techniques are sometimes used, there's always a risk that residual sensitive information could remain hidden within the training data. Also, some GenAI services might involve collaboration with third-party developers, introducing another potential point of data leakage if proper security measures aren't in place.

API Vulnerabilities

Many GenAI services connect with other systems through APIs. These APIs act as bridges, allowing data to flow between the GenAI service and other applications. If these APIs are not properly secured with encryption and access controls, they could become a point of entry for attackers to steal data.

Best Practices To Mitigate GenAI Data Leaks

Completely eliminating risk is impossible, but as security professionals, we can significantly reduce the chances of a data leak by implementing a multi-layered approach. Here are some best practices I recommend incorporating into your organization's GenAI security strategy:

Data Minimization

The golden rule of data security applies here as well—less is more. Don't provide GenAI services with more data than absolutely necessary. For each interaction, carefully evaluate the minimum amount of data required to achieve your desired outcome. Sensitive information like personally identifiable data (PII) should be avoided whenever possible.

Anonymization And Tokenization

When dealing with sensitive data that can't be entirely removed, consider anonymization or tokenization techniques. Anonymization involves replacing identifying details with non-descriptive values while tokenization substitutes sensitive data with random tokens that can be later decrypted with proper access controls.

Educating Your Users

Remember, your employees are the front-line defense against data breaches. Invest in user education and training programs to raise awareness about the potential security risks associated with GenAI interactions. Teach them how to properly handle sensitive information when working with these services.

Controling Who Has Access

Implement robust access controls to limit who can interact with GenAI services and the type of data they can access. Adhere to the principle of least privilege, providing users only the minimum permissions needed for their designated tasks. Multi-factor authentication can also add an extra layer of security for accessing sensitive data or advanced GenAI functionalities.

Monitoring And Logging Everything

Consider implementing a system for continuously monitoring your interactions with GenAI services. This allows you to detect any suspicious activity or potential data exfiltration attempts in real time. Also, log all GenAI interactions for audit purposes. These logs can be invaluable for forensic analysis in case of a suspected breach.

Strengthening API Security Practices

Finally, we can't overlook the importance of API security. Those bridges connecting GenAI services to other systems—the APIs—need robust security measures in place. Encryption of data, both in transit and at rest, is a must. Regular security audits of APIs are also crucial to identify and address any vulnerabilities before they can be exploited.

Bottom Line

Just because GenAI poses potential data leak risks doesn't mean we should shy away from this powerful technology. The benefits of GenAI, when used responsibly, are undeniable. By implementing the best practices outlined above, you can significantly mitigate these risks and leverage the full potential of GenAI for your organization.

Remember, data security is not a one-time fix. The key is to be continuously vigilant about the risks and proactively take steps to mitigate them. In doing so, we can get the full potential of GenAI while safeguarding our data and building trust in this exciting new technology.


Forbes Technology Council is an invitation-only community for world-class CIOs, CTOs and technology executives. Do I qualify?


Follow me on Twitter or LinkedInCheck out my website