Article Details

Original Article Text

Click to Toggle View

How to manage the security risks of generative AI tools. Over the past year, we’ve witnessed an explosive growth spurt in consumer-focused AI productivity tools that has once again transformed the way we work. Once the realm of data science and engineering teams, generative AI was packaged and delivered to the masses in 2023. Now, just over a year after the splashy arrival of ChatGPT, artificial intelligence is almost imperceptibly woven into the daily rhythms of modern work. For context, SaaS security and governance provider Nudge Security has discovered over 500 unique generative AI applications in customer environments to date, up from 300 in November of 2023 and 150 in July of 2023. Given this pace of adoption and the rapid introduction of new AI tools, IT and security teams are struggling just to gain an understanding of who’s even using GenAI tools in their organizations. And because many of these AI tools have embraced the SaaS go-to-market model with free tiers and trials, there’s little barrier to entry, and no way of discovering usage through procurement or expense reports. How can organizations mitigate these risks? As is often the case, you have two options for managing the risks of GenAI tools. You can try to block all access, or you can offer ways for employees to experiment with these tools, with visibility and guidance from IT & security teams. User behavior research shows that blocking access can often motivate employees to look for ways around security controls, ultimately undermining IT governance efforts. And, in order to block a particular tool, you typically need to know it exists in the first place, which is a challenge given that new GenAI tools are popping up almost daily. How Nudge Security can help Nudge Security is a SaaS management platform that discovers all SaaS accounts ever created by anyone in your organization, including GenAI tools. Within minutes of starting a free trial, you’ll have a continuously updated inventory of your AI footprint. As new AI apps are introduced, Nudge Security alerts admins, providing visibility into who's using what, and allowing you to guide employees towards best practices to mitigate AI security risks. With Nudge Security, you can: 1. Gain immediate visibility of your GenAI footprint If they’re anything like other workers, your employees are already using AI in the workplace—but they may not want to admit it. To understand the role AI tools play for your organization, you need to know what's already out there and stay on top of new tools as employees sign up for them. With Nudge Security, you can get an immediate inventory of all the AI tools your employees are using, and set up alerts to notify you when a new AI tool is introduced. Nudge Security automatically discovers AI tools and other SaaS applications in your environment, and categorizes them by type for easy filtering, including the free, paid, and trial accounts that you might not be able to discover by relying on procurement processes or combing through expense reports. 2. Assess AI tools with context on usage and security Nudge Security provides a summary view of each application to help you assess new AI tools quickly. You can see a short description of the app, find out how many accounts and integrations have been created by members of your organization, identify the original user, and check your users' security hygiene. Drilling into each tab in the menu provides even more information to support your evaluations. For each app, Nudge Security also provides security context that can help you evaluate new applications quickly and systematically, such as links to their terms of service and privacy policies, an overview of their security program, supported authentication methods, and more. For each vendor you’ll also see their breach history and get alerted of security incidents affecting the applications your employees are using so you can intervene swiftly to secure their accounts, integrations, and data. As you complete vendor security reviews, you can set a status for each application to “Approved”, “Acceptable” or “Unacceptable” and share an Application Directory with employees so they can choose from approved options. 4. Surface and review risky OAuth grants The ease of agreeing to an OAuth grant can entice users to hand over more access to GenAI tools than they might realize. That's why Nudge Security reveals the scopes each application has been granted and provides OAuth risk scores to help you identify risky OAuth grants quickly. You’ll have the context you need to help you understand exactly what access and permissions each user has granted and what it means for your organization, so you can intervene if an application has too much access. 5. Provide timely acceptable use guidance Given the viral spread of AI tools, you have the best chance of influencing users' behavior by reaching them immediately when they sign up for a new app. Nudge Security offers just-in-time interventions called “nudges”, so you can reach users immediately via email or Slack when they create a new account. As soon as a user signs up for an AI tool, you can nudge them to review and acknowledge your AI acceptable use policy, reaching them right when the information is most relevant and useful. Users can accept the policy directly from the nudge (and all responses are gathered to document acknowledgement of the acceptable use policy), or they can request help if they have questions. You can also nudge them toward using an alternative application that you've already vetted, or prompt them to take a more secure action like setting up multi-factor authentication. Fuel innovation while mitigating risks with Nudge Security Your business needs AI to be competitive, which means your users need help determining which GenAI tools are trustworthy and how to use them in a way that does not put corporate data at risk. Nudge Security helps you discover, secure and govern AI usage so you can assess the security of new tools efficiently and nudge your users towards secure practices. Learn more about how you can manage AI security risks with Nudge Security and start a 14-day trial. Sponsored and written by NudgeSecurity.

Daily Brief Summary

MISCELLANEOUS // Managing Security Risks with AI Tools in the Workplace

The adoption of generative AI (GenAI) tools in the workplace has surged rapidly, with businesses observing a significant rise in AI application usage from 150 in July 2023 to over 500 recently.

Many employees use GenAI tools without formal oversight due to free trials and SaaS models, complicating IT and security teams' tracking and management efforts.

Nudge Security offers a SaaS management platform that detects all SaaS and GenAI tools used across an organization, providing an immediate comprehensive inventory even for newly adopted tools.

The platform allows security teams to review and assess the security of these tools by offering insights into usage, user identity, and integration details with aligned security evaluations.

Nudge Security helps identify and manage risky permissions granted via OAuth, enhancing security by understanding and overseeing the scope of access each application has.

It supports IT governance by sending timely "nudges" to users when they adopt new AI tools, prompting them to acknowledge the organization’s AI usage policies and encouraging secure practices.

Through these mechanisms, Nudge Security enables businesses to maintain a balance between fostering innovation with new AI technologies and ensuring robust protection against associated security risks.