AI Is Coming: Meet the Startups Building Cyber Defenses for the Age of AI

Written by

Shivani Singh

Published on

Read

7 min

There’s a dark side to AI that’s underreported. It’s not deepfakes, mass surveillance, job displacement, or AI takeovers but rather every CISO’s (Chief Information Security Officer’s) worst nightmare. Since the launch of ChatGPT in 2022, employees have been feeding code, emails, and sensitive documents into generative AI tools, opening the floodgates to data leaks and security risks.

It’s not just third-party tools like ChatGPT that are raising concerns. Businesses are also developing their own AI tools using internal company data, which presents new challenges for CISOs. For example, an internal AI chatbot trained on Slack conversations could inadvertently expose private discussions if it has security vulnerabilities. As the CEO of Adversa AI, an AI Security startup, put it: “The PC sparked the first cybersecurity revolution, followed by the cloud and cloud security. Now, we’re entering the era of AI — and AI security is the natural next step.”

In this blog, we’ll map out the rapidly growing AI security ecosystem, highlighting where innovation is happening now and predicting where it’s headed next.

Overview

It’s no surprise that AI security innovation is gaining momentum. To address risks, companies like Apple, Samsung, and JPMorgan initially responded by banning AI tools altogether. However, as AI adoption accelerates, it has become clear that such blanket bans can stifle innovation.

Instead of avoiding AI, organizations are now focusing on securing it. This has given rise to a new wave of AI security startups tackling various AI security challenges, from providing visibility into AI usage within organizations to securing AI models before and after deployment.

For example, Protect AI offers both pre- and post-deployment scanning to safeguard large language models (LLMs), while Aurascape.ai helps businesses monitor AI usage across the firm and enforce security policies. The impact of these solutions is already evident. HiddenLayer, for instance, partnered with a global financial firm to uncover vulnerabilities in its fraud detection AI, saving the company millions. In 2024, Cisco acquired the AI security startup Robust Intelligence for $400 million to offer AI security solutions to its corporate customers, showing an appetite for products in the space.

The market map below gives an overview to the layers of AI security that innovators are producing and companies deploying. We’ll walk through each area to explain the issues and emerging solutions.

1. Visibility & Governance

The first step in securing AI is gaining visibility into how employees interact with AI tools: identifying which models they use, what data they input, and what they receive in return. Once CISOs have a clear view of AI usage within the organization, they can implement control measures (also known as governance) to manage AI access and usage. Governance applies to both third-party AI tools (like ChatGPT) and internally developed AI tools (such as an in-house chatbot).

Internal AI Models

For internally built AI, Role-Based Access Control (RBAC) is essential to restrict access to sensitive data. This is a security method that gives access to systems, applications, and data based on a user’s role, ensuring resource access is only given to those who need it for their job functions.

For example, if a company builds an AI-powered chatbot trained on internal Slack conversations, it’s crucial to ensure the security of the data layer that contains the private Slack conversations. This involves attaching metadata to the chat data so that only authorized users can access specific information – you don’t want just anyone viewing private Slack conversations.

A key innovation in this space is policy-as-code, which allows teams to define security policies using code rather than relying on manual processes (like writing directly to configuration files). Policy-as-code frameworks can include rules, conditions, and automated protections — such as blocking unauthorized access when breaches occur. CISOs can use programming languages like Python or Rego to define policies, making security faster and less error-prone.

Backward compatibility is also essential. New AI security tools must seamlessly integrate with existing enterprise system like Salesforce to prevent accidental data leaks and maintain consistent access control.

External AI Services

For third-party tools like ChatGPT, similar policy enforcement principles can apply. CISOs can set controls specifying which tools employees are allowed to use, define the types of data that can be input to the tools, and establish safeguards to prevent data leaks, often using policy-as-code or comparable frameworks.

While most AI security tools today focus on human interactions with AI, the scope is quickly evolving. APIs and AI agents are starting to access AI models too, so very soon we expect AI security solutions to secure these non-human AI interactions as well.

2. Pre-Deployment Testing

The next layer of AI security focuses on selecting the right AI model and preparing it for deployment within your organization. As of 2025, the landscape of AI models is rapidly expanding, with platforms like Hugging Face offering 1.5M models and 40,000 new models added every week. However, choosing the right model isn’t just about performance; it’s also about evaluating factors like bias and security.

Most AI models are trained on datasets sourced from the internet, making them susceptible to biases and misinformation. These biases can result in reputational damage and even legal repercussions. A prominent example occurred in 2024 when Google’s Gemini model sparked controversy by generating racially diverse images, even when users explicitly requested images of a specific race.

To mitigate such risks, many enterprises are investing in AI auditing and compliance. Similar to how companies use tools like Drata and Vanta to achieve SOC 2 certification for security and privacy, organizations are increasingly adopting AI model certification solutions to ensure their AI models are free from bias and toxicity. Additionally, businesses are partnering with vendors that assess AI models for security risks and ethical concerns, providing documentation for regulators and customers.

Another emerging trend is firms using AI to detect high-risk employee behaviors that could compromise security. For example, Dune Security’s platform identifies risky actions, such as clicking on phishing links or visiting unsecured websites, and flags them as potential threats.

Continuous Vulnerability Scanning

Continuous vulnerability scanning is gaining momentum as a component of AI security. Unlike traditional software, AI models cannot be effectively scanned using existing Endpoint Detection and Response (EDR) tools, which often fail to detect AI-specific threats such as payload attacks, prompt engineering, and exposure of personally identifiable information (PII).

For example, the AI model DeepSeek has demonstrated several vulnerabilities identified by AI security startups like Holistic AI and Hidden Layer. Certain prompts can inadvertently cause DeepSeek to reveal its training data, while structured prompts can trap the model in an endless loop — a phenomenon known as Adversarial Chain of Thought. A red-teaming report by Enkrypt AI revealed that DeepSeek was 11 times more likely to generate harmful content, 4 times more toxic, and 4.5 times more likely to produce insecure code compared to leading models like GPT-4 and Claude-3. Consequently, new tools are emerging that provide continuous scanning of AI models, identifying vulnerabilities and highlighting areas that require remediation.

Currently, most solutions focus solely on protecting and identifying vulnerabilities within the AI model itself. However, the next wave of innovation is expected to go beyond model protection, aiming to secure the entire AI ecosystem. This would include the application, its data flows, containerized services, vector databases, and other integrated infrastructure components.

3. Post Deployment Defense

So, you’ve chosen your model, scanned for vulnerabilities, tested its performance, and are ready to deploy it to real users. But you’re not done yet. Once an AI system goes live, it becomes a prime target for attackers seeking to exploit its weaknesses in various ways.

Let’s break down a key list of attack vectors, though this list is not exhaustive.

  • Home

    Jailbreaking:

    A model that involves tricking it into generating harmful, unethical, or restricted content by bypassing its built-in safeguards. Attackers exploit weaknesses in how LLMs interpret prompts to override security controls. One common method is role-playing attacks, where an attacker instructs the AI to "act as an evil AI" or simulate a restricted scenario, deceiving it into producing content it would normally block. Another technique is character obfuscation, which uses spaces, special characters, or ASCII tricks to confuse content moderation filters and evade detection.
  • Home

    SQL Injection Attacks:

    SQL Injection attacks is another well-known web security threat that also applies to AI models interfacing with databases. Suppose an AI-powered chatbot or tool is given direct access to a database without proper input sanitization. In that case, an attacker can insert malicious SQL queries to bypass authentication and access confidential data.
  • Home

    Remote Code Execution (RCE) Attacks:

    RCE attacks can impact AI models that support code generation or execution, such as AI-powered coding assistants. If a model is configured to execute code provided by a user, an attacker could inject malicious scripts into a prompt, trick the model into running the scripts on the host system, and gain control over the underlying infrastructure of the AI application. For example, a coding assistant that allows executing Python snippets could be manipulated to run a command that grants the attacker remote access to a system.

Many solutions are emerging to defend against these attack vectors. One promising approach is the AI firewall, which acts like a security guard for your AI models.

AI Firewall

Just as Web Application Firewalls (WAFs) monitor and protect web applications from harmful traffic, an AI firewall monitors and filters the data flowing in and out of an AI model to detect suspicious activity or malicious inputs that could compromise the model. One example is Robust Intelligence’s AI Firewall, which validates model inputs and outputs in real time, protecting against various threats, including prompt injection, PII extraction, and hallucination. Robust Intelligence’s AI firewall is integrated with MongoDB Atlas Vector Search, allowing MongoDB’s 40,000+ customers to use any commercial or open-source AI solutions safely.

Robust Intelligence’s AI security solutions have gained significant traction, with its products being used by JPMorgan Chase, Expedia, Deloitte, Cisco, and even the U.S. Department of Defense. The adoption highlights the growing need for reliable AI security solutions today, and we expect the demand to grow.

Looking Ahead

As AI security evolves, some key questions are shaping its future:

  • Home

    Who will lead AI security?

    Startups are leading the charge in innovation today, but it remains to be seen whether they will maintain their dominance or if established security giants will catch up.
  • Home

    Where will AI security live?

    Today, many AI security solutions reside at the software and application layer. However, future innovations may shift toward the platform, infrastructure, or hardware level. It remains to be seen which of these layers will ultimately dominate.
  • Home

    Will AI security emerge as a stand-alone product?

    Early conversations with CISOs and founders suggest that AI security may be distinct enough to emerge as its own category, just like Identity & Access Management, Network Security, Application Security, Cloud Security. Whether this will ultimately happen remains to be seen.

AI is not slowing down, and neither are the threats. The future belongs to those who build defenses as fast as AI evolves.

We’d love to hear from you. If you’re interested in AI security or have insights on where the space is headed, please connect.

📩 Vedant Agrawal – [email protected]

 📩 Shivani Singh – [email protected]

ABOUT THE AUTHORS

Shivani Singh
Shivani Singh
Senior Associate

Shivani Singh is an early and growth-stage investor at Alumni Ventures, focusing on the SaaS, Cybersecurity, and Healthtech sectors. Previously, Shivani was a Senior Product Manager at Google, where she led key product initiatives across Google Cloud, Google Meet, and the Google Pixel teams, including the launch of Google Meet for Enterprise during the COVID-19 pandemic. Shivani holds a Bachelor's degree in Computer Engineering from the University of Illinois Urbana Champaign and an MBA from Harvard Business School.

Vedant Agrawal
Vedant Agrawal
Premji Invest

Vedant Agrawal leads early-stage investing at Premji Invest with a focus on AI apps and infra. He’s been involved in investments like Hippocratic, Galileo, Writer, Poolside, Hugging Face, and Runway. Previously, he was an early-stage investor at Sequoia Capital India, and was part of the founding team of Surge, Sequoia’s founder acceleration program (successor being Sequois Arc in the US). Post that, he helped launch and scale a new ed-tech vertical at Scaler Academy to $20M in annualized bookings in nine months. He began his career as a Business Analyst at McKinsey & Company. Vedant holds an MBA from Harvard Business School and a B.Tech from IIT Madras.

Kshiteej Prasad
Kshiteej Prasad
Alumni Ventures Scout

Kshiteej Prasad is an MBA candidate at Harvard Business School with a background in global operations at Procter & Gamble and early-stage investing at Bessemer Venture Partners. He’s passionate about backing bold ideas across industrial, climate, fintech, and cybersecurity verticals, and spends his free time chasing squash rallies and better thesis ideas.

This communication is from Alumni Ventures, a for-profit venture capital company that is not affiliated with or endorsed by any school. It is not personalized advice, and AV only provides advice to its client funds. This communication is neither an offer to sell, nor a solicitation of an offer to purchase, any security. Such offers are made only pursuant to the formal offering documents for the fund(s) concerned, and describe significant risks and other material information that should be carefully considered before investing. For additional information, please see here. Example portfolio companies are provided for illustrative purposes only and are not necessarily indicative of any AV fund or the outcomes experienced by any investor. Example portfolio companies shown are not available to future investors, except potentially in the case of follow-on investments. Venture capital investing involves substantial risk, including risk of loss of all capital invested. This communication includes forward-looking statements, generally consisting of any statement pertaining to any issue other than historical fact, including without limitation predictions, financial projections, the anticipated results of the execution of any plan or strategy, the expectation or belief of the speaker, or other events or circumstances to exist in the future. Forward-looking statements are not representations of actual fact, depend on certain assumptions that may not be realized, and are not guaranteed to occur. Any forward-looking statements included in this communication speak only as of the date of the communication. AV and its affiliates disclaim any obligation to update, amend, or alter such forward-looking statements, whether due to subsequent events, new information, or otherwise.