Insight

2025 AI security solutions Radar

Published November 13, 2025

  • Cybersecurity

Key takeaways

  • AI security market is maturing and consolidating with major acquisitions
  • Number of solutions increased from 88 to 94; new category “AI Firewall & Response” added
  • Companies face a choice: specialized solutions vs. cloud-native AI security
  • Agentic AI introduces broader and more complex security risks

2025 AI security solutions Radar

The AI security market is entering a new phase

After several years of excitement and exploration, we are now witnessing a clear consolidation of the AI security solutions market. The AI security sector is entering a phase of maturity, as reflected in the evolution of our radar. Since our 2024 edition of the AI security solutions radar, five major acquisitions have taken place:

  • Cisco acquired Robust Intelligence in September 2024
  • SAS acquired Hazy in November 2024
  • H Company acquired Mithril Security at the end of 2024
  • Nvidia acquired Gretel in March 2025
  • Palo Alto announced its intention to acquire ProtectAI in April 2025

These motions reflect a clear desire by major IT players to secure their positions by absorbing key technology startups.

Simultaneously, our new mapping lists 94 solutions, compared to 88 in the October 2024 edition. Fifteen new solutions have entered the radar, while eight have been removed. These removals are mainly due to discontinued offerings or strategic repositioning: some startups failed to gain market traction, while others shifted focus to broader AI applications beyond cybersecurity.

Finally, a paradigm shift is underway: solutions are moving beyond a mere stacking of technical blocks and evolving into integrated defense architectures, designed to meet the long-term needs of large organizations. Interoperability, scalability, and alignment with the needs of large enterprises are becoming the new standards. AI cybersecurity is now asserting itself as a global strategy, no longer just a collection of ad hoc responses.

To reflect this evolution, we have updated our own mapping by creating a new category, AI Firewall & Response, which results from the merger of our Machine Learning Detection & Response and Secure Chat/LLM Firewall categories.

Best of breed or good enough? The integration dilemma

With the growing integration of AI security components into the offerings of major Cloud Providers (Microsoft Azure, AWS, Google Cloud), a strategic question arises: Should we favor expert solutions or rely on the native capabilities of hyperscalers?

  • Specialized solutions offer technical depth and targeted coverage, complementing existing security
  • Integrated components are easier to deploy, interoperable with existing infrastructure, and often sufficient for standard use cases

This is not about choosing one over the other but about shedding light on the possibilities. Here is an overview of some security levers available through hyperscaler offerings.

This approach goes beyond securing data at rest or in transit: it aims to protect computations in progress, using secure enclaves. It ensures a high level of confidentiality throughout the lifecycle of AI models, sensitive data, or proprietary algorithms, by preventing any unauthorized access.

Agentic AI: a cross-cutting risk, a distributed security approach

Among the trends drawing increasing attention from cybersecurity experts, agentic AI is gaining ground. These systems, capable of making decisions, planning actions, and interacting with complex environments, actually combine two types of vulnerabilities:

  • those of traditional IT systems
  • and those specific to AI models

The result: an expanded attack area and potentially critical consequences. If misconfigured, an agent could access sensitive files, execute malicious code, or trigger unexpected side effects in a production environment.

An aggravating factor adds to this: the emergence of the Model Context Protocol (MCP), a standard currently being adopted that allows LLMs to interact in a standardized way with third-party tools and services (email, calendar, drive…). While it facilitates the rise of agents, it also introduces new attack vectors:

  • Exposure or theft of authentication tokens
  • Lack of authentication mechanisms for tools
  • Possibility of prompt injection attacks in seemingly harmless content
  • Or even compromise of an MCP server granting access to all connected services

Beyond technical vulnerabilities, the unpredictable behavior of agentic AI introduces a new layer of complexity. Because actions directly stem from AI model outputs, a misinterpretation or planning error can lead to major deviations from the original intent.

In this context, securing agentic AI does not fall under a single category. It requires cross-cutting coverage, mobilizing all components of our radar: robustness evaluation, monitoring, data protection, explainability, filtering, and risk management. And this is precisely what we’re seeing in the market: the first responses to agentic AI security do not come from new players, but from additional features integrated into existing solutions. An emerging issue, then, but one already being addressed.

Our recommendations: which AI security components should be prioritized?

Given the evolution of threats, the growing complexity of AI systems (especially agents), and the diversity of available solutions, we recommend focusing efforts on three major categories of security, which complement each other.

Monitoring AI systems has become essential. Indeed, an AI can evolve unpredictably, degrade over time, or begin generating problematic responses without immediate detection. This is especially critical in the case of agentic AI, whose behavior can have a direct operational impact if left unchecked.

In the face of this volatility, it is crucial to detect weak signals in real time (prompt injection attempts, behavioral drift, emerging biases, etc.). That’s why it’s preferable to rely on expert solutions dedicated to detection and response, which offer specific analyses and alert mechanisms tailored to these threats.

Conclusion: building a Security Strategy for Enterprise AI

As artificial intelligence becomes deeply embedded in enterprise operations, securing AI systems is no longer optional; it is a strategic imperative. The rapid evolution of threats, the emergence of agentic AI, and the increasing complexity of models call for a shift from reactive responses to proactive, integrated security strategies.

Organizations need to move beyond fragmented approaches and embrace a holistic framework that brings together robustness testing, continuous monitoring, and ethical safeguards. The rise of integrated defense architectures and the convergence of AI security domains reflect a maturing market, one that is now ready to support enterprise-scale deployments.

The challenge is clear: find the right balance between specialized tools and native cloud capabilities, ensure broad and consistent coverage, and keep AI systems trustworthy, resilient, and aligned with business goals.

 

 

We thank Anthony APRUZZESE for his valuable contribution to the writing of this article.

  • Gérôme Billois

    Partner – France, Paris

    Wavestone

    LinkedIn
  • Sleh-Eddine Choura

    Manager – France, Paris

    Wavestone

    Linkedin
  • Aliénor Fougerat

    Senior Consultant

    Wavestone

    LinkedIn
  • Yashveen Jootun

    Consultant – France, Paris

    Wavestone

    Linkedin