Innovation Hub

Featured Posts

Insights
xx
min read

Securing AI Agents: The Questions That Actually Matter

Insights
xx
min read

The Hidden Risk of Agentic AI: What Happens Beyond the Prompt

Insights
xx
min read

Why Autonomous AI Is the Next Great Attack Surface

Get all our Latest Research & Insights

Explore our glossary to get clear, practical definitions of the terms shaping AI security, governance, and risk management.

Research

Research
xx
min read

LiteLLM Supply Chain Attack

Research
xx
min read

Exploring the Security Risks of AI Assistants like OpenClaw

Research
xx
min read

Agentic ShadowLogic

Research
xx
min read

MCP and the Shift to AI Systems

Videos

Report and Guides

Report and Guide
xx
min read

2026 AI Threat Landscape Report

Register today to receive your copy of the report on March 18th and secure your seat for the accompanying webinar on April 8th.

Report and Guide
xx
min read

Securing AI: The Technology Playbook

A practical playbook for securing, governing, and scaling AI applications for Tech companies.

Report and Guide
xx
min read

Securing AI: The Financial Services Playbook

A practical playbook for securing, governing, and scaling AI systems in financial services.

HiddenLayer AI Security Research Advisory

CVE-2026-3071

Flair Vulnerability Report

An arbitrary code execution vulnerability exists in the LanguageModel class due to unsafe deserialization in the load_language_model method. Specifically, the method invokes torch.load() with the weights_only parameter set to False, which causes PyTorch to rely on Python’s pickle module for object deserialization.

CVE-2025-62354

Allowlist Bypass in Run Terminal Tool Allows Arbitrary Code Execution During Autorun Mode

When in autorun mode, Cursor checks commands sent to run in the terminal to see if a command has been specifically allowed. The function that checks the command has a bypass to its logic allowing an attacker to craft a command that will execute non-allowed commands.

CVE-2025-62353

Path Traversal in File Tools Allowing Arbitrary Filesystem Access

A path traversal vulnerability exists within Windsurf’s codebase_search and write_to_file tools. These tools do not properly validate input paths, enabling access to files outside the intended project directory, which can provide attackers a way to read from and write to arbitrary locations on the target user’s filesystem.

SAI-ADV-2025-012

Data Exfiltration from Tool-Assisted Setup

Windsurf’s automated tools can execute instructions contained within project files without asking for user permission. This means an attacker can hide instructions within a project file to read and extract sensitive data from project files (such as a .env file) and insert it into web requests for the purposes of exfiltration.

In the News

News
XX
min read
HiddenLayer Unveils New Agentic Runtime Security Capabilities for Securing Autonomous AI Execution

News
XX
min read
HiddenLayer Releases the 2026 AI Threat Landscape Report, Spotlighting the Rise of Agentic AI and the Expanding Attack Surface of Autonomous Systems

News
XX
min read
HiddenLayer’s Malcolm Harkins Inducted into the CSO Hall of Fame

Insights
xx
min read

AI Coding Assistants at Risk

From autocomplete to full-blown code generation, AI-powered development tools like Cursor are transforming the way software is built. They’re fast, intuitive, and trusted by some of the world’s most recognized brands, such as Samsung, Shopify, monday.com, US Foods, and more.

Insights
xx
min read

OpenSSF Model Signing for Safer AI Supply Chains

The future of artificial intelligence depends not just on powerful models but also on our ability to trust them. As AI models become the backbone of countless applications, from healthcare diagnostics to financial systems, their integrity and security have never been more important. Yet the current AI ecosystem faces a fundamental challenge: How does one prove that the model to be deployed is exactly what the creator intended? Without layered verification mechanisms, organizations risk deploying compromised, tampered, or maliciously modified models, which could lead to potentially catastrophic consequences.

Insights
xx
min read

Structuring Transparency for Agentic AI

As generative AI evolves into more autonomous, agent-driven systems, the way we document and govern these models must evolve too. Traditional methods of model documentation, built for static, prompt-based models, are no longer sufficient. The industry is entering a new era where transparency isn't optional, it's structural.

Insights
xx
min read

Built-In AI Model Governance

A large financial institution is preparing to deploy a new fraud detection model. However, progress has stalled.

Insights
xx
min read

Life at HiddenLayer: Where Bold Thinkers Secure the Future of AI

At HiddenLayer, we’re not just watching AI change the world—we’re building the safeguards that make it safer. As a remote-first company focused on securing machine learning systems, we’re operating at the edge of what’s possible in tech and security. That’s exciting. It’s also a serious responsibility. And we’ve built a team that shows up every day ready to meet that challenge.

Insights
xx
min read

Integrating HiddenLayer’s Model Scanner with Databricks Unity Catalog

As machine learning becomes more embedded in enterprise workflows, model security is no longer optional. From training to deployment, organizations need a streamlined way to detect and respond to threats that might lurk inside their models. The integration between HiddenLayer’s Model Scanner and Databricks Unity Catalog provides an automated, frictionless way to monitor models for vulnerabilities as soon as they are registered. This approach ensures continuous protection without slowing down your teams.

Insights
xx
min read

Behind the Build: HiddenLayer’s Hackathon

At HiddenLayer, innovation isn’t a buzzword; it’s a habit. One way we nurture that mindset is through our internal hackathon: a time-boxed, creativity-fueled event where employees step away from their day-to-day roles to experiment, collaborate, and solve real problems. Whether it’s optimizing a workflow or prototyping a tool that could transform AI security, the hackathon is our space for bold ideas.

Insights
xx
min read

The AI Security Playbook

As AI rapidly transforms business operations across industries, it brings unprecedented security vulnerabilities that existing tools simply weren’t designed to address. This article reveals the hidden dangers lurking within AI systems, where attackers leverage runtime vulnerabilities to exploit model weaknesses, and introduces a comprehensive security framework that protects the entire AI lifecycle. Through the real-world journey of Maya, a data scientist, and Raj, a security lead, readers will discover how HiddenLayer’s platform seamlessly integrates robust security measures from development to deployment without disrupting innovation. In a landscape where keeping pace with adversarial AI techniques is nearly impossible for most organizations, this blueprint for end-to-end protection offers a crucial advantage before the inevitable headlines of major AI breaches begin to emerge.

Insights
xx
min read

Governing Agentic AI

Artificial intelligence is evolving rapidly. We’re moving from prompt-based systems to more autonomous, goal-driven technologies known as agentic AI. These systems can take independent actions, collaborate with other agents, and interact with external systems—all with limited human input. This shift introduces serious questions about governance, oversight, and security.

Insights
xx
min read

AI Policy in the U.S.

Artificial intelligence (AI) has rapidly evolved from a cutting-edge technology into a foundational layer of modern digital infrastructure. Its influence is reshaping industries, redefining public services, and creating new vectors of economic and national competitiveness. In this environment, we need to change the narrative of “how to strike a balance between regulation and innovation” to “how to maximize performance across all dimensions of AI development”.

Insights
xx
min read

RSAC 2025 Takeaways

RSA Conference 2025 may be over, but conversations are still echoing about what’s possible with AI and what’s at risk. This year’s theme, “Many Voices. One Community,” reflected the growing understanding that AI security isn’t a challenge one company or sector can solve alone. It takes shared responsibility, diverse perspectives, and purposeful collaboration.

Insights
xx
min read

Universal Bypass Discovery: Why AI Systems Everywhere Are at Risk

HiddenLayer researchers have developed the first single, universal prompt injection technique, post-instruction hierarchy, that successfully bypasses safety guardrails across nearly all major frontier AI models. This includes models from OpenAI (GPT-4o, GPT-4o-mini, and even the newly announced GPT-4.1), Google (Gemini 1.5, 2.0, and 2.5), Microsoft (Copilot), Anthropic (Claude 3.7 and 3.5), Meta (Llama 3 and 4 families), DeepSeek (V3, R1), Qwen (2.5 72B), and Mixtral (8x22B).

Webinars

Offensive and Defensive Security for Agentic AI

Webinars

How to Build Secure Agents

Webinars

Beating the AI Game, Ripple, Numerology, Darcula, Special Guests from Hidden Layer… – Malcolm Harkins, Kasimir Schulz – SWN #471

Webinars

HiddenLayer Webinar: 2024 AI Threat Landscape Report

Webinars

HiddenLayer Model Scanner

Webinars

HiddenLayer Webinar: A Guide to AI Red Teaming

Webinars

HiddenLayer Webinar: Accelerating Your Customer's AI Adoption

Webinars

HiddenLayer: AI Detection Response for GenAI

Webinars

HiddenLayer Webinar: Women Leading Cyber

research
xx
min read

LiteLLM Supply Chain Attack

research
xx
min read

Exploring the Security Risks of AI Assistants like OpenClaw

research
xx
min read

Agentic ShadowLogic

research
xx
min read

MCP and the Shift to AI Systems

research
xx
min read

The Lethal Trifecta and How to Defend Against It

research
xx
min read

EchoGram: The Hidden Vulnerability Undermining AI Guardrails

research
xx
min read

Same Model, Different Hat

research
xx
min read

The Expanding AI Cyber Risk Landscape

research
xx
min read

The First AI-Powered Cyber Attack

research
xx
min read

Prompts Gone Viral: Practical Code Assistant AI Viruses

research
xx
min read

Persistent Backdoors

research
xx
min read

Visual Input based Steering for Output Redirection (VISOR)

Report and Guide
xx
min read

2026 AI Threat Landscape Report

Report and Guide
xx
min read

Securing AI: The Technology Playbook

Report and Guide
xx
min read

Securing AI: The Financial Services Playbook

Report and Guide
xx
min read

AI Threat Landscape Report 2025

Report and Guide
xx
min read

HiddenLayer Named a Cool Vendor in AI Security

Report and Guide
xx
min read

A Step-By-Step Guide for CISOS

Report and Guide
xx
min read

AI Threat landscape Report 2024

Report and Guide
xx
min read

HiddenLayer and Intel eBook

Report and Guide
xx
min read

Forrester Opportunity Snapshot

Report and Guide
xx
min read

Gartner® Report: 3 Steps to Operationalize an Agentic AI Code of Conduct for Healthcare CIOs

news
xx
min read

HiddenLayer Unveils New Agentic Runtime Security Capabilities for Securing Autonomous AI Execution

news
xx
min read

HiddenLayer Releases the 2026 AI Threat Landscape Report, Spotlighting the Rise of Agentic AI and the Expanding Attack Surface of Autonomous Systems

news
xx
min read

HiddenLayer’s Malcolm Harkins Inducted into the CSO Hall of Fame

news
xx
min read

HiddenLayer Selected as Awardee on $151B Missile Defense Agency SHIELD IDIQ Supporting the Golden Dome Initiative

news
xx
min read

HiddenLayer Announces AWS GenAI Integrations, AI Attack Simulation Launch, and Platform Enhancements to Secure Bedrock and AgentCore Deployments

news
xx
min read

HiddenLayer Joins Databricks’ Data Intelligence Platform for Cybersecurity

news
xx
min read

HiddenLayer Appoints Chelsea Strong as Chief Revenue Officer to Accelerate Global Growth and Customer Expansion

news
xx
min read

HiddenLayer Listed in AWS “ICMP” for the US Federal Government

news
xx
min read

New TokenBreak Attack Bypasses AI Moderation with Single-Character Text Changes

news
xx
min read

Beating the AI Game, Ripple, Numerology, Darcula, Special Guests from Hidden Layer… – Malcolm Harkins, Kasimir Schulz – SWN #471

news
xx
min read

All Major Gen-AI Models Vulnerable to ‘Policy Puppetry’ Prompt Injection Attack

news
xx
min read

One Prompt Can Bypass Every Major LLM’s Safeguards

SAI Security Advisory

Pickle Load in Serialized Profile Load

An attacker can create a maliciously crafted Ydata-profiling report containing malicious code and share it with a victim. When the victim loads the report, the code will be executed on their system.

SAI Security Advisory

Model Deserialization Leads to Code Execution

An attacker can create a malicious crafted model containing an OperatorFuncNode, and share it with a victim. If the victim is using Python 3.11 or later and loads the malicious model arbitrary code will execute on their system.

SAI Security Advisory

Command Injection in CaptureDependency Function

A command injection vulnerability exists inside the capture_dependencies function of the AWS Sagemakers util file. If a user used the util function when creating their code, an attacker can leverage the vulnerability to run arbitrary commands on a system running the code by injecting a system command into the string passed to the function.

SAI Security Advisory

Command Injection in Capture Dependency

An attacker can inject a malicious pickle object into a numpy file and share it with a victim user. When the victim uses the NumpyDeserializer.deserialize function of the base_deserializers python file to load it, the allow_pickle optional argument can be set to ‘false’ and passed to np.load, leading to the safe loading of the file. However, by default the optional parameter was set to true, so if this is not specifically changed by the victim, this will result in the loading and execution of the malicious pickle object.

SAI Security Advisory

R-bitrary Code Execution Through Deserialization Vulnerability

An attacker could leverage the R Data Serialization format to insert arbitrary code into an RDS formatted file, or an R package as an RDX or RDB component, which will be executed when referenced or called with ReadRDS. This is because of the lazy evaluation process used in the unserialize function of the R programming language.

SAI Security Advisory

Out of bounds read due to lack of string termination in assert

An attacker can create a malicious onnx model which fails an assert statement in a way that an error string equal to or greater than 2048 characters is printed out and share it with a victim. When the victim tries to load the onnx model a string is created which leaks program memory.

SAI Security Advisory

Path sanitization bypass leading to arbitrary read

An attacker can create a malicious onnx model containing paths to externally located tensors and share it with a victim. When the victim tries to load the externally located tensors a directory traversal attack can occur leading to an arbitrary read on a victim’s system leading to information disclosure.

SAI Security Advisory

Web Server Renders User HTML Leading to XSS

An attacker can provide a URL rather than uploading an image to the Debug Samples tab of an Experiment. If the URL has the extension .html, the web server retrieves the HTML page, which is assumed to contain trusted data. The HTML is marked as safe and rendered on the page, resulting in arbitrary JavaScript running in any user’s browser when they view the samples tab.

SAI Security Advisory

Cross-Site Request Forgery in ClearML Server

An attacker can craft a malicious web page that triggers a CSRF when visited. When a user browses to the malicious web page a request is sent which can allow an attacker to fully compromise a user’s account.

SAI Security Advisory

Improper Auth Leading to Arbitrary Read-Write Access

An attacker can, due to lack of authentication, arbitrarily upload, delete, modify, or download files on the fileserver, even if the files belong to another user.

SAI Security Advisory

Path Traversal on File Download

An attacker can upload or modify a dataset containing a link pointing to an arbitrary file and a target file path. When a user interacts with this dataset, such as when using the Dataset.squash method, the file is written to the target path on the user’s system.

SAI Security Advisory

Pickle Load on Artifact Get

An attacker can create a pickle file containing arbitrary code and upload it as an artifact to a Project via the API. When a victim user calls the get method within the Artifact class to download and load a file into memory, the pickle file is deserialized on their system, running any arbitrary code it contains.

Stay Ahead of AI Security Risks

Get research-driven insights, emerging threat analysis, and practical guidance on securing AI systems—delivered to your inbox.

By submitting this form, you agree to HiddenLayer's Terms of Use and acknowledge our Privacy Statement.

Thanks for your message!

We will reach back to you as soon as possible.

Oops! Something went wrong while submitting the form.