GCP Vertex AI Security Blind Spots Exposed

You deploy an AI agent in GCP's Vertex AI thinking it's your trusty sidekick. Turns out, it might be spilling your secrets to attackers. Unit 42's research just blew the lid off this sneaky vulnerability.

Diagram showing malicious AI agent extracting credentials from GCP Vertex AI service account

Key Takeaways

  • Default P4SA permissions in Vertex AI enable privilege escalation and data exfil from compromised agents.
  • Google updated docs post-disclosure, but core permission models need overhaul for true safety.
  • AI agents amplify cloud risks — audit now to avoid turning helpers into hackers.

Everyone figured Google’s Vertex AI was the gold standard for building autonomous AI agents — smoothly, enterprise-ready, plug it into your workflows and watch the magic. Developers were buzzing about Agent Engine and the ADK kit, dreaming of agents handling complex tasks without constant hand-holding. But here’s the gut punch: researchers from Palo Alto Networks’ Unit 42 just proved those agents can flip into double agents, siphoning data and punching holes in your GCP setup.

This GCP Vertex AI security blind spot isn’t some edge case. It’s baked into default permissions. And it changes everything — suddenly, that ‘autonomous’ AI everyone’s chasing? Might be your worst security nightmare.

Look, I’ve covered Silicon Valley’s cloud hype for two decades. Remember the AWS S3 bucket fiascoes back in 2017? Millions of records exposed because devs trusted defaults too much. Déjà vu. Unit 42 deployed a malicious agent using Google’s own tools, exploited the Per-Project, Per-Product Service Agent (P4SA) — which Google manages — and boom, grabbed creds for another service agent. From there? Privileged access to consumer project data, restricted images, even source code in a producer project.

“We were able to achieve privileged access to data in a consumer project, and to restricted images and source code within a producer project that is part of Google’s infrastructure.”

That’s straight from their report. Chilling, right? They packaged it as a pickle file — yeah, Python’s serialization format — deployed via Agent Engine, then hit Google’s metadata service to extract JSON creds. One call to the agent, and it’s game over.

How Did They Pull Off the Double Agent Trick?

It starts simple. You init Vertex AI with your project ID, location, staging bucket. Define tools — in this case, a ‘get_service_agent_credentials’ function laced with malice. Slap together an Agent named ‘my_double_agent’ using Gemini 2.0 Flash, give it instructions to exfiltrate data, and deploy.

Code’s public now, but Google tweaked the workflow post-discovery, so it won’t run as-is. Still, the P4SA had excessive default permissions. That service account — [email protected] — let them impersonate and escalate. Extract creds, act on behalf, rifle through buckets and infra.

But — and this is my cynical veteran take — who’s really surprised? Google’s been playing fast and loose with service accounts since day one. Remember the 2020 App Engine flaws? Attackers chaining perms for full takeovers. History repeats because no one slows down for security when VC cash flows.

The kicker? Unit 42 shared with Google, who updated docs on resource use. Good on them for collaborating. But docs aren’t code. And defaults? Still a gamble.

Short para for punch: Trust no agent.

Now, dig deeper. These agents interact with services, make decisions independently. Enterprise workflows mean permissions galore — storage, compute, secrets. A compromised one? It’s insider threat on steroids. Exfiltrate data quietly while pretending to serve coffee orders or whatever benign task you assigned.

Unit 42’s setup used ADK’s google_search tool, but the real weapon was that metadata query: http://metadata.google.internal/computeMetadata/v1/instance/?recursive=true. Grabs everything. JSON spew of creds, details. Figures in their report show it plain as day.

Can Your Vertex AI Agents Be Weaponized Right Now?

Yes — or close enough. Google’s changes mean new deploys might dodge this exact path, but the permission model? Fundamentally flawed until audited. P4SA scopes are per-project, per-product, but defaults grant too much. Attackers love that.

Think about it. You’re a dev rushing an agent live for Black Friday sales forecasting. Miss a perm tweak? Double agent steals customer PII. Or worse, in finance — trades data to rivals. We’ve seen precursors: LangChain vulns letting prompt injection steal keys. Vertex AI scales that to GCP-wide chaos.

My unique insight, absent from Unit 42’s piece: This echoes the SolarWinds supply chain hack, but bottom-up. Not nation-states slipping in; your own tools turning rogue. Prediction? By 2025, agentic AI breaches hit headlines monthly. Who’s making money? Not you — Palo Alto, hawking Prisma AIRS, Cortex Cloud Identity Security, their AI-SPM. Article ends with a sales pitch, naturally.

Google’s PR spin calls it ‘comprehensive.’ Please. It’s a platform with blind spots big as the Valley’s ego. They revised docs — bravo — but where’s the least-privilege enforcement by default? Where’s agent sandboxing?

Why Does This Matter for Cloud Security Teams?

Because AI agents are the new normal. Vertex AI pushes them hard: integrate, automate, scale. But autonomy breeds risk. Delegate tasks, grant perms — attackers pivot from one compromised agent to your whole org.

Unit 42 offers assessments, incident response. Smart. But prevention? Scan your P4SAs, revoke defaults, monitor metadata calls. Tools like their Cortex lineup help, sure. Open-source? GCloud IAM audits, agent tracing.

Cynical aside — Google’s ecosystem locks you in deeper. Fix one hole, three more for agents pop up. I’ve seen it with Kubernetes perms, Anthos sprawl.

One sentence wonder: Time to audit, yesterday.

Wrapping the thread: This isn’t anti-Google rabble-rousing. It’s a wake-up. Hype sold safe AI; reality delivers double agents. Enterprises, pause deploys. Demand better scoping. And Google? Ship secure defaults, not docs.


🧬 Related Insights

  • Read more:
  • Read more:

Frequently Asked Questions

What is a double agent in GCP Vertex AI?

It’s a deployed AI agent that looks legit but secretly exfiltrates data or escalates privileges via misconfigured service accounts like P4SA.

How to fix Vertex AI permission risks?

Audit P4SAs, enable least-privilege, monitor agent tools and metadata queries, follow Google’s updated docs — but verify with tools like IAM Recommender.

Is Vertex AI safe after Google’s update?

Safer, but not bulletproof. Defaults still risky; test your setups against Unit 42’s methods.

Aisha Patel
Written by

Former ML engineer turned writer. Covers computer vision and robotics with a practitioner perspective.

Frequently asked questions

What is a double agent in GCP Vertex AI?
It's a deployed AI agent that looks legit but secretly exfiltrates data or escalates privileges via misconfigured service accounts like P4SA.
How to fix Vertex AI permission risks?
Audit P4SAs, enable least-privilege, monitor agent tools and metadata queries, follow Google's updated docs — but verify with tools like IAM Recommender.
Is Vertex AI safe after Google's update?
Safer, but not bulletproof. Defaults still risky; test your setups against Unit 42's methods.

Worth sharing?

Get the best AI stories of the week in your inbox — no noise, no spam.

Originally reported by Palo Alto Unit 42

Stay in the loop

The week's most important stories from theAIcatchup, delivered once a week.