The employee just wanted to streamline their workflow. Context AI's Office Suite promised to help build presentations and documents using AI agents - a tempting offer for anyone drowning in productivity demands. They clicked "Allow All" permissions without a second thought. It was a small, third-party AI tool. What could go wrong?
Everything, as it turns out.
On April 19, 2026, Vercel - the cloud platform powering millions of websites including Next.js deployments - announced a devastating security breach. Hackers had compromised internal systems and accessed customer credentials. The attack vector wasn't a sophisticated zero-day exploit or a nation-state APT campaign. It was a compromised OAuth token from a third-party AI productivity tool called Context AI.
The breach represents everything security professionals have been warning about: the invisible risks of AI tool sprawl, the fragility of OAuth-based integrations, and how a single employee's app installation can cascade into a multi-million dollar supply chain disaster.
The Attack Chain: From Game Exploits to Enterprise Breach
Phase 1: The Initial Compromise at Context AI
The story begins not with Vercel, but with a Context AI employee who downloaded game exploits - likely seeking cheats or modifications for a video game. This innocent-seeming action triggered a malware infection that deployed Lumma Stealer, an information-stealing trojan increasingly popular among cybercriminals.
Lumma Stealer extracted a treasure trove of credentials from the compromised machine:
- Google Workspace credentials
- Multiple application login details
- OAuth tokens for Context AI's services
- Internal system access keys
According to Hudson Rock's analysis of the stolen data, the attackers gained access to Context AI's AWS environment and - critically - the OAuth tokens that Context AI's consumer users had granted to the AI Office Suite application.
Phase 2: The OAuth Token Jackpot
Context AI's AI Office Suite was designed to let users "work with AI agents to build presentations, documents, and spreadsheets." The key feature: integration with external applications via OAuth, allowing the AI to access and manipulate data across a user's Google Workspace.
When users installed the Context AI Chrome extension and granted permissions, they weren't just giving access to their own data. They were handing over OAuth tokens that could potentially access any Google Workspace where the app was authorized - including enterprise environments with thousands of users.
Here's where Vercel enters the picture. At least one Vercel employee signed up for Context AI's AI Office Suite using their Vercel enterprise Google account and granted "Allow All" permissions. Context AI's security bulletin notes that "Vercel's internal OAuth configurations appear to have allowed this action to grant these broad permissions in Vercel's enterprise Google Workspace."
Phase 3: Pivoting to Vercel's Crown Jewels
With the compromised OAuth token, attackers gained access to the Vercel employee's Google Workspace account. From there, they pivoted into Vercel's internal systems.
The attackers accessed environment variables that were not marked as "sensitive" in Vercel's systems. While Vercel encrypts sensitive environment variables to prevent reading, non-sensitive variables were stored in plaintext - a design decision that proved catastrophic.
These environment variables contained:
- API keys for customer applications
- Database credentials
- Internal service tokens
- Deployment secrets
Vercel's security bulletin acknowledges the attackers were "highly sophisticated based on their operational velocity and detailed understanding of Vercel's systems." The company has engaged Mandiant and law enforcement for the investigation.
Phase 4: The BreachForums Listing
The attackers weren't content with just accessing Vercel's systems. They claimed to have obtained a Vercel database access key and portions of source code. On BreachForums, a well-known cybercriminal marketplace, the threat actors posted the stolen Vercel database for sale at $2 million.
The listing claimed to include:
- Customer API keys
- Source code
- Database data from Vercel's internal systems
While the threat actor initially claimed to represent the ShinyHunters hacking group, ShinyHunters later told BleepingComputer they were not involved in this incident - suggesting either false attribution or an independent operator using the group's name for credibility.
The Scope: Hundreds of Organizations at Risk
Direct Vercel Impact
Vercel has contacted customers whose "non-sensitive" environment variables were compromised. The company emphasizes that:
- Next.js and Turbopack projects were not affected
- npm packages published by Vercel remain secure
- Services remain operational
- No evidence that "sensitive" environment variables were accessed
However, the full scope remains unknown. Vercel continues investigating "whether and what data was exfiltrated" and warns that the incident may affect "hundreds of users across many organizations" beyond just Vercel itself.
The Context AI Blast Radius
Context AI's security bulletin reveals the compromise extends beyond Vercel. The company admitted that "the unauthorized actor also likely compromised OAuth tokens for some of our consumer users."
The Context AI Office Suite Chrome extension (ID: omddlmnhcofjbnbflmjginpjjblphbgk) was removed from the Chrome Web Store on March 27, 2026 - suspiciously close to when the breach occurred. The OAuth app ID (110671459871-30f1spbu0hptbs60cb4vsmv79i7bbvqj.apps.googleusercontent.com) has been published as an indicator of compromise.
Any organization whose employees used Context AI's tools with Google Workspace accounts could potentially be affected. The supply chain implications are staggering.
Why This Breach Matters: The AI Tool Sprawl Problem
The Shadow AI Crisis
This breach exemplifies the shadow AI problem facing enterprises in 2026. Employees increasingly adopt AI-powered productivity tools without security review, often granting broad permissions to integrate with corporate systems.
Context AI was a small startup offering AI-powered document creation. It wasn't on any enterprise software approved list. Yet one employee's decision to try it created a pathway for attackers into a billion-dollar company's infrastructure.
The attack highlights how AI tool adoption has outpaced security governance. While organizations focus on securing sanctioned AI deployments, unsanctioned "shadow AI" tools create invisible attack vectors.
OAuth's Permission Problem
OAuth was designed to enable secure delegated access. In practice, it has become a security nightmare:
Overly Broad Permissions: Users routinely grant applications far more access than needed. Context AI's "Allow All" permission granted access to entire Google Workspace environments, not just specific documents.
Permission Persistence: OAuth tokens remain valid until explicitly revoked. Even after Context AI discovered the breach, existing tokens could still be used by attackers.
Lack of Enterprise Controls: Many organizations lack visibility into which OAuth applications employees have authorized, making incident response nearly impossible.
Third-Party Risk: When you grant an OAuth permission, you're not just trusting the application - you're trusting their entire security infrastructure. Context AI's breach became Vercel's breach through this transitive trust relationship.
The Supply Chain Domino Effect
Modern software development relies on complex supply chains of interconnected services. Vercel's breach demonstrates how a compromise at one link - Context AI - can cascade to every organization that trusted that link.
The attack pattern is becoming increasingly common:
- Compromise a vendor with broad OAuth access
- Harvest OAuth tokens from the vendor's infrastructure
- Use tokens to pivot into customer environments
- Exploit customer environments to access their customers
This "island hopping" through supply chain relationships allows attackers to maximize impact while minimizing detection risk.
Lessons from the Vercel Breach
For Security Teams
Audit OAuth Applications Immediately
Google Workspace administrators should immediately audit third-party app access for the compromised Context AI OAuth app ID: 110671459871-30f1spbu0hptbs60cb4vsmv79i7bbvqj.apps.googleusercontent.com
Check for the Context AI Chrome extension ID: omddlmnhcofjbnbflmjginpjjblphbgk
Any user who authorized this application should be considered potentially compromised.
Implement OAuth Governance
- Maintain an inventory of authorized OAuth applications
- Block high-risk applications by default
- Require security review for new OAuth authorizations
- Implement regular access reviews for existing OAuth grants
- Set up alerts for new OAuth authorizations in your environment
Review Environment Variable Security
Vercel's breach highlights the danger of storing credentials in "non-sensitive" environment variables. Security teams should:
- Audit all environment variables for credential exposure
- Implement secrets management solutions instead of environment variables
- Encrypt all credentials at rest
- Rotate credentials regularly
- Monitor for credential access anomalies
Assume Supply Chain Compromise
If your organization uses Vercel or Context AI, assume compromise until proven otherwise:
- Rotate all API keys and credentials
- Review access logs for anomalous activity
- Check for unauthorized deployments or configuration changes
- Verify integrity of deployed applications
For Developers
Pin Dependencies to Specific Versions
Ox Security recommends pinning Vercel-maintained npm packages to specific versions to prevent future supply chain attacks:
- Next.js
- SWR (stale-while-revalidate)
- Turborepo
- AI SDK
This prevents attackers from publishing malicious updates to compromised packages.
Rotate Secrets Immediately
Vercel CEO Guillermo Rauch advised customers to rotate any keys and credentials marked as "non-sensitive." Even if you weren't directly notified, proactive rotation is prudent.
Enable Multi-Factor Authentication
Vercel has emphasized the importance of 2FA for all accounts. Enable at least two authentication methods:
- Authenticator app
- Passkey
- Hardware security key
For Business Leaders
Invest in Shadow AI Governance
The Vercel breach demonstrates that shadow AI isn't just a productivity issue - it's a critical security risk. Organizations need:
- Clear policies on AI tool usage
- Security review processes for new AI tools
- Employee education on OAuth permission risks
- Technical controls to prevent unauthorized AI tool installation
Reevaluate Third-Party Risk Management
Traditional vendor security assessments often miss OAuth-based risks. Update your third-party risk program to include:
- OAuth permission scope analysis
- Assessment of vendor access to your infrastructure
- Monitoring for vendor security incidents
- Incident response plans for vendor compromises
Budget for Supply Chain Security
Supply chain attacks are becoming the norm, not the exception. Security budgets must account for:
- Supply chain monitoring tools
- Vendor security assessments
- Incident response capabilities
- Cyber insurance covering supply chain incidents
The Bigger Picture: AI as an Attack Vector
The Vercel breach represents a broader trend: AI tools becoming primary attack vectors. Context AI wasn't compromised because it was a high-value target - it was compromised because it provided a pathway to high-value targets.
Attackers are increasingly targeting:
AI Development Tools: Code assistants, AI IDEs, and development platforms that integrate with source control and deployment pipelines.
AI Productivity Suites: Document creation, presentation builders, and workflow automation tools that request broad workspace access.
AI Infrastructure Providers: Model hosting services, vector databases, and AI middleware that sit between applications and data.
Each of these represents a potential supply chain compromise waiting to happen.
FAQ: The Vercel-Context AI Breach
What exactly happened in the Vercel breach?
Attackers compromised Context AI, a third-party AI productivity tool, and stole OAuth tokens. These tokens allowed access to a Vercel employee's Google Workspace account, which was then used to access Vercel's internal systems and customer credentials stored in non-sensitive environment variables.
Was my Vercel deployment affected?
Vercel has contacted customers whose credentials were compromised. If you didn't receive notification, your credentials likely weren't directly accessed. However, all Vercel customers should rotate credentials as a precaution and monitor for suspicious activity.
What is Context AI?
Context AI is a startup that built AI-powered productivity tools including the "AI Office Suite" - a workspace for creating presentations, documents, and spreadsheets using AI agents. The company disclosed a security incident in March 2026 involving unauthorized access to their AWS environment and OAuth tokens.
How did the attackers get in?
The attack chain was: (1) Context AI employee downloaded game exploits, (2) Lumma Stealer malware infected their machine, (3) Credentials and OAuth tokens were stolen, (4) Attackers used OAuth tokens to access Vercel employee's Google account, (5) Attackers pivoted to Vercel internal systems.
What data was stolen?
Attackers accessed non-sensitive environment variables containing API keys, tokens, and database credentials. They claimed to have obtained a Vercel database access key and source code, which they listed for sale at $2 million on BreachForums.
Is the threat actor ShinyHunters?
The threat actor initially claimed to represent ShinyHunters, but ShinyHunters told BleepingComputer they were not involved. The true identity of the attackers remains unknown.
How can I check if my organization was affected?
Google Workspace administrators should check for the compromised OAuth app ID: 110671459871-30f1spbu0hptbs60cb4vsmv79i7bbvqj.apps.googleusercontent.com
Also check for the Context AI Chrome extension ID: omddlmnhcofjbnbflmjginpjjblphbgk
What should I do if we used Context AI?
Treat your environment as potentially compromised. Immediately:
- Revoke the Context AI OAuth application
- Rotate all credentials that may have been accessible
- Review access logs for anomalous activity
- Check for unauthorized data access or exfiltration
How can organizations prevent similar breaches?
Key preventive measures include:
- Implementing OAuth governance and monitoring
- Blocking high-risk third-party applications
- Encrypting all credentials at rest
- Regular credential rotation
- Shadow AI governance programs
- Supply chain security monitoring
Are Vercel's open source projects safe?
Vercel has confirmed that Next.js, Turbopack, and other open source projects were not affected. npm packages published by Vercel have been verified as uncompromised through collaboration with GitHub, Microsoft, npm, and Socket.
What is Vercel doing to prevent future breaches?
Vercel has implemented several security enhancements:
- Environment variable creation now defaults to "sensitive"
- Improved team-wide management of environment variables
- Enhanced activity logging with better filtering
- Clearer team invite emails
- Engagement with Mandiant for incident response
Conclusion: The New Reality of AI Supply Chain Risk
The Vercel breach is a wake-up call for every organization using AI tools. The attack demonstrates how a single employee's decision to try a productivity app can cascade into a multi-million dollar security incident affecting hundreds of organizations.
As AI tools proliferate, so do the attack vectors they create. Every OAuth authorization is a potential supply chain compromise. Every third-party AI integration is a trust relationship that can be weaponized.
The security community has long warned about shadow IT. Shadow AI is the same problem amplified by the speed of AI adoption and the broad permissions these tools require. Organizations that don't implement governance now will find themselves the next Vercel - wondering how a small AI tool they never approved led to a devastating breach.
The lesson is clear: In the age of AI-powered productivity, security can't be an afterthought. It must be embedded in every tool evaluation, every OAuth authorization, and every employee training program.
Because the next Context AI is already out there. And someone in your organization is about to click "Allow All."
Don't let that click become your $2 million mistake.
Stay ahead of emerging threats. Subscribe to the Hexon.bot newsletter for weekly AI security insights.
Related Reading: