Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Still buying AI transfer stations on Taobao? Claude Code source code whistleblower: At least dozens have been poisoned
Claude Code source code leak incident whistleblower’s latest research reveals hidden cybersecurity risks in commercial AI intermediary stations. Real-world tests find that some intermediary stations steal credentials, wallet private keys, or inject malicious code, turning them into nodes for supply-chain attacks.
Claude Code source code leak incident whistleblower reveals cybersecurity risks of AI intermediary stations
Recently, a research paper titled “Your Agent Is Mine” has been published. One of the authors is Chaofan Shou, the whistleblower who was the first to expose the Claude Code source code leak incident.
This paper is the first to conduct a systematic security threat study of third-party API routers for large language models (LLMs)—commonly known as intermediary stations—and reveals that these types of intermediary stations may become nodes in supply-chain attacks.
What is an AI intermediary station?
Because calling LLMs consumes a large number of tokens and results in high computing costs, AI intermediary stations can help customers significantly reduce expenses by caching and reusing repeated problem background explanations.
At the same time, intermediary stations also have an automatic model allocation function: they can dynamically switch between different billing standards and performance models based on the difficulty level of a user’s questions, and they can automatically switch to backup models when a single model server goes offline to ensure overall service connection stability.
Intermediary stations are especially popular in China, because the country cannot directly access certain specific overseas AI products, and with enterprises’ demand for localized billing, intermediary stations have become an important bridge connecting upstream model providers and downstream developers. Platforms such as OpenRouter and SiliconFlow also fall within this category.
However, despite appearing to lower costs and technical barriers, intermediary stations hide a significant cybersecurity risk behind the scenes.
Image source: Research paper reveals supply-chain attack risks of AI intermediary stations
AI intermediary stations have full access rights, making them supply-chain attack vulnerabilities
The paper points out that intermediary stations operate at the application layer of the network architecture and have full plaintext read access to JSON payload data during transmission.
Because there is no end-to-end encryption integrity verification between the client and the upstream model provider, the intermediary station can easily inspect and tamper with API keys, system prompt words, and tool invocation parameters in the model’s output.
The research team notes that as early as March 2026, the well-known open-source router LiteLLM was already targeted by a dependency confusion attack, allowing attackers to inject malicious code into the request processing pipeline, highlighting the vulnerability of this component.
Real-world tests show dozens of AI intermediary stations exhibit malicious behavior
The research team actually purchased 28 paid intermediary stations on platforms such as Taobao, Xianyu, and Shopify, and collected 400 free intermediary stations from public communities for in-depth testing. The real-world results found that a total of 1 paid intermediary station and 8 free intermediary stations would proactively inject malicious code.
Among the tested samples of free intermediary stations, 17 intermediary stations attempted to use AWS decoy credentials set up by the researchers, and 1 intermediary station directly stole encrypted cryptocurrency from the researchers’ Ethereum wallets.
The research data further shows that as long as intermediary stations reuse leaked upstream credentials, or route traffic to nodes with weaker security defenses, even intermediary stations that originally appeared normal can be forced into the same attack surface.
During the poisoning tests, the research team found that these implicated nodes processed more than 2.1 billion tokens in total, and exposed 99 real credentials across 440 sessions; furthermore, 401 sessions were in a fully autonomous operating state, enabling attackers to inject malicious payloads directly and easily, without needing complex trigger conditions.
Image source: Research paper tested over 400 intermediary stations; the results reveal that dozens of AI intermediary stations have malicious behavior
Four core attack methods unveiled
The paper categorizes the attack behaviors of malicious intermediary stations into two main core categories and two types of adaptive evasion variants.
To evade routine cybersecurity detection, attackers further evolved a dependency confusion-based target injection method: they specifically alter the package names in software installation commands, replacing the legitimate package with a malicious package of the same name or easily confused name that has been published in advance on a public registry, thereby establishing a persistent supply-chain backdoor within the target system.
Another method is conditional delivery, where malicious actions are triggered only under specific conditions—for example, when the number of requests exceeds 50, or when a system is detected to be operating in a fully autonomous mode (YOLO mode)—so that the attack can evade security review tests limited to a finite number of attempts.
Three feasible defense measures
In response to supply-chain attacks involving poisoning in AI intermediary stations, the paper proposes three feasible defense measures:
Call on upstream model providers to establish cryptographic verification mechanisms
Although client-side defense mechanisms can reduce some risks at this stage, they cannot fundamentally resolve the vulnerability in source identity authentication. As long as the intermediary station’s modification behavior does not trigger abnormal alerts on the client side, attackers can still easily change the semantics of program execution and carry out sabotage.
To fully ensure the security of the AI agent ecosystem, the final reliance must be on response mechanisms that support cryptographic verification provided by upstream model providers. Only by rigorously cryptographically binding the model-produced results to the final execution commands on the client can end-to-end data integrity be ensured and supply-chain risks from intermediary tampering be comprehensively prevented.
Further reading:
OpenAI’s Mixpanel got into trouble! Leads to some users’ personal data leaking—watch out for phishing emails
A copy-paste error caused $50 million to evaporate! Crypto address poisoning scams are back—how to prevent them