Futures
Access hundreds of perpetual contracts
CFD
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Promotions
AI
Gate AI
Your all-in-one conversational AI partner
Gate AI Bot
Use Gate AI directly in your social App
GateClaw
Gate Blue Lobster, ready to go
Gate for AI Agent
AI infrastructure, Gate MCP, Skills, and CLI
Gate Skills Hub
10K+ Skills
From office tasks to trading, the all-in-one skill hub makes AI even more useful.
GateRouter
Smartly choose from 40+ AI models, with 0% extra fees
Tether's "Psycho-Historical" Farce: Printing $USDT to Fund AI Small Models—Can They Fool You This Time?
Buddy, today let’s talk about something interesting. Tether, the one that prints $USDT, has recently launched an AI project called QVAC.
You read that right, a stablecoin giant suddenly diving into artificial intelligence. And they’re starting off by using science fiction novels as theoretical backing, calling their foundational large model “rooted in the principles of psychohistory.”
What’s the concept of psychohistory? In Asimov’s “Foundation,” Hari Seldon uses mathematics and statistics to predict human group behavior, aiming to shorten the dark age after the empire’s collapse. Basically, it’s packaging grand narratives around commercial ambitions.
Market observers point out that this move by Tether is essentially replicating their stablecoin approach into the AI field.
Let’s see how Tether makes money. $USDT converts global offshore dollar demand into reserve assets mainly in U.S. short-term government bonds. Q1 2026 data shows a net profit of $1.04 billion, with reserve buffers of $8.23 billion, and directly and indirectly holding about $141 billion in U.S. short-term Treasuries.
How do they spend this money? In January this year, Tether spent to buy 8,888 BTC, turning interest income into Bitcoin holdings. Now, they want to extend this asset allocation logic into AI.
The idea is straightforward: use the cash flow from stablecoins to fund AI infrastructure. Transitioning from private dollar liquidity issuers to private digital infrastructure builders.
QVAC’s positioning is an “infinite stable intelligent platform,” mainly focusing on decentralized intelligent systems that run locally. In other words, enabling you to run AI locally, keep your data, and not hand over control to centralized servers.
This concept is consistent with $USDT’s philosophy. Fund flows are permissionless, user data is under their control, and AI runs locally nearby.
What’s the cost? Convenience. Cloud-based large models are more powerful, but come with platform risks, pricing risks, regulatory risks, network latency, and data routing issues. Local AI models sacrifice some performance but gain ownership, privacy, and stable availability.
This trade-off is familiar to the crypto community. Self-hosting isn’t as convenient as exchanges—until an exchange collapses, then you realize its value. The same applies to local AI.
QVAC has chosen a completely different track: deployability, privacy protection, low latency, composability, and independence from a single platform.
Technically, it doesn’t compete with OpenAI or Google in model capability but aims to be the underlying OS for edge AI.
What is VAC Fabric? Tether says it leverages Vulkan and Metal backends to enable model fine-tuning on mainstream consumer hardware, including Android phones, Apple chips, and ordinary computers. It also uses dynamic chunking to handle mobile device VRAM limitations.
Sounds impressive, but the core question is: why would developers use your framework? The open-source AI ecosystem already has many mature components: Llama, Qwen, Mistral, Gemma, DeepSeek, llama.cpp, Ollama, and more.
QVAC bets that developers need a complete edge framework that handles model loading, inference, speech recognition, OCR, translation, image generation, retrieval, P2P distribution, and local fine-tuning through a unified interface.
That’s a big gamble.
Then, QVAC released its first benchmark product: MedPsy. It’s a medical health language model, with versions of 1.7 billion and 4 billion parameters.
Official data is quite explosive. MedPsy-1.7B scores an average of 62.62 across seven medical benchmarks, surpassing Google’s MedGemma-1.5-4B-it’s 51.20, despite having less than half the parameters. MedPsy-4B scores an average of 70.54, slightly ahead of MedGemma-27B-text-it’s 69.95, with only about a quarter of its parameters.
HealthBench results show an even bigger gap: MedPsy-4B scores 74.00, while MedGemma-27B only gets 65.00.
If these results can be independently reproduced, it proves one thing: in specific verticals, lightweight small models can challenge massive cloud-based systems.
Note, I emphasize “if.”
Currently, all impressive data comes from QVAC’s own releases. The key question is external validation: Are the training data contaminated? How broad is the coverage? How are prompts constructed? How much do teacher models influence results?
The training process is clear: using the Tongyi Qianwen 3 as the backbone, multi-stage supervised fine-tuning, generating over 30 million synthetic data points, with Baichuan M3-235B large model serving as the teacher for long-text reasoning supervision.
Quantization deployment is well handled. The official has released GGUF quantized versions, with Q4_K_M quantization reducing model size by 69%, with less than 1 point average loss. The 4-billion-parameter model is only 2.72GB, and the 1.7-billion-parameter version is just 1.28GB, making it easy to deploy locally.
They also openly state: for text interaction only, limited to English, not suitable for clinical emergencies, and may hallucinate.
The medical field has a strong demand for local inference. In scenarios like medical records and diagnostic assistance, data cannot leave hospitals. MedPsy’s approach is correct, but only external researchers reproducing benchmarks and testing in real clinical workflows can truly prove its strength.
Ultimately, it’s a battle between convenience and autonomous control.
Cloud AI is extremely easy to use: users open an app, input commands, get results—no need to worry about model weights, device VRAM, or quantization parameters. The platform handles all the complex tech, enabling rapid growth.
QVAC requires users to take on more operational responsibility, but gains offline availability, data privacy, and independence from APIs.
Tether’s ultimate logic is coherent: funds ($USDT), computing power (QVAC), and intelligent agents follow the same sovereign design paradigm.
Of course, the decentralized narrative isn’t perfect.
Inference layer is decentralized: users download models themselves, run locally, and sensitive data stays on their devices. Relying on Holepunch network architecture, supporting delegated inference and decentralized model distribution.
But governance remains centralized: Tether fully funds, manages naming, marketing, and the flagship applications, model system, and SDK roadmap are led by a single company.
The entire ecosystem still needs to gradually establish a distributed control mechanism in areas like default node registration, release channels, security standards, model access, and community governance.
Right now, QVAC’s credibility entirely depends on third-party reproduction results.
If external benchmarks for MedPsy can be replicated, Tether will truly realize its “smart asset reserve” concept: lightweight, open-source, locally deployable models for vertical domains, enough to compete with cloud giants in highly sensitive sectors.
Even if benchmark scores narrow or reverse, the infrastructure value of QVAC remains valid; only the model performance narrative will weaken.
The industry’s ultimate question remains: does extreme convenience lead to concentrated power, or does autonomous control require operational costs?
Asimov’s psychohistory studies the evolution of complex large systems under pressure. Tether reinterprets it: how infrastructure can resist centralized monopolies.
The sci-fi narrative is grand, the technology is still in early stages, but the overall strategic logic is clear and coherent.
Relying on the continuous cash flow of the world’s largest stablecoin, Tether is building an AI architecture centered on local operation, peer-to-peer networks, open-source tools, and lightweight edge models.
No one doubts that stablecoin giants have the strength to deploy AI.
The real question is: can QVAC develop sufficiently powerful models and infrastructure to make users willing to accept moderate operational burdens for local autonomous control?
MedPsy is the first quantifiable threshold.
Third-party reproduction results will determine whether QVAC’s psychohistory narrative is a sci-fi metaphor or a foundational architecture that truly enters the mainstream edge AI track.
Follow me for more real-time analysis and insights into the crypto market!