Futures
Access hundreds of perpetual contracts
CFD
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Promotions
AI
Gate AI
Your all-in-one conversational AI partner
Gate AI Bot
Use Gate AI directly in your social App
GateClaw
Gate Blue Lobster, ready to go
Gate for AI Agent
AI infrastructure, Gate MCP, Skills, and CLI
Gate Skills Hub
10K+ Skills
From office tasks to trading, the all-in-one skill hub makes AI even more useful.
GateRouter
Smartly choose from 40+ AI models, with 0% extra fees
Xiaomi Open Sources OmniVoice: A voice cloning model covering 646 languages, trained solely on open-source data to outperform commercial systems
According to Beating Monitoring, Xiaomi AI Laboratory’s new generation Kaldi team has open-sourced OmniVoice, a zero-shot voice cloning TTS (text-to-speech) model supporting 646 languages. It can clone voice timbre with just a few seconds of reference audio and works across languages: given a Chinese recording, the model can speak Japanese, Korean, or other languages with the same voice. All code, weights, and training data are open source under the Apache-2.0 license.
Architecturally, OmniVoice takes a minimalist approach. The entire model consists of a single bidirectional Transformer that directly maps from text to multi-codebook acoustic tokens (discrete sound encodings), without the two-stage pipeline of first converting to semantic tokens then to acoustic tokens. Two key design choices support this simple structure: a full-codebook random masking strategy to improve training efficiency, and initialization with pre-trained parameters from large language models to enhance pronunciation accuracy. Inference runs 40 times faster than real-time, using PyTorch directly without additional optimization.
All training data comes from 50 open-source speech datasets, totaling 580k hours after noise reduction and quality filtering. Low-resource languages are trained with dynamic upsampling to ensure effective training. In tests across 24 languages, OmniVoice’s voice similarity and intelligibility surpass several commercial systems. In tests across 102 languages, intelligibility approaches or even exceeds that of real recordings. Even languages with less than 10 hours of training data can be synthesized effectively.
In addition to voice cloning, the model supports customizing voice timbre via text descriptions (such as “male, middle-aged, very low pitch” or “female, young, Sichuan dialect”), automatic noise reduction from reference audio, insertion of tone symbols like laughter or sighs, and pronunciation correction for Chinese and English polyphones and proper nouns.