Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Promotions
AI
Gate AI
Your all-in-one conversational AI partner
Gate AI Bot
Use Gate AI directly in your social App
GateClaw
Gate Blue Lobster, ready to go
Gate for AI Agent
AI infrastructure, Gate MCP, Skills, and CLI
Gate Skills Hub
10K+ Skills
From office tasks to trading, the all-in-one skill hub makes AI even more useful.
GateRouter
Smartly choose from 40+ AI models, with 0% extra fees
The training cost is reduced by 16 times, and the ultimate compression is 42 times! Open source text generation image model
Original source: AIGC Open Community
Stable Diffusion is currently one of the most powerful open source text-generated image diffusion models, but it has a big disadvantage for small and medium-sized enterprises and individual developers that do not have A100 or H100, requiring high training costs.
In order to solve this pain point, the Wuerstchen open source model adopts a new technical architecture to achieve 42 times ultimate compression while ensuring image quality. ** Taking the 512x512 size training image as an example, Stable Diffusion1.4 requires 150,000 hours of GPU training time, while Wuerstchen only requires 9,000 hours, and the training cost is reduced by 16 times**.
Even if the image resolution is as high as 1536, Wuerstchen only requires 24,602 hours, and the training cost is still 6 times cheaper than Stable Diffusion.
Therefore, this open source product is conducive to developers who do not have huge computing power to try the diffusion model, and at the same time, they can explore better training methods on this basis.
Github:
paper:
Wuerstchen brief introduction
The Wuerstchen diffusion model adopts a method that works in the highly compressed latent space of the image. This is one of the reasons why its training cost is lower than Stable Diffusion.
Compressing data can reduce the cost of training and inference by orders of magnitude. For example, training on 1024×1024 images is definitely much more expensive than training on 32×32 images. Usually the compression range used in the industry is about 4-8 times.
And Wuerstchen pushed compression to the extreme through a brand-new technical architecture, achieving 42 times space compression, which is an unprecedented technological breakthrough! Because once the compression exceeds 16 times, ordinary methods cannot achieve image reconstruction at all.
Wuerstchen extreme compression principle
Wuerstchen’s extreme compression method is divided into three stages: A, B, and C: Stage A) performs initial training, and uses vector quantization generative adversarial network (VQGAN) to create a discretized latent space and map the data to a pre-set This compact representation of points in a defined, smaller set helps model learning and inference speed;
Phase B) further compresses, using an encoder to project the image into a more compact space, and a decoder to try to reconstruct the latent representation of the VQGAN from the encoded image.
And a label predictor based on the Paella model is used to accomplish this task. This model is based on the representation of the encoded image and can be trained using a smaller number of sampling steps, which is a huge help in improving computing power efficiency.
Wuerstchen accepted image training data with resolutions between 1024x1024 and 1536x1536, and the output image quality is very stable. Even non-equivalent images such as 1024x2048 can still achieve good results.
Wuerstchen generates picture display
According to the case presented by Wuerstchen, the model’s ability to understand text is very good, and the quality effect it generates is comparable to the strongest open source diffusion models such as Stable Diffusion.
Real photo of an eagle wearing a white coat