💞 #Gate Square Qixi Celebration# 💞
Couples showcase love / Singles celebrate self-love — gifts for everyone this Qixi!
📅 Event Period
August 26 — August 31, 2025
✨ How to Participate
Romantic Teams 💑
Form a “Heartbeat Squad” with one friend and submit the registration form 👉 https://www.gate.com/questionnaire/7012
Post original content on Gate Square (images, videos, hand-drawn art, digital creations, or copywriting) featuring Qixi romance + Gate elements. Include the hashtag #GateSquareQixiCelebration#
The top 5 squads with the highest total posts will win a Valentine's Day Gift Box + $1
GPT-4 is a mixed model of 8x220 billion parameters? This gossip went crazy today
Source: Heart of the Machine
"The parameter volume of GPT-4 is as high as 100 trillion." I believe many people still remember the "heavy" news that swiped the screen at the beginning of this year and a chart that was spread virally.
In fact, many people believe and spread such rumors because the AI community has been increasing the parameter size of AI models in recent years. The Switch Transformer released by Google in January 2021 raised the parameters of the AI large model to 1.6 trillion. Since then, many institutions have successively launched their own trillion-parameter large models. Based on this, people have every reason to believe that GPT-4 will be a huge model with trillions of parameters, and 100 trillion parameters is not impossible.
Although Sam Altman's rumor helped us get rid of a wrong answer, the OpenAI team behind him has been tight-lipped about the real parameters of GPT-4, and even the official technical report of GPT-4 did not disclose any information.
Until recently, this mystery was suspected to be pierced by the "genius hacker" George Hotz.
George Hotz is famous for cracking the iPhone at the age of 17 and hacking the Sony PS3 at the age of 21. He is currently the boss of a company (comma.ai) that develops automatic driving assistance systems.
He was recently interviewed by an AI tech podcast called Latent Space. In the interview, he talked about GPT-4, saying that GPT-4 is actually a hybrid model. Specifically, it uses an ensemble system of 8 expert models, each with 220 billion parameters (slightly more than GPT-3's 175 billion parameters), and these models have been trained on different data and task distribution training.
As for the future trend, he believes that people will train smaller models and improve performance through long-term fine-tuning and discovering various tricks. He mentioned that compared with the past, the training effect has been significantly improved, although the computing resources have not changed, which shows that the improvement of the training method has played a big role.
At present, George Hotz's "breaking news" about GPT-4 has been widely spread on Twitter.