Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Promotions
AI
Gate AI
Your all-in-one conversational AI partner
Gate AI Bot
Use Gate AI directly in your social App
GateClaw
Gate Blue Lobster, ready to go
Gate for AI Agent
AI infrastructure, Gate MCP, Skills, and CLI
Gate Skills Hub
10K+ Skills
From office tasks to trading, the all-in-one skill hub makes AI even more useful.
GateRouter
Smartly choose from 40+ AI models, with 0% extra fees
Ultraman Preview: OpenAI's new cybersecurity model GPT 5.5-Cyber to be unveiled in a few days, clashing with Claude Mythos
OpenAI has announced it will roll out a dedicated cybersecurity model, GPT-5.5-Cyber, to take on experts in the cybersecurity field with Anthropic’s tightly controlled defense strategies. Altman once predicted that a world-shaking cyberattack is highly likely to occur in 2026.
OpenAI GPT-5.5-Cyber to Make Its Debut in a Few Days
OpenAI CEO Sam Altman (Sam Altman) hinted earlier today (4/30) that in the coming days, it will launch the new-generation cybersecurity model GPT-5.5-Cyber for experts in the cybersecurity field to use. He said the team will work with the ecosystem and the government to find reliable access methods to ensure the security of enterprises and critical infrastructure.
In April this year, Altman, during an interview with Axios founder Mike Allen, also predicted that a cyberattack highly likely to rock the world could happen in 2026.
Whether his remarks accurately reflect the evolving threat landscape has continued to be debated, and recently Anthropic launched the Claude Mythos model that can autonomously identify software vulnerabilities—further intensifying the debate and drawing attention from the U.S. government.
OpenAI Plans to Push Cybersecurity Tools to Governments at All Levels
The divergence between OpenAI and Anthropic in their defense approach reflects a broader debate in the AI industry.
According to CNN, until recently, OpenAI’s cybersecurity trust access program was limited to a small number of partners, and it is now opening permissions to all reviewed government levels—from federal agencies to local governments—so that approved organizations can use specialized versions of models with fewer protective restrictions.
Sasha Baker, OpenAI’s director of national security policy, said that OpenAI does not believe it should be the sole decision-maker determining tool access permissions and the top priorities.
Two Major AI Giants Disagree on Defense Strategies: Democratization vs. Strict Control
Anthropic’s Mythos model can identify and exploit software vulnerabilities. Based on potential harm, the company has been gradually promoting it through a tightly controlled “Glass Wing” program, and it is working with government representatives.
On security, Anthropic argues that only a slow and cautious approach can reduce an arms race spurred by hackers using AI, but OpenAI plans to open up its models broadly.
Baker said network defense capabilities must be democratized so that everyone benefits; restricting access to only the top 50 companies in Fortune is not enough. She stressed that this is an opportunity for companies to patch vulnerabilities before they fall into the hands of malicious actors.
Image source: Getty Images/ANTHONY WALLACE/AFP Sasha Baker, Director of National Security Policy at OpenAI
OpenAI Actively Cooperates with the U.S. to Form an Action Plan for the Intelligence Era
Recently, OpenAI held a hands-on workshop in Washington, where Baker revealed that attendees included representatives from the Pentagon, the White House, the U.S. Department of Homeland Security, and DARPA, who jointly tested the security capabilities of the new model. They also plan to return to Washington in a few weeks to collect feedback.
In addition, OpenAI is rolling out an action plan to coordinate cyber defense between governments and enterprises in the intelligence era. The company expects that over the next few days it will introduce new security features for ChatGPT accounts and provide tools to help the public improve their personal cybersecurity habits.
Devil or Savior? AI Giants Play the Doomsday-Crisis Card
However, many AI companies frequently warn that the technology could lead to a doomsday scenario, which has sparked doubts in academia.
In an interview with BBC, Shannon Vallor, a professor of ethics at the University of Edinburgh, said that the companies’ fear-marketing strategies have worked: they have turned their products into things that end the world, without harming the companies or limiting their power. Instead, it makes the public feel that the only entities capable of providing protection are these companies themselves.
She said that utopia and apocalypse are two sides of the same coin: “No matter which scenario it is, the scale is too vast and steeped in myth, making it feel as though mechanisms like regulation, governance, or legal proceedings are simply unable to do anything.”
This leads people to believe that the only thing they can do is wait for results—to see whether these technologies ultimately become devils that bring an end to civilization, or saviors that grant us utopia. Even the name “Mythos” (myth) seems designed to inspire a sense of religious awe.
Further reading:
Need New Policies in the AI Era! OpenAI Unveils 4 Major Proposals: A Three-Day Weekend and a Robot Tax