Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Launchpad
Be early to the next big token project
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Cursor releases Composer 2 technical details: Based on Kimi K2.5, with model updates every five hours
How I Clarified This Matter
I reviewed the official arXiv paper, blog posts, and discussions on social media, focusing mainly on two questions: What are the model architecture and capability boundaries of Composer 2? How is the training feedback loop based on production data and the five-hour update cycle specifically implemented?
Official materials clarified a few things: The foundational model comes from Moonshot AI’s Kimi K2.5; further pre-training and large-scale reinforcement learning were conducted based on that; the training method is similar to PULSE, claiming to achieve efficient cross-datacenter training at a scale of 1T parameters.
There was a small episode: Cursor initially did not disclose the identity of the foundational model, only revealing it after being questioned by the community, explaining that the self-developed training part accounted for about 75% of the computing power. This indicates they are taking a hybrid approach of “open-source/external base + self-developed overlay.”
What Happened
Why This Matter is Worth Attention
My view: Real-time reinforcement learning directly moves the “training-deployment” cycle into the production environment, significantly shortening the feedback loop and resulting in quantifiable online benefits.
Regarding production data vs. synthetic data:
Regarding engineering rhythm:
Regarding competition:
Data and Controversy
In terms of functionality: Supports semantic search, shell execution, multi-step tasks, suitable for long conversations and complex coding workflows.
Training scale: Adopted PULSE’s method, achieving cross-data center training at a scale of 1T parameters, emphasizing throughput and cost efficiency.
Disclosure controversy: The foundational model was initially not stated as Kimi, and was only acknowledged after being questioned. The official emphasized that self-developed training accounted for about 75%.
Impact on the Industry
Risks and Limitations
Importance Assessment
My judgment: This is an “early but effective” engineering paradigm. The most direct beneficiaries are developers and team leaders: The sooner they establish a production data loop and high-frequency evaluation deployment process, the more they can gain an advantage in product iteration speed and cost-effectiveness.