Qwen team has launched the open‑weight Qwen3.5‑397B‑A17B model, introducing major advances in multimodal performance, reinforcement learning, and training efficiencyQwen team has launched the open‑weight Qwen3.5‑397B‑A17B model, introducing major advances in multimodal performance, reinforcement learning, and training efficiency

Qwen Rolls Out New Vision‑Language Model To Advance Coding, Reasoning, And Multimodal AI Performance

2026/02/16 22:30
3 min read
Qwen Rolls Out New Vision‑Language Model To Advance Coding, Reasoning, And Multimodal AI Performance

Alibaba Cloud’s Qwen team has introduced the first model in its new Qwen3.5 series, unveiling the open‑weight Qwen3.5‑397B‑A17B. 

Positioned as a native vision‑language system, the model delivers strong performance across reasoning, coding, agent tasks, and multimodal understanding, reflecting a significant advance in the company’s large‑scale AI development efforts. 

The model is built on a hybrid architecture that combines linear attention through Gated Delta Networks with a sparse mixture‑of‑experts design, enabling high efficiency during inference. Although the full system contains 397 billion parameters, only 17 billion are activated for each forward pass, allowing it to maintain high capability while reducing computational cost. The release also expands language and dialect coverage from 119 to 201, broadening accessibility for users and developers worldwide.

Qwen3.5 Marks A Major Leap In Reinforcement Learning And Pretraining Efficiency

The Qwen3.5 series introduces substantial gains over Qwen3, driven largely by extensive reinforcement learning scaling across a wide range of environments. Rather than optimizing for narrow benchmarks, the team focused on increasing task difficulty and generalizability, resulting in improved agent performance across evaluations such as BFCL‑V4, VITA‑Bench, DeepPlanning, Tool‑Decathlon, and MCP‑Mark. Additional results will be detailed in an upcoming technical report.

Pretraining improvements span power, efficiency, and versatility. Qwen3.5 is trained on a significantly larger volume of visual‑text data with strengthened multilingual, STEM, and reasoning content, enabling it to match the performance of earlier trillion‑parameter models. Architectural upgrades—including higher‑sparsity MoE, hybrid attention, stability refinements, and multi‑token prediction—deliver major throughput gains, particularly at extended context lengths of 32k and 256k tokens. The model’s multimodal capabilities are strengthened through early text‑vision fusion and expanded datasets covering images, STEM materials, and video, while a larger 250k vocabulary improves encoding and decoding efficiency across most languages.

The infrastructure behind Qwen3.5 is designed for efficient multimodal training. A heterogeneous parallelism strategy separates vision and language components to avoid bottlenecks, while sparse activation enables near‑full throughput even on mixed text‑image‑video workloads. A native FP8 pipeline reduces activation memory by roughly half and increases training speed by more than 10 percent, maintaining stability at massive token scales. 

Reinforcement learning is supported by a fully asynchronous framework capable of handling models of all sizes, improving hardware utilization, load balancing, and fault recovery. Techniques such as FP8 end‑to‑end training, speculative decoding, rollout router replay, and multi‑turn rollout locking help maintain consistency and reduce gradient staleness. The system is built to support large‑scale agent workflows, enabling seamless multi‑turn interactions and broad generalization across environments.

Users can interact with Qwen3.5 through Qwen Chat, which offers Auto, Thinking, and Fast modes depending on the task. The model is also available through Alibaba Cloud’s ModelStudio, where advanced features such as reasoning, web search, and code execution can be enabled through simple parameters. Integration with third‑party coding tools allows developers to adopt Qwen3.5 into existing workflows with minimal friction.

According to the Qwen team, Qwen3.5 establishes a foundation for universal digital agents through its hybrid architecture and native multimodal reasoning. Future development will focus on system‑level integration, including persistent memory for cross‑session learning, embodied interfaces for real‑world interaction, self‑directed improvement mechanisms, and economic awareness for long‑term autonomous operation. The objective is to move beyond task‑specific assistants toward coherent, persistent agents capable of managing complex, multi‑day objectives with reliable, human‑aligned judgment.

The post Qwen Rolls Out New Vision‑Language Model To Advance Coding, Reasoning, And Multimodal AI Performance appeared first on Metaverse Post.

Market Opportunity
Major Logo
Major Price(MAJOR)
$0,07965
$0,07965$0,07965
+1,89%
USD
Major (MAJOR) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.
Tags:

You May Also Like

Today’s Biggest Crypto Movers: Winners & Losers

Today’s Biggest Crypto Movers: Winners & Losers

Today's Biggest Crypto Movers: Winners & Losers Crypto Market Update 📊 Today's top movers show mixed signals. Let's dive into which cryptos made the biggest waves
Share
Blockchainmagazine2026/02/18 13:00
Altcoins Poised to Benefit from SEC’s New ETF Listing Standards

Altcoins Poised to Benefit from SEC’s New ETF Listing Standards

The post Altcoins Poised to Benefit from SEC’s New ETF Listing Standards appeared on BitcoinEthereumNews.com. On Wednesday, the US SEC (Securities and Exchange Commission) took a landmark step in crypto regulation, approving generic listing standards for spot crypto ETFs (exchange-traded funds). This new framework eliminates the case-by-case 19b-4 approval process, streamlining the path for multiple digital asset ETFs to enter the market in the coming weeks. Grayscale’s Multi-Crypto Milestone Sponsored Grayscale secured a first-mover advantage as its Digital Large Cap Fund (GDLC) received approval under the new listing standards. Products that will be traded under the ticker GDLC include Bitcoin, Ethereum, XRP, Solana, and Cardano. “Grayscale Digital Large Cap Fund $GDLC was just approved for trading along with the Generic Listing Standards. The Grayscale team is working expeditiously to bring the FIRST multi-crypto asset ETP to market with Bitcoin, Ethereum, XRP, Solana, and Cardano,” wrote Grayscale CEO Peter Mintzberg. The approval marks the US’s first diversified, multi-crypto ETP, signaling a shift toward broader portfolio products rather than single-asset ETFs. Bloomberg’s Eric Balchunas explained that around 12–15 cryptocurrencies now qualify for spot ETF consideration. However, this is contingent on the altcoins having established futures trading on Coinbase Derivatives for at least six months. Sponsored This includes well-known altcoins like Dogecoin (DOGE), Litecoin (LTC), and Chainlink (LINK), alongside the majors already included in Grayscale’s GDLC. Altcoins in the Spotlight Amid New Era of ETF Eligibility Several assets have already met the key condition, regulated futures trading on Coinbase. For example, Solana futures launched in February 2024, making the token eligible as of August 19. “The SEC approved generic ETF listing standards. Assets with a regulated futures contract trading for 6 months qualify for a spot ETF. Solana met this criterion on Aug 19, 6 months after SOL futures launched on Coinbase Derivatives,” SolanaFloor indicated. Sponsored Crypto investors and communities also identified which tokens stand to gain. Chainlink…
Share
BitcoinEthereumNews2025/09/18 13:46
Stake vs BC.Game: Best Crypto Casino in 2026?

Stake vs BC.Game: Best Crypto Casino in 2026?

Choosing the best crypto casino isn’t simple. Compare Stake and BC.Game with our detailed breakdown to see which platform fits your style best. The post Stake vs
Share
Bitcoinchaser2026/02/18 12:40