AI News 10月09日 17:38
思科发布新款AI数据中心路由器,应对互联瓶颈
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

思科推出8223路由系统,旨在解决AI工作负载跨多个数据中心连接的挑战。该系统搭载的Silicon One P200芯片提供业界领先的51.2 Tb/s带宽和深度缓冲能力,以应对AI训练产生的大规模、爆发式流量。相较于传统路由器,8223系统在速度、智能缓冲和能效方面表现更优,并支持长距离连接,为AI基础设施的规模化部署提供关键支持。该技术已获得微软、阿里云等巨头的初步应用,并因其可编程性为未来AI网络演进预留了空间。

💡 **AI基础设施规模化挑战催生“Scale-Across”新模式:** 随着AI模型训练和复杂AI系统运行对计算资源的需求激增,单个数据中心已难以满足其空间、电力和散热的极限。为此,行业正转向“Scale-Across”模式,将AI工作负载分布到地理位置分散的多个数据中心。然而,这导致数据中心间的互联成为关键瓶颈,传统网络设备难以应对AI工作负载特有的海量、爆发式流量模式。

🚀 **思科8223系统专为AI互联设计,突破性能瓶颈:** 思科发布的8223路由系统采用全新的Silicon One P200芯片,提供高达51.2 Tb/s的带宽和200亿+包/秒的处理能力,并具备深度缓冲技术。这使其能够有效吸收AI训练产生的流量峰值,防止网络拥堵,确保GPU集群的连续高效运行。同时,该系统在三机架单元(3RU)的紧凑设计下实现了高能效,并支持长达1000公里的800G相干光连接,为大规模分布式AI部署提供强大支撑。

🌐 **多方巨头布局AI数据中心互联,市场竞争激烈:** 思科并非唯一进入AI数据中心互联领域的厂商,Broadcom和Nvidia此前已推出各自的解决方案。思科凭借其在企业和运营商网络的深厚积累、成熟的Silicon One产品组合以及与主要超大规模云服务提供商的合作关系,在该市场中具备竞争优势。其8223系统支持SONiC和未来IOS XR,提供灵活的部署选项,以应对客户避免供应商锁定的需求。

🔒 **可编程性和安全性为未来AI网络奠定基础:** 8223系统搭载的可编程P200芯片能够适应不断发展的AI网络协议和标准,无需硬件更换即可进行升级,从而保护了客户的长期投资。此外,该系统还内置了针对未来量子计算威胁的线路速率加密功能,并通过与思科观测平台集成,提供全面的网络监控能力,确保分布式AI基础设施的安全和可靠运行。

Cisco has entered an increasingly competitive race to dominate AI data centre interconnect technology, becoming the latest major player to unveil purpose-built routing hardware for connecting distributed AI workloads across multiple facilities.

The networking giant unveiled its 8223 routing system on October 8, introducing what it claims is the industry’s first 51.2 terabit per second fixed router specifically designed to link data centres running AI workloads. 

At its core sits the new Silicon One P200 chip, representing Cisco’s answer to a challenge that’s increasingly constraining the AI industry: what happens when you run out of room to grow.

A three-way battle for scale-across supremacy?

For context, Cisco isn’t alone in recognising this opportunity. Broadcom fired the first salvo in mid-August with its “Jericho 4” StrataDNX switch/router chips, which began sampling and also offered 51.2 Tb/sec of aggregate bandwidth backed by HBM memory for deep packet buffering to manage congestion.

Two weeks after Broadcom’s announcement, Nvidia unveiled its Spectrum-XGS scale-across network—a notably cheeky name given that Broadcom’s “Trident” and “Tomahawk” switch ASICs belong to the StrataXGS family.

Nvidia secured CoreWeave as its anchor customer but provided limited technical details about the Spectrum-XGS ASICs. Now Cisco is rolling out its own components for the scale-across networking market, setting up a three-way competition among networking heavyweights.

The problem: AI is too big for one building

To understand why multiple vendors are rushing into this space, consider the scale of modern AI infrastructure. Training large language models or running complex AI systems requires thousands of high-powered processors working in concert, generating enormous amounts of heat and consuming massive amounts of electricity. 

Data centres are hitting hard limits—not just on available space, but on how much power they can supply and cool.

“AI compute is outgrowing the capacity of even the largest data centre, driving the need for reliable, secure connection of data centres hundreds of miles apart,” said Martin Lund, Executive Vice President of Cisco’s Common Hardware Group.

The industry has traditionally addressed capacity challenges through two approaches: scaling up (adding more capability to individual systems) or scaling out (connecting more systems within the same facility). 

But both strategies are reaching their limits. Data centres are running out of physical space, power grids can’t supply enough electricity, and cooling systems can’t dissipate the heat fast enough.

This forces a third approach: “scale-across,” distributing AI workloads across multiple data centres that might be in different cities or even different states. However, this creates a new problem—the connections between these facilities become critical bottlenecks.

Why traditional routers fall short

AI workloads behave differently from typical data centre traffic. Training runs generate massive, bursty traffic patterns—periods of intense data movement followed by relative quiet. If the network connecting data centres can’t absorb these surges, everything slows down, wasting expensive computing resources and, critically, time and money.

Traditional routing equipment wasn’t designed for this. Most routers prioritise either raw speed or sophisticated traffic management, but struggle to deliver both simultaneously while maintaining reasonable power consumption. For AI data centre interconnect applications, organisations need all three: speed, intelligent buffering, and efficiency.

Cisco’s answer: The 8223 system

Cisco’s 8223 system represents a departure from general-purpose routing equipment. Housed in a compact three-rack-unit chassis, it delivers 64 ports of 800-gigabit connectivity—currently the highest density available in a fixed routing system. More importantly, it can process over 20 billion packets per second and scale up to three Exabytes per second of interconnect bandwidth.

The system’s distinguishing feature is deep buffering capability, enabled by the P200 chip. Think of buffers as temporary holding areas for data—like a reservoir that catches water during heavy rain. When AI training generates traffic surges, the 8223’s buffers absorb the spike, preventing network congestion that would otherwise slow down expensive GPU clusters sitting idle waiting for data.

Power efficiency is another critical advantage. As a 3RU system, the 8223 achieves what Cisco describes as “switch-like power efficiency” while maintaining routing capabilities—crucial when data centres are already straining power budgets.

The system also supports 800G coherent optics, enabling connections spanning up to 1,000 kilometres between facilities—essential for geographic distribution of AI infrastructure.

Industry adoption and real-world applications

Major hyperscalers are already deploying the technology. Microsoft, an early Silicon One adopter, has found the architecture valuable across multiple use cases. 

Dave Maltz, technical fellow and corporate vice president of Azure Networking at Microsoft, noted that “the common ASIC architecture has made it easier for us to expand from our initial use cases to multiple roles in DC, WAN, and AI/ML environments.”

Alibaba Cloud plans to use the P200 as a foundation for expanding its eCore architecture. Dennis Cai, vice president and head of network Infrastructure at Alibaba Cloud, stated the chip “will enable us to extend into the Core network, replacing traditional chassis-based routers with a cluster of P200-powered devices.”

Lumen is also exploring how the technology fits into its network infrastructure plans. Dave Ward, chief technology officer and product officer at Lumen, said the company is “exploring how the new Cisco 8223 technology may fit into our plans to enhance network performance and roll out superior services to our customers.”

Programmability: Future-proofing the investment

One often-overlooked aspect of AI data centre interconnect infrastructure is adaptability. AI networking requirements are evolving rapidly, with new protocols and standards emerging regularly. 

Traditional hardware typically requires replacement or expensive upgrades to support new capabilities. The P200’s programmability addresses this challenge. 

Organisations can update the silicon to support emerging protocols without replacing hardware—important when individual routing systems represent significant capital investments and AI networking standards remain in flux.

Security considerations

Connecting data centres hundreds of miles apart introduces security challenges. The 8223 includes line-rate encryption using post-quantum resilient algorithms, addressing concerns about future threats from quantum computing. Integration with Cisco’s observability platforms provides detailed network monitoring to identify and resolve issues quickly.

Can Cisco compete?

With Broadcom and Nvidia already staking their claims in the scale-across networking market, Cisco faces established competition. However, the company brings advantages: a long-standing presence in enterprise and service provider networks, the mature Silicon One portfolio launched in 2019, and relationships with major hyperscalers already using its technology.

The 8223 ships initially with open-source SONiC support, with IOS XR planned for future availability. The P200 will be available across multiple platform types, including modular systems and the Nexus portfolio. 

This flexibility in deployment options could prove decisive as organisations seek to avoid vendor lock-in while building out distributed AI infrastructure.

Whether Cisco’s approach becomes the industry standard for AI data centre interconnect remains to be seen, but the fundamental problem all three vendors are addressing—efficiently connecting distributed AI infrastructure—will only grow more pressing as AI systems continue scaling beyond single-facility limits. 

The real winner may ultimately be determined not by technical specifications alone, but by which vendor can deliver the most complete ecosystem of software, support, and integration capabilities around their silicon.

See also:

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security Expo, click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.

The post Can Cisco’s new AI data centre router tackle the industry’s biggest infrastructure bottleneck? appeared first on AI News.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

思科 AI数据中心 网络互联 路由器 Silicon One Cisco AI Data Center Network Interconnect Router Silicon One
相关文章