AI News 14小时前
华为发布CloudMatrix 384 AI 集群,挑战英伟达
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

华为发布了其CloudMatrix 384 AI芯片集群,一个用于AI学习的新系统,该系统采用由光纤连接的Ascend 910C处理器集群。其分布式架构在资源利用率和芯片内时间方面优于传统的GPU硬件设置。华为表示,尽管受到美国制裁,该新框架仍使其成为“英伟达市场主导地位的强劲挑战者”。然而,使用该框架需要数据工程师适应新的工作流程,转向华为支持的MindSpore等框架,并可能需要转换或重新训练现有模型,因为MindSpore与PyTorch/TensorFlow在语法、训练流程和函数调用上存在差异。文章还介绍了MindIR用于模型部署、CANN(计算架构神经网络)以及MindSpore的GRAPH_MODE和PYNATIVE_MODE执行模式,并指出华为的ModelArts平台与Ascend硬件和MindSpore框架紧密集成。尽管华为的工具正在发展,可能不如NVIDIA生态系统成熟,但它为组织提供了减少对英伟达依赖的选择,并提供迁移指南和支持。

🚀 **华为CloudMatrix 384 AI 集群的发布** 华为推出了CloudMatrix 384 AI芯片集群,这是一个基于Ascend 910C处理器并采用光纤连接的分布式系统,旨在提升AI学习效率。该系统通过其分布式架构,在资源利用和处理速度上超越了传统的GPU硬件,旨在挑战英伟达在AI芯片市场的领先地位,即使在面临美国制裁的情况下。

🔧 **工作流程与框架的转变** 使用华为的AI技术需要数据工程师适应新的工作流程,重点在于采用华为支持的深度学习框架,如MindSpore。这通常意味着需要将现有的PyTorch或TensorFlow模型转换为MindSpore格式,或使用MindSpore API进行重新训练,因为MindSpore在语法、训练流程和函数调用上与主流框架存在显著差异,例如算子行为和默认权重初始化方法的不同。

💡 **MindSpore生态系统与部署** MindSpore提供了MindIR(MindSpore中间表示)作为模型部署的格式,类似于Nvidia的NIM。模型训练完成后,可以通过`mindspore.export`工具导出为MindIR格式,然后使用Ascend芯片的推理API进行部署。MindSpore在训练和推理逻辑上比PyTorch/TensorFlow更明确,要求预处理与训练输入匹配,并优化静态图执行。此外,MindSpore提供了GRAPH_MODE(编译执行,性能优化)和PYNATIVE_MODE(即时执行,易于调试)两种执行模式,允许团队在开发灵活性和部署性能之间取得平衡。

🛠️ **CANN与ModelArts的集成** 华为的CANN(计算架构神经网络)提供了一套专门针对Ascend软件的工具和库,功能上类似于NVIDIA的CUDA。CANN的性能分析和调试工具对于优化Ascend硬件上的模型表现至关重要。华为的云端AI开发和部署平台ModelArts与Ascend硬件和MindSpore框架深度集成,支持从数据标注到模型部署和监控的全流程,为用户提供了与AWS SageMaker和Google Vertex AI类似的体验,但针对华为的AI处理器进行了优化。

Huawei has released its CloudMatrix 384 AI chip cluster, a new system for AI learning. It employs clusters of Ascend 910C processors, joined via optical links. The distributed architecture means the system can outperform traditional hardware GPU setups, particularly in terms of resource use and on-chip time, despite the individual Ascend chips being less powerful than those of competitors.

Huawei’s new framework positions the tech giant as a “formidable challenger to Nvidia’s market-leading position, despite ongoing US sanctions,” the company claims.

To use the new Huawei framework for AI, data engineers will need to adapt their workflows, using frameworks that support Huawei’s Ascend processors, such MindSpore, which are available from Huawei and its partners

Framework transition: From PyTorch/TensorFlow to MindSpore

Unlike NVIDIA’S ecosystem, which predominantly uses frameworks like PyTorch and TensorFlow (engineered to take full advantage of CUDA), Huawei’s Ascend processors perform best when used with MindSpore, a deep learning framework developed by the company.

If data engineers already have models built in PyTorch or TensorFlow, they will likely need to convert models to the MindSpore format or retrain them using the MindSpore API.

It is worth noting that MindSpore uses different syntax, training pipelines and function calls from PyTorch or TensorFlow, so a degree of re-engineering will be necessary to replicate the results from model architectures and training pipelines. For instance, individual operator behaviour varies, such as padding modes in convolution and pooling layers. There are also differences in default weight initialisation methods.

Using MindIR for model deployment

MindSpore employs MindIR (MindSpore Intermediate Representation), a close analogue to Nvidia NIM. According to MindSpore’s official documentation, once a model has been trained in MindSpore, it can be exported using the mindspore.export utility, which converts the trained network into the MindIR format.

Detailed by DeepWiki’s guide, deploying a model for inference typically involves loading the exported MindIR model and then running predictions using MindSpore’s inference APIs for Ascend chips, which handle model de-serialisation, allocation, and execution.

MindSpore separates training and inference logic more explicitly than PyTorch or TensorFlow. Therefore, all preprocessing needs to match training inputs, and static graph execution must be optimised. MindSpore Lite or Ascend Model Zoo are recommended for additional hardware-specific tuning.

Adapting to CANN (Compute Architecture for Neural Networks)

Huawei’s CANN features a set of tools and libraries tailored for Ascend software, paralleling NVIDIA’s CUDA in functionality. Huawei recommends using CANN’s profiling and debugging tools to monitor and improve model performance on Ascend hardware.

Execution Modes: GRAPH_MODE vs.PYNATIVE_MODE

MindSpore provides two execution modes:

For initial development, PYNATIVE_MODE is recommended for simpler iterative testing and debugging. When models are ready to be deployed, switching to GRAPH_MODE can help achieve maximum efficiency on Ascend hardware. Switching between modes lets engineering teams balance development flexibility with deployment performance.

Code should be adjusted for each mode. For instance, when in GRAPH_MODE, it’s best to avoid Python-native control flow where possible.

Deployment environment: Huawei ModelArts

As you might expect, Huawei’s ModelArts, the company’s cloud-based AI development and deployment platform, is tightly integrated with Huawei’s Ascend hardware and the MindSpore framework. While it is comparable to platforms like AWS SageMaker and Google Vertex AI, it is optimised for Huawei’s AI processors.

Huawei says ModelArts supports the full pipeline from data labelling and preprocessing to model training, deployment, and monitoring. Each stage of the pipeline is available via API or the web interface.

In summary

Adapting to MindSpore and CANN may necessitate training and time, particularly for teams accustomed to NVIDIA’s ecosystem, with data engineers needing to understand various new processes. These include how CANN handles model compilation and optimisation for Ascend hardware, adjusting tooling and automation pipelines designed initially for NVIDIA GPUs, and learning new APIs and workflows specific to MindSpore.

Although Huawei’s tools are evolving, they lack the maturity, stability, and broader ecosystem support that frameworks like PyTorch with CUDA offer. However, Huawei hopes that migrating to its processes and infrastructure will pay off in terms of results, and let organisations reduce reliance on US-based Nvidia.

Huawei’s Ascend processors may be powerful and designed for AI workloads, but they have only limited distribution in some countries. Teams outside Huawei’s core markets may struggle to test or deploy models on Ascend hardware, unless they use partner platforms, like ModelArts, that offer remote access.

Fortunately, Huawei provides extensive migration guides, support, and resources to support any transition.

(Image source: “Huawei P9” by 405 Mi16 is licensed under CC BY-NC-ND 2.0.)

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and co-located with other leading technology events. Click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.

The post Re-engineering for better results: The Huawei AI stack appeared first on AI News.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

华为 AI芯片 CloudMatrix 384 Ascend 910C MindSpore CANN ModelArts 人工智能 深度学习 Huawei AI Chips MindSpore CANN ModelArts Artificial Intelligence Deep Learning
相关文章