VMware Hands-on Labs Blog 09月29日 10:48
vSphere与AI/ML集成实践
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了如何将vSphere与AI/ML技术结合,通过GPU加速、Hugging Face编码辅助和Ray分布式计算,提升虚拟化环境的AI/ML工作负载效率。文章介绍了使用NVIDIA vGPU在vSphere中加速机器学习工作负载的实践,以及Hugging Face插件如何优化Visual Studio IDE的编码体验,并展示了Ray在vSphere上的分布式AI/ML工作负载扩展方案。

💻 vSphere与GPU结合:通过NVIDIA vGPU技术,在vSphere虚拟机中实现近乎裸金属性能的GPU加速,为数据科学家和ML从业者提供高效的虚拟机部署方案。

🖥️ Hugging Face编码辅助:Hugging Face插件为Visual Studio IDE提供实时代码建议,利用ML算法预测代码,显著提升开发效率和代码质量,特别适合Python等语言的开发者。

🌐 Ray分布式计算:VMware与Ray合作,通过开源插件在vSphere上运行Ray,实现AI/ML工作负载的跨服务器分布式处理,支持模型训练和推理的无服务器式扩展,提升资源利用率和计算能力。

🔧 实验室实践:文章提供了多个实验室模块,涵盖从GPU加速配置到Hugging Face插件使用,再到Ray集群部署的详细步骤,帮助读者动手实践并深入理解相关技术。

🚀 性能优化:通过结合vSphere的虚拟化优势和GPU的强大算力,以及Ray的分布式调度能力,实现AI/ML工作负载的快速部署、高效训练和灵活扩展,满足不同场景下的性能需求。

Yes! We all know vSphere, but do you really? Do you know how virtualization meets artificial intelligence; and scalability meets efficiency? Let’s dive deep into the world of AI and ML with vSphere with GPUs and some new partnerships between VMware with Hugging Face and Ray.

Imagine harnessing the power of vSphere’s virtualization expertise with cutting-edge GPU cards. Add to that the seamless integration with a coding assistant from Hugging Face’s AI magic and the dynamic scaling abilities of Ray. The best thing of all is that you can get hands-on and test the labsolutely amazing products and integrations with VMware Hands-on Labs! The labs described here are all related to vSphere for AI/ML work as the main topic, so let’s dive in!

Accelerate AI/ML in vSphere Using GPUs

Let’s start with a focus on the core infrastructure for vSphere with GPUs. A GPU within a vSphere virtual machine can deliver near bare-metal performance. You can get all the resources and storage benefits without sacrificing the scalability of VMs.

This lab will be particularly useful for people like data scientists and/or ML practitioners who want to deliver virtual machines to their end users with GPU acceleration already added to them.

In this lab, you will learn how you can accelerate Machine Learning Workloads on vSphere using GPUs. The acceleration is done through the combination of the management benefits of VMware vSphere with the power of NVIDIA GPUs.

Lab Modules
Module 1 – Introduction to AI/ML in vSphere Using GPUs
Module 2 – Run Machine Learning Workloads Using NVIDIA vGPU
Module 3 – Using GPUs in Pass-through Mode
Module 4 – Configure a container using NVIDIA vGPU in vSphere with Tanzu
Module 5 – Choosing the Right Profile for your Workload (vGPU vs MIG vGPU)

Dive into the lab: https://userui.learningplatform.vmware.com/HOL/catalogs/lab/14433

Redefining Coding Dynamics: Hugging Face’s Plugin for Visual Studio IDE

Let’s now dive deeper into one powerful AI/ML use case that demonstrates how Generative AI can help with the coding/application development process – and how that can be incorporated into a developer tool like an Integrated Development Environment (IDE).

Imagine a tool tailored specifically to your company’s coding style, offering real-time code suggestions in a matter of seconds. Introducing the Hugging Face plugin for Visual Studio IDE. This plug-in talks to a server with GPUs which produces the code completion. The technology uses Machine Learning algorithms to predict the next line of code, providing developers with a new tool to accelerate and streamline their coding process.

Many data scientists are in fact developers in Python or other languages, so what they have on their desktop for daily use is an IDE or a Jupyter Notebook. This lab will show how their work can be made easier and more effective.

Lab Modules
Module 1 – Introduction to AI Autocomplete with Visual Studio Plugin

Take the lab: https://userui.learningplatform.vmware.com/HOL/catalogs/lab/14355

Scale AI Workloads with Ray on VMware

Now, if you already have a coding assistant and your virtualized environment is a 10 out of 10 in performance, let’s take it to the next level. We want to distribute these AI/ML workloads over several servers.

VMware now has a partnership with Anyscale, the creators of Ray, to create an open-source plug-in to run Ray on vSphere using Virtual Machines. Ray is an optimized workload scheduler for ML workloads that brings a serverless-style scaling to training and inferencing workloads. Ray provides the ability to split an AI/ML model training job across multiple physical and virtual machines and control that from a central point, called a “head node” and VMware has contributed to upstream Ray to make this possible on vSphere 8. 

The head node manages the cluster and scales the number of worker nodes within it. These distributed worker nodes are responsible for training, fine-tuning, and serving models.

Take the lab and test the solution for yourself! 

Lab Modules
Module 1 – Provision a Ray Cluster on vSphere and Execute “Hello World”
Module 2 – Train an XGBoost model on Ray and Destroy Ray Cluster

Test the solution now: https:// userui.learningplatform.vmware.com/HOL/catalogs/lab/14435

For more information on our partnerships and our GenAI offerings, go to https://news.vmware.com/releases/vmware-explore-2023-private-ai-foundation

If you have a comment or request, contact us at discovery-request@vmware.com
Stay connected to VMware via Twitter, Facebook, and YouTube

The post Integrating AI & ML in vSphere appeared first on VMware Hands-on Labs Blog.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

vSphere AI 机器学习 GPU加速 Hugging Face Ray 虚拟化 分布式计算
相关文章