首页 > 系统相关 >VMware Private AI Foundation with NVIDIA - 生成式人工智能解决方案

VMware Private AI Foundation with NVIDIA - 生成式人工智能解决方案

时间:2024-10-17 08:53:08浏览次数:7  
标签:Foundation Aria AI 生成式 NVIDIA GPU VMware

VMware Private AI Foundation with NVIDIA - 生成式人工智能解决方案

通过 NVIDIA 的加速计算以及 VMware Cloud Foundation 的虚拟基础架构管理和云管理来运行生成式 AI 工作负载

请访问原文链接:https://sysin.org/blog/vmware-private-ai-foundation-nvidia/ 查看最新版。原创作品,转载请保留出处。

作者主页:sysin.org


VMware Private AI Foundation with NVIDIA

通过联合生成式 AI 平台解锁生成式人工智能并释放生产力。解决隐私、选择、成本、性能和合规性问题。

Architecture

阅读博客

解锁新一代人工智能并释放生产力

  • 阅读解决方案简介

  • 阅读 IDC 白皮书

    icon-privacy1.png

  • 实现隐私、安全和合规性

    使用人工智能服务的架构方法来实现企业数据的隐私、安全和控制。

    icon-agility2.png

  • 获得加速性能

    借助 VMware Cloud Foundation 和 NVIDIA AI Enterprise 中的集成软件和硬件功能,从生成式 AI 模型中获取最佳性能。

    icon-automation.png

  • 简化生成式 AI 部署并优化成本

    利用矢量数据库、深度学习虚拟机等特殊功能,获得简化的部署体验和显着的成本效率。

构建和部署私有且安全的生成式 AI 模型

vmw-icon-data-center-extension.svg

  • 引导式部署

    通过工作负载域和相关组件的引导式部署 (sysin),显着提高部署速度

    icon-datacenter.png

  • 用于启用 RAG 工作流程的矢量数据库

    通过 PostgreSQL 上的 pgvector 支持的矢量数据库,实现数据快速查询和实时更新,以增强 LLMs 的输出。

    icon-app-volumes.png

  • 目录设置向导

    通过精心策划和优化的 AI 基础设施目录项,简化复杂项目的基础设施配置。

    icon-virtualization.png

  • GPU 监控

    通过跨集群和主机查看 GPU 资源利用率来简化 GPU 使用,从而获得优化的性能和成本。

    icon-solution-developer.png

  • 深度学习虚拟机模板

    使用预配置的深度学习虚拟机提高环境的一致性。

    icon-microservices.png

  • NVIDIA Nemo Retriever

    通过一系列 NVIDIA CUDA-X生成式 AI 微服务增强 RAG 功能 (sysin),使组织能够将自定义模型无缝连接到不同的业务数据。

    icon-lightweight-v2.png

  • NVIDIA NIM Operator

    使用 NVIDIA AI 工作流程示例简化 RAG 应用程序部署到生产中,无需重写代码。

    icon-container4.png

  • NVIDIA NIM

    通过一组易于使用的微服务实现大规模无缝 AI 推理,这些微服务旨在加速生成式 AI 在企业中的部署。

icon-usage-meter.png

  • NVIDIA GPU Operator

    自动管理将 GPU 与 Kubernetes 结合使用所需的软件的生命周期。提高 GPU 性能、利用率和遥测。

系统架构

System Architecture of VMware Private AI Foundation with NVIDIA

VMware Private AI Foundation with NVIDIA runs on top of VMware Cloud Foundation adding support for AI workloads in VI workload domains with vSphere IaaS control plane provisioned by using kubectl and VMware Aria Automation .

Example Architecture for VMware Private AI Foundation with NVIDIA

sysin

Component Description
GPU-enabled ESXi hosts ESXi hosts that configured in the following way: Have an NVIDIA GPU that is supported for VMware Private AI Foundation with NVIDIA. The GPU is shared between workloads by using the time slicing or Multi-Instance GPU (MIG) mechanism. Have the NVIDIA vGPU host manager driver installed so that you can use vGPU profiles based on MIG or time slicing.
Supervisor One or more vSphere clusters enabled for vSphere IaaS control plane so that you can run virtual machines and containers on vSphere by using the Kubernetes API. A Supervisor is a Kubernetes cluster itself, serving as the control plane to manage workload clusters and virtual machines.
Harbor registry A local image registry in a disconnected environment where you host the container images downloaded from the NVIDIA NGC catalog.
NSX Edge cluster A cluster of NSX Edge nodes that provides 2-tier north-south routing for the Supervisor and the workloads it runs.The Tier-0 gateway on the NSX Edge cluster is in active-active mode.
NVIDIA Operators NVIDIA GPU Operator. Automates the management of all NVIDIA software components needed to provision GPU to containers in a Kubernetes cluster. NVIDIA GPU Operator is deployed on a TKG cluster. NVIDIA Network Operator. NVIDIA Network Operator also helps configuring the right mellanox drivers for containers using virtual functions for high speed networking, RDMA and GPUDirect.Network Operator works together with the GPU Operator to enable GPUDirect RDMA on compatible systems. NVIDIA Network Operator is deployed on a TKG cluster.
Vector database A PostgreSQL database that has the pgvector extension enabled so that you can use it in Retrieval Augmented Generation (RAG) AI workloads.
NVIDIA Licensing Portal NVIDIA Delegated License Service (DLS) You use the NVIDIA Licensing Portal to generate a client configuration token to assign a license to the guest vGPU driver in the deep learning virtual machine and the GPU Operators on TKG clusters. In a disconnected environment or to have your workloads getting license information without using an Internet connection, you host the NVIDIA licenses locally on a Delegated License Service (DLS) appliance.
Content library Content libraries store the images for the deep learning virtual machines and for the Tanzu Kubernetes releases. You use these images for AI workload deployment within the VMware Private AI Foundation with NVIDIA environment. In a connected environment, content libraries pull their content from VMware managed public content libraries. In a disconnected environment, you must upload the required images manually or pull them from an internal content library mirror server.
NVIDIA GPU Cloud (NGC) catalog A portal for GPU-optimized containers for AI, and machine learning that are tested and ready to run on supported NVIDIA GPUs on premises on top of VMware Private AI Foundation with NVIDIA.

As a cloud administrator (sysin), you use the management components in VMware Cloud Foundation

Management Component Description
SDDC Manager You use SDDC Manager for the following tasks: Deploy a GPU-enabled VI workload domain that is based vSphere Lifecycle Manager images and add clusters to it. Deploy an NSX Edge cluster in VI workload domains for use by Supervisor instances and in the management domain for the VMware Aria Suite components of VMware Private AI Foundation with NVIDIA. Deploy a VMware Aria Suite Lifecycle instance which is integrated with the SDDC Manager repository.
VI Workload Domain vCenter Server You use this vCenter Server instance to enable and configure a Supervisor.
VI Workload Domain NSX Manager SDDC Manager uses this NSX Manager to deploy and update NSX Edge clusters.
VMware Aria Suite Lifecycle You use VMware Aria Suite Lifecycle to deploy and update VMware Aria Automation and VMware Aria Operations.
VMware Aria Automation You use VMware Aria Automation to add self-service catalog items for deploying AI workloads for DevOps engineers and data scientists.
VMware Aria Operations You use VMware Aria Operations for monitoring the GPU consumption in the GPU-enabled workload domains.
VMware Data Services Manager You use VMware Data Services Manager to create vector databases, such as a PostgreSQL database with pgvector extension.

VMware 相关组件

VMware Components in VMware Private AI Foundation with NVIDIA

VMware Cloud Foundation 5.2

The functionality of the VMware Private AI Foundation with NVIDIA solution is available across several software components.

  • VMware Cloud Foundation 5.2
  • VMware Aria Automation 8.18
  • VMware Aria Operations 8.18
  • VMware Data Services Manager 2.1

VMware Cloud Foundation 5.1

The functionality of the VMware Private AI Foundation with NVIDIA solution is available across several software components.

  • VMware Cloud Foundation 5.1.1
  • VMware Aria Automation 8.16.2 and VMware Aria Automation 8.17
  • VMware Aria Operations 8.16 and VMware Aria Operations 8.17.1
  • VMware Data Services Manager 2.0.x

准备好开始了吗?

联系 VMware

VMware Private AI Foundation with NVIDIA 支持两种用例:

  • 开发用例
    云管理员和 DevOps 工程师可以以深度学习虚拟机的形式配置 AI 工作负载,包括检索增强生成 (RAG)。数据科学家可以使用这些深度学习虚拟机进行人工智能开发。
  • 生产用例
    云管理员可以为 DevOps 工程师提供具有 NVIDIA 环境的 VMware Private AI Foundation,以便在 vSphere IaaS 控制平面上的 Tanzu Kubernetes Grid (TKG) 集群上调配生产就绪的 AI 工作负载。

相关产品:

更多:VMware 产品下载汇总

标签:Foundation,Aria,AI,生成式,NVIDIA,GPU,VMware
From: https://www.cnblogs.com/sysin/p/18450330

相关文章

  • STM32F103+Air780 OTA升级测试说明
     测试1,单片机通过串口1和GPRS模块通信; 单片机PA8引脚作为复位模组使用;串口2做日志打印(115200)(单片机)PA9  ----  (Air780 )RX;(单片机)PA10  ----  (Air780 )TX;(单片机)PA8  ----  (Air780 )RST2,打开这节例程3,可以使用下载器先下载Bo......
  • 分享字节跳动的免费AI编程助手(豆包MarsCode)
    邀请新用户注册登录&使用即可赚火星币,凭对应火星币可赢取京东卡和现金,最高可达1万元。接下来来手把手教你如何参与活动,推广赢好礼!合伙人推广赢好礼1.成为合伙人STEP1:点击https://www.marscode.cn/events/s/iBEnTPtB/,进入活动页面,点击【点击赚钱】按钮。 STEP2:活动页面......
  • langchain multi modal support
    Howtopassmultimodaldatadirectlytomodelshttps://python.langchain.com/v0.2/docs/how_to/multimodal_inputs/ message=HumanMessage(content=[{"type":"text","text":"describetheweatherinthisimag......
  • 利用LangGraph和Waii实现你的chat2db!
    0前言在数据分析领域快速发展的今天,通过自然语言与数据交互的能力变得越来越有价值。对话式分析旨在使复杂数据结构对没有专业技能的用户更易于访问。LangGraph是个框架,用于构建使用语言模型的状态化、多代理应用程序。Waii提供文本到SQL和文本到图表的功能,使用户能够通过......
  • 电子商务行业的产品推荐系统变革:AI知识库引领智能化与个性化
    在数字化浪潮的席卷下,电子商务行业正以前所未有的速度发展,而产品推荐系统作为电商平台与用户之间的桥梁,其智能化与个性化水平的高低,直接关系到用户体验、转化率乃至整个平台的竞争力。近年来,随着人工智能(AI)技术的飞速发展,特别是AI知识库的构建与应用,电商行业的产品推荐系统正......
  • AI程序开发体验之CRM系统
    一、CRM系统的核心功能CRM系统是企业用于管理与客户关系的软件,它通过集成多种功能来提升客户满意度和企业效率。以下是一些CRM系统的核心功能:客户数据管理:这是CRM系统的基础功能,允许企业存储和组织客户的联系信息、购买历史和偏好等数据,以便快速访问并提供个性化服务。销......
  • 【AI大模型】从 RAG 1.0到RAG 2.0,这次做对了什么?
    RAG是目前最流行的补充生成式人工智能模型的方式,最近RAG的开创者提出了新的上下文语言模型(CLM),他们称之为“RAG2.0”。今天让我们一块来从RAG目前的原理和缺点出发,看看他们所提出的RAG2.0是否能够为行业带来新的希望。LLM的时间有效性您应该知道,所有独立的大型......
  • 27K star!有没有显卡都能搞,Langchain-Chatchat 快速基于LLM构建本地智能知识库
    觉得搞一个AI的智能问答知识库很难吗?那是你没有找对方向和工具,今天我们分享一个开源项目,帮助你快速构建基于Langchain和LLM的本地知识库问答,在GitHub已经获得27Kstar,它就是:Langchain-Chatchat......
  • C#线程---await
    简介:前面异步编程的不足:它允许我们以模块化的方式设计程序,来组合不同的异步操作。遗憾的是:一:当阅读此类程序时仍然非常难理解程序的实际执行顺序。在大型程序中将会有许多相互依赖的任务和后续操作,用于运行其他后续操作的后续操作,处理异常的后续操作,并且它们都出现在程序代......
  • 拥抱AI-海关机场护照查验-护照查验接口-护照识别真伪核验
    在全球化日益加深的今天,国际旅行已成为人们生活中不可或缺的一部分。无论是商务出行还是休闲旅游,护照作为国际通行的身份证明文件,其真实性和有效性至关重要。为了确保国家安全和旅客的顺利通关,海关机场在护照查验过程中采用了先进的护照识别真伪核验技术。什么是护照查验......