Nvidia 第一季度收入為 71.9 億美元,環比增長 19%,同比下降 13%,其中數據中心收入達到創紀錄的 42.8 億美元,環比增長 18%,同比增長 14%。
Nvidia 預計第二季度收入為 110 億美元,上下浮動 2%。
首席執行官黃仁勳討論了構建大型語言模型並將其提煉成更小尺寸以用於各種設備的重要性,以及網絡和軟件在加速計算中的重要性。
Nvidia 正在大幅增加其供應以滿足對加速計算的激增需求,並在未來幾個季度推出一波產品。
使用警語:中文譯文來源為 Google 翻譯,僅供參考,實際內容請以英文原文為主
Operator
Operator
Good afternoon. My name is David, and I'll be your conference operator today. At this time, I'd like to welcome everyone to NVIDIA's First Quarter Earnings Call. Today's conference is being recorded. (Operator Instructions) Simona Jankowski, you may begin your conference.
下午好。我叫大衛,今天我將擔任你們的會議接線員。此時此刻,歡迎大家參加 NVIDIA 第一季度財報電話會議。今天的會議正在錄製中。 (操作員說明)Simona Jankowski,你可以開始你的會議了。
Simona Jankowski - VP of IR
Simona Jankowski - VP of IR
Thank you. Good afternoon, everyone, and welcome to NVIDIA's conference call for the first quarter of fiscal 2024. With me today from NVIDIA are Jensen Huang, President and Chief Executive Officer; and Colette Kress, Executive Vice President and Chief Financial Officer. I'd like to remind you that our call is being webcast live on NVIDIA's Investor Relations website. The webcast will be available for replay until the conference call to discuss our financial results for the second quarter of fiscal 2024.
謝謝。大家下午好,歡迎參加 NVIDIA 2024 財年第一季度電話會議。今天來自 NVIDIA 的是總裁兼首席執行官黃仁勳;執行副總裁兼首席財務官 Colette Kress。我想提醒您,我們的電話會議正在 NVIDIA 的投資者關係網站上進行網絡直播。在討論我們 2024 財年第二季度財務業績的電話會議之前,可以重播網絡廣播。
The content of today's call is NVIDIA's property. It can't be reproduced or transcribed without our prior written consent. During this call, we may make forward-looking statements based on current expectations. These are subject to a number of significant risks and uncertainties, and our actual results may differ materially. For a discussion of factors that could affect our future financial results and business, please refer to the disclosure in today's earnings release, our most recent Forms 10-K and 10-Q and the reports that we may file on Form 8-K with the Securities and Exchange Commission. All our statements are made as of today, May 24, 2023, based on information currently available to us. Except as required by law, we assume no obligation to update any such statements.
今天電話會議的內容是 NVIDIA 的財產。未經我們事先書面同意,不得複製或轉錄。在此次電話會議中,我們可能會根據當前預期做出前瞻性陳述。這些都受到許多重大風險和不確定性的影響,我們的實際結果可能存在重大差異。有關可能影響我們未來財務業績和業務的因素的討論,請參閱今天的收益發布中的披露、我們最近的 10-K 和 10-Q 表格以及我們可能在 8-K 表格上提交的報告證券交易委員會。我們的所有聲明都是根據我們目前可獲得的信息,截至今天,即 2023 年 5 月 24 日作出的。除非法律要求,否則我們不承擔更新任何此類聲明的義務。
During this call, we will discuss non-GAAP financial measures. You can find a reconciliation of these non-GAAP financial measures to GAAP financial measures in our CFO commentary, which is posted on our website.
在這次電話會議中,我們將討論非 GAAP 財務指標。您可以在我們網站上發布的 CFO 評論中找到這些非 GAAP 財務指標與 GAAP 財務指標的對賬。
And with that, let me turn the call over to Colette.
就這樣,讓我把電話轉給科萊特。
Colette M. Kress - Executive VP & CFO
Colette M. Kress - Executive VP & CFO
Thanks, Simona. Q1 revenue was $7.19 billion, up 19% sequentially and down 13% year-on-year. Strong sequential growth was driven by record data center revenue with our gaming and professional visualization platforms emerging from channel inventory corrections.
謝謝,西蒙娜。第一季度收入為 71.9 億美元,環比增長 19%,同比下降 13%。強勁的連續增長是由創紀錄的數據中心收入推動的,我們的遊戲和專業可視化平台從渠道庫存調整中脫穎而出。
Starting with Data Center. Record revenue of $4.28 billion was up 18% sequentially and up 14% year-on-year on strong growth of our accelerated computing platform worldwide. Generative AI is driving exponential growth in compute requirements and a fast transition to NVIDIA accelerated computing, which is the most versatile, most energy-efficient and the lowest TCO approach to train and deploy AI. Generative AI drove significant upside in demand for our products, creating opportunities and broad-based global growth across our markets.
從數據中心開始。由於我們在全球加速計算平台的強勁增長,創紀錄的收入達到 42.8 億美元,環比增長 18%,同比增長 14%。生成式 AI 正在推動計算需求呈指數級增長,并快速過渡到 NVIDIA 加速計算,這是訓練和部署 AI 的最通用、最節能且 TCO 最低的方法。生成式人工智能推動了對我們產品需求的顯著增長,為我們的市場創造了機會和基礎廣泛的全球增長。
Let me give you some color across our 3 major customer categories: cloud service providers or CSPs, consumer Internet companies and enterprises. First, CSPs around the world are racing to deploy our flagship Hopper and Ampere architecture GPUs to meet the surge in interest from both enterprise and consumer AI applications for training and inference. Multiple CSPs announced the availability of H100 on their platforms, including private previews at Microsoft Azure, Google Cloud and Oracle Cloud Infrastructure, upcoming offerings at AWS and general availability at emerging GPU-specialized cloud providers like CoreWeave and Lambda. In addition to enterprise AI adoption, these CSPs are serving strong demand for H100 from generative AI pioneers.
讓我為您介紹一下我們的 3 個主要客戶類別:雲服務提供商或 CSP、消費者互聯網公司和企業。首先,世界各地的 CSP 競相部署我們的旗艦 Hopper 和 Ampere 架構 GPU,以滿足企業和消費者 AI 應用程序對訓練和推理的興趣激增。多個 CSP 宣佈在其平台上提供 H100,包括 Microsoft Azure、Google Cloud 和 Oracle Cloud Infrastructure 的私人預覽版、AWS 即將推出的產品以及新興 GPU 專用雲提供商(如 CoreWeave 和 Lambda)的全面可用性。除了企業採用 AI 之外,這些 CSP 還滿足生成 AI 先驅對 H100 的強烈需求。
Second, consumer Internet companies are also at the forefront of adopting generative AI and deep-learning-based recommendation systems, driving strong growth. For example, Meta has now deployed its H100-powered Grand Teton AI supercomputer for its AI production and research teams.
其次,消費互聯網公司也走在採用生成式人工智能和基於深度學習的推薦系統的前沿,推動強勁增長。例如,Meta 現在為其 AI 生產和研究團隊部署了由 H100 驅動的 Grand Teton AI 超級計算機。
Third, enterprise demand for AI and accelerated computing is strong. We are seeing momentum in verticals such as automotive, financial services, health care and telecom where AI and accelerated computing are quickly becoming integral to customers' innovation road maps and competitive positioning. For example, Bloomberg announced it has a $50 billion parameter model, BloombergGPT, to help with financial natural language processing tasks such as sentiment analysis, named entity recognition, news classification and question answering.
第三,企業對人工智能和加速計算的需求旺盛。我們看到了汽車、金融服務、醫療保健和電信等垂直領域的發展勢頭,人工智能和加速計算正迅速成為客戶創新路線圖和競爭定位不可或缺的一部分。例如,彭博社宣布擁有價值 500 億美元的參數模型 BloombergGPT,以幫助處理金融自然語言處理任務,例如情感分析、命名實體識別、新聞分類和問答。
Auto insurance company CCC Intelligent Solutions is using AI for estimating repairs. And AT&T is working with us on AI to improve fleet dispatches so their field technicians can better serve customers. Among other enterprise customers using NVIDIA AI are Deloitte for logistics and customer service, and Amgen for drug discovery and protein engineering.
汽車保險公司 CCC Intelligent Solutions 正在使用 AI 進行維修估算。 AT&T 正在與我們合作開發人工智能以改進車隊調度,以便他們的現場技術人員能夠更好地為客戶服務。使用 NVIDIA AI 的其他企業客戶包括從事物流和客戶服務的 Deloitte,以及從事藥物發現和蛋白質工程的 Amgen。
This quarter, we started shipping DGX H100, our Hopper generation AI system, which customers can deploy on-prem. And with the launch of DGX Cloud through our partnership with Microsoft Azure, Google Cloud and Oracle Cloud Infrastructure, we deliver the promise of NVIDIA DGX to customers from the cloud. Whether the customers deploy DGX on-prem or via DGX Cloud, they get access to NVIDIA AI software, including NVIDIA-based command, end-to-end AI frameworks and pretrained models. We provide them with the blueprint for building and operating AI, spanning our expertise across systems, algorithms, data processing and training methods.
本季度,我們開始發貨 DGX H100,這是我們的 Hopper 一代 AI 系統,客戶可以在本地部署該系統。通過我們與 Microsoft Azure、Google Cloud 和 Oracle Cloud Infrastructure 的合作推出 DGX Cloud,我們從雲端向客戶兌現了 NVIDIA DGX 的承諾。無論客戶是在本地部署 DGX 還是通過 DGX Cloud 部署,他們都可以訪問 NVIDIA AI 軟件,包括基於 NVIDIA 的命令、端到端 AI 框架和預訓練模型。我們為他們提供構建和操作 AI 的藍圖,涵蓋我們在系統、算法、數據處理和培訓方法方面的專業知識。
We also announced NVIDIA AI Foundations, which are model foundry services available on DGX Cloud that enable businesses to build, refine and operate custom large language models and generative AI models trained with their own proprietary data created for unique domain-specific tasks. They include NVIDIA NeMo for large language models, NVIDIA Picasso for images, video and 3D, and NVIDIA BioNeMo for life sciences. Each service has 6 elements: pretrained models, frameworks for data processing and curation, proprietary knowledge-based vector databases, systems for fine-tuning, aligning and guard railing, optimized inference engines, and support from NVIDIA experts to help enterprises fine-tune models for their custom use cases.
我們還宣布了 NVIDIA AI Foundations,這是 DGX Cloud 上提供的模型鑄造服務,使企業能夠構建、改進和操作自定義大型語言模型和生成 AI 模型,這些模型使用為獨特的領域特定任務創建的自己的專有數據進行訓練。它們包括用於大型語言模型的 NVIDIA NeMo,用於圖像、視頻和 3D 的 NVIDIA Picasso,以及用於生命科學的 NVIDIA BioNeMo。每項服務有 6 個要素:預訓練模型、數據處理和管理框架、基於知識的專有矢量數據庫、用於微調、對齊和護欄的系統、優化的推理引擎,以及來自 NVIDIA 專家的支持,以幫助企業微調模型對於他們的自定義用例。
ServiceNow, a leading enterprise services platform is an early adopter of DGX Cloud and NeMo. They are developing custom large language models trained on data specifically for the ServiceNow platform. Our collaboration will let ServiceNow create new enterprise-grade generative AI offerings with the thousands of enterprises worldwide running on the ServiceNow platform, including for IT departments, customer service teams, employees and developers.
領先的企業服務平台 ServiceNow 是 DGX Cloud 和 NeMo 的早期採用者。他們正在開發專門針對 ServiceNow 平台的數據訓練的自定義大型語言模型。我們的合作將使 ServiceNow 與全球數千家在 ServiceNow 平台上運行的企業(包括 IT 部門、客戶服務團隊、員工和開發人員)一起創建新的企業級生成人工智能產品。
Generative AI is also driving a step function increase in inference workloads. Because of their size and complexities, these workloads require acceleration. The latest MLPerf industry benchmark released in April showed NVIDIA's inference platforms deliver performance that is orders of magnitude ahead of the industry with unmatched versatility across diverse workloads. To help customers deploy generative AI applications at scale, at GTC, we announced 4 major new inference platforms that leverage the NVIDIA AI software stack. These include L4 Tensor Core GPU for AI video, L40 for Omniverse and graphics rendering, H100 NBL for large language models and the Grace Hopper Superchip for LLMs, and also recommendation systems and vector databases. Google Cloud is the first CSP to adopt our L4 inference platform with the launch of its G2 virtual machines for generative AI inference and other workloads, such as Google Cloud Dataproc, Google AlphaFold and Google Cloud's Immersive Stream, which render 3D and AR experiences. In addition, Google is integrating our Triton Inference Server with Google Kubernetes engine and its cloud-based Vertex AI platform.
生成式 AI 也在推動推理工作負載的階躍函數增加。由於它們的規模和復雜性,這些工作負載需要加速。 4 月份發布的最新 MLPerf 行業基準測試表明,NVIDIA 的推理平台提供領先行業幾個數量級的性能,並且在各種工作負載上具有無與倫比的多功能性。為了幫助客戶大規模部署生成式 AI 應用程序,我們在 GTC 上宣布了 4 個利用 NVIDIA AI 軟件堆棧的主要新推理平台。其中包括用於 AI 視頻的 L4 Tensor Core GPU、用於 Omniverse 和圖形渲染的 L40、用於大型語言模型的 H100 NBL 和用於 LLM 的 Grace Hopper Superchip,以及推薦系統和矢量數據庫。谷歌云是第一個採用我們的 L4 推理平台的 CSP,推出了用於生成 AI 推理和其他工作負載的 G2 虛擬機,例如 Google Cloud Dataproc、Google AlphaFold 和谷歌云的沉浸式流,它們呈現 3D 和 AR 體驗。此外,谷歌正在將我們的 Triton 推理服務器與穀歌 Kubernetes 引擎及其基於雲的 Vertex AI 平台相集成。
In networking, we saw strong demand at both CSPs and enterprise customers for generative AI and accelerated computing, which require high-performance networking like NVIDIA's Mellanox networking platforms. Demand relating to general purpose CPU infrastructure remains soft. As generative AI applications grow in size and complexity, high-performance networks become essential for delivering accelerated computing at data center scale to meet the enormous demand of both training and inferencing. Our 400-gig Quantum-2 InfiniBand platform is the gold standard for AI-dedicated infrastructure, with broad adoption across major cloud and consumer Internet platforms such as Microsoft Azure.
在網絡方面,我們看到 CSP 和企業客戶對生成 AI 和加速計算的強烈需求,這需要像 NVIDIA 的 Mellanox 網絡平台這樣的高性能網絡。與通用 CPU 基礎設施相關的需求仍然疲軟。隨著生成式 AI 應用程序的規模和復雜性不斷增加,高性能網絡對於在數據中心規模上提供加速計算以滿足訓練和推理的巨大需求變得至關重要。我們的 400-gig Quantum-2 InfiniBand 平台是 AI 專用基礎設施的黃金標準,在主要的雲和消費者互聯網平台(例如 Microsoft Azure)中得到廣泛採用。
With the combination of in-network computing technology and the industry's only end-to-end data center scale optimized software stack, customers routinely enjoy a 20% increase in throughput for their sizable infrastructure investment. For multi-tenant cloud transitioning to support generative AI, our high-speed Ethernet platform with BlueField-3 DPUs and Spectrum-4 Ethernet switching offers the highest available Ethernet network performance. BlueField-3 is in production and has been adopted by multiple hyperscale and CSP customers, including Microsoft Azure, Oracle Cloud, CoreWeave, Baidu and others. We look forward to sharing more about our 400-gig spectrum for accelerated AI networking platform next week at the Computex Conference in Taiwan.
通過結合網絡內計算技術和業界唯一的端到端數據中心規模優化軟件堆棧,客戶通常可以將其大規模基礎設施投資的吞吐量提高 20%。為了支持生成式 AI 的多租戶雲過渡,我們配備 BlueField-3 DPU 和 Spectrum-4 以太網交換的高速以太網平台提供了最高的可用以太網網絡性能。 BlueField-3 已投入生產,並已被多個超大規模和 CSP 客戶採用,包括 Microsoft Azure、Oracle Cloud、CoreWeave、百度等。我們期待下週在台灣的 Computex 大會上分享更多關於我們用於加速 AI 網絡平台的 400-gig 頻譜的信息。
Lastly, our Grace data center CPU is sampling with customers. At this week's International Supercomputing Conference in Germany, the University of Bristol announced a new supercomputer based on the NVIDIA Grace CPU Superchip, which is 6x more energy efficient than the previous supercomputer. This adds to the growing momentum for Grace with both CPU-only and CPU-GPU opportunities across AI and cloud and supercomputing applications. The coming wave of BlueField-3, Grace and Grace Hopper Superchips will enable a new generation of super energy-efficient accelerated data centers.
最後,我們的 Grace 數據中心 CPU 正在向客戶提供樣品。在本周於德國舉行的國際超級計算大會上,布里斯託大學宣布了一款基於 NVIDIA Grace CPU 超級芯片的新型超級計算機,其能效比之前的超級計算機高 6 倍。這增加了 Grace 在人工智能、雲和超級計算應用程序中的純 CPU 和 CPU-GPU 機會的增長勢頭。即將到來的 BlueField-3、Grace 和 Grace Hopper 超級芯片浪潮將使新一代超高能效加速數據中心成為可能。
Now let's move to Gaming. Gaming revenue of $2.24 billion was up 22% sequentially and down 38% year-on-year. Strong sequential growth was driven by sales of the 40 Series GeForce RTX GPUs for both notebooks and desktops. Overall end demand was solid and consistent with seasonality, demonstrating resilience against a challenging consumer spending backdrop. The GeForce RTX 40 Series GPU laptops are off to a great start, featuring 4 NVIDIA inventions: RTX Path Tracing, DLSS 3 AI rendering, Reflex ultra-low latency rendering and Max-Q energy-efficient technologies. They deliver tremendous gains in industrial design, performance and battery life for gamers and creators.
現在讓我們轉向遊戲。博彩收入為 22.4 億美元,環比增長 22%,同比下降 38%。用於筆記本電腦和台式機的 40 系列 GeForce RTX GPU 的銷售推動了強勁的環比增長。整體終端需求穩健且符合季節性,顯示出在充滿挑戰的消費者支出背景下的韌性。 GeForce RTX 40 系列 GPU 筆記本電腦開局良好,擁有 4 項 NVIDIA 發明:RTX 路徑追踪、DLSS 3 AI 渲染、Reflex 超低延遲渲染和 Max-Q 節能技術。它們為遊戲玩家和創作者帶來了工業設計、性能和電池壽命方面的巨大進步。
Unlike our desktop offerings, 40 Series laptops support the NVIDIA Studio platform for software technologies, including acceleration for creative data science and AI workflows, and Omniverse, giving content creators unmatched tools and capabilities. In desktop, we ramped the RTX 4070, which joined the previously launched RTX 4090, 4080 and the 4070 Ti GPUs. The RTX 4070 is nearly 3x faster than the RTX 2070 and offers our large installed base a spectacular upgrade.
與我們的台式機產品不同,40 系列筆記本電腦支持適用於軟件技術的 NVIDIA Studio 平台,包括創意數據科學和人工智能工作流程的加速,以及為內容創作者提供無與倫比的工具和功能的 Omniverse。在台式機方面,我們推出了 RTX 4070,它加入了之前推出的 RTX 4090、4080 和 4070 Ti GPU。 RTX 4070 比 RTX 2070 快近 3 倍,為我們龐大的安裝基礎提供了驚人的升級。
Last week, we launched the 60 family, RTX 4060 and 4060 Ti, bringing our newest architecture to the world's core gamers starting at just $299. These GPUs for the first time provide 2x the performance of the latest gaming console at mainstream price points. The 4060 Ti is available starting today, while the 4060 will be available in July.
上週,我們推出了 60 系列、RTX 4060 和 4060 Ti,將我們最新的架構帶給全球核心遊戲玩家,起價僅為 299 美元。這些 GPU 首次以主流價位提供 2 倍於最新遊戲機的性能。 4060 Ti 從今天開始上市,而 4060 將於 7 月上市。
Generative AI will be transformative to gaming and content creation from development to runtime. At the Microsoft Build Developer Conference earlier this week, we showcased how Windows PCs and workstations with NVIDIA RTX GPUs will be AI powered at their core. NVIDIA and Microsoft have collaborated on end-to-end software engineering, spanning from the Windows operating system to the NVIDIA graphics drivers and NeMo LLM framework to help make Windows on NVIDIA RTX Tensor Core GPUs a supercharged platform for generative AI.
生成式人工智能將從開發到運行時對遊戲和內容創作產生變革。在本週早些時候的 Microsoft Build 開發者大會上,我們展示了配備 NVIDIA RTX GPU 的 Windows PC 和工作站將如何以人工智能為核心。 NVIDIA 和 Microsoft 在端到端軟件工程方面展開合作,從 Windows 操作系統到 NVIDIA 圖形驅動程序和 NeMo LLM 框架,以幫助使基於 NVIDIA RTX Tensor Core GPU 的 Windows 成為生成人工智能的增壓平台。
Last quarter, we announced a partnership with Microsoft to bring Xbox PC games to GeForce NOW. The first game from this partnership, Gears 5, is now available with more set to be released in the coming months. There are now over 1,600 games on GeForce NOW, the richest content available on any cloud gaming service.
上個季度,我們宣布與微軟合作,將 Xbox PC 遊戲帶到 GeForce NOW。該合作夥伴關係的第一款遊戲 Gears 5 現已上市,未來幾個月將發布更多遊戲。現在 GeForce NOW 上有超過 1,600 款遊戲,是所有云遊戲服務中內容最豐富的遊戲。
Moving to Pro Visualization. Revenue of $295 million was up 31% sequentially and down 53% year-on-year. Sequential growth was driven by stronger workstation demand across both mobile and desktop form factors, with strength in key verticals such as public sector, health care and automotive. We believe the channel inventory correction is behind us. The ramp of our Ada Lovelace GPU architecture in workstations kicks off with major product cycle. At GTC, we announced 6 new RTX GPUs for laptops and desktop workstations with further rollouts planned in the coming quarters.
轉向專業可視化。收入 2.95 億美元,環比增長 31%,同比下降 53%。連續增長是由移動和桌面形式因素對工作站的強勁需求推動的,公共部門、醫療保健和汽車等關鍵垂直領域的實力強勁。我們認為渠道庫存調整已經過去。我們的 Ada Lovelace GPU 架構在工作站中的提升隨著主要產品週期開始。在 GTC 上,我們宣布了 6 款適用於筆記本電腦和台式機工作站的全新 RTX GPU,併計劃在未來幾個季度推出更多產品。
Generative AI is a major new workload for NVIDIA-powered workstation. Our collaboration with Microsoft transforms Windows into the ideal platform for creators and designers harnessing generative AI to elevate their creativity and productivity. At GTC, we announced NVIDIA Omniverse Cloud and NVIDIA fully managed service, running in Microsoft Azure. That includes the full suite of Omniverse applications and NVIDIA OVX infrastructure. Using this full stack cloud environment, customers can design, develop, deploy and manage industrial metaverse applications. NVIDIA Omniverse Cloud will be available starting in the second half of this year. Microsoft and NVIDIA will also connect Office 365 applications with Omniverse.
生成式 AI 是 NVIDIA 驅動的工作站的主要新工作負載。我們與 Microsoft 的合作將 Windows 轉變為創作者和設計師利用生成式 AI 提升創造力和生產力的理想平台。在 GTC 上,我們宣布了在 Microsoft Azure 中運行的 NVIDIA Omniverse Cloud 和 NVIDIA 全託管服務。其中包括全套 Omniverse 應用程序和 NVIDIA OVX 基礎架構。使用這個全棧雲環境,客戶可以設計、開發、部署和管理工業元宇宙應用程序。 NVIDIA Omniverse Cloud 將於今年下半年開始提供。微軟和 NVIDIA 還將 Office 365 應用程序與 Omniverse 連接起來。
Omniverse Cloud is being used by companies to digitize their workflows from design and engineering to smart factories and 3D content generation for marketing. The automotive industry has been a leading early adopter of Omniverse, including companies such as BMW Group, Geely Lotus, General Motors and Jaguar Land Rover.
公司正在使用 Omniverse Cloud 將其工作流程數字化,從設計和工程到智能工廠和用於營銷的 3D 內容生成。汽車行業一直是 Omniverse 的領先早期採用者,包括寶馬集團、吉利蓮花、通用汽車和捷豹路虎等公司。
Moving to Automotive. Revenue was $296 million, up 1% sequentially and up 114% from a year ago. Our strong year-on-year growth was driven by the ramp of the NVIDIA DRIVE Orin across a number of new energy vehicles. As we announced in March, our automotive design win pipeline over the next 6 years now is down at $14 billion, up from $11 billion a year ago, giving us visibility into continued growth over the coming years.
轉向汽車。營收為 2.96 億美元,環比增長 1%,同比增長 114%。我們強勁的同比增長是由 NVIDIA DRIVE Orin 在許多新能源汽車上的增長推動的。正如我們在 3 月份宣布的那樣,我們未來 6 年的汽車設計贏得管道從一年前的 110 億美元下降到 140 億美元,這讓我們可以預見未來幾年的持續增長。
Sequentially, growth moderated as some NEV customers in China are adjusting their production schedules to reflect slower-than-expected demand growth. We expect this dynamic to linger for the rest of the calendar year. During the quarter, we expanded our partnership with BYD, the world's leading manufacturer of NEVs. Our new design win will extend BYD's use of the DRIVE Orin to its next-generation, high-volume Dynasty and Ocean series of vehicles set to start production in calendar 2024.
隨後,隨著中國一些新能源汽車客戶正在調整生產計劃以反映低於預期的需求增長,增長放緩。我們預計這種動態將持續到今年餘下的時間。在本季度,我們擴大了與全球領先的新能源汽車製造商比亞迪的合作夥伴關係。我們贏得的新設計將把比亞迪對 DRIVE Orin 的使用擴展到將於 2024 年投產的下一代大批量王朝和海洋系列汽車。
Moving to the rest of the P&L. GAAP gross margins was 64.6%. Non-GAAP gross margins were 66.8%. Gross margins have now largely recovered to prior peak levels as we have absorbed higher costs and offset them by innovating and delivering higher-valued products as well as products incorporating more and more software. Sequentially, GAAP operating expenses were down 3% and non-GAAP operating expenses were down 1%. We have held OpEx at roughly the same level over the last -- past 4 quarters while working through the inventory corrections in Gaming and Professional Visualization. We now expect to increase investments in the business while also delivering operating leverage.
轉到損益表的其餘部分。 GAAP 毛利率為 64.6%。非美國通用會計準則毛利率為 66.8%。由於我們吸收了更高的成本並通過創新和提供更高價值的產品以及包含越來越多軟件的產品來抵消這些成本,毛利率現在已基本恢復到之前的峰值水平。按美國通用會計準則計算,營業費用環比下降 3%,非美國通用會計準則營業費用下降 1%。在過去 4 個季度中,我們將 OpEx 保持在大致相同的水平,同時進行遊戲和專業可視化的庫存更正。我們現在希望增加對業務的投資,同時提供運營槓桿。
We returned $99 million to shareholders in the form of cash dividends. At the end of Q1, we have approximately $7 billion remaining under our share repurchase authorization through December 2023.
我們以現金股息的形式向股東返還了 9900 萬美元。在第一季度末,到 2023 年 12 月,我們的股票回購授權剩餘約 70 億美元。
Let me turn to the outlook for the second quarter of fiscal '24. Total revenue is expected to be $11 billion, plus or minus 2%. We expect this sequential growth to largely be driven by Data Center, reflecting a steep increase in demand related to generative AI and large language models. This demand has extended our Data Center visibility out a few quarters, and we have procured substantially higher supply for the second half of the year.
讓我談談 24 財年第二季度的展望。總收入預計為 110 億美元,上下浮動 2%。我們預計這種連續增長將主要由數據中心推動,反映出與生成人工智能和大型語言模型相關的需求急劇增長。這種需求使我們的數據中心可見性延長了幾個季度,我們為今年下半年採購了大幅增加的供應量。
GAAP and non-GAAP gross margins are expected to be 68.6% and 70%, respectively, plus or minus 50 basis points. GAAP and non-GAAP operating expenses are expected to be approximately $2.71 billion and $1.9 billion, respectively. GAAP and non-GAAP other income and expenses are expected to be an income of approximately $90 million, excluding gains and losses from nonaffiliated investments.
GAAP 和非 GAAP 毛利率預計分別為 68.6% 和 70%,上下浮動 50 個基點。 GAAP 和非 GAAP 運營費用預計分別約為 27.1 億美元和 19 億美元。 GAAP 和非 GAAP 其他收入和支出預計收入約為 9000 萬美元,不包括非附屬投資的收益和損失。
GAAP and non-GAAP tax rates are expected to be 14%, plus or minus 1%, excluding any discrete items. Capital expenditures are expected to be approximately $300 million to $350 million. Further financial details are included in the CFO commentary and other information available on our IR website.
GAAP 和非 GAAP 稅率預計為 14%,上下浮動 1%,不包括任何離散項目。資本支出預計約為 3 億至 3.5 億美元。更多財務細節包含在首席財務官的評論中以及我們 IR 網站上提供的其他信息中。
In closing, let me highlight some of the upcoming events. Jensen will give the Computex keynote address in person in Taipei this coming Monday, May 29, local time, which will be Sunday evening in the U.S. In addition, we will be attending the BofA Global Technology Conference in San Francisco on June 6 and Rosenblatt Virtual Technology Summit on the age of AI on June 7 and the New Street Future of Transportation Virtual Conference on June 12. Our earnings call to discuss the results of our second quarter fiscal '24 is scheduled for Wednesday, August 23.
最後,讓我強調一些即將舉行的活動。 Jensen 將於當地時間 5 月 29 日星期一親自在台北發表主題演講,這將是美國的星期天晚上。此外,我們將參加 6 月 6 日在舊金山舉行的美國銀行全球技術會議和 Rosenblatt Virtual 6 月 7 日的人工智能時代技術峰會和 6 月 12 日的新街未來交通虛擬會議。我們的財報電話會議定於 8 月 23 日星期三舉行,討論我們第二季度財年 '24 的結果。
Well, that covers our opening remarks. We're now going to open the call for questions. Operator, would you please poll for questions?
好吧,這涵蓋了我們的開場白。我們現在要開始提問。接線員,請你投票提問好嗎?
Operator
Operator
(Operator Instructions)
(操作員說明)
We'll take our first question from Toshiya Hari with Goldman Sachs.
我們將與高盛一起接受 Toshiya Hari 的第一個問題。
Toshiya Hari - MD
Toshiya Hari - MD
Congrats on the strong results and incredible outlook. Just one question on Data Center. Colette, you mentioned the vast majority of the sequential increase in revenue this quarter will come from Data Center. I was curious what the construct is there, if you can speak to what the key drivers are from April to July. And perhaps more importantly, you talked about visibility into the second half of the year. I'm guessing it's more of a supply problem at this point. What kind of sequential growth beyond the July quarter can your supply chain support at this point?
祝賀您取得了出色的成績和令人難以置信的前景。只有一個關於數據中心的問題。 Colette,你提到本季度收入的絕大部分環比增長將來自數據中心。我很好奇那裡的構造是什麼,如果你能談談 4 月到 7 月的關鍵驅動因素是什麼。也許更重要的是,您談到了下半年的可見性。我猜現在更多的是供應問題。在這一點上,您的供應鏈可以支持 7 月季度之後的什麼樣的連續增長?
Colette M. Kress - Executive VP & CFO
Colette M. Kress - Executive VP & CFO
Okay. So a lot of different questions there, so let me see if I can start, and I'm sure Jensen will have some following up comments. So when we talk about our sequential growth that were expected between Q1 and Q2, our generative AI large language models are driving this surge in demand, and it's broad-based across both our consumer Internet companies, our CSPs, our enterprises and our AI start-ups. It is also interest in both of our architectures, both of our Hopper latest architecture as well as our Ampere architecture. This is not surprising as we generally often sell both of our architectures at the same time.
好的。那裡有很多不同的問題,讓我看看我是否可以開始,我相信 Jensen 會有一些後續評論。因此,當我們談論我們在第一季度和第二季度之間預期的連續增長時,我們的生成 AI 大型語言模型正在推動這種需求激增,並且它廣泛存在於我們的消費者互聯網公司、我們的 CSP、我們的企業和我們的 AI 開始-UPS。它也對我們的兩種架構感興趣,包括我們的 Hopper 最新架構以及我們的 Ampere 架構。這並不奇怪,因為我們通常會同時出售我們的兩種架構。
This is also a key area where deep recommendators are driving growth. And we also expect to see growth both in our computing as well as in our networking business. So those are some of the key things that we have baked in when we think about the guidance that we have provided to Q2. We also surfaced in our opening remarks that we are working on both supply today for this quarter, but we have also procured a substantial amount of supply for the second half. We have some significant supply chain flow to serve our significant customer demand that we see, and this is demand that we see across a wide range of different customers.
這也是深度推薦推動增長的關鍵領域。我們還希望在我們的計算和網絡業務中看到增長。因此,當我們考慮為第二季度提供的指導時,這些是我們考慮的一些關鍵事項。我們還在開場白中提到,我們今天正在處理本季度的供應,但我們也為下半年採購了大量供應。我們有一些重要的供應鏈流程來滿足我們所看到的重要客戶需求,這是我們在廣泛的不同客戶中看到的需求。
They are building platforms for some of the largest enterprises but also setting things up at the CSPs and the large consumer Internet companies. So we have visibility right now for our data center demand that has probably extended out a few quarters, and this led us to working on quickly procuring that substantial supply for the second half.
他們正在為一些最大的企業構建平台,同時也在 CSP 和大型消費互聯網公司中進行設置。因此,我們現在可以看到可能已經持續幾個季度的數據中心需求,這促使我們努力為下半年快速採購大量供應。
I'm going to pause there and see if Jensen wants to add a little bit more.
我要在那兒暫停一下,看看 Jensen 是否想再補充一點。
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
I thought that was great, Colette. Thank you.
我覺得那很棒,科萊特。謝謝。
Operator
Operator
Next, we'll go to C.J. Muse with Evercore ISI.
接下來,我們將與 Evercore ISI 一起前往 C.J. Muse。
Christopher James Muse - Senior MD, Head of Global Semiconductor Research & Senior Equity Research Analyst
Christopher James Muse - Senior MD, Head of Global Semiconductor Research & Senior Equity Research Analyst
I guess with Data Center essentially doubling quarter-on-quarter, 2 natural kind of questions that relate to 1 another come to mind. Number one, where are we in terms of driving acceleration into servers to support AI? And as part of that, as you deal with longer cycle times with TSMC and your other partners, how are you thinking about managing the commitments there with where you want to manage your lead times in the coming years to best kind of match that supply and demand?
我猜數據中心基本上是每季度翻一番,我會想到 2 個相互關聯的自然問題。第一,我們在推動服務器加速以支持 AI 方面進展如何?作為其中的一部分,當您與台積電和您的其他合作夥伴處理更長的周期時間時,您如何考慮管理那裡的承諾以及您希望在未來幾年管理您的交貨時間的地方,以最好地匹配供應和要求?
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
Yes, C.J., thanks for the question. I'll start backwards. The -- remember, we were in full production of both Ampere and Hopper when the ChatGPT moment came, and it helped everybody crystallize how to transition from the technology of large language models to a product and service based on a chatbot. The integration of guardrails and alignment systems with reinforcement learning human feedback, knowledge vector databases for proprietary knowledge, connection to search, all of that came together in a really wonderful way. And it's the reason why I call it the iPhone moment. All the technology came together and helped everybody realize what an amazing product it can be and what capabilities it can have.
是的,C.J.,謝謝你的提問。我會倒著開始。 - 請記住,當 ChatGPT 時刻到來時,我們正在全面生產 Ampere 和 Hopper,它幫助每個人明確瞭如何從大型語言模型技術過渡到基於聊天機器人的產品和服務。護欄和對齊系統與強化學習人類反饋的集成、專有知識的知識向量數據庫、與搜索的連接,所有這些都以一種非常美妙的方式結合在一起。這就是我稱之為 iPhone 時刻的原因。所有的技術匯集在一起,幫助每個人意識到它可以是多麼了不起的產品以及它可以擁有什麼功能。
And so we were already in full production. NVIDIA's supply chain flow and our supply chain is very significant as you know. And we build supercomputers in volume, and these are giant systems and we build them in volume. It includes, of course, the GPUs, but on our GPUs, the system boards have 35,000 other components. And the networking and the fiber optics and the incredible transceivers and the NICs, the SmartNICs, the switches, all of that has to come together in order for us to stand up a data center. And so we were already in full production when the moment came. We had to really significantly increase our procurement substantially for the second half, as Colette said.
所以我們已經全面投入生產。如您所知,NVIDIA 的供應鏈流程和我們的供應鏈非常重要。我們批量構建超級計算機,這些都是巨大的系統,我們批量構建它們。它當然包括 GPU,但在我們的 GPU 上,系統板有 35,000 個其他組件。網絡、光纖、令人難以置信的收發器和 NIC、SmartNIC、交換機,所有這些都必須結合在一起,才能讓我們建立一個數據中心。因此,當那一刻到來時,我們已經全面投入生產。正如 Colette 所說,我們不得不在下半年大幅增加採購量。
Now let me talk about the bigger picture and why the entire world's data centers are moving toward accelerated computing. It's been known for some time, and you've heard me talk about it, that accelerated computing is a full stack problem but -- it is full stack challenged. But if you could successfully do it in a large number of application domain that's taken us 15 years, it's sufficiently that almost the entire data center's major applications could be accelerated. You could reduce the amount of energy consumed and the amount of cost for a data center substantially by an order of magnitude. It costs a lot of money to do it because you have to do all the software and everything and you have to build all the systems and so on and so forth, but we've been at it for 15 years.
現在讓我談談大局以及為什麼整個世界的數據中心都在朝著加速計算的方向發展。眾所周知,加速計算是一個全棧問題,而且你也聽過我談論它,但是——它是全棧挑戰。但是如果你能在我們用了15年的大量應用領域成功地做到這一點,就足以讓幾乎整個數據中心的主要應用都得到加速。您可以將數據中心的能耗和成本大幅降低一個數量級。這樣做要花很多錢,因為你必須做所有的軟件和一切,你必須構建所有的系統等等,但我們已經做了 15 年了。
And what happened is when generative AI came along, it triggered a killer app for this computing platform that's been in preparation for some time. And so now we see ourselves in 2 simultaneous transitions. The world's $1 trillion data center is nearly populated entirely by CPUs today. And I -- $1 trillion, $250 billion a year, it's growing of course. But over the last 4 years, call it $1 trillion worth of infrastructure installed, and it's all completely based on CPUs and dumb NICs. It's basically unaccelerated.
發生的事情是,當生成式 AI 出現時,它為這個已經準備了一段時間的計算平台觸發了一個殺手級應用程序。所以現在我們看到自己處於 2 個同時轉換中。今天,世界上價值 1 萬億美元的數據中心幾乎完全由 CPU 組成。我——1 萬億美元,每年 2500 億美元,當然還在增長。但在過去的 4 年裡,安裝了價值 1 萬億美元的基礎設施,而且它完全基於 CPU 和啞 NIC。基本上是不加速的。
In the future, it's fairly clear now with this -- with generative AI becoming the primary workload of most of the world's data centers generating information, it is very clear now that -- and the fact that accelerated computing is so energy efficient, that the budget of a data center will shift very dramatically towards accelerated computing, and you're seeing that now. We're going through that moment right now as we speak, while the world's data center CapEx budget is limited. But at the same time, we're seeing incredible orders to retool the world's data centers.
在未來,這一點現在已經很清楚了——隨著生成人工智能成為世界上大多數數據中心生成信息的主要工作負載,現在很清楚——加速計算是如此節能這一事實,以至於數據中心的預算將非常顯著地轉向加速計算,你現在就看到了。就在我們說話的時候,我們正在經歷那個時刻,而世界數據中心的資本支出預算是有限的。但與此同時,我們看到了令人難以置信的重組全球數據中心的訂單。
And so I think you're starting -- you're seeing the beginning of, call it, a 10-year transition to basically recycle or reclaim the world's data centers and build it out as accelerated computing. You have a pretty dramatic shift in the spend of a data center from traditional computing and to accelerated computing with SmartNICs, smart switches, of course, GPUs and the workload is going to be predominantly generative AI.
所以我認為你正在開始 - 你正在看到一個 10 年過渡的開始,基本上回收或回收世界數據中心並將其構建為加速計算。數據中心的支出從傳統計算轉向使用 SmartNIC、智能交換機,當然還有 GPU 的加速計算,工作負載將主要是生成 AI。
Operator
Operator
We'll move to our next question, Vivek Arya with BofA Securities.
我們將轉到下一個問題,美國銀行證券公司的 Vivek Arya。
Vivek Arya - MD in Equity Research & Research Analyst
Vivek Arya - MD in Equity Research & Research Analyst
Colette, just wanted to clarify, does visibility mean data center sales can continue to grow sequentially in Q3 and Q4? Or do they sustain at Q2 level? So I just wanted to clarify that. And then, Jensen, my question is that given this very strong demand environment, what does it do to the competitive landscape? Does it invite more competition in terms of custom ASICs? Does it invite more competition in terms of other GPU solutions or other kinds of solutions? What -- how do you see the competitive landscape change over the next 2 to 3 years?
Colette,只是想澄清一下,可見性是否意味著數據中心銷售額可以在第三季度和第四季度繼續連續增長?還是維持在 Q2 水平?所以我只想澄清一下。然後,詹森,我的問題是,鑑於這種非常強勁的需求環境,它對競爭格局有何影響?它會在定制 ASIC 方面引發更多競爭嗎?它會在其他 GPU 解決方案或其他類型的解決方案方面引發更多競爭嗎?什麼——您如何看待未來 2 到 3 年的競爭格局變化?
Colette M. Kress - Executive VP & CFO
Colette M. Kress - Executive VP & CFO
Yes, Vivek, thanks for the question. Let me see if I can add a little bit more color. We believe that the supply that we will have for the second half of the year will be substantially larger than H1. So we are expecting not only the demand that we just saw in this last quarter, the demand that we have in Q2 for our forecast but also planning on seeing something in the second half of the year. We just have to be careful here, but we're not here to guide on the second half. But yes, we do plan a substantial increase in the second half compared to the first half.
是的,Vivek,謝謝你的提問。讓我看看我是否可以添加更多顏色。我們相信,我們今年下半年的供應量將大大超過上半年。因此,我們不僅期待我們在上個季度看到的需求,我們在第二季度的預測需求,而且還計劃在今年下半年看到一些東西。我們在這裡必須小心,但我們不是來指導下半場的。但是,是的,與上半年相比,我們確實計劃在下半年大幅增加。
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
Regarding competition, we have competition from every direction. Start-ups, really, really well funded and innovative start-ups, countless of them all over the world. We have competitions from existing semiconductor companies. We have competition from CSPs with internal projects, and many of you know about most of these. And so we're mindful of competition all the time, and we get competition all the time.
關於競爭,我們有來自各個方向的競爭。初創企業,真的,真的是資金雄厚的創新型初創企業,在世界各地數不勝數。我們有來自現有半導體公司的競爭。我們有來自 CSP 的內部項目競爭,你們中的許多人都知道其中的大部分。因此,我們一直都在關注競爭,而且我們一直都有競爭。
NVIDIA's value proposition at the core is we are the lowest cost solution. We're the lowest TCO solution. And the reason for that is because accelerated computing is 2 things that I talk about often, which is it's a full stack problem. It's a full stack challenge. You have to engineer all of the software and all the libraries and all the algorithms, integrate them into and optimize the frameworks and optimize it for the architecture of not just one chip but the architecture of an entire data center but all the way into the frameworks, all the way into the models.
NVIDIA 的核心價值主張是我們是成本最低的解決方案。我們是 TCO 最低的解決方案。這樣做的原因是因為加速計算是我經常談論的兩件事,這是一個全棧問題。這是一個全棧挑戰。你必須設計所有的軟件、所有的庫和所有的算法,將它們集成到框架中並對其進行優化,不僅針對一個芯片的架構,而且針對整個數據中心的架構進行優化,而是一直優化到框架中,一直到模型。
And the amount of engineering and distributed computing, fundamental computer science work is really quite extraordinary. It is the hardest computing as we know. And so number one, it's a full stack challenge and you have to optimize it across the whole thing and across just a mind-blowing number of stacks. We have 400 acceleration libraries. As you know, the amount of libraries and frameworks that we accelerate is pretty mind blowing.
工程和分佈式計算的數量,基礎計算機科學工作確實非常了不起。這是我們所知道的最難的計算。因此,第一,這是一個完整的堆棧挑戰,你必須在整個過程中以及在數量驚人的堆棧中對其進行優化。我們有 400 個加速庫。如您所知,我們加速的庫和框架的數量非常驚人。
The second part is that generative AI is a large-scale problem, and it's a data center scale problem. It's another way of thinking that the computer is the data center or the data center is the computer. It's not the chip. It's the data center, and it's never happened like us before. And in this particular environment, your networking operating system, your distributed computing engines, your understanding of the architecture of the networking gear, the switches and the computing systems, the computing fabric, that entire system is your computer, and that's what you're trying to operate. And so in order to get the best performance, you have to understand full stack and understand data center scale. And that's what accelerated computing is.
第二部分是生成式人工智能是一個大規模問題,是一個數據中心規模的問題。計算機是數據中心或數據中心是計算機是另一種思維方式。這不是芯片。這是數據中心,以前從未像我們這樣發生過。在這個特定的環境中,你的網絡操作系統,你的分佈式計算引擎,你對網絡設備架構的理解,交換機和計算系統,計算結構,整個系統就是你的計算機,這就是你試圖操作。因此,為了獲得最佳性能,您必須了解全棧並了解數據中心規模。這就是加速計算。
The second thing is that utilization, which talks about the amount of the types of applications that you can accelerate and the versatility of your architecture, keeps that utilization high. If you can do one thing and doing one thing only incredibly fast, then your data center is largely underutilized, and it's hard to scale that out. NVIDIA's universal GPU and the fact that we accelerate so many stacks makes our utilization incredibly high. And so number one is throughput, and that's software-intensive problems and data center architecture problem. The second is digitalization versatility problem.
第二件事是利用率,它討論了可以加速的應用程序類型的數量和架構的多功能性,使利用率保持高水平。如果您可以做一件事並且只以難以置信的速度做一件事,那麼您的數據中心在很大程度上未得到充分利用,並且很難將其擴展。 NVIDIA 的通用 GPU 以及我們加速這麼多堆棧的事實使我們的利用率非常高。因此,第一是吞吐量,這是軟件密集型問題和數據中心架構問題。二是數字化通用性問題。
And the third is just data center expertise. We've built 5 data centers of our own, and we've helped companies all over the world build data centers. And we integrate our architecture into all the world's clouds. From the moment of delivery of the product to the standing up and the deployment, the time to operations of a data center is measured not -- it can -- if you're not good at it and not proficient at it, it could take months. Standing up a supercomputer, let's see, some of the largest supercomputers in the world were installed about 1.5 years ago, and now they're coming online.
第三是數據中心專業知識。我們已經建立了5個自己的數據中心,我們已經幫助世界各地的公司建立數據中心。我們將我們的架構集成到世界上所有的雲中。從交付產品的那一刻到站起來和部署,數據中心的運營時間不是——它可以——如果你不擅長它並且不精通它,它可能需要個月。站在一台超級計算機上,讓我們看看,世界上一些最大的超級計算機是大約 1.5 年前安裝的,現在它們正在上線。
And so it's not unheard of to see a delivery to operations of about a year. Our delivery to operation's measured in weeks. And that's -- we've taken data centers and supercomputers, and we've turned it into products. And the expertise of the team in doing that is incredible.
因此,交付運營大約一年的情況並非聞所未聞。我們交付運營的時間以周為單位。那就是——我們採用了數據中心和超級計算機,並將其轉化為產品。團隊在這方面的專業知識令人難以置信。
And so our value proposition is in final analysis. All of this technology translates into infrastructure, the highest throughput and the lowest possible cost. And so I think our market is, of course, very, very competitive, very large, but the challenge is really, really great.
因此,我們的價值主張歸根結底。所有這些技術都轉化為基礎設施、最高吞吐量和盡可能低的成本。所以我認為我們的市場當然非常非常有競爭力,非常大,但挑戰真的非常大。
Operator
Operator
Next, we go to Aaron Rakers with Wells Fargo.
接下來,我們將前往 Aaron Rakers 和 Wells Fargo。
Aaron Christopher Rakers - MD of IT Hardware & Networking Equipment and Senior Equity Analyst
Aaron Christopher Rakers - MD of IT Hardware & Networking Equipment and Senior Equity Analyst
Congrats on the quarter. As we kind of think about unpacking the various different growth drivers of the Data Center business going forward, I'm curious, Colette, of just how we should think about the monetization effect of software considering that the expansion of your cloud service agreements continues to grow. I'm curious of what -- where do you think we're at in terms of that approach, in terms of the AI enterprise software suite and other drivers of software-only revenue going forward?
祝賀這個季度。當我們考慮分析未來數據中心業務的各種不同增長驅動因素時,我很好奇,Colette,考慮到您的雲服務協議不斷擴展,我們應該如何考慮軟件的貨幣化效應生長。我很好奇 - 你認為我們在這種方法方面,在 AI 企業軟件套件和其他純軟件收入驅動因素方面處於什麼位置?
Colette M. Kress - Executive VP & CFO
Colette M. Kress - Executive VP & CFO
Thanks for the question. Software is really important to our accelerated platforms. Not only do we have a substantial amount of software that we are including in our newest architecture and essentially all products that we have, we're now with many different models to help customers start their work in generative AI and accelerated computing.
謝謝你的問題。軟件對我們的加速平台非常重要。我們不僅擁有包含在我們最新架構中的大量軟件和我們擁有的幾乎所有產品,我們現在還擁有許多不同的模型來幫助客戶開始他們在生成 AI 和加速計算方面的工作。
So anything that we have here from DGX Cloud on providing those services, helping them build models or, as you've discussed, the importance of NVIDIA AI Enterprise, essentially, that operating system for AI, so all things should continue to grow as we go forward, both the architecture and the infrastructure as well as the -- both availability of the software and our ability to monetize that with it as well. I'll turn it over to Jensen if needs to comment.
因此,我們從 DGX Cloud 那裡獲得的關於提供這些服務、幫助他們構建模型的任何東西,或者正如您所討論的那樣,NVIDIA AI Enterprise 的重要性,本質上是 AI 操作系統,所以隨著我們的發展,所有東西都應該繼續增長向前邁進,架構和基礎設施以及 - 軟件的可用性以及我們通過它獲利的能力。如果需要發表評論,我會將其轉交給 Jensen。
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
Yes. We can see in real time the growth of generative AI in CSPs, both for training the models, refining the models as well as deploying the models. As Colette said earlier, inference is now a major driver of accelerated computing because generative AI is used so capably in so many applications already.
是的。我們可以實時看到 CSP 中生成人工智能的增長,包括訓練模型、改進模型以及部署模型。正如 Colette 之前所說,推理現在是加速計算的主要驅動力,因為生成 AI 已經在如此多的應用程序中得到瞭如此有效的使用。
There are 2 segments that require a new stack of software, and the 2 segments are enterprise and industrials. Enterprise requires a new stack of software because many enterprises need to have all the capabilities that we've talked about, whether it's large language models, the ability to adapt them for your proprietary use case and your proprietary data and alignment to your own principles and your own operating domains. You want to have the ability to be able to do that in a high-performance computing sandbox, and we call that DGX Cloud, and to create that model.
有兩個部分需要新的軟件堆棧,這兩個部分是企業和工業。企業需要一個新的軟件堆棧,因為許多企業需要擁有我們討論過的所有功能,無論是大型語言模型,還是針對您的專有用例和您的專有數據調整它們的能力,以及與您自己的原則保持一致的能力,您自己的操作域。你希望能夠在高性能計算沙箱中做到這一點,我們稱之為 DGX 雲,並創建該模型。
Then you want to deploy your chatbot or your AI in any cloud because you have services and you have agreements with multiple cloud vendors and depending on the applications, you might deploy it on various clouds. And for the enterprise, we have NVIDIA AI Foundations for helping you create custom models and we have NVIDIA AI Enterprise. NVIDIA AI Enterprise is the only accelerated stack -- GPU accelerated stack in the world that is enterprise safe and enterprise supported. There are constant patching that you have to do. There are 4,000 different packages that build up NVIDIA AI Enterprise and represents the operating engine -- end-to-end operating engine of the entire AI workflow.
然後你想在任何云中部署你的聊天機器人或你的人工智能,因為你有服務並且你與多個雲供應商有協議並且根據應用程序,你可以將它部署在各種雲上。對於企業,我們有 NVIDIA AI Foundations 來幫助您創建自定義模型,我們還有 NVIDIA AI Enterprise。 NVIDIA AI Enterprise 是世界上唯一的加速堆棧——GPU 加速堆棧,它是企業安全和企業支持的。您必須不斷進行修補。有 4,000 個不同的包構建了 NVIDIA AI Enterprise 並代表了運行引擎——整個 AI 工作流的端到端運行引擎。
It's the only one of its kind from data ingestion, data processing. Obviously, in order to train an AI model, you have a lot of data you have to process and package up and curate and align. And there's just a whole bunch of stuff that you have to do to the data to prepare it for training. That amount of data could consume some 40%, 50%, 60% of your computing time. And so data processing is a very big deal. And then the second aspect of it is training the model, refining the model. And the third is deploying model for inferencing.
它是同類產品中唯一一款來自數據攝取、數據處理的產品。顯然,為了訓練 AI 模型,您必須處理、打包、整理和對齊大量數據。您需要對數據做一大堆事情來為訓練做準備。這些數據量可能會消耗您大約 40%、50% 和 60% 的計算時間。因此,數據處理是一件大事。然後它的第二個方面是訓練模型,改進模型。第三是部署推理模型。
NVIDIA AI Enterprise supports and patches and security patches continuously all of those 4,000 packages of software. And for an enterprise that wants to deploy their engines just like they want to deploy Red Hat Linux, this is incredibly complicated software. In order to deploy that in every cloud and as well as on-prem, it has to be secure. It has to be supported. And so NVIDIA AI Enterprise is the second part.
NVIDIA AI Enterprise 不斷為所有這 4,000 個軟件包提供補丁和安全補丁。對於想要像部署 Red Hat Linux 一樣部署引擎的企業來說,這是一個極其複雜的軟件。為了在每個雲和本地部署它,它必須是安全的。它必須得到支持。因此,NVIDIA AI Enterprise 是第二部分。
The third is Omniverse. Just as people are starting to realize that you need to align an AI to ethics, the same for robotics, you need to align the AI for physics, and aligning an AI for ethics includes a technology called reinforcement learning human feedback. In the case of industrial applications and robotics, it's reinforcement learning Omniverse feedback. And Omniverse is a vital engine for software-defined and robotic applications and industries. And so Omniverse also needs to be a cloud service platform.
第三個是全宇宙。正如人們開始意識到您需要使 AI 符合道德規範一樣,機器人技術也是如此,您需要使 AI 符合物理學規範,而使 AI 符合倫理規範包括一種稱為強化學習人類反饋的技術。就工業應用和機器人技術而言,它是強化學習 Omniverse 反饋。 Omniverse 是軟件定義和機器人應用程序和行業的重要引擎。所以Omniverse也需要做一個雲服務平台。
And so our software stack, the 3 software stacks, AI Foundations, AI Enterprise and Omniverse, runs in all of the world's clouds that we have partnerships, DGX Cloud partnerships with. Azure, we have partnerships on both AI as well as Omniverse. With GCP and Oracle, we have great partnerships in DGX Cloud for AI, and AI Enterprise is integrated into all 3 of them.
因此,我們的軟件堆棧,即 3 個軟件堆棧,AI Foundations、AI Enterprise 和 Omniverse,在我們有合作夥伴關係的全球所有云中運行,DGX Cloud 合作夥伴關係。 Azure,我們在 AI 和 Omniverse 方面都有合作夥伴關係。通過 GCP 和 Oracle,我們在 DGX Cloud for AI 方面建立了良好的合作夥伴關係,並且 AI Enterprise 已集成到所有這三個合作夥伴中。
And so I think the -- in order for us to extend the reach of AI beyond the cloud and into the world's enterprise and into the world's industries, you need 2 new types of -- you need new software stacks in order to make that happen. And by putting it in the cloud, integrated into the world CSP clouds, it's a great way for us to partner with the sales and the marketing team and the leadership team of all the cloud vendors.
所以我認為 - 為了讓我們將人工智能的範圍擴展到雲之外,進入世界企業和世界各行業,你需要兩種新類型 - 你需要新的軟件堆棧才能實現這一目標.通過將其放入雲中,集成到世界 CSP 雲中,這是我們與所有云供應商的銷售和營銷團隊以及領導團隊合作的好方法。
Operator
Operator
Next, we'll go to Timothy Arcuri with UBS.
接下來,我們將與瑞銀一起去 Timothy Arcuri。
Timothy Michael Arcuri - MD and Head of Semiconductors & Semiconductor Equipment
Timothy Michael Arcuri - MD and Head of Semiconductors & Semiconductor Equipment
I had a question and then I had a clarification as well. So the question, first, is, Jensen, on the InfiniBand versus Ethernet argument, can you sort of speak to that debate and maybe how you see it playing out? I know you need the low latency of InfiniBand for AI. But can you sort of talk about the attach rate of your InfiniBand solutions to what you're shipping on the core compute side and maybe whether that's similarly crowding out Ethernet like you are with -- on the compute side? And then the clarification, Colette, is that there wasn't a share buyback despite you still having about $7 billion on the share repo authorization. Was that just timing?
我有一個問題,然後我也得到了澄清。所以第一個問題是,Jensen,關於 InfiniBand 與以太網的爭論,你能談談這場辯論,也許你是如何看待它的結果的?我知道您需要 InfiniBand 的低延遲 AI。但是,您能否談談您的 InfiniBand 解決方案與您在核心計算端交付的產品的連接率,以及這是否像您在計算端所使用的那樣同樣排擠以太網?然後澄清一下,科萊特,儘管你仍然有大約 70 億美元的股票回購授權,但沒有股票回購。那隻是時機嗎?
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
Colette, how about you go first? You take the question first.
科萊特,你先走如何?你先回答問題。
Colette M. Kress - Executive VP & CFO
Colette M. Kress - Executive VP & CFO
Sure. That is correct. We have $7 billion available in our current authorization for repurchases. We did not repurchase anything in this last quarter, but we do repurchase opportunistically and we'll consider that as we go forward as well.
當然。那是對的。我們目前的回購授權中有 70 億美元可供使用。我們在上個季度沒有回購任何東西,但我們確實有機會回購,我們也會在前進的過程中考慮這一點。
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
InfiniBand and Ethernet are -- target different applications in a data center. They both have their place. InfiniBand had a record quarter. We're going to have a giant record year. And InfiniBand has a really -- NVIDIA's Quantum InfiniBand has an exceptional road map. It's going to be really incredible.
InfiniBand 和以太網針對數據中心中的不同應用程序。他們都有自己的位置。 InfiniBand 有一個創紀錄的季度。我們將迎來創紀錄的一年。 InfiniBand 有一個真正的——NVIDIA 的 Quantum InfiniBand 有一個特殊的路線圖。這將是非常不可思議的。
The 2 networks are very different. InfiniBand is designed for an AI factory if you will. If that data center is running a few applications for a few people for a specific use case and is doing it continuously, and that infrastructure costs you, pick a number, $500 million, the difference between InfiniBand and Ethernet could be 15%, 20% in overall throughput. And if you spent $500 million on an infrastructure and the difference is 10% to 20%, and it's $100 million, InfiniBand's basically free. That's the reason why people use it.
這兩個網絡非常不同。如果您願意,InfiniBand 是為 AI 工廠設計的。如果那個數據中心針對特定用例為幾個人運行幾個應用程序並且持續運行,並且該基礎設施花費了你,選擇一個數字,5 億美元,InfiniBand 和以太網之間的差異可能是 15%、20%在總吞吐量。如果你在基礎設施上花費了 5 億美元,差異是 10% 到 20%,而且是 1 億美元,InfiniBand 基本上是免費的。這就是人們使用它的原因。
InfiniBand is effectively free. The difference in data center throughput is just -- it's too great to ignore. And you're using it for that one application. And so however, if your data center is a cloud data center and it's multi-tenant, it's a bunch of little jobs, a bunch of little jobs and is shared by millions of people, then Ethernet is really the right answer. There's a new segment in the middle where the cloud is becoming a generative AI cloud. It's not an AI factory per se, but it's still a multi-tenant cloud, but it wants to run generative AI workloads.
InfiniBand 實際上是免費的。數據中心吞吐量的差異太大了,不容忽視。您正在將它用於那個應用程序。因此,如果您的數據中心是一個雲數據中心並且是多租戶的,它是一堆小工作,一堆小工作並且由數百萬人共享,那麼以太網確實是正確的答案。中間有一個新的部分,雲正在成為一個生成的 AI 雲。它本身不是 AI 工廠,但它仍然是一個多租戶雲,但它想要運行生成 AI 工作負載。
This new segment is a wonderful opportunity. And at Computex -- I referred to it at the last GTC. At Computex, we're going to announce a major product line for this segment, which is for Ethernet-focused generative AI application type of clouds. But InfiniBand is doing fantastically, and we're doing record numbers quarter-on-quarter, year-on-year.
這個新細分市場是一個絕佳的機會。在 Computex 上——我在上屆 GTC 上提到了它。在 Computex 上,我們將宣布該細分市場的主要產品線,該產品線適用於以以太網為中心的生成 AI 應用類型的雲。但 InfiniBand 的表現非常出色,而且我們的季度環比和年環比都創下了歷史新高。
Operator
Operator
Next, we'll go to Stacy Rasgon with Bernstein Research.
接下來,我們將前往 Bernstein Research 的 Stacy Rasgon。
Stacy Aaron Rasgon - Senior Analyst
Stacy Aaron Rasgon - Senior Analyst
I had a question on inference versus training for generative AI. So you're talking about inference as being a very large opportunity. I guess 2 subparts of that. Is that because inference basically scales with like the usage versus like training is more of a one and done? And can you give us some sort of -- even if it's just like qualitatively, like do you think inference is bigger than training or vice versa? Like if it's bigger, how much bigger? Is it like the opportunity, is it 5x? Is it 10x? Is there anything you can give us on those 2 workloads within generative AI? Would be helpful.
我有一個關於生成 AI 的推理與訓練的問題。所以你說的推理是一個非常大的機會。我猜是其中的 2 個子部分。那是因為推理基本上與使用量成比例,而不是像訓練更像是一個完成的?你能給我們一些——即使它只是定性的,就像你認為推理比訓練大還是反之亦然?如果它更大,那麼大多少?就像機會一樣,是 5 倍嗎?是10倍嗎?關於生成式 AI 中的這 2 個工作負載,您有什麼可以給我們的嗎?會有幫助的。
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
Yes, I'll work backwards. You're never done with training. You're always -- every time you deploy, you're collecting new data. When you collect new data, you train with the new data. And so you're never done training. You're never done producing and processing a vector database that augments the large language model. You're never done with vectorizing all of the collected structured, unstructured data that you have. And so whether you're building a recommender system, a large language model, a vector database, these are probably the 3 major applications of -- the 3 core engines, if you will, of the future of computing as well as a bunch of other stuff.
是的,我會逆向工作。你永遠不會完成訓練。您總是——每次部署時,您都在收集新數據。當您收集新數據時,您將使用新數據進行訓練。所以你永遠不會完成訓練。您永遠不會完成生成和處理用於擴充大型語言模型的矢量數據庫。您永遠不會完成矢量化所有收集到的結構化和非結構化數據的工作。因此,無論您是在構建推薦系統、大型語言模型、向量數據庫,這些可能是未來計算的 3 個主要應用程序——如果您願意的話,這 3 個核心引擎以及一堆其他的東西。
But obviously, these are very -- 3 very important ones. They are always, always running. You're going to see that more and more companies realize they have a factory for intelligence -- an intelligence factory. And in that particular case, it's largely dedicated to training and processing data and vectorizing data and learning representation of the data, so on and so forth.
但顯然,這些是非常 - 3 個非常重要的。他們總是,總是在奔跑。你會看到越來越多的公司意識到他們有一個智能工廠——智能工廠。在那種特殊情況下,它主要致力於訓練和處理數據、矢量化數據和學習數據的表示,等等。
The inference part of it are APIs that are either open APIs that can be connected to all kinds of applications, APIs that are integrated into workflows but APIs of all kinds. There will be hundreds of APIs in a company. Some of them, they built themselves. Some of them, part that -- many of them could come from companies like ServiceNow and Adobe that we're partnering with in AI Foundations. And they'll create a whole bunch of generative AI APIs that companies can then connect into their workflows or use as an application. And of course, there'll be a whole bunch of Internet service companies.
它的推理部分是 API,這些 API 要么是可以連接到各種應用程序的開放 API,要么是集成到工作流中的 API,要么是各種 API。一個公司會有上百個API。其中一些是他們自己建造的。其中一些,部分 - 其中許多可能來自我們在 AI Foundations 中與之合作的 ServiceNow 和 Adobe 等公司。他們將創建一大堆生成式 AI API,公司可以將這些 API 連接到他們的工作流程中或用作應用程序。當然,還會有一大堆互聯網服務公司。
And so I think you're seeing for the very first time, simultaneously, a very significant growth in the segment of AI factories as well as a market that -- a segment that really didn't exist before but now it's growing exponentially practically by the week for AI inference with APIs. The simple way to think about it in the end is that the world has a $1 trillion of data center installed and it used to be 100% CPUs. In the future, we know -- we've heard it in enough places, and I think this year's ISC keynote was actually about the end of Moore's Law.
所以我認為你是第一次同時看到 AI 工廠部分和市場的非常顯著的增長——這個部分以前並不存在,但現在它實際上呈指數增長使用 API 進行 AI 推理的一周。最終的簡單思考方式是,世界上安裝了 1 萬億美元的數據中心,並且曾經是 100% 的 CPU。在未來,我們知道——我們已經在足夠多的地方聽到過,我認為今年的 ISC 主題演講實際上是關於摩爾定律的終結。
We've seen it in a lot of places now that you can't reasonably scale out data centers with general purpose computing and that accelerated computing is the path forward. And now it's got a killer app. It's called generative AI. And so the easiest way to think about that is your $1 trillion infrastructure. Every quarter's capital, CapEx budget would lean very heavily into generative AI, into accelerated computing infrastructure, everywhere from the number of GPUs that would be used in the CapEx budget to the accelerated switches and accelerated networking chips that connect them all.
我們現在已經在很多地方看到它,您無法通過通用計算合理地擴展數據中心,而加速計算是前進的道路。現在它有了一個殺手級應用程序。它被稱為生成人工智能。因此,考慮這一點的最簡單方法是您的 1 萬億美元基礎設施。每個季度的資本、資本支出預算都將大量依賴於生成人工智能、加速計算基礎設施,無處不在,從資本支出預算中使用的 GPU 數量到連接它們的加速交換機和加速網絡芯片。
The easiest way to think about that is, over the next 4, 5, 10 years, most of that $1 trillion and then compensating, adjusting for all the growth in data center still, it will be largely generative AI. And so that's probably the easiest way to think about that, and that's training as well as inference.
最簡單的思考方式是,在接下來的 4、5、10 年裡,這 1 萬億美元中的大部分,然後補償,調整數據中心的所有增長,主要是生成人工智能。因此,這可能是最簡單的思考方式,即訓練和推理。
Operator
Operator
Next, we'll go to Joseph Moore with Morgan Stanley.
接下來,我們將與摩根士丹利一起去約瑟夫摩爾。
Joseph Lawrence Moore - Executive Director
Joseph Lawrence Moore - Executive Director
I wanted to follow up on that in terms of the focus on inference. It's pretty clear that this is a really big opportunity around large language models. But the cloud customers are also talking about trying to reduce cost per query by very significant amounts. Can you talk about the ramifications of that for you guys? Is that where some of the specialty inference products that you launched at GTC come in? And just how are you going to help your customers get the cost per query down?
我想在對推理的關注方面跟進這一點。很明顯,這是圍繞大型語言模型的一個非常大的機會。但是雲客戶也在談論嘗試以非常顯著的數量降低每次查詢的成本。你能談談這對你們的影響嗎?這就是你們在 GTC 上推出的一些專業推理產品的用武之地嗎?您將如何幫助您的客戶降低每次查詢的成本?
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
Yes, that's a great question. Whether you're -- whether -- you start by building a large language model, and you use that large language model, very large version, and you could distill them into medium, small and tiny size. And the tiny sized ones, you could put in your phone and your PC and so on and so forth. And they all have good -- they all have -- it seems surprising, but they all can do the same thing. But obviously, the zero shot or the generalizability of the large language model, the biggest one is much more versatile and it can do a lot more amazing things.
是的,這是一個很好的問題。無論你是——是否——你從構建一個大型語言模型開始,然後你使用那個大型語言模型,非常大的版本,你可以將它們提煉成中型、小型和微型。還有小尺寸的,你可以放進你的手機和你的電腦等等。他們都有好處——他們都有——這似乎令人驚訝,但他們都可以做同樣的事情。但很明顯,零鏡頭或者說大語言模型的泛化能力,最大的一個是通用的多,可以做很多更神奇的事情。
And the large one would teach the smaller ones how to be good AIs, and so you use the large one to generate prompts to align the smaller ones and so on and so forth. And so you start by building very large ones, and then you also have to train a whole bunch of smaller ones. That's exactly the reason why we have so many different sizes of our inference. You saw that I announced L4; L40; H100 NVL, which also has H100. And then it has -- and then we have H100 HGX, and then we have H100 multinode with NVLink. And so there's a -- you could have model sizes of any kind that you like.
大的會教小的如何成為好的人工智能,所以你用大的來生成提示來對齊小的,等等。所以你從構建非常大的開始,然後你還必須訓練一大堆較小的。這正是我們有這麼多不同規模的推理的原因。你看到我宣布了L4; L40; H100 NVL,其中還有H100。然後它有——然後我們有 H100 HGX,然後我們有帶 NVLink 的 H100 多節點。所以有一個 - 你可以擁有任何你喜歡的模型尺寸。
The other thing that's important is these are models, but they're connected ultimately to applications. And the applications could have image in, video out, video in, text out, image in, proteins out, text in, 3D out, video in, in the future, 3D graphics out. So the input and the output requires a lot of pre and postprocessing. The pre and postprocessing can't be ignored. And this is one of the things that most of the specialized chip arguments fall apart. And it's because the length -- the model itself is only, call it, 25% of the data -- of the overall processing of inference. The rest of it is about preprocessing, postprocessing, security, decoding, all kinds of things like that.
另一件重要的事情是這些是模型,但它們最終連接到應用程序。應用程序可以有圖像輸入、視頻輸出、視頻輸入、文本輸出、圖像輸入、蛋白質輸出、文本輸入、3D 輸出、視頻輸入,在未來,3D 圖形輸出。所以輸入和輸出需要大量的預處理和後處理。預處理和後處理不容忽視。這是大多數專業芯片爭論失敗的原因之一。這是因為整個推理過程的長度——模型本身只佔數據的 25%。其餘部分是關於預處理、後處理、安全、解碼,以及諸如此類的各種事情。
And so I think the -- we -- the multi-modality aspect of inference, the multidiversity of inference that it's going to be done in the cloud on-prem, it's going to be done in multi-cloud. That's the reason why we have AI Enterprise in all the clouds. It's going to be done on-premise. It's the reason why we have a great partnership with Dell we just announced the other day called Project Helix. That's going to be integrated into third-party services. That's the reason why we have a great partnership with ServiceNow and Adobe because they're going to be creating a whole bunch of generative AI capabilities. And so there's a -- the diversity and the reach of generative AI is so broad, you need to have some very fundamental capabilities like what I just described in order to really address the whole space of it.
所以我認為 - 我們 - 推理的多模態方面,推理的多樣性,它將在本地雲中完成,它將在多雲中完成。這就是我們在所有云中擁有 AI Enterprise 的原因。它將在內部完成。這就是為什麼我們與戴爾建立了良好的合作夥伴關係,我們前幾天剛剛宣布了名為 Project Helix 的原因。這將被集成到第三方服務中。這就是我們與 ServiceNow 和 Adobe 建立良好合作夥伴關係的原因,因為他們將創建一大堆生成 AI 功能。因此,生成式人工智能的多樣性和影響範圍如此之廣,你需要具備一些非常基本的能力,就像我剛才描述的那樣,才能真正解決它的整個領域。
Operator
Operator
Next, we'll go to Harlan Sur with JPMorgan.
接下來,我們將與摩根大通一起前往哈蘭蘇爾。
Harlan Sur - Senior Analyst
Harlan Sur - Senior Analyst
Congratulations on the strong results and execution. I really appreciate more of the focus or some of the focus today on your networking products. I mean it's really an integral part to sort of maximize the full performance of your compute platforms. I think your data center networking business is driving about $1 billion of revenues per quarter, plus or minus. That's 2.5x growth from 3 years ago, right, when you guys acquired Mellanox, so very strong growth.
祝賀您取得了出色的成績和執行力。我真的很感激今天對你們網絡產品的更多關注或部分關注。我的意思是它確實是一個不可或缺的部分,可以最大限度地發揮您的計算平台的全部性能。我認為您的數據中心網絡業務每季度帶來約 10 億美元的收入,上下浮動。這是 3 年前的 2.5 倍增長,當你們收購 Mellanox 時,增長非常強勁。
But given the very high attach of your InfiniBand Ethernet solutions, your accelerated compute platforms, is the networking run rate stepping up in line with your compute shipments? And then what is the team doing to further unlock more networking bandwidth going forward just to keep pace with the significant increase in compute complexity, data sets, requirements for lower latency, better traffic predictability and so on?
但是,考慮到您的 InfiniBand 以太網解決方案、您的加速計算平台的連接率非常高,網絡運行速度是否與您的計算出貨量同步增長?那麼,為了跟上計算複雜性、數據集、更低延遲的要求、更好的流量可預測性等方面顯著增加的步伐,團隊正在做什麼來進一步釋放更多的網絡帶寬?
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
Yes. Harlan, I really appreciate that. Nearly everybody who thinks about AI, they think about that chip, that accelerator chip. And in fact, this is the whole point nearly completely. And I've mentioned before that accelerated computing is about the stack, about the software. And networking, remember, we announced very, very early on this networking stack called DOCA, and we have an acceleration library called Magnum IO. These 2 pieces of software are some of the crown jewels of our company. Nobody ever talks about it because it's hard to understand, but it makes it possible for us to connect tens of thousands of GPUs.
是的。哈倫,我真的很感激。幾乎每個想到 AI 的人都會想到那個芯片,那個加速器芯片。事實上,這幾乎完全是重點。我之前提到過,加速計算與堆棧有關,與軟件有關。還有網絡,請記住,我們很早就宣布了這個名為 DOCA 的網絡堆棧,並且我們有一個名為 Magnum IO 的加速庫。這兩個軟件是我們公司的一些皇冠上的明珠。沒有人談論它,因為它很難理解,但它使我們有可能連接數以萬計的 GPU。
How do you connect tens of thousands of GPUs if the operating system of the data center, which is the infrastructure, is not insanely great? And so that's the reason why we're so obsessed about networking in the company. And one of the great things that we have, we have Mellanox, as you know quite well, was the world's highest performance and the unambiguous leader in high-performance networking, is the reason why our 2 companies are together.
如果作為基礎設施的數據中心的操作系統不是非常出色,您如何連接數万個 GPU?這就是為什麼我們如此痴迷於公司網絡的原因。我們擁有的一件偉大的事情,我們擁有 Mellanox,正如您所知,它是世界上性能最高的,並且是高性能網絡領域的明確領導者,這就是我們兩家公司走到一起的原因。
You also see that our network expands starting from NVLink, which is a computing fabric with really super low latency, and it communicates using memory references, not network packaged. And then we take NVLink. We connect it inside multiple GPUs, and I've described going beyond the GPU. And I'll talk a lot more about that at Computex in a few days. And then that gets connected to InfiniBand, which includes the NIC and the SmartNIC, BlueField-3 that we're in full production with and the switches. All of the fiber optics that are optimized end to end, these things are running at incredible line rates.
您還看到我們的網絡從 NVLink 開始擴展,它是一種具有真正超低延遲的計算結構,它使用內存引用而不是網絡打包進行通信。然後我們採用 NVLink。我們將它連接到多個 GPU 中,我已經描述了超越 GPU 的情況。幾天后我將在 Computex 上詳細討論這個問題。然後連接到 InfiniBand,其中包括 NIC 和 SmartNIC,我們正在全面生產的 BlueField-3 和交換機。所有端到端優化的光纖,這些東西都以令人難以置信的線路速率運行。
And then beyond that, if you want to connect the smart AI factory -- this AI factory into your computing fabric, we have a brand-new type of Ethernet that we'll be announcing at Computex. And so the -- this whole area of the computing fabric extending -- connecting all of these GPUs and computing units together all the way through the networking, through the switches, the software stack is insanely complicated. And so we're -- I'm delighted you understand it, and -- but we don't break it out particularly because we think of the whole thing as a computing platform as it should be.
除此之外,如果你想將智能人工智能工廠——這個人工智能工廠連接到你的計算結構中,我們有一種全新的以太網類型,我們將在 Computex 上宣布。因此——計算結構的整個區域擴展——通過網絡、交換機將所有這些 GPU 和計算單元連接在一起,軟件堆棧非常複雜。所以我們 - 我很高興你能理解它,而且 - 但我們不會特別打破它,因為我們認為整個事情應該是一個計算平台。
We sell it to all of the world's data centers as components so that they can integrate it into whatever style or architecture that they would like and we can still run our software stack. That's the reason why we break it up. It's way more complicated the way that we do it, but it makes it possible for NVIDIA's computing architecture to be integrated into anybody's data center in the world from cloud of all different kinds to on-prem of all different kinds, all the way out to the edge to 5G. And so this way of doing it is really, really complicated, but it gives us incredible reach.
我們將它作為組件出售給世界上所有的數據中心,這樣他們就可以將它集成到他們想要的任何風格或架構中,而我們仍然可以運行我們的軟件堆棧。這就是我們分手的原因。我們做這件事的方式要復雜得多,但它使 NVIDIA 的計算架構可以集成到世界上任何人的數據中心,從各種不同類型的雲到各種類型的內部部署,一直到邊緣到 5G。所以這種做事的方式真的非常複雜,但它給了我們不可思議的影響力。
Operator
Operator
And our last question will come from Matt Ramsay with TD Cowen.
我們的最後一個問題將來自 Matt Ramsay 和 TD Cowen。
Matthew D. Ramsay - MD & Senior Research Analyst
Matthew D. Ramsay - MD & Senior Research Analyst
Congratulations, Jensen, and to the whole team. One of the things I wanted to dig into a little bit is the DGX Cloud offering. You guys have been working on this for some time behind the scenes, where you sell in the hardware to your hyperscale partners and then lease it back for your own business. And the rest of us kind of found out about it publicly a few months ago.
祝賀 Jensen,並祝賀整個團隊。我想深入研究的其中一件事是 DGX Cloud 產品。你們已經在幕後為此工作了一段時間,您將硬件出售給您的超大規模合作夥伴,然後將其租回用於您自己的業務。我們其他人幾個月前就公開發現了這件事。
And as we look forward over the next number of quarters, as Colette discussed, to high visibility in the Data Center business, maybe you could talk a little bit about the mix you're seeing of hyperscale customers buying for their own first-party internal workloads versus their own sort of third party, their own customers versus what of that big upside in Data Center going forward is systems that you're selling in with potential to support your DGX Cloud offerings and what you've learned since you've launched it about the potential of that business.
正如 Colette 所討論的,我們期待在接下來的幾個季度中數據中心業務的高知名度,也許你可以談談你看到的超大規模客戶為自己的第一方內部購買的組合工作負載與他們自己的第三方、他們自己的客戶與數據中心未來的巨大優勢是您銷售的系統有可能支持您的 DGX 雲產品以及自推出以來您學到的東西關於該業務的潛力。
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
Yes. Thanks, Matt. It's -- without being too specific about numbers, but the ideal scenario, the ideal mix is something like 10% NVIDIA DGX Cloud and 90% the CSPs' clouds. And the reason -- and our DGX Cloud is the NVIDIA stack. It's the pure NVIDIA stack. It is architected the way we like and achieves the best possible performance. It gives us the ability to partner very deeply with the CSPs to create the highest-performing infrastructure, number one.
是的。謝謝,馬特。它是 - 沒有太具體的數字,但理想的場景,理想的組合是 10% 的 NVIDIA DGX 雲和 90% 的 CSP 雲。原因——我們的 DGX 雲是 NVIDIA 堆棧。這是純 NVIDIA 堆棧。它按照我們喜歡的方式構建,並實現最佳性能。它使我們能夠與 CSP 深入合作,創建性能最高的基礎設施,排名第一。
Number two, it allows us to partner with the CSPs to create markets. Like for example, we're partnering with Azure to bring Omniverse Cloud to the world's industries. And the world has never had a system like that, the computing stack with all the generative AI stuff and all the 3D stuff and the physics stuff, incredibly large database and really high-speed networks and low-latency networks, that kind of a virtual -- industrial virtual world has never existed before. And so we partnered with Microsoft to create Omniverse Cloud inside Azure Cloud.
第二,它使我們能夠與 CSP 合作創造市場。例如,我們正在與 Azure 合作,將 Omniverse Cloud 引入全球各行各業。世界上從來沒有過這樣的系統,計算堆棧包含所有生成的 AI 東西、所有 3D 東西和物理東西、難以置信的大數據庫和真正的高速網絡和低延遲網絡,那種虛擬——工業虛擬世界前所未有。因此,我們與 Microsoft 合作,在 Azure Cloud 中創建 Omniverse Cloud。
And so it allows us, number two, to create new applications together and develop new markets together. And we go to market as one team. And we benefit by getting customers on our computing platform, and they benefit by having us in their cloud, number one. But number two, the amount of data and services and security services and all of the amazing things that Azure and GCP and OCI have, they can instantly have access to that through Omniverse Cloud.
因此,它允許我們,第二,一起創建新的應用程序並一起開發新的市場。我們作為一個團隊進入市場。我們通過讓客戶使用我們的計算平台而受益,他們通過將我們置於他們的雲中而受益,排名第一。但第二,數據和服務以及安全服務的數量以及 Azure、GCP 和 OCI 擁有的所有令人驚奇的東西,他們可以通過 Omniverse Cloud 立即訪問這些東西。
And so it's a huge win-win. And for the customers, the way that NVIDIA's cloud works for these early applications, they could do it anywhere. So one standard stack runs in all the clouds. And if they would like to take their software and run it on the CSP's cloud themselves and manage it themselves, we're delighted by that because NVIDIA AI Enterprise, NVIDIA AI Foundations and long term, this is going to take a little longer, but NVIDIA Omniverse will run in the CSP's clouds. Okay? So our goal really is to drive architecture, to partner deeply in creating new markets and the new applications that we're doing and provide our customers with the flexibilities to run NVIDIA everywhere, including on-prem.
所以這是一個巨大的雙贏。對於客戶來說,NVIDIA 的雲為這些早期應用程序工作的方式,他們可以在任何地方進行。因此,一個標準堆棧在所有云中運行。如果他們願意使用他們的軟件並自己在 CSP 的雲上運行並自己管理它,我們對此感到很高興,因為 NVIDIA AI Enterprise、NVIDIA AI Foundations 和長期來看,這將需要更長的時間,但是NVIDIA Omniverse 將在 CSP 的雲中運行。好的?因此,我們的目標實際上是推動架構,深入合作以創造新市場和我們正在做的新應用程序,並為我們的客戶提供在任何地方(包括本地)運行 NVIDIA 的靈活性。
And so that -- those were the primary reasons for it. And it's worked out incredibly. Our partnership with the 3 CSPs and that we currently have DGX Cloud in and their sales force and marketing teams, their leadership teams is really quite spectacular. It works great.
因此,這些是它的主要原因。結果令人難以置信。我們與 3 個 CSP 的合作夥伴關係以及我們目前擁有 DGX Cloud 以及他們的銷售團隊和營銷團隊,他們的領導團隊確實非常出色。它很好用。
Operator
Operator
I'll now turn it back over to Jensen Huang for closing remarks.
我現在將它轉回給 Jensen Huang 作結束語。
Jensen Huang;Co-Founder, CEO, President & Director
Jensen Huang;Co-Founder, CEO, President & Director
The computer industry is going through 2 simultaneous transitions, accelerated computing and generative AI. CPU scaling has slowed, yet computing demand is strong and now with generative AI, supercharged. Accelerated computing, a full stack and data center scale approach that NVIDIA pioneered is the best path forward. There's $1 trillion installed in the global data center infrastructure based on the general purpose computing method of the last era. Companies are now racing to deploy accelerated computing for the generative AI era.
計算機行業正在經歷兩個同時發生的轉變,即加速計算和生成 AI。 CPU 擴展速度放緩,但計算需求強勁,現在隨著生成 AI 的出現,性能得到提升。加速計算、NVIDIA 開創的全堆棧和數據中心規模方法是最好的前進道路。基於上個時代的通用計算方法,全球數據中心基礎設施中安裝了 1 萬億美元。公司現在競相為生成人工智能時代部署加速計算。
Over the next decade, most of the world's data centers will be accelerated. We are significantly increasing our supply to meet their surging demand. Large language models can learn information encoded in many forms. Guided by large language models, generative AI models can generate amazing content with models to fine-tune, guardrail, align to guiding principles and to ground facts, generative AI is emerging from labs and is on its way to industrial applications.
未來十年,全球大部分數據中心都將加速發展。我們正在顯著增加供應以滿足他們不斷增長的需求。大型語言模型可以學習以多種形式編碼的信息。在大型語言模型的指導下,生成式 AI 模型可以生成令人驚嘆的內容,並通過模型進行微調、保護、與指導原則和基本事實保持一致,生成式 AI 正在從實驗室中出現,並正在走向工業應用。
As we scale with cloud and Internet service providers, we are also building platforms for the world's largest enterprises. Whether within one of our CSP partners or on-prem with Dell Helix, whether on a leading enterprise platform like ServiceNow and Adobe or bespoke with NVIDIA AI Foundations, we can help enterprises leverage their domain expertise and data to harness generative AI securely and safely.
隨著我們與雲和互聯網服務提供商一起擴展,我們也在為世界上最大的企業構建平台。無論是在我們的 CSP 合作夥伴之一內部還是與 Dell Helix 的本地部署,無論是在 ServiceNow 和 Adobe 等領先的企業平台上還是與 NVIDIA AI Foundations 定制,我們都可以幫助企業利用他們的領域專業知識和數據來安全可靠地利用生成 AI。
We are ramping a wave of products in the coming quarters, including H100, our Grace and Grace Hopper Superchips and our BlueField-3 and Spectrum-4 networking platform. They are all in production. They will help deliver data center scale computing that is also energy efficient and sustainable computing. Join us next week at Computex, and we'll show you what's next. Thank you.
我們將在未來幾個季度推出一系列產品,包括 H100、我們的 Grace 和 Grace Hopper 超級芯片以及我們的 BlueField-3 和 Spectrum-4 網絡平台。它們都在生產中。它們將幫助提供數據中心規模的計算,同時也是節能和可持續的計算。下週加入我們的 Computex,我們將向您展示接下來的內容。謝謝。
Operator
Operator
This concludes today's conference call. You may now disconnect.
今天的電話會議到此結束。您現在可以斷開連接。