DGX Cloud – NVIDIA Technical Blog News and tutorials for developers, data scientists, and IT admins 2025-03-27T16:00:00Z http://www.open-lab.net/blog/feed/ Wen Jie Ong <![CDATA[Accelerating the Future of Transportation with SES AI��s NVIDIA-Powered Innovation for Electric Vehicles]]> http://www.open-lab.net/blog/?p=97805 2025-03-25T17:36:45Z 2025-03-25T16:00:00Z Electric vehicles (EVs) are transforming transportation, but challenges such as cost, longevity, and range remain barriers to widespread adoption. At the heart...]]> Electric vehicles (EVs) are transforming transportation, but challenges such as cost, longevity, and range remain barriers to widespread adoption. At the heart...Decorative image of a car at night.

Electric vehicles (EVs) are transforming transportation, but challenges such as cost, longevity, and range remain barriers to widespread adoption. At the heart of these challenges lies battery technology��specifically, the electrolyte, a critical component that enables energy storage and delivery. The electrolyte��s properties directly impact a battery��s charging speed, power output, stability��

Source

]]>
1
Vishal Ganeriwala <![CDATA[Seamlessly Scale AI Across Cloud Environments with NVIDIA DGX Cloud Serverless Inference]]> http://www.open-lab.net/blog/?p=97192 2025-03-20T17:07:54Z 2025-03-18T21:22:51Z NVIDIA DGX Cloud Serverless Inference is an auto-scaling AI inference solution that enables application deployment with speed and reliability. Powered by NVIDIA...]]> NVIDIA DGX Cloud Serverless Inference is an auto-scaling AI inference solution that enables application deployment with speed and reliability. Powered by NVIDIA...

NVIDIA DGX Cloud Serverless Inference is an auto-scaling AI inference solution that enables application deployment with speed and reliability. Powered by NVIDIA Cloud Functions (NVCF), DGX Cloud Serverless Inference abstracts multi-cluster infrastructure setups across multi-cloud and on-premises environments for GPU-accelerated workloads. Whether managing AI workloads��

Source

]]>
0
Emily Potyraj <![CDATA[Measure and Improve AI Workload Performance with NVIDIA DGX Cloud Benchmarking]]> http://www.open-lab.net/blog/?p=97548 2025-03-20T17:07:42Z 2025-03-18T21:21:17Z As AI capabilities advance, understanding the impact of hardware and software infrastructure choices on workload performance is crucial for both technical...]]> As AI capabilities advance, understanding the impact of hardware and software infrastructure choices on workload performance is crucial for both technical...

As AI capabilities advance, understanding the impact of hardware and software infrastructure choices on workload performance is crucial for both technical validation and business planning. Organizations need a better way to assess real-world, end-to-end AI workload performance and the total cost of ownership rather than just comparing raw FLOPs or hourly cost per GPU.

Source

]]>
0
Hao Wang <![CDATA[Petabyte-Scale Video Processing with NVIDIA NeMo Curator on NVIDIA DGX Cloud]]> http://www.open-lab.net/blog/?p=97031 2025-03-20T17:07:03Z 2025-03-18T19:22:51Z With the rise of physical AI, video content generation has surged exponentially. A single camera-equipped autonomous vehicle can generate more than 1 TB of...]]> With the rise of physical AI, video content generation has surged exponentially. A single camera-equipped autonomous vehicle can generate more than 1 TB of...NeMo Video Curator icon in a workflow diagram.

With the rise of physical AI, video content generation has surged exponentially. A single camera-equipped autonomous vehicle can generate more than 1 TB of video daily, while a robotics-powered manufacturing facility may produce 1 PB of data daily. To leverage this data for training and fine-tuning world foundation models (WFMs), you must first process it efficiently.

Source

]]>
3
Emily Potyraj <![CDATA[NVIDIA DGX Cloud Introduces Ready-To-Use Templates to Benchmark AI Platform Performance]]> http://www.open-lab.net/blog/?p=95558 2025-02-20T15:54:23Z 2025-02-11T17:00:00Z In the rapidly evolving landscape of AI systems and workloads, achieving optimal model training performance extends far beyond chip speed. It requires a...]]> In the rapidly evolving landscape of AI systems and workloads, achieving optimal model training performance extends far beyond chip speed. It requires a...Three icons in a row, including DGX in the middle.

In the rapidly evolving landscape of AI systems and workloads, achieving optimal model training performance extends far beyond chip speed. It requires a comprehensive evaluation of the entire stack, from compute to networking to model framework. Navigating the complexities of AI system performance can be difficult. There are many application changes that you can make��

Source

]]>
0
Martin Cimmino <![CDATA[Continued Pretraining of State-of-the-Art LLMs for Sovereign AI and Regulated Industries with iGenius and NVIDIA DGX Cloud]]> http://www.open-lab.net/blog/?p=95012 2025-01-23T19:54:22Z 2025-01-16T12:00:00Z In recent years, large language models (LLMs) have achieved extraordinary progress in areas such as reasoning, code generation, machine translation, and...]]> In recent years, large language models (LLMs) have achieved extraordinary progress in areas such as reasoning, code generation, machine translation, and...Stack diagram for LLM Megatron Core.

In recent years, large language models (LLMs) have achieved extraordinary progress in areas such as reasoning, code generation, machine translation, and summarization. However, despite their advanced capabilities, foundation models have limitations when it comes to domain-specific expertise such as finance or healthcare or capturing cultural and language nuances beyond English.

Source

]]>
0
Brad Nemire <![CDATA[NVIDIA Project DIGITS, A Grace Blackwell AI Supercomputer On Your Desk]]> http://www.open-lab.net/blog/?p=94765 2025-01-23T19:54:30Z 2025-01-09T18:19:00Z Powered by the new GB10 Grace Blackwell Superchip, Project DIGITS can tackle large generative AI models of up to 200B parameters.]]> Powered by the new GB10 Grace Blackwell Superchip, Project DIGITS can tackle large generative AI models of up to 200B parameters.

Powered by the new GB10 Grace Blackwell Superchip, Project DIGITS can tackle large generative AI models of up to 200B parameters.

Source

]]>
5
Niels Bantilan <![CDATA[Democratizing AI Workflows with Union.ai and NVIDIA DGX Cloud]]> http://www.open-lab.net/blog/?p=81110 2024-05-08T17:57:05Z 2024-04-24T01:12:42Z GPUs were initially specialized for rendering 3D graphics in video games, primarily to accelerate linear algebra calculations. Today, GPUs have become one of...]]> GPUs were initially specialized for rendering 3D graphics in video games, primarily to accelerate linear algebra calculations. Today, GPUs have become one of...Decorative image of different workflows against a grey background.

GPUs were initially specialized for rendering 3D graphics in video games, primarily to accelerate linear algebra calculations. Today, GPUs have become one of the critical components of the AI revolution. We now rely on these workhorses to fulfill deep learning workloads, crunching through massive and complex semi-structured datasets. However, as demand for AI-based solutions has��

Source

]]>
0
Mehran Maghoumi <![CDATA[Scale and Curate High-Quality Datasets for LLM Training with NVIDIA NeMo Curator]]> http://www.open-lab.net/blog/?p=80168 2025-02-17T05:28:15Z 2024-03-27T18:00:00Z Enterprises are using large language models (LLMs) as powerful tools to improve operational efficiency and drive innovation. NVIDIA NeMo microservices aim to...]]> Enterprises are using large language models (LLMs) as powerful tools to improve operational efficiency and drive innovation. NVIDIA NeMo microservices aim to...

Enterprises are using large language models (LLMs) as powerful tools to improve operational efficiency and drive innovation. NVIDIA NeMo microservices aim to make building and deploying models more accessible to enterprises. An important step for building any LLM system is to curate the dataset of tokens to be used for training or customizing the model. However, curating a suitable dataset��

Source

]]>
0
Ike Nnoli <![CDATA[Generative AI for Digital Human Technologies and New AI-powered NVIDIA RTX Lighting]]> http://www.open-lab.net/blog/?p=79707 2024-12-09T16:51:28Z 2024-03-19T17:00:00Z At GDC 2024, NVIDIA announced that leading AI application developers such as Inworld AI are using NVIDIA digital human technologies to accelerate the deployment...]]> At GDC 2024, NVIDIA announced that leading AI application developers such as Inworld AI are using NVIDIA digital human technologies to accelerate the deployment...Still image from Covert Protocol game demo.

At GDC 2024, NVIDIA announced that leading AI application developers such as Inworld AI are using NVIDIA digital human technologies to accelerate the deployment of generative AI-powered game characters alongside updated NVIDIA RTX SDKs that simplify the creation of beautiful worlds. You can incorporate the full suite of NVIDIA digital human technologies or individual microservices into��

Source

]]>
0
Amanda Saunders <![CDATA[NVIDIA NIM Offers Optimized Inference Microservices for Deploying AI Models at Scale]]> http://www.open-lab.net/blog/?p=79467 2024-06-03T15:44:17Z 2024-03-18T22:00:00Z The rise in generative AI adoption has been remarkable. Catalyzed by the launch of OpenAI��s ChatGPT in 2022, the new technology amassed over 100M users within...]]> The rise in generative AI adoption has been remarkable. Catalyzed by the launch of OpenAI��s ChatGPT in 2022, the new technology amassed over 100M users within...An illustration representing NVIDIA NIM.

The rise in generative AI adoption has been remarkable. Catalyzed by the launch of OpenAI��s ChatGPT in 2022, the new technology amassed over 100M users within months and drove a surge of development activities across almost every industry. By 2023, developers began POCs using APIs and open-source community models from Meta, Mistral, Stability, and more. Entering 2024��

Source

]]>
0
Alan Nafiiev <![CDATA[Accelerating Drug Discovery at Receptor.AI with NVIDIA BioNeMo Cloud APIs]]> http://www.open-lab.net/blog/?p=77569 2024-05-08T17:57:29Z 2024-02-14T21:00:00Z The quest for new, effective treatments for diseases that remain stubbornly resistant to current therapies is at the heart of drug discovery. This traditionally...]]> The quest for new, effective treatments for diseases that remain stubbornly resistant to current therapies is at the heart of drug discovery. This traditionally...

The quest for new, effective treatments for diseases that remain stubbornly resistant to current therapies is at the heart of drug discovery. This traditionally long and expensive process has been radically improved by AI techniques like deep learning, empowered by the rise of accelerated computing. Receptor.AI, a London-based drug discovery company and NVIDIA Inception member��

Source

]]>
0
Tanya Lenz <![CDATA[Webinar: Accelerate AV Development with NVIDIA DGX Cloud and NVIDIA AI Enterprise]]> http://www.open-lab.net/blog/?p=72286 2024-05-08T17:57:52Z 2023-10-30T20:00:00Z Learn how to leverage NVIDIA AI-powered infrastructure and software to accelerate AV development for maximum efficiency.]]> Learn how to leverage NVIDIA AI-powered infrastructure and software to accelerate AV development for maximum efficiency.

Learn how to leverage NVIDIA AI-powered infrastructure and software to accelerate AV development for maximum efficiency.

Source

]]>
0
Joe Handzik <![CDATA[High-Performance Storage on NVIDIA DGX Cloud with Oracle Cloud Infrastructure]]> http://www.open-lab.net/blog/?p=63551 2024-05-08T17:58:47Z 2023-04-18T18:43:47Z The incredible advances of accelerated computing are powered by data. The role of data in accelerating AI workloads is crucial for businesses looking to stay...]]> The incredible advances of accelerated computing are powered by data. The role of data in accelerating AI workloads is crucial for businesses looking to stay...Data center

The incredible advances of accelerated computing are powered by data. The role of data in accelerating AI workloads is crucial for businesses looking to stay ahead of the curve in the current fast-paced digital environment. Speeding up access to that data is yet another way that NVIDIA accelerates entire AI workflows. NVIDIA DGX Cloud caters to a wide variety of market use cases.

Source

]]>
0
���˳���97caoporen����