William Hill – NVIDIA Technical Blog News and tutorials for developers, data scientists, and IT admins 2025-01-22T18:31:27Z http://www.open-lab.net/blog/feed/ William Hill <![CDATA[NVIDIA Hackathon Winners Share Strategies for RAPIDS-Accelerated ML Workflows]]> http://www.open-lab.net/blog/?p=94393 2025-01-22T18:31:27Z 2024-12-20T18:00:00Z Approximately 220 teams gathered at the Open Data Science Conference (ODSC) West this year to compete in the NVIDIA hackathon, a 24-hour machine learning (ML)...]]>

Approximately 220 teams gathered at the Open Data Science Conference (ODSC) West this year to compete in the NVIDIA hackathon, a 24-hour machine learning (ML) competition. Data scientists and engineers designed models that were evaluated based on accuracy and processing speed. The top three teams walked away with prize packages that included NVIDIA RTX Ada Generation GPUs, Google Colab credits…

Source

]]>
William Hill <![CDATA[Just Released: Torch-TensorRT v2.4.0]]> http://www.open-lab.net/blog/?p=89229 2024-09-19T17:50:49Z 2024-09-19T17:50:46Z Includes C++ runtime support in Windows Support, Enhanced Dynamic Shape support in Converters, PyTorch 2.4, CUDA 12.4, TensorRT 10.1, Python 3.12.]]>

Includes C++ runtime support in Windows Support, Enhanced Dynamic Shape support in Converters, PyTorch 2.4, CUDA 12.4, TensorRT 10.1, Python 3.12.

Source

]]>
William Hill <![CDATA[Achieving High Mixtral 8x7B Performance with NVIDIA H100 Tensor Core GPUs and NVIDIA TensorRT-LLM]]> http://www.open-lab.net/blog/?p=84749 2024-08-07T23:50:14Z 2024-07-02T18:00:00Z As large language models (LLMs) continue to grow in size and complexity, the performance requirements for serving them quickly and cost-effectively continue to...]]>

As large language models (LLMs) continue to grow in size and complexity, the performance requirements for serving them quickly and cost-effectively continue to grow. Delivering high LLM inference performance requires an efficient parallel computing architecture and a flexible and highly optimized software stack. Recently, NVIDIA Hopper GPUs running NVIDIA TensorRT-LLM inference software set…

Source

]]>
William Hill <![CDATA[Supercharge Generative AI Development with Firebase Genkit, Optimized by NVIDIA RTX GPUs]]> http://www.open-lab.net/blog/?p=82771 2024-05-30T19:55:50Z 2024-05-20T16:27:39Z At Google I/O 2024, Google announced Firebase Genkit, a new open-source framework for developers to add generative AI to web and mobile applications using...]]>

At Google I/O 2024, Google announced Firebase Genkit, a new open-source framework for developers to add generative AI to web and mobile applications using models like Google Gemini, Google Gemma. With Firebase Genkit, you can build apps that integrate intelligent agents, automate customer support, use semantic search, and convert unstructured data into insights. Genkit also includes a developer UI…

Source

]]>
William Hill <![CDATA[NVIDIA TensorRT 10.0 Upgrades Usability, Performance, and AI Model Support]]> http://www.open-lab.net/blog/?p=82402 2024-05-30T19:55:57Z 2024-05-14T15:00:00Z NVIDIA today announced the latest release of NVIDIA TensorRT, an ecosystem of APIs for high-performance deep learning inference. TensorRT includes inference...]]>

NVIDIA today announced the latest release of NVIDIA TensorRT, an ecosystem of APIs for high-performance deep learning inference. TensorRT includes inference runtimes and model optimizations that deliver low latency and high throughput for production applications. This post outlines the key features and upgrades of this release, including easier installation, increased usability…

Source

]]>
���˳���97caoporen����