NeMo Curator

Oct 15, 2024
Train Highly Accurate LLMs with the Zyda-2 Open 5T-Token Dataset Processed with NVIDIA NeMo Curator
Open-source datasets have significantly democratized access to high-quality data, lowering the barriers of entry for developers and researchers to train...
5 MIN READ

Oct 15, 2024
DataStax Announces New AI Development Platform, Built with NVIDIA AI
As enterprises increasingly adopt AI technologies, they face a complex challenge of efficiently developing, securing, and continuously improving AI applications...
6 MIN READ

Oct 04, 2024
Just Released: NVIDIA NeMo Curator Improvements for Accelerating Data Curation
NeMo Curator now supports images, enabling you to process data for training accurate generative AI models.
1 MIN READ

Sep 10, 2024
Streamlining Data Processing for Domain Adaptive Pretraining with NVIDIA NeMo Curator
Domain-adaptive pretraining (DAPT) of large language models (LLMs) is an important step towards building domain-specific models. These models demonstrate...
16 MIN READ

Jul 31, 2024
Curating Custom Datasets for LLM Parameter-Efficient Fine-Tuning with NVIDIA NeMo Curator
In a recent post, we discussed how to use NVIDIA NeMo Curator to curate custom datasets for pretraining or continuous training use cases of large language...
11 MIN READ

Jul 23, 2024
Supercharging Llama 3.1 across NVIDIA Platforms
Meta's Llama collection of large language models are the most popular foundation models in the open-source community today, supporting a variety of use cases....
8 MIN READ

Jul 23, 2024
Customize Generative AI Models for Enterprise Applications with Llama 3.1
The newly unveiled Llama 3.1 collection of 8B, 70B, and 405B large language models (LLMs) is narrowing the gap between proprietary and open-source models. Their...
10 MIN READ

Jul 10, 2024
Curating Non-English Datasets for LLM Training with NVIDIA NeMo Curator
Data curation plays a crucial role in the development of effective and fair large language models (LLMs). High-quality, diverse training data directly...
12 MIN READ

May 29, 2024
Generative AI Agents Developer Contest: Top Tips for Getting Started
Join our contest that runs through June 17 and showcase your innovation using cutting-edge generative AI-powered applications using NVIDIA and LangChain...
3 MIN READ

May 21, 2024
Curating Custom Datasets for LLM Training with NVIDIA NeMo Curator
Data curation is the first, and arguably the most important, step in the pretraining and continuous training of large language models (LLMs) and small language...
14 MIN READ

May 17, 2024
Training Localized Multilingual LLMs with NVIDIA NeMo, Part 2
In Part 1, we discussed how to train a monolingual tokenizer and merge it with a pretrained LLM’s tokenizer to form a multilingual tokenizer. In this post, we...
8 MIN READ

May 17, 2024
Training Localized Multilingual LLMs with NVIDIA NeMo, Part 1
In today's globalized world, the ability of AI systems to understand and communicate in diverse languages is increasingly crucial. Large language models (LLMs)...
14 MIN READ

May 13, 2024
Customizing Neural Machine Translation Models with NVIDIA NeMo, Part 2
In the first post, we walked through the prerequisites for a neural machine translation example from English to Chinese, running the pretrained model with NeMo,...
11 MIN READ

Mar 27, 2024
Develop Custom Enterprise Generative AI with NVIDIA NeMo
Generative AI is transforming computing, paving new avenues for humans to interact with computers in natural, intuitive ways. For enterprises, the prospect of...
14 MIN READ

Mar 27, 2024
Scale and Curate High-Quality Datasets for LLM Training with NVIDIA NeMo Curator
Enterprises are using large language models (LLMs) as powerful tools to improve operational efficiency and drive innovation. NVIDIA NeMo microservices aim to...
6 MIN READ

Mar 27, 2024
Fine-Tune and Align LLMs Easily with NVIDIA NeMo Customizer
As large language models (LLMs) continue to gain traction in enterprise AI applications, the demand for custom models that can understand and integrate specific...
5 MIN READ