Toutes les actualités, formations et événements
Adding New Knowledge to LLMs
11 / 03 / 2025,
Actualité
About this Course
Large Language Models (LLMs) are powerful, but their knowledge is often general-purpose and may lack the specific, up-to-date, or specialized information required for enterprise applications. The "Adding Knowledge to LLMs" workshop provides a comprehensive, hands-on guide to the essential techniques for augmenting and customizing LLMs.
This workshop takes you on a complete journey from raw data to a fine-tuned, optimized model. You will begin by learning how to curate high-quality datasets and generate synthetic data with NVIDIA NeMo Curator. Next, you will dive deep into the crucial process of model evaluation, using benchmarks, LLM-as-a-judge, and the NeMo Evaluator to rigorously assess model performance. With a solid foundation in evaluation, you will then explore a suite of powerful customization techniques, including Continued Pretraining to inject new knowledge, Supervised Fine-Tuning to teach new skills, and Direct Preference Optimization (DPO) to align model behavior with human preferences.
Finally, you will learn to make your customized models efficient for real-world deployment by exploring essential optimization techniques like quantization, pruning, and knowledge distillation using TensorRT-LLM and the NeMo framework. The workshop culminates in a hands-on assessment where you will apply your new skills to align an LLM to a specific conversational style, solidifying your ability to tailor models for any application.
Learning Objectives
By participating in this workshop, participants will be equipped to:
- Curate high-quality datasets and generate synthetic data using NVIDIA NeMo Curator.
- Rigorously evaluate LLM performance with benchmarks (MMLU), LLM-as-a-judge, and the NeMo Evaluator.
- Inject new domain-specific knowledge into LLMs using Continued Pretraining (CPT).
- Teach LLMs new skills and align them to specific tasks with Supervised Fine-Tuning (SFT).
- Align model behavior to human preferences for style, tone, and safety using Direct Preference Optimization (DPO).
- Compress and optimize LLMs for efficient deployment using Quantization, Pruning, and Knowledge Distillation with TensorRT-LLM and NeMo.
- Apply end-to-end model customization workflows to solve real-world problems
Course Details
Duration: 08:00
Price:
Level: Technical - Intermediate
Subject: Generative AI/LLM
Language: English
Course Prerequisites:
- Familiarity with Python programming and Jupyter notebooks.
- Basic understanding of Large Language Models and their applications.
- Conceptual knowledge of deep learning and neural networks.
Tools, libraries, frameworks used: Python, NVIDIA NeMo, NVIDIA TensorRT-LLM, Docker, MLflow
Topics Covered
In service of teaching and demonstrating how to add knowledge to and customize LLMs for enterprise use, this workshop will cover the following topics and technologies:
- Data Curation and Synthetic Data Generation
- Advanced LLM Evaluation Techniques (including LLM-as-a-Judge and ELO)
- Continued Pretraining (CPT) for Knowledge Injection
- Supervised Fine-Tuning (SFT) for Skill Acquisition
- Direct Preference Optimization (DPO) for Behavioral Alignment
- Model Optimization: Quantization, Pruning, and Knowledge Distillation
- NVIDIA NeMo Framework, NeMo Curator, NeMo Evaluator, and NeMo-RL
- TensorRT-LLM for High-Performance Inference
Course Outline
The table below is a suggested timeline for the course. Please coordinate with the instructor for the best pacing and emphasis.
| 1. Data Curation and Synthetic Data Generation |
|
| 2. Evaluating Large Language Models |
|
| 3. Customizing LLMs |
|
| 4. Optimizing LLMs for Deployment |
|
| 5. Interactive Assessment |
|