Sergio Perez & Harshita Seth
Adding knowledge to open-source LLMs
#1about 4 minutes
Understanding the LLM training pipeline and knowledge gaps
LLMs are trained through pre-training and alignment, but require new knowledge to stay current, adapt to specific domains, and acquire new skills.
#2about 5 minutes
Adding domain knowledge with continued pre-training
Continued pre-training adapts a foundation model to a specific domain by training it further on specialized, unlabeled data using self-supervised learning.
#3about 6 minutes
Developing skills and reasoning with supervised fine-tuning
Supervised fine-tuning uses instruction-based datasets to teach models specific tasks, chat capabilities, and complex reasoning through techniques like chain of thought.
#4about 8 minutes
Aligning models with human preferences using reinforcement learning
Preference alignment refines model behavior using reinforcement learning, evolving from complex RLHF with reward models to simpler methods like DPO.
#5about 2 minutes
Using frameworks like NeMo RL to simplify model alignment
Frameworks like the open-source NeMo RL abstract away the complexity of implementing advanced alignment algorithms like reinforcement learning.
Related jobs
Jobs that call for the skills explored in this talk.
Matching moments
02:26 MIN
Understanding the core capabilities of large language models
Data Privacy in LLMs: Challenges and Best Practices
01:12 MIN
Introducing InstructLab for accessible LLM fine-tuning
Unlocking the Power of AI: Accessible Language Model Tuning for All
08:05 MIN
How large language models are trained
Inside the Mind of an LLM
05:18 MIN
Addressing the core challenges of large language models
Accelerating GenAI Development: Harnessing Astra DB Vector Store and Langflow for LLM-Powered Apps
02:55 MIN
Addressing the key challenges of large language models
Large Language Models ❤️ Knowledge Graphs
03:42 MIN
Using large language models as a learning tool
Google Gemini: Open Source and Deep Thinking Models - Sam Witteveen
02:21 MIN
The training process of large language models
Google Gemini: Open Source and Deep Thinking Models - Sam Witteveen
03:17 MIN
Using RAG to enrich LLMs with proprietary data
RAG like a hero with Docling
Featured Partners
Related Videos
Inside the Mind of an LLM
Emanuele Fabbiani
Unlocking the Power of AI: Accessible Language Model Tuning for All
Cedric Clyburn & Legare Kerrison
LLMOps-driven fine-tuning, evaluation, and inference with NVIDIA NIM & NeMo Microservices
Anshul Jindal
Self-Hosted LLMs: From Zero to Inference
Roberto Carratalá & Cedric Clyburn
Exploring LLMs across clouds
Tomislav Tipurić
Give Your LLMs a Left Brain
Stephen Chin
Three years of putting LLMs into Software - Lessons learned
Simon A.T. Jiménez
Large Language Models ❤️ Knowledge Graphs
Michael Hunger
Related Articles
View all articles.png?w=240&auto=compress,format)
.gif?w=240&auto=compress,format)
.png?w=240&auto=compress,format)

From learning to earning
Jobs that call for the skills explored in this talk.





Understanding Recruitment Group
Barcelona, Spain
Remote
Node.js
Computer Vision
Machine Learning


NLP People
Municipality of Valencia, Spain
Intermediate
GIT
Linux
NoSQL
NumPy
Keras
+11

NLP People
Municipality of Valencia, Spain
Intermediate
NumPy
Keras
Pandas
PyTorch
Tensorflow
+3

CGI Group Inc.
Köln, Germany
Senior
Data analysis
Natural Language Processing