A3AIARC102: AI Fine Tuning and Data Preparation for Pre-Trained Models

Become an EPIC Affiliate

To view the class schedule you need to become an Affiliate

  • Largest “Guaranteed To Run” public technical training schedules available
  • Easy to become an Affiliate – no charge or fee
Become an EPIC Affiliate

already an Affiliate?  Login

About this Course

You will develop the skills to gather, clean, and organize data for fine-tuning pre-trained LLMs and Generative AI models. Through a combination of lectures and hands-on labs, you will use Python to fine-tune open-source Transformer models. Gain practical experience with LLM frameworks, learn essential training techniques, and explore advanced topics such as quantization. During the hands-on labs, you will access a GPU-accelerated server for practical experience with industry-standard tools and frameworks.

Audience Profile

• Project Managers
• Architects
• Developers
• Data Acquisition Specialists

At Course Completion

• Clean and Curate Data for AI Fine-Tuning
• Establish guidelines for obtaining RAW Data
• Go from Drowning in Data to Clean Data
• Fine-Tune AI Models with PyTorch
• Understand AI architecture: Transformer model
• Describe tokenization and word embeddings
• Install and use AI frameworks like Llama-3
• Perform LoRA and QLoRA Fine-Tuning
• Explore model quantization and fine-tuning
• Deploy and Maximize AI Model Performance

Outline

Data Curation for AI
💬 Lecture: Curating Data for AI
💻 Lecture + Lab: Gathering Raw Data
💻 Lecture + Lab: Data Cleaning and Preparation
💻 Lecture + Lab: Data Labeling
💻 Lecture + Lab: Data Organization
💬 Lecture: Premade Datasets for Fine Tuning
💻 Lecture + Lab: Obtain and Prepare Premade Datasets
Deep Learning
💬 Lecture: What is Intelligence?
💬 Lecture: Generative AI
💬 Lecture: The Transformer Model
💬 Lecture: Feed Forward Neural Networks
💻 Lecture + Lab: Tokenization
💻 Lecture + Lab: Word Embeddings
💻 Lecture + Lab: Positional Encoding
Pre-trained LLM
💬 Lecture: A History of Neural Network Architectures
💬 Lecture: Introduction to the LLaMa.cpp Interface
💬 Lecture: Preparing A100 for Server Operations
💻 Lecture + Lab: Operate LLaMa3 Models with LLaMa.cpp
💻 Lecture + Lab: Selecting Quantization Level to Meet Performance and Perplexity Requirements
Fine Tuning
💬 Lecture: Fine-Tuning a Pre-Trained LLM
💬 Lecture: PyTorch
💻 Lecture + Lab: Basic Fine Tuning with PyTorch
💻 Lecture + Lab: LoRA Fine-Tuning LLaMa3 8B
💻 Lecture + Lab: QLoRA Fine-Tuning LLaMa3 8B
Operating Fine-Tuned Model
💬 Lecture: Running the llama.cpp Package
💻 Lecture + Lab: Deploy Llama API Server
💻 Lecture + Lab: Develop LLaMa Client Application
💻 Lecture + Lab: Write a Real-World AI Application using the Llama API

Prerequisites

• Python or Equivalent Experience
• Familiarity with Linux