You can now do FP8 reinforcement learning locally! (<5GB VRAM)

Posted by danielhanchen@reddit | LocalLLaMA | View on Reddit | 61 comments

You can now do FP8 reinforcement learning locally! (<5GB VRAM)

Hey r/LocalLlama! We're getting close to our last release of 2025! Thanks so much for all the support this year. The DeepSeek team back in Jan showcased how powerful FP8 RL can be with GRPO. Well, you can now try it on your local hardware using only 5GB VRAM! RTX 50x, 40x series all work!

Why should you do FP8 training?
NVIDIA's research finds FP8 training can match BF16 accuracy whilst getting 1.6x faster inference time. We collabed with TorchAO from PyTorch to introduce FP8 RL training, making FP8 GRPO possible on home GPUs with no accuracy loss!

You can read our blogpost for our findings and more: https://docs.unsloth.ai/new/fp8-reinforcement-learning

Llama 3.2 1B FP8 Colab Notebook: https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Llama_FP8_GRPO.ipynb

In the notebook, you can plug in any of our previous reward functions or RL environment examples, including our auto kernel creation and our 2048 game notebooks. To enable fp8:

import os; os.environ['UNSLOTH_VLLM_STANDBY'] = "1" # Saves 30% VRAM
from unsloth import FastLanguageModel
model, tokenizer = FastLanguageModel.from_pretrained(
    model_name = "unsloth/Qwen3-8B",
    max_seq_length = 2048,
    load_in_4bit = False, # False for LoRA 16bit
    fast_inference = True, # Enable vLLM fast inference
    max_lora_rank = 32,
    load_in_fp8 = True, # Float8 RL / GRPO!
)

Hope you all have a lovely Thanksgiving, a lovely rest of the week and I'll be here to answer any and all questions! =)