Skip to content
#

rft

Here are 12 public repositories matching this topic...

The course teaches how to fine-tune LLMs using Group Relative Policy Optimization (GRPO)—a reinforcement learning method that improves model reasoning with minimal data. Learn RFT concepts, reward design, LLM-as-a-judge evaluation, and deploy jobs on the Predibase platform.

  • Updated Jun 13, 2025
  • Jupyter Notebook

Improve this page

Add a description, image, and links to the rft topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the rft topic, visit your repo's landing page and select "manage topics."

Learn more