🏋️
Hugging Face Model Trainer
availableTrain and fine-tune language models using TRL on Hugging Face Jobs — SFT, DPO, GRPO, reward modeling, and GGUF conversion.
ai-mlWhat This Skill Does
Train and fine-tune language models using TRL on Hugging Face Jobs — SFT, DPO, GRPO, reward modeling, and GGUF conversion.
Capabilities
- Fine-Tuning Methods — SFT, DPO, GRPO, and reward modeling
- Cloud GPU Training — Train on HF Jobs with proper hardware selection
- GGUF Conversion — Convert models for local deployment
- Monitoring — Track training progress with Trackio
How It Works
- Select a base model and training method
- Prepare and validate your training dataset
- Submit the training job to HF Jobs with selected hardware
- Monitor progress and download the fine-tuned model
Example Usage
"Fine-tune Llama with SFT on my instruction dataset"
"Run DPO training for preference alignment"
"Convert my model to GGUF for local inference"