LLaMA-Factory
Fine-tune 100+ LLMs with LoRA/QLoRA and a web UI on Clore.ai GPUs using LLaMA-Factory
Server Requirements
Parameter
Minimum
Recommended
Quick Deploy on CLORE.AI
Variable
Example
Description
Step-by-Step Setup
1. Rent a GPU Server on CLORE.AI
Task
VRAM
Recommended GPU
2. SSH into Your Server
3. Create Working Directories
4. Pull the Docker Image
5. Launch LLaMA-Factory
6. Access the Web Interface
Usage Examples
Example 1: LoRA Fine-Tuning via Web UI (LLaMA Board)
Example 2: CLI-Based QLoRA Fine-Tuning
Example 3: Upload Custom Dataset
Example 4: DPO (Direct Preference Optimization)
Example 5: Inference with Fine-Tuned Model
Configuration
Key Training Parameters
Parameter
Typical Value
Description
Supported Fine-tuning Methods
Method
Memory Use
Quality
When to Use
Multi-GPU DeepSpeed Training
Performance Tips
1. Optimal QLoRA Settings by GPU
2. Flash Attention 2 for Longer Contexts
3. Gradient Checkpointing
4. Choose the Right LoRA Target
5. Freeze Top Layers for Fast Adaptation
6. Monitor with TensorBoard
Troubleshooting
Problem: "CUDA out of memory" during training
Problem: Training loss not decreasing
Problem: Slow training speed
Problem: Model not found in web UI
Problem: Dataset format errors
Problem: WebUI port not accessible
Links
Clore.ai GPU Recommendations
Use Case
Recommended GPU
Est. Cost on Clore.ai
Last updated
Was this helpful?