Get started with Unsloth Studio
A guide for getting started with the fine-tuning studio, data recipes, model exporting, and chat.
Unsloth Studio is a local, browser-based GUI for fine-tuning LLMs without writing any code. It wraps the training pipeline in a clean interface that handles model loading, dataset formatting, hyperparameter configuration, and live training monitoring.
StudioData RecipeExportChatVideo
Setup Unsloth Studio
First, launch Unsloth Studio using either a local install or a cloud option. Follow the install instructions for your setup, or use our free Colab notebook. For a local setup, run:
uunsloth studio -H 0.0.0.0 -p 8888Then open http://localhost:8888 in your browser.
On first launch you will need to create a password to secure your account and sign in again later.
You’ll then see a brief onboarding wizard to choose a model, dataset, and basic settings. You can skip it at any time and configure everything manually.

Studio - Quickstart
Unsloth Studio homepage has 4 main areas: Model, Dataset, Parameters, and Training/Config
Easy setup for models and data from Hugging Face or local files
Flexible training choices like QLoRA, LoRA, or full fine-tuning, with defaults filled in
Helpful config tools for splits, column mapping, hyperparameters and YAML configs
Great training visibility with live progress, GPU stats, charts, startup status

1. Select model and method
Model Type
Select the modality that matches your use-case:
Text
Chat, instruction following, completion
Vision
Image + text (VLMs)
Audio
Speech / audio understanding
Embeddings
Sentence embeddings, retrieval
Training Method
Three methods are available, toggled with a pill selector:
QLoRA
4-bit quantized base model + LoRA adapter
Lowest
LoRA
Full-precision base model + LoRA adapter
Medium
Full Fine-tuning
All weights are trained
Highest
Type any Hugging Face model name or search the Hub directly from the combobox. Local models stored in ~/.unsloth/studio/models and your Hugging Face cache also appear in the list.
GGUF-format models are excluded from training - they are inference-only.
When you pick a model the Studio automatically fetches its configuration from the backend and pre-fills sensible defaults for all hyperparameters.
HuggingFace Token
Paste your Hugging Face access token here if the model is gated (e.g. Llama, Gemma). The token is validated in real-time and an error is shown inline if it is invalid.
2. Dataset
Source
Switch between two tabs to choose where your data comes from:
HuggingFace Hub - live search against the Hub. The last-updated date is shown for each result.
Local - drag-and-drop or click to upload a file unstructured or structured files like:
PDF,DOCX,JSONL,JSON,CSV, orParquetformat. Previously uploaded datasets appear in a list that refreshes automatically.
Prompt Studio how to interpret and format your data:
auto
Let Unsloth detect the format automatically
alpaca
instruction / input / output columns
chatml
OpenAI-style messages array
sharegpt
ShareGPT-style conversations
Splits and Slicing
Subset - automatically populated from the dataset card.
Train split / Eval split - choose which splits to use. Setting an eval split enables the Eval Loss chart during training.
Dataset slice - optionally restrict training to a row range (start index / end index) for quick experiments.
Column Mapping
If the Studio cannot automatically map your dataset columns to the correct roles a Dataset Preview dialog opens. It shows sample rows and lets you assign each column to instruction, input, output, image, etc. Suggested mappings are pre-filled where possible.
3. Hyperparameters
Parameters are grouped into collapsible sections. You can view our detailed LoRA hyperparameters guide here:
🧠Hyperparameters GuideMax Steps
0
0 means use Epochs instead
Context Length
2048
Options: 512 → 32768
Learning Rate
2e-4
LoRA Settings
(Hidden when Full Fine-tuning is selected)
Rank
16
Slider 4–128
Alpha
32
Slider 4–256
Dropout
0.05
LoRA Variant
LoRA
LoRA / RS-LoRA / LoftQ
Target Modules
All on
q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
For Vision models with an image dataset, four additional checkboxes appear:
Finetune Vision Layers
Finetune Language Layers
Finetune Attention Modules
Finetune MLP Modules
Training Hyperparameters
Organized into three tabs:
| Parameter | Default | |---|---| | Epochs | 3 | | Batch Size | 4 | | Gradient Accumulation | 8 | | Weight Decay | 0.01 | | Optimizer | AdamW 8-bit |
| Parameter | Default | |---|---| | LR Scheduler | linear | | Warmup Steps | 5 | | Gradient Checkpointing | unsloth | | Random Seed | 3407 | | Save Steps | 0 | | Eval Steps | 0 | | Packing | false | | Train on Completions | false |
| Parameter | Default | |---|---| | Enable W&B | false | | W&B Project | llm-finetuning | | Enable TensorBoard | false | | TensorBoard Dir | runs | | Log Frequency | 10 |
Unsloth Gradient Checkpointing: unsloth uses Unsloth's custom memory-efficient implementation, which can reduce VRAM usage significantly compared to the standard PyTorch option. It is the recommended default.
4. Training and Config
The bottom-right card has three config management buttons and the Start Training button.
Upload
Load a previously saved .yaml config file
Save
Export the current config to YAML
Reset
Revert all parameters to the model's defaults
The Start Training button stays disabled until a model and dataset are both configured. Validation errors appear inline - for example, setting eval steps without choosing an eval split, or pairing a text-only model with a vision dataset.
Loading Screen
After you click Start Training, a full-page overlay appears while the backend prepares everything.

The overlay shows an animated terminal with live phase updates:
Blue: Downloading model / dataset
Amber: Loading model / dataset
Blue: Configuring
Green: Training
You can cancel at any time using the × button in the corner. A confirmation dialog will appear before anything is stopped.
Training Progress and Observability
Once the first training step arrives the overlay dismisses and the live training view is revealed. The fine-tuning process is complete when steps reach 100% on the progress bar. You can view the elapsed time and tokens.

Status Panel
The left column shows:
Epoch - current fractional epoch (e.g.
Epoch 1.23)Progress bar - step-based, with percentage
Key metrics:
Loss - training loss to 4 decimal places
LR - current learning rate in scientific notation
Grad Norm - gradient norm
Model - the model being trained
Method -
QLoRA/LoRA/Full
Timing row - elapsed time, ETA, steps per second, and total tokens processed
GPU Monitor
The right column shows live GPU stats polled every few seconds:
Utilization - percentage bar
Temperature - °C bar
VRAM - used / total GB
Power - draw / limit in watts
Stopping Training
Use the Stop Training button in the top-right of the progress card. A dialog gives you two choices:
Stop & Save - saves a checkpoint before stopping
Cancel - stops immediately with no checkpoint

Each chart has settings (gear icon) with:
Viewing window
Last N steps slider
EMA Smoothing
0.6
Show Raw
On
Show Smoothed
On
Show Average line
On
Scale (per series)
Linear / Log
Outlier clipping
No clip / p99 / p95

Config Files
All training configurations can be saved and reloaded as YAML files. Files are named automatically as:

The YAML is structured into three sections:
This makes it easy to reproduce runs, share configurations, or version-control your experiments.
Data Recipes - Quickstart
Unsloth Data Recipes lets you upload documents like PDFs or CSVs files and transforms them into useable datasets. Create and edit datasets visually via a graph-node workflow.
The recipes page is the main entry point. Recipes are stored locally in the browser, so you come back to saved work later. From here, you can create a blank recipe or open a guided learning recipe.

Data Recipes follows the same basic path. You open the recipes page, create or pick a recipe, build the workflow in the editor, validate it run a preview, then run the full dataset once the output looks right. Add seed data and generation blocks, validate the workflow, preview sample output, then run a full dataset build. Unsloth Data Recipes is powered by NVIDIA DataDesigner.
At a glance a usual workflow should look like this:
Open the recipes page.
Create a new recipe or open an existing one.
Add blocks to define your dataset workflow.
Click Validate to catch configuration issues early.
Run a preview to inspect sample rows quickly.
Run a full dataset build when the recipe is ready.
Review progress and output live in graph or in Executions view for mode details.
Select the resulting dataset in Studio and fine tune a model.
Export - Quickstart
Use Unsloth Studio 'Export' to export, save, or convert models to GGUF, Safetensors, or LoRA for deployment, sharing, or local inference in Unsloth, llama.cpp, Ollama, vLLM, and more. Export a trained checkpoint or convert any existing model.

You can read our detailed tutorial / guide about exporting models with Unsloth Studio here:
Model Export Chat - Quickstart
Unsloth Studio lets you run models 100% offline on your computer. Run model formats like GGUF and safetensors from Hugging Face or from your local files.
Download + Run any model like GGUFs, fine-tuned adapters, safetensors etc.
Compare different model outputs side-by-side
Upload documents, images, and audio in your prompts
Tune inference settings like: temperature, top-p, top-k and system prompt

You can read our detailed tutorial / guide about running models with Unsloth Studio here:
Studio Chat Video Tutorial
Here are 2 video tutorials to get you started with Unsloth Studio!
The Unsloth Studio versions shown in the videos are old and are not reflective of the current version.
Here is a video tutorial created by NVIDIA to get you started with Studio:
Here is our complete step-by-step video tutorial, from installation to using Studio:
Advanced Settings
CLI Commands
The Unsloth CLI (cli.py) provides the following commands:
Project Structure
API Reference
All endpoints require a valid JWT Authorization: Bearer <token> header (except /api/auth/* and /api/health).
GET
/api/health
Health check
GET
/api/system
System info (GPU, CPU, memory)
POST
/api/auth/signup
Create account (requires setup token on first run)
POST
/api/auth/login
Login and receive JWT tokens
POST
/api/auth/refresh
Refresh an expired access token
GET
/api/auth/status
Check if auth is initialized
POST
/api/train/start
Start a training job
POST
/api/train/stop
Stop a running training job
POST
/api/train/reset
Reset training state
GET
/api/train/status
Get current training status
GET
/api/train/metrics
Get training metrics (loss, LR, steps)
GET
/api/train/stream
SSE stream of real-time training progress
GET
/api/models/
List available models
POST
/api/inference/chat
Send a chat message for inference
GET
/api/datasets/
List / manage datasets
Last updated
Was this helpful?

