# Models

- [Gemma 4 - How to Run Locally](/docs/models/gemma-4.md): Run Google’s new Gemma 4 models locally, including E2B, E4B, 26B A4B, and 31B.
- [Gemma 4 Fine-tuning Guide](/docs/models/gemma-4/train.md): Train Gemma 4 by Google with Unsloth.
- [Qwen3.5 - How to Run Locally](/docs/models/qwen3.5.md): Run the new Qwen3.5 LLMs including Medium: Qwen3.5-35B-A3B, 27B, 122B-A10B, Small: Qwen3.5-0.8B, 2B, 4B, 9B and 397B-A17B on your local device!
- [Qwen3.5 Fine-tuning Guide](/docs/models/qwen3.5/fine-tune.md): Learn how to fine-tune Qwen3.5 LLMs with Unsloth.
- [Qwen3.5 GGUF Benchmarks](/docs/models/qwen3.5/gguf-benchmarks.md): See how Unsloth Dynamic GGUFs perform + analysis of perplexity, KL divergence & MXFP4.
- [GLM-5.1 - How to Run Locally](/docs/models/glm-5.1.md): Run the new GLM-5.1 model by Z.ai on your own local device!
- [NVIDIA Nemotron 3 Nano - How To Run Guide](/docs/models/nemotron-3.md): Run & fine-tune NVIDIA Nemotron 3 Nano locally on your device!
- [NVIDIA Nemotron-3-Super: How To Run Guide](/docs/models/nemotron-3/nemotron-3-super.md): Run & fine-tune NVIDIA Nemotron-3-Super-120B-A12B locally on your device!
- [Qwen3-Coder-Next: How to Run Locally](/docs/models/qwen3-coder-next.md): Guide to run Qwen3-Coder-Next locally on your device!
- [GLM-4.7-Flash: How To Run Locally](/docs/models/glm-4.7-flash.md): Run & fine-tune GLM-4.7-Flash locally on your device!
- [MiniMax-M2.5: How to Run Guide](/docs/models/minimax-m25.md): Run MiniMax-M2.5 locally on your own device!
- [Kimi K2.5: How to Run Locally Guide](/docs/models/kimi-k2.5.md): Guide on running Kimi-K2.5 on your own local device!
- [gpt-oss: How to Run Guide](/docs/models/gpt-oss-how-to-run-and-fine-tune.md): Run & fine-tune OpenAI's new open-source models!
- [gpt-oss Reinforcement Learning](/docs/models/gpt-oss-how-to-run-and-fine-tune/gpt-oss-reinforcement-learning.md)
- [Tutorial: How to Train gpt-oss with RL](/docs/models/gpt-oss-how-to-run-and-fine-tune/gpt-oss-reinforcement-learning/tutorial-how-to-train-gpt-oss-with-rl.md): Learn to train OpenAI gpt-oss with GRPO to autonomously beat 2048 locally or on Colab.
- [Tutorial: How to Fine-tune gpt-oss](/docs/models/gpt-oss-how-to-run-and-fine-tune/tutorial-how-to-fine-tune-gpt-oss.md): Learn step-by-step how to train OpenAI gpt-oss locally with Unsloth.
- [Long Context gpt-oss Training](/docs/models/gpt-oss-how-to-run-and-fine-tune/long-context-gpt-oss-training.md)
- [Large language model (LLMs) Tutorials](/docs/models/tutorials.md): Discover the latest LLMs and learn how to run and fine-tune models locally for optimal performance with Unsloth.
- [GLM-5: How to Run Locally Guide](/docs/models/tutorials/glm-5.md): Run the new GLM-5 model by Z.ai on your own local device!
- [Qwen3 - How to Run & Fine-tune](/docs/models/tutorials/qwen3-how-to-run-and-fine-tune.md): Learn to run & fine-tune Qwen3 locally with Unsloth + our Dynamic 2.0 quants
- [Qwen3-VL: How to Run Guide](/docs/models/tutorials/qwen3-how-to-run-and-fine-tune/qwen3-vl-how-to-run-and-fine-tune.md): Learn to fine-tune and run Qwen3-VL locally with Unsloth.
- [Qwen3-2507: Run Locally Guide](/docs/models/tutorials/qwen3-how-to-run-and-fine-tune/qwen3-2507.md): Run Qwen3-30B-A3B-2507 and 235B-A22B Thinking and Instruct versions locally on your device!
- [Qwen3-Coder: How to Run Locally](/docs/models/tutorials/qwen3-coder-how-to-run-locally.md): Run Qwen3-Coder-30B-A3B-Instruct and 480B-A35B locally with Unsloth Dynamic quants.
- [Gemma 3 - How to Run Guide](/docs/models/tutorials/gemma-3-how-to-run-and-fine-tune.md): How to run Gemma 3 effectively with our GGUFs on llama.cpp, Ollama, Open WebUI and how to fine-tune with Unsloth!
- [Gemma 3n: How to Run & Fine-tune](/docs/models/tutorials/gemma-3-how-to-run-and-fine-tune/gemma-3n-how-to-run-and-fine-tune.md): Run Google's new Gemma 3n locally with Dynamic GGUFs on llama.cpp, Ollama, Open WebUI and fine-tune with Unsloth!
- [DeepSeek-OCR 2: How to Run & Fine-tune Guide](/docs/models/tutorials/deepseek-ocr-2.md): Guide on how to run and fine-tune DeepSeek-OCR-2 locally.
- [GLM-4.7: How to Run Locally Guide](/docs/models/tutorials/glm-4.7.md): A guide on how to run Z.ai GLM-4.7 model on your own local device!
- [How to Run Qwen-Image-2512 Locally in ComfyUI](/docs/models/tutorials/qwen-image-2512.md): Step-by-step tutorial for running Qwen-Image-2512 on your local device with ComfyUI.
- [Run Qwen-Image-2512 in stable-diffusion.cpp Tutorial](/docs/models/tutorials/qwen-image-2512/stable-diffusion.cpp.md): Tutorial for using Qwen-Image-2512 in stable-diffusion.cpp.
- [Devstral 2 - How to Run Guide](/docs/models/tutorials/devstral-2.md): Guide for local running Mistral Devstral 2 models: 123B-Instruct-2512 and Small-2-24B-Instruct-2512.
- [Ministral 3 - How to Run Guide](/docs/models/tutorials/ministral-3.md): Guide for Mistral Ministral 3 models, to run or fine-tune locally on your device
- [DeepSeek-OCR: How to Run & Fine-tune](/docs/models/tutorials/deepseek-ocr-how-to-run-and-fine-tune.md): Guide on how to run and fine-tune DeepSeek-OCR locally.
- [Kimi K2 Thinking: Run Locally Guide](/docs/models/tutorials/kimi-k2-thinking-how-to-run-locally.md): Guide on running Kimi-K2-Thinking and Kimi-K2 on your own local device!
- [GLM-4.6: Run Locally Guide](/docs/models/tutorials/glm-4.6-how-to-run-locally.md): A guide on how to run Z.ai GLM-4.6 and GLM-4.6V-Flash model on your own local device!
- [Qwen3-Next: Run Locally Guide](/docs/models/tutorials/qwen3-next.md): Run Qwen3-Next-80B-A3B-Instruct and Thinking versions locally on your device!
- [FunctionGemma: How to Run & Fine-tune](/docs/models/tutorials/functiongemma.md): Learn how to run and fine-tune FunctionGemma locally on your device and phone.
- [DeepSeek-V3.1: How to Run Locally](/docs/models/tutorials/deepseek-v3.1-how-to-run-locally.md): A guide on how to run DeepSeek-V3.1 and Terminus on your own local device!
- [DeepSeek-R1-0528: How to Run Locally](/docs/models/tutorials/deepseek-r1-0528-how-to-run-locally.md): A guide on how to run DeepSeek-R1-0528 including Qwen3 on your own local device!
- [Liquid LFM2.5: How To Run & Fine-tune](/docs/models/tutorials/lfm2.5.md): Run and fine-tune LFM2.5 Instruct and Vision locally on your device!
- [Magistral: How to Run & Fine-tune](/docs/models/tutorials/magistral-how-to-run-and-fine-tune.md): Meet Magistral - Mistral's new reasoning models.
- [IBM Granite 4.0](/docs/models/tutorials/ibm-granite-4.0.md): How to run IBM Granite-4.0 with Unsloth GGUFs on llama.cpp, Ollama and how to fine-tune!
- [Llama 4: How to Run & Fine-tune](/docs/models/tutorials/llama-4-how-to-run-and-fine-tune.md): How to run Llama 4 locally using our dynamic GGUFs which recovers accuracy compared to standard quantization.
- [Grok 2](/docs/models/tutorials/grok-2.md): Run xAI's Grok 2 model locally!
- [Devstral: How to Run & Fine-tune](/docs/models/tutorials/devstral-how-to-run-and-fine-tune.md): Run and fine-tune Mistral Devstral 1.1, including Small-2507 and 2505.
- [How to Run Local LLMs with Docker: Step-by-Step Guide](/docs/models/tutorials/how-to-run-llms-with-docker.md): Learn how to run Large Language Models (LLMs) with Docker & Unsloth on your local device.
- [DeepSeek-V3-0324: How to Run Locally](/docs/models/tutorials/deepseek-v3-0324-how-to-run-locally.md): How to run DeepSeek-V3-0324 locally using our dynamic quants which recovers accuracy
- [DeepSeek-R1: How to Run Locally](/docs/models/tutorials/deepseek-r1-how-to-run-locally.md): A guide on how you can run our 1.58-bit Dynamic Quants for DeepSeek-R1 using llama.cpp.
- [DeepSeek-R1 Dynamic 1.58-bit](/docs/models/tutorials/deepseek-r1-how-to-run-locally/deepseek-r1-dynamic-1.58-bit.md): See performance comparison tables for Unsloth's Dynamic GGUF Quants vs Standard IMatrix Quants.
- [Phi-4 Reasoning: How to Run & Fine-tune](/docs/models/tutorials/phi-4-reasoning-how-to-run-and-fine-tune.md): Learn to run & fine-tune Phi-4 reasoning models locally with Unsloth + our Dynamic 2.0 quants
- [QwQ-32B: How to Run effectively](/docs/models/tutorials/qwq-32b-how-to-run-effectively.md): How to run QwQ-32B effectively with our bug fixes and without endless generations + GGUFs.
