# Inference & Deployment

You can also run your fine-tuned models by using [Unsloth's 2x faster inference](https://unsloth.ai/docs/basics/inference-and-deployment/unsloth-inference).

<table data-card-size="large" data-view="cards"><thead><tr><th></th><th data-hidden data-card-target data-type="content-ref"></th><th data-hidden data-type="content-ref"></th></tr></thead><tbody><tr><td><a href="../../new/studio#run-models-locally">Unsloth Studio Chat</a></td><td><a href="../new/studio/chat">chat</a></td><td></td></tr><tr><td><a href="inference-and-deployment/saving-to-gguf">llama.cpp - Saving to GGUF</a></td><td><a href="inference-and-deployment/saving-to-gguf">saving-to-gguf</a></td><td><a href="inference-and-deployment/saving-to-gguf">saving-to-gguf</a></td></tr><tr><td><a href="inference-and-deployment/vllm-guide">vLLM</a></td><td><a href="inference-and-deployment/vllm-guide">vllm-guide</a></td><td><a href="inference-and-deployment/vllm-guide">vllm-guide</a></td></tr><tr><td><a href="inference-and-deployment/saving-to-ollama">Ollama</a></td><td><a href="inference-and-deployment/saving-to-ollama">saving-to-ollama</a></td><td><a href="inference-and-deployment/saving-to-ollama">saving-to-ollama</a></td></tr><tr><td><a href="inference-and-deployment/lm-studio">LM Studio</a></td><td><a href="inference-and-deployment/lm-studio">lm-studio</a></td><td></td></tr><tr><td><a href="inference-and-deployment/sglang-guide">SGLang</a></td><td><a href="inference-and-deployment/sglang-guide">sglang-guide</a></td><td><a href="inference-and-deployment/vllm-guide/vllm-engine-arguments">vllm-engine-arguments</a></td></tr><tr><td><a href="inference-and-deployment/troubleshooting-inference">Troubleshooting</a></td><td><a href="inference-and-deployment/troubleshooting-inference">troubleshooting-inference</a></td><td><a href="inference-and-deployment/troubleshooting-inference">troubleshooting-inference</a></td></tr><tr><td><a href="inference-and-deployment/llama-server-and-openai-endpoint">llama-server &#x26; OpenAI endpoint</a></td><td><a href="inference-and-deployment/llama-server-and-openai-endpoint">llama-server-and-openai-endpoint</a></td><td></td></tr><tr><td><a href="tool-calling-guide-for-local-llms">Tool Calling</a></td><td><a href="tool-calling-guide-for-local-llms">tool-calling-guide-for-local-llms</a></td><td></td></tr><tr><td><a href="inference-and-deployment/deploy-llms-phone">Run LLMs on your Phone</a></td><td><a href="inference-and-deployment/deploy-llms-phone">deploy-llms-phone</a></td><td></td></tr></tbody></table>
