# 推論とデプロイ

微調整済みモデルは次の方法でも実行できます [Unslothの2倍高速な推論](https://unsloth.ai/docs/jp/ji-ben/inference-and-deployment/unsloth-inference).

<table data-card-size="large" data-view="cards"><thead><tr><th></th><th data-hidden data-card-target data-type="content-ref"></th><th data-hidden data-type="content-ref"></th></tr></thead><tbody><tr><td><a href="../../xin-zhe/studio#run-models-locally">Unsloth Studio Chat</a></td><td><a href="../xin-zhe/studio/chat">chat</a></td><td></td></tr><tr><td><a href="inference-and-deployment/saving-to-gguf">llama.cpp - GGUFへの保存</a></td><td><a href="inference-and-deployment/saving-to-gguf">saving-to-gguf</a></td><td><a href="inference-and-deployment/saving-to-gguf">saving-to-gguf</a></td></tr><tr><td><a href="inference-and-deployment/vllm-guide">vLLM</a></td><td><a href="inference-and-deployment/vllm-guide">vllm-guide</a></td><td><a href="inference-and-deployment/vllm-guide">vllm-guide</a></td></tr><tr><td><a href="inference-and-deployment/saving-to-ollama">Ollama</a></td><td><a href="inference-and-deployment/saving-to-ollama">saving-to-ollama</a></td><td><a href="inference-and-deployment/saving-to-ollama">saving-to-ollama</a></td></tr><tr><td><a href="inference-and-deployment/lm-studio">LM Studio</a></td><td><a href="inference-and-deployment/lm-studio">lm-studio</a></td><td></td></tr><tr><td><a href="inference-and-deployment/sglang-guide">SGLang</a></td><td><a href="inference-and-deployment/sglang-guide">sglang-guide</a></td><td><a href="inference-and-deployment/vllm-guide/vllm-engine-arguments">vllm-engine-arguments</a></td></tr><tr><td><a href="inference-and-deployment/troubleshooting-inference">トラブルシューティング</a></td><td><a href="inference-and-deployment/troubleshooting-inference">troubleshooting-inference</a></td><td><a href="inference-and-deployment/troubleshooting-inference">troubleshooting-inference</a></td></tr><tr><td><a href="inference-and-deployment/llama-server-and-openai-endpoint">llama-server &#x26; OpenAIエンドポイント</a></td><td><a href="inference-and-deployment/llama-server-and-openai-endpoint">llama-server-and-openai-endpoint</a></td><td></td></tr><tr><td><a href="tool-calling-guide-for-local-llms">ツール呼び出し</a></td><td><a href="tool-calling-guide-for-local-llms">tool-calling-guide-for-local-llms</a></td><td></td></tr><tr><td><a href="inference-and-deployment/deploy-llms-phone">スマートフォンでLLMを実行する</a></td><td><a href="inference-and-deployment/deploy-llms-phone">deploy-llms-phone</a></td><td></td></tr></tbody></table>
