mirror of https://github.com/vllm-project/vllm.git
37 lines
2.5 KiB
Markdown
37 lines
2.5 KiB
Markdown
---
|
|
title: TPU
|
|
---
|
|
[](){ #tpu-supported-models }
|
|
|
|
# TPU Supported Models
|
|
## Text-only Language Models
|
|
|
|
| Model | Architecture | Supported |
|
|
|-----------------------------------------------------|--------------------------------|-----------|
|
|
| mistralai/Mixtral-8x7B-Instruct-v0.1 | MixtralForCausalLM | 🟨 |
|
|
| mistralai/Mistral-Small-24B-Instruct-2501 | MistralForCausalLM | ✅ |
|
|
| mistralai/Codestral-22B-v0.1 | MistralForCausalLM | ✅ |
|
|
| mistralai/Mixtral-8x22B-Instruct-v0.1 | MixtralForCausalLM | ❌ |
|
|
| meta-llama/Llama-3.3-70B-Instruct | LlamaForCausalLM | ✅ |
|
|
| meta-llama/Llama-3.1-8B-Instruct | LlamaForCausalLM | ✅ |
|
|
| meta-llama/Llama-3.1-70B-Instruct | LlamaForCausalLM | ✅ |
|
|
| meta-llama/Llama-4-* | Llama4ForConditionalGeneration | ❌ |
|
|
| microsoft/Phi-3-mini-128k-instruct | Phi3ForCausalLM | 🟨 |
|
|
| microsoft/phi-4 | Phi3ForCausalLM | ❌ |
|
|
| google/gemma-3-27b-it | Gemma3ForConditionalGeneration | 🟨 |
|
|
| google/gemma-3-4b-it | Gemma3ForConditionalGeneration | ❌ |
|
|
| deepseek-ai/DeepSeek-R1 | DeepseekV3ForCausalLM | ❌ |
|
|
| deepseek-ai/DeepSeek-V3 | DeepseekV3ForCausalLM | ❌ |
|
|
| RedHatAI/Meta-Llama-3.1-8B-Instruct-quantized.w8a8 | LlamaForCausalLM | ✅ |
|
|
| RedHatAI/Meta-Llama-3.1-70B-Instruct-quantized.w8a8 | LlamaForCausalLM | ✅ |
|
|
| Qwen/Qwen3-8B | Qwen3ForCausalLM | ✅ |
|
|
| Qwen/Qwen3-32B | Qwen3ForCausalLM | ✅ |
|
|
| Qwen/Qwen2.5-7B-Instruct | Qwen2ForCausalLM | ✅ |
|
|
| Qwen/Qwen2.5-32B | Qwen2ForCausalLM | ✅ |
|
|
| Qwen/Qwen2.5-14B-Instruct | Qwen2ForCausalLM | ✅ |
|
|
| Qwen/Qwen2.5-1.5B-Instruct | Qwen2ForCausalLM | 🟨 |
|
|
|
|
✅ Runs and optimized.
|
|
🟨 Runs and correct but not optimized to green yet.
|
|
❌ Does not pass accuracy test or does not run.
|