SambaStack supports a variety of models that can be deployed to both on-premises and hosted environments. Contact your system administrator to determine which models are available on your deployment. You can also use the Model list API command to view which models are deployed and available for your use.Documentation Index
Fetch the complete documentation index at: https://docs-preprod.sambanova.ai/docs/llms.txt
Use this file to discover all available pages before exploring further.
Deployment options
When deploying models in SambaStack, administrators can select from various context length and batch size combinations.- Smaller batch sizes provide higher token throughput (tokens/second).
- Larger batch sizes provide better concurrency for multiple users.
Supported models
The table below lists supported models, context lengths, batch sizes, and features.| Developer/Model ID | Type | Context length (batch size) | Features and optimizations | View on Hugging Face |
|---|---|---|---|---|
| Meta | ||||
Meta-Llama-3.3-70B-Instruct | Text | View
| View
| Model card |
Meta-Llama-3.1-8B-Instruct | Text | View
| View
| Model card |
Llama-4-Maverick-17B-128E-Instruct | Image, Text | View
| View
| Model card |
| DeepSeek | ||||
DeepSeek-R1-0528 | Reasoning, Text | View
| View
| Model card |
DeepSeek-R1-Distill-Llama-70B | Reasoning, Text | View
| View
| Model card |
DeepSeek-V3-0324 | Text | View
| View
| Model card |
DeepSeek-V3.1 | Reasoning, Text | View
| View
| Model card |
| OpenAI | ||||
Whisper-Large-v3 | Audio | View
| View
| Model card |
| Qwen | ||||
Qwen3-32B | Reasoning, Text | View
| View
| Model card |
| Tokyotech-llm | ||||
Llama-3.3-Swallow-70B-Instruct-v0.4 | Text | View
| View
| Model card |
| Other | ||||
E5-Mistral-7B-Instruct | Embedding | View
| View
| Model card |
