A powerful large language model for text generation and reasoning optimized for any device
DeepSeek R1 is a powerful large language model (LLM) designed for text generation, reasoning, and code generation tasks. With InferX, you can run DeepSeek R1 on any device using the same API - from edge devices to powerful servers.
DeepSeek R1 is included with InferX:
Parameter | Description | Default | Range |
---|---|---|---|
max_tokens | Maximum number of tokens to generate | 256 | 1-2048 |
temperature | Controls randomness (higher = more random) | 0.8 | 0.0-2.0 |
top_p | Nucleus sampling parameter | 0.95 | 0.0-1.0 |
top_k | Limits vocabulary to top k tokens | 40 | 1-100 |
InferX optimizes DeepSeek R1 for your hardware:
Hardware | Tokens/Second | Memory Usage |
---|---|---|
Jetson AGX Orin | ~15 | ~8GB |
RTX 4090 | ~50 | ~12GB |
Intel i7 CPU | ~5 | ~6GB |
A powerful large language model for text generation and reasoning optimized for any device
DeepSeek R1 is a powerful large language model (LLM) designed for text generation, reasoning, and code generation tasks. With InferX, you can run DeepSeek R1 on any device using the same API - from edge devices to powerful servers.
DeepSeek R1 is included with InferX:
Parameter | Description | Default | Range |
---|---|---|---|
max_tokens | Maximum number of tokens to generate | 256 | 1-2048 |
temperature | Controls randomness (higher = more random) | 0.8 | 0.0-2.0 |
top_p | Nucleus sampling parameter | 0.95 | 0.0-1.0 |
top_k | Limits vocabulary to top k tokens | 40 | 1-100 |
InferX optimizes DeepSeek R1 for your hardware:
Hardware | Tokens/Second | Memory Usage |
---|---|---|
Jetson AGX Orin | ~15 | ~8GB |
RTX 4090 | ~50 | ~12GB |
Intel i7 CPU | ~5 | ~6GB |