DeepSeek-R1-Distill-1.5B is the most compact model in the DeepSeek-R1 distilled model family, built on the Qwen2.5-1.5B architecture. Despite its small size, it has inherited fundamental reasoning skills from its teacher model — DeepSeek-R1. The model was fine-tuned on data generated by DeepSeek-R1, allowing it to significantly outperform other open-source models of similar size across multiple reasoning benchmarks.
Technically, the 1.5B version is optimized for operation on devices with limited computing resources, such as laptops, mobile devices, and edge servers. It delivers fast response times and low power consumption, making it ideal for offline applications and integration into end-user products with strict latency requirements.
In terms of use cases, DeepSeek-R1-Distill-1.5B excels at basic text analysis, short-answer generation, automation of routine tasks (such as processing customer support requests), and educational applications where a compact yet capable model is needed. Despite its size limitations, the model performs well on tasks involving simple logical operations. This makes it an excellent choice for applications where speed and minimal resource consumption are critical.
Model Name | Context | Type | GPU | TPS | Status | Link |
---|
There are no public endpoints for this model yet.
Rent your own physically dedicated instance with hourly or long-term monthly billing.
We recommend deploying private instances in the following scenarios:
Name | vCPU | RAM, MB | Disk, GB | GPU | |||
---|---|---|---|---|---|---|---|
16 | 32768 | 160 | 1 | $0.41 | Launch | ||
16 | 16384 | 160 | 1 | $0.46 | Launch | ||
16 | 32768 | 160 | 1 | $0.53 | Launch | ||
16 | 32768 | 160 | 2 | $0.57 | Launch | ||
16 | 24576 | 160 | 1 | $0.88 | Launch | ||
16 | 32768 | 160 | 1 | $1.15 | Launch | ||
12 | 65536 | 160 | 1 | $1.20 | Launch | ||
16 | 65536 | 160 | 1 | $1.59 | Launch | ||
16 | 65536 | 160 | 1 | $2.58 | Launch | ||
16 | 65536 | 160 | 1 | $5.11 | Launch |
Name | vCPU | RAM, MB | Disk, GB | GPU | |||
---|---|---|---|---|---|---|---|
16 | 32768 | 160 | 1 | $0.41 | Launch | ||
16 | 16384 | 160 | 1 | $0.46 | Launch | ||
16 | 32768 | 160 | 1 | $0.53 | Launch | ||
16 | 32768 | 160 | 2 | $0.57 | Launch | ||
16 | 24576 | 160 | 1 | $0.88 | Launch | ||
16 | 32768 | 160 | 1 | $1.15 | Launch | ||
12 | 65536 | 160 | 1 | $1.20 | Launch | ||
16 | 65536 | 160 | 1 | $1.59 | Launch | ||
16 | 65536 | 160 | 1 | $2.58 | Launch | ||
16 | 65536 | 160 | 1 | $5.11 | Launch |
Name | vCPU | RAM, MB | Disk, GB | GPU | |||
---|---|---|---|---|---|---|---|
16 | 32768 | 160 | 1 | $0.41 | Launch | ||
16 | 16384 | 160 | 1 | $0.46 | Launch | ||
16 | 32768 | 160 | 1 | $0.53 | Launch | ||
16 | 32768 | 160 | 2 | $0.57 | Launch | ||
16 | 24576 | 160 | 1 | $0.88 | Launch | ||
16 | 32768 | 160 | 1 | $1.15 | Launch | ||
12 | 65536 | 160 | 1 | $1.20 | Launch | ||
16 | 65536 | 160 | 1 | $1.59 | Launch | ||
16 | 65536 | 160 | 1 | $2.58 | Launch | ||
16 | 65536 | 160 | 1 | $5.11 | Launch |
Contact our dedicated neural networks support team at nn@immers.cloud or send your request to the sales department at sale@immers.cloud.