Model
/
LLM
/
DeepSeek

DeepSeek R1

Open-source reasoning model rivaling OpenAI-o1, excelling in math, code, reasoning, and cost efficiency.

GMI Cloud is excited to announce that we are now hosting a dedicated DeepSeek-R1 inference endpoint, on optimized, US-based hardware. What's DeepSeek-R1? Read our initial takeaways here.

Technical details:

Model Provider:
DeepSeek
Type:
Chat
Parameters:
685B
Deployment:
Serverless (MaaS) or Dedicated Endpoint
Quantization:
FP16
Context Length:
Up to 128,000 tokens

Distilled models offering:

  • DeepSeek-R1-Distill-Llama-70B
  • DeepSeek-R1-Distill-Qwen-32B
  • DeepSeek-R1-Distill-Qwen-14B
  • DeepSeek-R1-Distill-Llama-8B
  • DeepSeek-R1-Distill-Qwen-7B
  • DeepSeek-R1-Distill-Qwen-1.5B

Try our token-free service with unlimited usage!

Reach out for access to our dedicated endpoint Here.