Open-source reasoning model rivaling OpenAI-o1, excelling in math, code, reasoning, and cost efficiency.

GMI Cloud is excited to announce that we are now hosting a dedicated DeepSeek-R1 inference endpoint, on optimized, US-based hardware. What's DeepSeek-R1? Read our initial takeaways here.
Reach out for access to our dedicated endpoint Here.
Get quick answers to common queries in our FAQs.
GMI Cloud hosts a dedicated DeepSeek-R1 inference endpoint on optimized, US-based hardware. You can use it either as Serverless (MaaS) for on-demand access or as a Dedicated Endpoint if you want an isolated deployment.
The page lists: Type: Chat • Parameters: 685B • Quantization: FP16 • Context length: up to 128,000 tokens.
Yes. The page advertises a token-free service with unlimited usage. Access is by request—reach out to get the dedicated endpoint.
GMI Cloud lists multiple distilled options:
DeepSeek-R1-Distill-Llama-70B
DeepSeek-R1-Distill-Qwen-32B
DeepSeek-R1-Distill-Qwen-14B
DeepSeek-R1-Distill-Llama-8B
DeepSeek-R1-Distill-Qwen-7B
DeepSeek-R1-Distill-Qwen-1.5B
The page describes DeepSeek R1 as an open-source reasoning model that excels in math, code, and reasoning with a focus on cost efficiency, making it suitable for chat-style reasoning tasks where long context and structured problem solving matter.
The page states that DeepSeek R1 is an open-source reasoning model rivaling OpenAI-o1, highlighting strong math, coding, and reasoning capabilities while emphasizing efficiency.