DeepSeek R1 is now available on Nvidia, AWS, and Github as available models on Hugging Face shoot past 3,000
DeepSeek's open source model sees rapid adoption
- Nvidia integrates DeepSeek-R1 as a NIM microservice
- AWS supports DeepSeek-R1 with a focus on scalable and cost-efficient AI deployment
- Microsoft also has future local deployment plans for DeepSeek
Having taken the AI world by storm in recent weeks, DeepSeek has now made significant strides in expanding the accessibility of its advanced reasoning models.
The company has announced its flagship DeepSeek R1 model is now available on multiple platforms, including Nvidia, AWS, and GitHub.
DeepSeek’s open source nature allows developers to build models based on its architecture, and, at press time, there are 3,374 DeepSeek-based models available collaborative AI-model development platform Hugging Face.
Nvidia, AWS, Github & Azure now offers DeepSeek
On AWS, DeepSeek-R1 models are now accessible through Amazon Bedrock which simplifies API integration and Amazon SageMaker which enables advanced customization and training, supported by AWS Trainium and Inferentia for optimized cost efficiency.
AWS also offers DeepSeek-R1-Distill, a lighter version, through Amazon Bedrock Custom Model Import. This serverless deployment simplifies infrastructure management while maintaining scalability.
Nvidia has also integrated DeepSeek-R1 as a NIM microservice, leveraging its Hopper architecture and FP8 Transformer Engine acceleration to deliver real-time, high-quality responses.
The model, which features 671 billion parameters and a 128,000-token context length, utilizes test-time scaling for improved accuracy.
Are you a pro? Subscribe to our newsletter
Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!
It also benefits from Nvidia’s Hopper architecture, using FP8 Transformer Engine acceleration and NVLink connectivity. Running on an HGX H200 system, DeepSeek-R1 can generate up to 3,872 tokens per second.
Microsoft’s Azure AI Foundry and GitHub have further expanded DeepSeek’s reach, offering developers a secure and scalable platform to integrate AI into their workflows.
Microsoft has also implemented extensive safety measures, including content filtering and automated assessments. The company claims that it plans to offer distilled versions of DeepSeek-R1 for local deployment on Copilot+ PCs in the future.
DeepSeek-R1 took the world by storm by offering a powerful, cost-efficient AI model with advanced reasoning capabilities and has dethroned popular AI models like ChatGPT.
R1 was reportedly trained for just $6 million, with its most advanced versions being about 95% cheaper to train than comparable models from NVIDIA and Microsoft.
You may also like
Efosa has been writing about technology for over 7 years, initially driven by curiosity but now fueled by a strong passion for the field. He holds both a Master's and a PhD in sciences, which provided him with a solid foundation in analytical thinking. Efosa developed a keen interest in technology policy, specifically exploring the intersection of privacy, security, and politics. His research delves into how technological advancements influence regulatory frameworks and societal norms, particularly concerning data protection and cybersecurity. Upon joining TechRadar Pro, in addition to privacy and technology policy, he is also focused on B2B security products. Efosa can be contacted at this email: udinmwenefosa@gmail.com
You must confirm your public display name before commenting
Please logout and then login again, you will then be prompted to enter your display name.