When you purchase through links on our site, we may earn an affiliate commission. This doesn’t affect our editorial independence.
Users of Alibaba Cloud can access DeepSeek’s AI model by signing into their PAI Model Gallery, where they can find DeepSeek’s most recent models as well as other “distilled” versions.
Alibaba Group Holding’s cloud computing services division on Monday made DeepSeek‘s AI models available on its platform. The company has joined a plethora of Big Tech companies that have made the Chinese-made, open-source AI system available to its customers.
In a statement posted on its WeChat handle, the company said “On [our] platform, users can achieve the whole process from training to deployment to inference with zero coding.”
According to the statement, the platform streamlines the model-development procedure, adding that it is “delivering a quicker, more effective, and more convenient AI development and application experience to developers and enterprise users.”
Alibaba Cloud users can log into their PAI Model Gallery, select DeepSeek’s AI models and use them to power their own text-generating and reasoning apps. The PAI Model gallery is a collection of open-source large language models (LLMs).
The platform features the most cutting-edge AI models from the Hangzhou-based start-up, DeepSeek-V3 and DeepSeek-R1, which are praised for having been created for a fraction of the price and processing power that large AI tech companies typically spend to create LLMs. Additionally, it offers distilled versions of those models, such as DeepSeek-R1-Distill-Qwen-7B.
What is LLM, Open-Source, Distilled-AI Model, and Cloud Computing?
LLM: According to Investopedia, a Large Language Model (LLM) is a deep learning algorithm that’s equipped to summarize, translate, predict, and generate text to convey ideas and concepts. Large language models rely on substantively large datasets to perform those functions. AI systems such as OpenAI’s ChatGPT are powered by LLMs.
Open Source: A software program is open source when its code is made publicly available, enabling outside developers to alter or distribute its architecture, repair broken links, or expand its functionality.
Distilled AI: In artificial intelligence (AI), model distillation is a process that moves knowledge from a big, complicated model (the “teacher”) to a smaller, more effective model (the “student”). During this process, the student model is trained to reproduce the teacher’s outputs. This helps the smaller model work more efficiently while maintaining almost all of the original model’s performance.
Cloud Computing: Cloud computing technology allows businesses to manage or distribute a variety of software and other digital resources over the internet as an on-demand service.
Big Techs Adopting DeepSeek
- Chinese video-game maker, Tencent Holdings announced on Sunday that it has also begun to support DeepSeek’s R1 reasoning model on its cloud computing platform.
- Leading AI chip manufacturer, Nvidia announced that DeepSeek-R1 is now accessible to users of its NIM microservice as of Thursday. The company claims that the model offers “leading accuracy,” “high inference efficiency,” and “state-of-the-art reasoning capabilities” for tasks involving logical inference, reasoning, math, coding, and language comprehension.
- Last week, Microsoft announced that it had featured DeepSeek R1 on its cloud-computing platform, Azure and its developer platform GitHub.
- Also, developers can now create softwares using DeepSeek’s R1 on Amazon Web Services.