First came LLaMA, then came Vicuna, Alpaca and Falcon — all LLaMA-based open source models. Recently, the winds of open source have shifted towards China and there has been a surge in models topping Hugging Face’s Open LLM leaderboard.
Recently, it was China-based DeepSeek, a 67 billion parameter open source model that is based on LLaMA and outperformed Llama 2, Claude-2, and Grok-1 on various metrics. DeepSeek functions as a research laboratory driven by the “mission to explore the mysteries surrounding artificial general intelligence (AGI) with a spirit of curiosity”. San Francisco-based Abacus AI, an AI-assisted data science platform to build ML systems and AI agents, is ready to host the model on its platform.
Deepseek is not the only Chinese open source LLM to be making waves, in fact, the leaderboard is trending with many China-based open source LLMs such as Tigerbot-70b-chat-v2, Yi-34B model, Qwen-72B and a smaller Qwen-1.8B. Tigerbot-70b-chat-v2, developed by Tiger Research and available on GitHub and Hugging Face, is built on top of Llama 2 70B architecture.
Keep reading with a 7-day free trial
Subscribe to Sector 6 | The Newsletter of AIM to keep reading this post and get 7 days of free access to the full post archives.