Don’t Overlook China’s Open Source LLMs

A version of a Chinese LLM tops the open LLM leaderboard.

Created Using DALL-E

Next Week in The Sequence:

  • Edge 369: Our series about LLM reasoning continues with the recently published Chain-of-Code(CoC) method. We review the original CoC paper by Google DeepMind and the super popular Embedchain framework.

  • Edge 370: We dive the new AlphaGeometry model created by Google DeepMind that is able to solve geometry problems at the level of a math olympiad gold medalist.

You can subscribe below!

TheSequence is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.

📝 Editorial: Don’t Overlook China’s Open Source LLMs

If you visit the open LLM leaderboard today, you might encounter an unfamiliar model at the top of the charts: Smaug-72B. Open-sourced by Abacus AI, this model is a fine-tuned version of another model, Qwen-72B, which Alibaba released a few months ago. The Qwen family of open-source LLMs has scored incredibly high across some of the top open-source benchmarks, showcasing the latest examples of Chinese innovation in the open-source generative AI space. While open-source LLMs are typically associated with Western models like LLaMA or Mistral, the pace of high-quality releases from China is nothing short of remarkable. Here are a few examples:

Smaug was technically developed by an American company but as a fine-tuned version of a Chinese model. From what I can tell, most open-source Chinese LLMs share strong architectural commonalities with models like Llama or Mistral; however, there hasn’t been any major innovation from an architectural standpoint. Nonetheless, the quality is undeniable. While many skeptics of open-source generative AI regularly cited China as a major concern, they fail to recognize the contributions that Chinese research labs and startups will make to the space. It would be interesting to see how regulation plays a role in the evolution of open-source LLMs in China and Western countries. For now, don’t overlook the Chinese open-source LLMs. They are very impressive.


🎥 Watch Now: Building Plaid’s ML Fraud Detection Application

Want to learn about Plaid’s ML platform journey? In this on-demand recording, Plaid Software Engineer Renault Young shared the technical challenges they faced, how they set up the data foundations they needed to start building an ML platform, what they used to look for patterns in transaction data in real time, and more. Today, Signal is Plaid’s biggest ML application and analyzes 1000+ risk factors per ACH transaction. 

The on-demand recording is now available for you to watch and share with your colleagues! 


🔎 ML Research

Specialized SLMs

Apple Research published a paper evaluating small language model architectures based on inference, specialization and training budgets. The paper evaluates different architectures such as hyper-networks or mixture of experts to achieve different levels of specializations based on budget constraints —> Read more.

Chain-of-Abstraction

Meta AI Research published a paper detailing Chain of Abstraction(CoA), a method that combines reasoning and tool learning in LLMs. CoA creates abstract placeholders in reasoning chains and then fills htem with specific knowledge using tools —> Read more.

Mastering Chess Without Search

Researchers from Google DeepMind published a paper proposing a 270 million parameter transformer model that was able to play chess at a grandmaster level. The model challenges traditional approaches to chess that relied on massive game datasets and complex heuristics —> Read more.

Self-Discover

Google DeepMind published a paper introducing Self-Discover, a framework to tackle complex reasoning problems with LLMs. The framework includes reasoning modules such as critical and step-by-step thinking as well as the building blocks to compose those modules into sophisticated reasoning chains —> Read more.

AI Controller Interface

Microsoft Research released a prototype of AI Controller Interface (AICI), a framework to implement controllers that constraint the outputs of LLMs. AICI’s architecture allows the implementation of custom logic blocks the during the token decoding process and still maintaining the state of the LLM —> Read more.

🤖 Cool AI Tech Releases

Smaug-72B

Abacus AI released Smaug-72B which sits at the top of the open LLM leaderboard —> Read more.

Gemini Advanced

Google rebranded Bard as Gemini and introduced Gemini Advanced with native integration for Google Docs and Gmail —> Read more.

TensorFlow GNN

Google released TensorFlow GNN, a new framework for graph neural networks in TensorFlow —> Read more.

Imagen 2

Google released Imagen 2, its powerful text-to-image model, across several of its AI products —> Read more.

SVD 1.1

Stability AI announced the release of SVD 1.1, a new version of its video generation model optimized for consistency —> Read more.

📡AI Radar

TheSequence is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.