UC Berkeley
Open Weights

koala-13b

Released Apr 2023

Arena AI
#265
Context2K
Parameters13B

Koala-13B is a dialogue-oriented language model developed by researchers at the Berkeley AI Research (BAIR) lab. It is a fine-tuned version of Meta's LLaMA-13B, specifically designed to demonstrate that high-quality, curated dialogue data can enable smaller models to perform competitively with much larger proprietary systems. The project emphasized the importance of data quality over sheer volume in the development of conversational AI.

Training and Data

The model was trained using the EasyLM framework, which utilizes JAX and Flax for efficient scaling. Its training dataset consisted of a diverse range of high-quality dialogue sources, including user-shared conversations from ShareGPT, human-curated dialogues, and existing instruction-following collections such as Stanford's Alpaca, the Human ChatGPT Comparison Corpus (HC3), and the Open Instruction Generalist (OIG) dataset. This focus on curated interactions allowed the model to better capture the nuances of human-assistant dialogue.

Performance and Evaluation

In human evaluation studies conducted by the Berkeley team, Koala-13B was compared against other open-source and proprietary models. The results indicated that human evaluators frequently preferred Koala's responses over those of the original Alpaca model and found its performance to be competitive with ChatGPT in a significant number of conversational scenarios. The research highlighted that a 13-billion parameter model could achieve strong conversational proficiency when trained on a carefully selected mix of publicly available data.

Rankings & Comparison