← Previous · All Episodes · Next →
Hala Technical Report: Building Arabic-Centric Instruction & Translation Models at Scale Episode 1157

Hala Technical Report: Building Arabic-Centric Instruction & Translation Models at Scale

· 21:36

|

🤗 Upvotes: 67 | cs.CL, cs.AI, cs.LG

Authors:
Hasan Abed Al Kader Hammoud, Mohammad Zbeeb, Bernard Ghanem

Title:
Hala Technical Report: Building Arabic-Centric Instruction & Translation Models at Scale

Arxiv:
http://arxiv.org/abs/2509.14008v1

Abstract:
We present Hala, a family of Arabic-centric instruction and translation models built with our translate-and-tune pipeline. We first compress a strong AR$\leftrightarrow$EN teacher to FP8 (yielding $\sim$2$\times$ higher throughput with no quality loss) and use it to create high-fidelity bilingual supervision. A lightweight language model LFM2-1.2B is then fine-tuned on this data and used to translate high-quality English instruction sets into Arabic, producing a million-scale corpus tailored to instruction following. We train Hala models at 350M, 700M, 1.2B, and 9B parameters, and apply slerp merging to balance Arabic specialization with base-model strengths. On Arabic-centric benchmarks, Hala achieves state-of-the-art results within both the "nano" ($\leq$2B) and "small" (7-9B) categories, outperforming their bases. We release models, data, evaluation, and recipes to accelerate research in Arabic NLP.


Subscribe

Listen to Daily Paper Cast using one of many popular podcasting apps or directories.

Apple Podcasts Spotify Overcast Pocket Casts YouTube
← Previous · All Episodes · Next →