LLM Reference
SEA-LION

SEA-LION

About

The SEA-LION family of large language models is an open-source initiative by AI Singapore aimed at enhancing the understanding of Southeast Asian languages and cultures in the global AI landscape. These models prioritize linguistic diversity by representing over 13 languages from the region, including major ones like English, Chinese, and Indonesian, as well as lesser-studied languages like Javanese and Sudanese. The latest version, SEA-LION v3, incorporates Google's Gemma 2 architecture and contains 9 billion parameters trained on an extensive dataset of 200 billion tokens. This effort is part of a collaboration with Sony Research to refine the model's accuracy, particularly for languages like Tamil, ensuring representation and equity in language processing technologies worldwide.

Models(2)

Details

ResearcherAI Singapore
Models2