LLM Reference

GLM-4 9B

About

The GLM-4 9B is a robust open-source large language model crafted by Zhipu AI at Tsinghua University. Utilizing a transformer architecture, it is part of the GLM (General Language Model) series. This model is equipped with 9.4 billion parameters and has been benchmarked to outperform similar models, such as Llama-3-8B. Its bilingual training dataset (Chinese and English) enables multilingual capabilities. Noteworthy features include support for multi-round conversations, long context reasoning with the ability to handle up to a million tokens, multilingual support for 26 languages, and advanced functionalities like web browsing and code execution. However, it may exhibit biases from its training data and struggle with nuanced language understanding and common sense reasoning.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Specifications

FamilyGLM-4
Released2024-06-05
ArchitectureDecoder Only
Specializationgeneral