LLM Reference

InternLM2 1.8B

About

The InternLM2-1.8B model is a leading-edge language model with 1.8 billion parameters, tailored for a broad range of natural language processing tasks. It excels in handling ultra-long contexts of up to 200,000 characters, significantly boosting its capacity to process extensive text inputs. The model features a transformer architecture optimized for both supervised fine-tuning and reinforcement learning from human feedback, enhancing alignment with user instructions. Trained on over 2 trillion tokens from diverse sources, InternLM2-1.8B demonstrates notable proficiency in text generation, reasoning, and coding tasks. However, users should be mindful of its potential for unpredictable outputs and limited depth in specialized domains.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Specifications

FamilyInternLM2
Released2024-01-12
Parameters1.8B
ArchitectureDecoder Only
Specializationgeneral