SOLAR 10.7B Japanese
About
The SOLAR-10.7B Japanese language model boasts 10.7 billion parameters and is constructed on the Llama 2 architecture, integrating weights from Mistral 7B. It uses a technique called Depth Up-Scaling (DUS) to scale its depth in a straightforward manner compatible with standard frameworks. Tailored for single-turn conversations, the model excels at generating text and potentially code, rivaling even those with up to 30 billion parameters in certain benchmarks. Although its performance varies across tasks, and the details of its training data are limited, it remains a proficient tool in the realm of language models.
Capabilities
MultimodalFunction CallingTool UseJSON Mode
Providers(1)
| Provider | Input (per 1M) | Output (per 1M) | Type | |
|---|---|---|---|---|
| Upstage Console | — | — | Serverless |