Falcon 7B
About
Falcon-7B, developed by the Technology Innovation Institute, is a cutting-edge large language model boasting a decoder-only architecture with 7 billion parameters. It's trained on 1,500 billion tokens from the curated web dataset, RefinedWeb, enhancing its performance in language tasks. The model is equipped with advanced features like FlashAttention and multiquery attention, optimizing speed and memory usage. With 32 layers and rotary positional embeddings, it manages a sequence length of up to 2048 tokens efficiently. Renowned for tasks such as text generation, summarization, translation, and conversational AI, Falcon-7B is open-source under Apache 2.0, suitable even for consumer hardware, needing at least 16GB of memory for inference 236.
Capabilities
Providers(4)
| Provider | Input (per 1M) | Output (per 1M) | Type | |
|---|---|---|---|---|
| Azure OpenAI | — | — | Provisioned | |
| GCP Vertex AI | — | — | Serverless | |
| Cloudflare Workers AI | — | — | Serverless | |
| Alibaba Cloud PAI-EAS | — | — | Serverless |