LLM Reference

BERT Base

About

BERT Base, developed by Google AI Language, is a prominent large language AI model that utilizes a bidirectional transformer architecture to enhance natural language processing tasks. It processes the entire input sequence simultaneously, offering a comprehensive understanding of context, which significantly improves the model's accuracy compared to previous unidirectional models. With approximately 110 million parameters and a hidden size of 768 across 12 layers, BERT Base is equipped to handle various NLP tasks such as named entity recognition, sentiment analysis, and question answering. While it achieves impressive results, the model demands considerable computational resources for training and fine-tuning, and it can reflect biases from its training data, indicating the importance of careful data management.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Specifications

FamilyBERT
Parameters110M
ArchitectureDecoder Only
Specializationgeneral