by nreimers
Open source · 21k downloads · 3 likes
BERT Tiny L 2 H 128 A 2 is a compact and optimized version of the BERT model, designed for natural language processing tasks with limited resources. With just two layers, 128 hidden units, and two attention heads, it strikes a balance between performance and efficiency, making it ideal for applications requiring low memory or computational consumption. This model excels in text comprehension, classification, question answering, and semantic analysis while remaining accessible on less powerful devices. Its lightweight nature makes it particularly well-suited for embedded environments or projects where fast inference is critical. Despite its reduced size, it retains some of BERT’s contextual modeling capabilities, delivering reliable results for moderate precision needs.
This is the BERT-Medium model from Google: https://github.com/google-research/bert#bert. A BERT model with 2 layers, 128 hidden unit size, and 2 attention heads.