Arabic-ALBERT
August 26, 2020
Ali Safaya
Koç University, CS – PhD Student
Transformers, ALBERT
Thanks to Google for providing free TPU for the training process and for Huggingface for hosting these models on their servers
The models were pretrained on ~4.4 Billion words:
Notes on training data:
albert-base | albert-large | albert-xlarge | |
---|---|---|---|
Hidden Layers | 12 | 24 | 24 |
Attention heads | 12 | 16 | 32 |
Hidden size | 768 | 1024 | 2048 |
Note: More results on other downstream NLP tasks will be added soon. if you use any of these models, we would appreciate your feedback.
You can use these models by installing torch
or tensorflow
and Huggingface library transformers
. And you can use it directly by initializing it like this: