File size: 1,118 Bytes
86ea2e2
 
 
 
 
 
 
68fdde7
86ea2e2
 
2c61d75
 
 
 
fbe2b25
2c61d75
 
 
 
 
 
 
 
c1689f9
 
 
 
 
 
 
 
 
2c61d75
68fdde7
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
---
title: README
emoji: 🏃
colorFrom: gray
colorTo: purple
sdk: static
pinned: false
license: mit
---

# Model Description
BioTinyBERT is the result of training the [TinyBERT](https://huggingface.co/huawei-noah/TinyBERT_General_4L_312D) model in a continual learning fashion for 200k training steps using a total batch size of 192 on the PubMed dataset. 

# Initialisation
We initialise our model with the pre-trained checkpoints of the [TinyBERT](https://huggingface.co/huawei-noah/TinyBERT_General_4L_312D) model available on Huggingface.

# Architecture
This model uses 4 hidden layers with a hidden dimension size and an embedding size of 768 resulting in a total of 15M parameters. 

# Citation
If you use this model, please consider citing the following paper:

```bibtex
@article{rohanian2023effectiveness,
  title={On the effectiveness of compact biomedical transformers},
  author={Rohanian, Omid and Nouriborji, Mohammadmahdi and Kouchaki, Samaneh and Clifton, David A},
  journal={Bioinformatics},
  volume={39},
  number={3},
  pages={btad103},
  year={2023},
  publisher={Oxford University Press}
}
```