| --- |
| language: |
| - en |
| tags: |
| - biomedical |
| - bionlp |
| - entity linking |
| - embedding |
| - bert |
| --- |
| The GEBERT model pre-trained with GAT graph encoder. |
|
|
| The model was published at [CLEF 2023 conference](https://clef2023.clef-initiative.eu/). The source code is available at [github](https://github.com/Andoree/GEBERT). |
|
|
|
|
| Pretraining data: biomedical concept graph and concept names from the UMLS (2020AB release). |
|
|
| Base model: [microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext](https://huggingface.co/microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext). |
|
|
|
|
|
|
| ```bibtex |
| @inproceedings{sakhovskiy2023gebert, |
| author="Sakhovskiy, Andrey |
| and Semenova, Natalia |
| and Kadurin, Artur |
| and Tutubalina, Elena", |
| title="Graph-Enriched Biomedical Entity Representation Transformer", |
| booktitle="Experimental IR Meets Multilinguality, Multimodality, and Interaction", |
| year="2023", |
| publisher="Springer Nature Switzerland", |
| address="Cham", |
| pages="109--120", |
| isbn="978-3-031-42448-9" |
| } |
| ``` |