This is the official pre-trained model introduced in DNA language model GROVER learns sequence context in the human genome
from transformers import AutoTokenizer, AutoModelForMaskedLM
# Import the tokenizer and the model
tokenizer = AutoTokenizer.from_pretrained("PoetschLab/GROVER")
model = AutoModelForMaskedLM.from_pretrained("PoetschLab/GROVER")
Some preliminary analysis shows that sequence re-tokenization using Byte Pair Encoding (BPE) changes significantly if the sequence is less than 50 nucleotides long. Longer than 50 nucleotides, you should still be careful with sequence edges. We advice to add 100 nucleotides at the beginning and end of every sequence in order to guarantee that your sequence is represented with the same tokens as the original tokenization. We also provide the tokenized chromosomes with their respective nucleotide mappers (They are available in the folder tokenized chromosomes).
BibTeX entry and citation info
@article{sanabria2024dna,
title={DNA language model GROVER learns sequence context in the human genome},
author={Sanabria, Melissa and Hirsch, Jonas and Joubert, Pierre M and Poetsch, Anna R},
journal={Nature Machine Intelligence},
pages={1--13},
year={2024},
publisher={Nature Publishing Group UK London}
}
- Downloads last month
- 998
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.