Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

This is a microsoft/codebert-base-mlm model, trained for 1,000,000 steps (with batch_size=32) on Java code from the codeparrot/github-code-clean dataset, on the masked-language-modeling task.

It is intended to be used in CodeBERTScore: https://github.com/neulab/code-bert-score, but can be used for any other model or task.

For more information, see: https://github.com/neulab/code-bert-score

Citation

If you use this model for research, please cite:

@article{zhou2023codebertscore,
  url = {https://arxiv.org/abs/2302.05527},
  author = {Zhou, Shuyan and Alon, Uri and Agarwal, Sumit and Neubig, Graham},
  title = {CodeBERTScore: Evaluating Code Generation with Pretrained Models of Code},  
  publisher = {arXiv},
  year = {2023},
}
Downloads last month
35,133
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for neulab/codebert-java

Adapters
1 model