Edit model card

Model description

LegalBert is a BERT-base-cased model fine-tuned on a subset of the case.law corpus. Further details can be found in this paper:

A Dataset for Statutory Reasoning in Tax Law Entailment and Question Answering
Nils Holzenberger, Andrew Blair-Stanek and Benjamin Van Durme
Proceedings of the 2020 Natural Legal Language Processing (NLLP) Workshop, 24 August 2020

Usage

from transformers import AutoModel, AutoTokenizer
model = AutoModel.from_pretrained("jhu-clsp/LegalBert")
tokenizer = AutoTokenizer.from_pretrained("jhu-clsp/LegalBert")

Citation

@inproceedings{holzenberger20dataset,
  author    = {Nils Holzenberger and
               Andrew Blair{-}Stanek and
               Benjamin Van Durme},
  title     = {A Dataset for Statutory Reasoning in Tax Law Entailment and Question
               Answering},
  booktitle = {Proceedings of the Natural Legal Language Processing Workshop 2020
               co-located with the 26th {ACM} {SIGKDD} International Conference on
               Knowledge Discovery {\&} Data Mining {(KDD} 2020), Virtual Workshop,
               August 24, 2020},
  series    = {{CEUR} Workshop Proceedings},
  volume    = {2645},
  pages     = {31--38},
  publisher = {CEUR-WS.org},
  year      = {2020},
  url       = {http://ceur-ws.org/Vol-2645/paper5.pdf},
}
Downloads last month
14
Inference API
Unable to determine this model’s pipeline type. Check the docs .