--- license: mit pipeline_tag: text-classification --- ## Roberta for Justification analyst This model is a fine-tuned version of the Roberta architecture that has been trained specifically for sequence classification. The fine-tuning process involved using the PyTorch deep learning framework and specific hyperparameters (2-4e, 1-8 epsilon) with Adagrad optimizer. --- ## Example Usage To use the model, first load it in PyTorch: ```python import torch from transformers import RobertaForSequenceClassification, RobertaTokenizer # Load the fine-tuned model model = RobertaForSequenceClassification.from_pretrained('Dzeniks/justification-analyst') # Load the tokenizer tokenizer = RobertaTokenizer.from_pretrained('Dzeniks/justification-analyst') # Tokenize the input sequence input_text = "This is a sample input sequence" input = tokenizer.encode_plus(claim, evidence, return_tensors="pt") # Use the model to make a prediction model.eval() with torch.no_grad(): prediction = model(**x) predictions = torch.argmax(outputs[0], dim=1).item() ``` ## Classification Labels The model was trained on a dataset consisting of claims and evidence, where the goal was to classify each claim as either supporting, refuting, or not having enough information to make a decision. The labels used for this task are as follows: - Label 0: Supports - Label 1: Refutes - Label 2: Not enough information