from transformers import pipeline import torch from transformers import AutoTokenizer, AutoModelForSequenceClassification, TextClassificationPipeline device = "cuda:0" if torch.cuda.is_available() else "cpu" model_name = 'qanastek/XLMRoberta-Alexa-Intents-Classification' tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForSequenceClassification.from_pretrained(model_name) intent_classifier = TextClassificationPipeline(model=model, tokenizer=tokenizer) def perform_intent_classification(text): result = intent_classifier(text) return {"Intent": [result]}