{ "_name_or_path": "OpenGVLab/pvt_v2_b0", "architectures": [ "PvtV2ForImageClassification" ], "attention_probs_dropout_prob": 0.0, "depths": [ 2, 2, 2, 2 ], "drop_path_rate": 0.0, "hidden_act": "gelu", "hidden_dropout_prob": 0.0, "hidden_sizes": [ 32, 64, 160, 256 ], "id2label": { "0": "avanzada", "1": "avanzada humeda", "2": "leve", "3": "moderada", "4": "no dmae" }, "image_size": { "height": 224, "width": 224 }, "initializer_range": 0.02, "label2id": { "avanzada": 0, "avanzada humeda": 1, "leve": 2, "moderada": 3, "no dmae": 4 }, "layer_norm_eps": 1e-06, "linear_attention": false, "mlp_ratios": [ 8, 8, 4, 4 ], "model_type": "pvt_v2", "num_attention_heads": [ 1, 2, 5, 8 ], "num_channels": 3, "num_encoder_blocks": 4, "out_features": [ "stage1", "stage2", "stage3", "stage4" ], "out_indices": [ 0, 1, 2, 3 ], "patch_sizes": [ 7, 3, 3, 3 ], "problem_type": "single_label_classification", "qkv_bias": true, "reshape_last_stage": true, "sr_ratios": [ 8, 4, 2, 1 ], "stage_names": [ "stage1", "stage2", "stage3", "stage4" ], "strides": [ 4, 2, 2, 2 ], "torch_dtype": "float32", "transformers_version": "4.40.2" }