{ | |
"config": { | |
"alpha": 8, | |
"architecture": "lora", | |
"attn_matrices": [ | |
"q", | |
"v" | |
], | |
"composition_mode": "add", | |
"dropout": 0.0, | |
"init_weights": "lora", | |
"intermediate_lora": false, | |
"leave_out": [], | |
"output_lora": false, | |
"r": 16, | |
"selfattn_lora": true, | |
"use_gating": false | |
}, | |
"config_id": "141b248112091265", | |
"hidden_size": 768, | |
"model_class": "BertForTokenClassification", | |
"model_name": "indolem/indobert-base-uncased", | |
"model_type": "bert", | |
"name": "nerugm-lora", | |
"version": "0.2.0" | |
} |