asahi417's picture
add model
87524b6
raw
history blame
6.12 kB
{
"_name_or_path": "deberta-v3-large-tweebank-ner",
"architectures": [
"DebertaV2ForTokenClassification"
],
"attention_probs_dropout_prob": 0.1,
"crf_state_dict": {
"_constraint_mask": [
[
1.0,
1.0,
1.0,
1.0,
1.0,
0.0,
0.0,
0.0,
1.0,
0.0,
1.0
],
[
1.0,
1.0,
1.0,
1.0,
0.0,
1.0,
0.0,
0.0,
1.0,
0.0,
1.0
],
[
1.0,
1.0,
1.0,
1.0,
0.0,
0.0,
1.0,
0.0,
1.0,
0.0,
1.0
],
[
1.0,
1.0,
1.0,
1.0,
0.0,
0.0,
0.0,
1.0,
1.0,
0.0,
1.0
],
[
1.0,
1.0,
1.0,
1.0,
1.0,
0.0,
0.0,
0.0,
1.0,
0.0,
1.0
],
[
1.0,
1.0,
1.0,
1.0,
0.0,
1.0,
0.0,
0.0,
1.0,
0.0,
1.0
],
[
1.0,
1.0,
1.0,
1.0,
0.0,
0.0,
1.0,
0.0,
1.0,
0.0,
1.0
],
[
1.0,
1.0,
1.0,
1.0,
0.0,
0.0,
0.0,
1.0,
1.0,
0.0,
1.0
],
[
1.0,
1.0,
1.0,
1.0,
0.0,
0.0,
0.0,
0.0,
1.0,
0.0,
1.0
],
[
1.0,
1.0,
1.0,
1.0,
0.0,
0.0,
0.0,
0.0,
1.0,
0.0,
0.0
],
[
0.0,
0.0,
0.0,
0.0,
0.0,
0.0,
0.0,
0.0,
0.0,
0.0,
0.0
]
],
"end_transitions": [
0.9945968389511108,
-0.7898595333099365,
-1.4528149366378784,
-0.39835214614868164,
-0.19360840320587158,
-0.30653536319732666,
1.1348109245300293,
-0.5007272362709045,
0.2630143165588379
],
"start_transitions": [
0.163371279835701,
1.0295491218566895,
1.9030133485794067,
-0.93556809425354,
-0.142340287566185,
0.7478364109992981,
-1.244533896446228,
0.7007433176040649,
1.0970624685287476
],
"transitions": [
[
0.5219072699546814,
-0.6073024868965149,
0.21660777926445007,
-0.08061438798904419,
0.34963756799697876,
-0.09881197661161423,
-0.12142079323530197,
-0.12750038504600525,
0.19532112777233124
],
[
-0.18758216500282288,
0.19614772498607635,
-0.21055644750595093,
-0.40785694122314453,
0.1777714490890503,
0.12013483047485352,
0.3959467113018036,
0.15063655376434326,
-0.29076510667800903
],
[
0.5534647703170776,
0.012826390564441681,
-0.16149963438510895,
-0.5842018127441406,
-0.0018042281735688448,
-0.43679139018058777,
-0.23927611112594604,
-0.10121553391218185,
-0.0426090732216835
],
[
0.25016123056411743,
0.04673261195421219,
-0.41687679290771484,
0.22481179237365723,
0.10355432331562042,
0.20035138726234436,
-0.3028927445411682,
-0.007212463766336441,
0.31153398752212524
],
[
0.5856021642684937,
-0.10948202013969421,
0.1351092904806137,
0.4515761733055115,
0.26148682832717896,
-0.11256065219640732,
-0.15908482670783997,
-0.4827544391155243,
-0.4094712436199188
],
[
0.10443474352359772,
-0.39016222953796387,
-0.06021846458315849,
-0.11585196107625961,
-0.11244403570890427,
-0.6651127934455872,
-0.07179304957389832,
0.05838318169116974,
-0.007226672023534775
],
[
-0.30400359630584717,
0.49990174174308777,
-0.3571702837944031,
-0.5655873417854309,
-0.5289591550827026,
-0.26642906665802,
0.28862693905830383,
-0.23844502866268158,
-0.19457215070724487
],
[
-0.37449777126312256,
-0.09788568317890167,
-0.47199949622154236,
-0.2298676073551178,
0.0554109625518322,
-0.08147075772285461,
-0.3137938976287842,
0.3010720908641815,
0.3137664198875427
],
[
-0.09058953821659088,
0.2583652436733246,
0.367694228887558,
0.45001059770584106,
-0.34188705682754517,
-0.0004857222083956003,
0.14611537754535675,
-0.3392927646636963,
0.4324687719345093
]
]
},
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 1024,
"id2label": {
"0": "B-LOC",
"1": "B-MISC",
"2": "B-ORG",
"3": "B-PER",
"4": "I-LOC",
"5": "I-MISC",
"6": "I-ORG",
"7": "I-PER",
"8": "O"
},
"initializer_range": 0.02,
"intermediate_size": 4096,
"label2id": {
"B-LOC": 0,
"B-MISC": 1,
"B-ORG": 2,
"B-PER": 3,
"I-LOC": 4,
"I-MISC": 5,
"I-ORG": 6,
"I-PER": 7,
"O": 8
},
"layer_norm_eps": 1e-07,
"max_position_embeddings": 512,
"max_relative_positions": -1,
"model_type": "deberta-v2",
"norm_rel_ebd": "layer_norm",
"num_attention_heads": 16,
"num_hidden_layers": 24,
"pad_token_id": 0,
"pooler_dropout": 0,
"pooler_hidden_act": "gelu",
"pooler_hidden_size": 1024,
"pos_att_type": [
"p2c",
"c2p"
],
"position_biased_input": false,
"position_buckets": 256,
"relative_attention": true,
"share_att_key": true,
"torch_dtype": "float32",
"transformers_version": "4.11.3",
"type_vocab_size": 0,
"vocab_size": 128100
}