| { |
| "_name_or_path": "/home/jinjukim/Phonemes_Bridge_Dialects/output/UD_North_Sami-Giella/checkpoint-180", |
| "adapters": { |
| "adapters": {}, |
| "config_map": {}, |
| "fusion_config_map": {}, |
| "fusion_name_map": {}, |
| "fusions": {} |
| }, |
| "architectures": [ |
| "BertAdapterModel" |
| ], |
| "attention_probs_dropout_prob": 0.1, |
| "classifier_dropout": null, |
| "directionality": "bidi", |
| "hidden_act": "gelu", |
| "hidden_dropout_prob": 0.1, |
| "hidden_size": 768, |
| "id2label": { |
| "0": "_", |
| "1": "acl", |
| "2": "advcl", |
| "3": "advmod", |
| "4": "amod", |
| "5": "appos", |
| "6": "aux", |
| "7": "case", |
| "8": "cc", |
| "9": "ccomp", |
| "10": "clf", |
| "11": "compound", |
| "12": "conj", |
| "13": "cop", |
| "14": "csubj", |
| "15": "dep", |
| "16": "det", |
| "17": "discourse", |
| "18": "dislocated", |
| "19": "expl", |
| "20": "fixed", |
| "21": "flat", |
| "22": "goeswith", |
| "23": "iobj", |
| "24": "list", |
| "25": "mark", |
| "26": "nmod", |
| "27": "nsubj", |
| "28": "nummod", |
| "29": "obj", |
| "30": "obl", |
| "31": "orphan", |
| "32": "parataxis", |
| "33": "punct", |
| "34": "reparandum", |
| "35": "root", |
| "36": "vocative", |
| "37": "xcomp", |
| "38": "dup", |
| "39": "mwe", |
| "40": "name", |
| "41": "remnant" |
| }, |
| "initializer_range": 0.02, |
| "intermediate_size": 3072, |
| "label2id": { |
| "_": 0, |
| "acl": 1, |
| "advcl": 2, |
| "advmod": 3, |
| "amod": 4, |
| "appos": 5, |
| "aux": 6, |
| "case": 7, |
| "cc": 8, |
| "ccomp": 9, |
| "clf": 10, |
| "compound": 11, |
| "conj": 12, |
| "cop": 13, |
| "csubj": 14, |
| "dep": 15, |
| "det": 16, |
| "discourse": 17, |
| "dislocated": 18, |
| "dup": 38, |
| "expl": 19, |
| "fixed": 20, |
| "flat": 21, |
| "goeswith": 22, |
| "iobj": 23, |
| "list": 24, |
| "mark": 25, |
| "mwe": 39, |
| "name": 40, |
| "nmod": 26, |
| "nsubj": 27, |
| "nummod": 28, |
| "obj": 29, |
| "obl": 30, |
| "orphan": 31, |
| "parataxis": 32, |
| "punct": 33, |
| "remnant": 41, |
| "reparandum": 34, |
| "root": 35, |
| "vocative": 36, |
| "xcomp": 37 |
| }, |
| "layer_norm_eps": 1e-12, |
| "max_position_embeddings": 512, |
| "model_type": "bert", |
| "num_attention_heads": 12, |
| "num_hidden_layers": 12, |
| "pad_token_id": -1, |
| "pooler_fc_size": 768, |
| "pooler_num_attention_heads": 12, |
| "pooler_num_fc_layers": 3, |
| "pooler_size_per_head": 128, |
| "pooler_type": "first_token_transform", |
| "position_embedding_type": "absolute", |
| "prediction_heads": { |
| "default": { |
| "activation_function": "gelu", |
| "bias": true, |
| "embedding_size": 768, |
| "head_type": "masked_lm", |
| "label2id": null, |
| "layer_norm": true, |
| "layers": 2, |
| "shift_labels": false, |
| "vocab_size": 119547 |
| }, |
| "ud_UD_North_Sami-Giella": { |
| "head_type": "dependency_parsing", |
| "label2id": { |
| "_": 0, |
| "acl": 1, |
| "advcl": 2, |
| "advmod": 3, |
| "amod": 4, |
| "appos": 5, |
| "aux": 6, |
| "case": 7, |
| "cc": 8, |
| "ccomp": 9, |
| "clf": 10, |
| "compound": 11, |
| "conj": 12, |
| "cop": 13, |
| "csubj": 14, |
| "dep": 15, |
| "det": 16, |
| "discourse": 17, |
| "dislocated": 18, |
| "dup": 38, |
| "expl": 19, |
| "fixed": 20, |
| "flat": 21, |
| "goeswith": 22, |
| "iobj": 23, |
| "list": 24, |
| "mark": 25, |
| "mwe": 39, |
| "name": 40, |
| "nmod": 26, |
| "nsubj": 27, |
| "nummod": 28, |
| "obj": 29, |
| "obl": 30, |
| "orphan": 31, |
| "parataxis": 32, |
| "punct": 33, |
| "remnant": 41, |
| "reparandum": 34, |
| "root": 35, |
| "vocative": 36, |
| "xcomp": 37 |
| }, |
| "num_labels": 42 |
| } |
| }, |
| "torch_dtype": "float32", |
| "transformers_version": "4.47.1", |
| "type_vocab_size": 2, |
| "use_cache": true, |
| "vocab_size": 119547 |
| } |
|
|