{ "architectures": [ "ElectraForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "embedding_size": 128, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 256, "id2label": { "0": "O", "1": "PER-B", "10": "ID-I", "11": "PWD-B", "12": "PWD-I", "13": "ORG-B", "14": "ORG-I", "15": "PHN-B", "16": "PHN-I", "17": "CRD-B", "18": "CRD-I", "19": "ACC-B", "2": "PER-I", "20": "ACC-I", "21": "PSP-B", "22": "PSP-I", "23": "DLN-B", "24": "DLN-I", "3": "LOC-B", "4": "LOC-I", "5": "RRN-B", "6": "RRN-I", "7": "EMA-B", "8": "EMA-I", "9": "ID-B" }, "initializer_range": 0.02, "intermediate_size": 1024, "label2id": { "ACC-B": 19, "ACC-I": 20, "CRD-B": 17, "CRD-I": 18, "DLN-B": 23, "DLN-I": 24, "EMA-B": 7, "EMA-I": 8, "ID-B": 9, "ID-I": 10, "LOC-B": 3, "LOC-I": 4, "O": 0, "ORG-B": 13, "ORG-I": 14, "PER-B": 1, "PER-I": 2, "PHN-B": 15, "PHN-I": 16, "PSP-B": 21, "PSP-I": 22, "PWD-B": 11, "PWD-I": 12, "RRN-B": 5, "RRN-I": 6 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "electra", "num_attention_heads": 4, "num_hidden_layers": 12, "pad_token_id": 0, "summary_activation": "gelu", "summary_last_dropout": 0.1, "summary_type": "first", "summary_use_proj": true, "type_vocab_size": 2, "vocab_size": 35000 }