| { | |
| "_name_or_path": "qanastek/51-languages-classifier", | |
| "architectures": [ | |
| "XLMRobertaForSequenceClassification" | |
| ], | |
| "attention_probs_dropout_prob": 0.1, | |
| "bos_token_id": 0, | |
| "classifier_dropout": null, | |
| "eos_token_id": 2, | |
| "hidden_act": "gelu", | |
| "hidden_dropout_prob": 0.1, | |
| "hidden_size": 768, | |
| "id2label": { | |
| "0": "af-ZA", | |
| "1": "am-ET", | |
| "2": "ar-SA", | |
| "3": "az-AZ", | |
| "4": "bn-BD", | |
| "5": "cy-GB", | |
| "6": "da-DK", | |
| "7": "de-DE", | |
| "8": "el-GR", | |
| "9": "en-US", | |
| "10": "es-ES", | |
| "11": "fa-IR", | |
| "12": "fi-FI", | |
| "13": "fr-FR", | |
| "14": "he-IL", | |
| "15": "hi-IN", | |
| "16": "hu-HU", | |
| "17": "hy-AM", | |
| "18": "id-ID", | |
| "19": "is-IS", | |
| "20": "it-IT", | |
| "21": "ja-JP", | |
| "22": "jv-ID", | |
| "23": "ka-GE", | |
| "24": "km-KH", | |
| "25": "kn-IN", | |
| "26": "ko-KR", | |
| "27": "lv-LV", | |
| "28": "ml-IN", | |
| "29": "mn-MN", | |
| "30": "ms-MY", | |
| "31": "my-MM", | |
| "32": "nb-NO", | |
| "33": "nl-NL", | |
| "34": "pl-PL", | |
| "35": "pt-PT", | |
| "36": "ro-RO", | |
| "37": "ru-RU", | |
| "38": "sl-SL", | |
| "39": "sq-AL", | |
| "40": "sv-SE", | |
| "41": "sw-KE", | |
| "42": "ta-IN", | |
| "43": "te-IN", | |
| "44": "th-TH", | |
| "45": "tl-PH", | |
| "46": "tr-TR", | |
| "47": "ur-PK", | |
| "48": "vi-VN", | |
| "49": "zh-CN", | |
| "50": "zh-TW" | |
| }, | |
| "initializer_range": 0.02, | |
| "intermediate_size": 3072, | |
| "label2id": { | |
| "af-ZA": 0, | |
| "am-ET": 1, | |
| "ar-SA": 2, | |
| "az-AZ": 3, | |
| "bn-BD": 4, | |
| "cy-GB": 5, | |
| "da-DK": 6, | |
| "de-DE": 7, | |
| "el-GR": 8, | |
| "en-US": 9, | |
| "es-ES": 10, | |
| "fa-IR": 11, | |
| "fi-FI": 12, | |
| "fr-FR": 13, | |
| "he-IL": 14, | |
| "hi-IN": 15, | |
| "hu-HU": 16, | |
| "hy-AM": 17, | |
| "id-ID": 18, | |
| "is-IS": 19, | |
| "it-IT": 20, | |
| "ja-JP": 21, | |
| "jv-ID": 22, | |
| "ka-GE": 23, | |
| "km-KH": 24, | |
| "kn-IN": 25, | |
| "ko-KR": 26, | |
| "lv-LV": 27, | |
| "ml-IN": 28, | |
| "mn-MN": 29, | |
| "ms-MY": 30, | |
| "my-MM": 31, | |
| "nb-NO": 32, | |
| "nl-NL": 33, | |
| "pl-PL": 34, | |
| "pt-PT": 35, | |
| "ro-RO": 36, | |
| "ru-RU": 37, | |
| "sl-SL": 38, | |
| "sq-AL": 39, | |
| "sv-SE": 40, | |
| "sw-KE": 41, | |
| "ta-IN": 42, | |
| "te-IN": 43, | |
| "th-TH": 44, | |
| "tl-PH": 45, | |
| "tr-TR": 46, | |
| "ur-PK": 47, | |
| "vi-VN": 48, | |
| "zh-CN": 49, | |
| "zh-TW": 50 | |
| }, | |
| "layer_norm_eps": 1e-05, | |
| "max_position_embeddings": 514, | |
| "model_type": "xlm-roberta", | |
| "num_attention_heads": 12, | |
| "num_hidden_layers": 12, | |
| "output_past": true, | |
| "pad_token_id": 1, | |
| "position_embedding_type": "absolute", | |
| "problem_type": "single_label_classification", | |
| "torch_dtype": "float32", | |
| "transformers_version": "4.33.3", | |
| "type_vocab_size": 1, | |
| "use_cache": true, | |
| "vocab_size": 250002 | |
| } | |