Datasets:
model large_stringclasses 17
values | language large_stringlengths 8 8 |
|---|---|
gherbal-v2 | abv_Arab |
gherbal-v2 | acx_Arab |
gherbal-v2 | aeb_Arab |
gherbal-v2 | afb_Arab |
gherbal-v2 | apc_Arab |
gherbal-v2 | apd_Arab |
gherbal-v2 | arb_Arab |
gherbal-v2 | arq_Arab |
gherbal-v2 | ary_Arab |
gherbal-v2 | ary_Latn |
gherbal-v2 | arz_Arab |
gherbal-v2 | ayl_Arab |
gherbal-v2 | ayn_Arab |
gherbal-v2 | azj_Latn |
gherbal-v2 | bul_Cyrl |
gherbal-v2 | cmn_Hans |
gherbal-v2 | deu_Latn |
gherbal-v2 | eng_Latn |
gherbal-v2 | fra_Latn |
gherbal-v2 | hau_Latn |
gherbal-v2 | heb_Hebr |
gherbal-v2 | ibo_Latn |
gherbal-v2 | ind_Latn |
gherbal-v2 | ita_Latn |
gherbal-v2 | jpn_Jpan |
gherbal-v2 | kor_Hang |
gherbal-v2 | mey_Arab |
gherbal-v2 | nld_Latn |
gherbal-v2 | nya_Latn |
gherbal-v2 | pes_Arab |
gherbal-v2 | pol_Latn |
gherbal-v2 | por_Latn |
gherbal-v2 | ron_Latn |
gherbal-v2 | rus_Cyrl |
gherbal-v2 | som_Latn |
gherbal-v2 | spa_Latn |
gherbal-v2 | swh_Latn |
gherbal-v2 | tha_Thai |
gherbal-v2 | tur_Latn |
gherbal-v2 | ukr_Cyrl |
gherbal-v2 | urd_Arab |
gherbal-v2 | vie_Latn |
gherbal-v2 | wol_Latn |
gherbal-v2 | xho_Latn |
gherbal-v2 | yor_Latn |
gherbal-v2 | zsm_Latn |
gherbal-v3 | abv_Arab |
gherbal-v3 | ace_Arab |
gherbal-v3 | ace_Latn |
gherbal-v3 | acm_Arab |
gherbal-v3 | acq_Arab |
gherbal-v3 | acx_Arab |
gherbal-v3 | aeb_Arab |
gherbal-v3 | afb_Arab |
gherbal-v3 | apc_Arab |
gherbal-v3 | apd_Arab |
gherbal-v3 | arb_Arab |
gherbal-v3 | arq_Arab |
gherbal-v3 | ars_Arab |
gherbal-v3 | ary_Arab |
gherbal-v3 | ary_Latn |
gherbal-v3 | arz_Arab |
gherbal-v3 | ayl_Arab |
gherbal-v3 | ayn_Arab |
gherbal-v3 | azj_Latn |
gherbal-v3 | bam_Latn |
gherbal-v3 | ban_Latn |
gherbal-v3 | ben_Beng |
gherbal-v3 | bho_Deva |
gherbal-v3 | bjn_Arab |
gherbal-v3 | bjn_Latn |
gherbal-v3 | bug_Latn |
gherbal-v3 | bul_Cyrl |
gherbal-v3 | ceb_Latn |
gherbal-v3 | ckb_Arab |
gherbal-v3 | cmn_Hans |
gherbal-v3 | crh_Latn |
gherbal-v3 | deu_Latn |
gherbal-v3 | dik_Latn |
gherbal-v3 | dzo_Tibt |
gherbal-v3 | eng_Latn |
gherbal-v3 | fra_Latn |
gherbal-v3 | fur_Latn |
gherbal-v3 | fuv_Latn |
gherbal-v3 | gug_Latn |
gherbal-v3 | hat_Latn |
gherbal-v3 | hau_Latn |
gherbal-v3 | heb_Hebr |
gherbal-v3 | hin_Deva |
gherbal-v3 | hne_Deva |
gherbal-v3 | ibo_Latn |
gherbal-v3 | ind_Latn |
gherbal-v3 | ita_Latn |
gherbal-v3 | jav_Latn |
gherbal-v3 | jpn_Jpan |
gherbal-v3 | kas_Arab |
gherbal-v3 | kas_Deva |
gherbal-v3 | kir_Cyrl |
gherbal-v3 | knc_Arab |
gherbal-v3 | knc_Latn |
LID Benchmark
Comprehensive evaluation of 17 language identification models across 8 diverse benchmarks.
Built by Omneity Labs.
Subsets
| Config | Description | Rows |
|---|---|---|
results_summary |
One row per model × benchmark × scope with aggregate metrics | ~136 |
results_aggregate |
Detailed aggregate metrics per model × benchmark × scope | ~816 |
results_per_language |
Per-language accuracy for every model × benchmark × scope | ~57k |
results_speed |
Inference speed (samples/sec) per model × benchmark | ~136 |
model_languages |
Supported language codes declared by each model | ~4.7k |
results_individual |
Every individual prediction (model × benchmark × sample) | ~28M |
Models
gherbal-v1, gherbal-v2, gherbal-v3, gherbal-v4, nllb-lid, openlid-v1, openlid-v2, hplt-openlid-v3, fastlid-176, glotlid, franc, franc-all, franc-min, cld2, langdetect, langid, py3langid
Benchmarks
| Benchmark | Source |
|---|---|
| flores-devtest | openlanguagedata/flores_plus (devtest split) |
| flores-dev | openlanguagedata/flores_plus (dev split) |
| madar | Madar |
| gherbal-multi | sawalni-ai/gherbal-multi |
| atlasia-lid | atlasia/Arabic-LID-Leaderboard |
| wili-2018 | wili_2018 |
| commonlid | commoncrawl/CommonLID |
| bouquet | facebook/bouquet |
Methodology
All predictions are normalized to ISO 639-3 + Script (ISO 15924) codes using babelcode. Metrics: accuracy, macro-F1, weighted-F1, precision, recall — computed under multiple scopes (full, self, v1–v4).
Interactive App
Explore results interactively: LID Benchmark Leaderboard
Citation
If you use this benchmark data in your research, please reference:
- Omneity Labs LID Benchmark: https://huggingface.co/datasets/omneity-labs/lid-benchmark
- Gherbal model: https://www.omneitylabs.com/models/gherbal
- Evaluation benchmarks: See individual benchmark datasets linked above.
Author
License
The evaluation results in this dataset are released under Apache 2.0. The underlying benchmark datasets retain their original licenses.
- Downloads last month
- 148