Entity-Aware Machine Translation Leaderboard
Overview
This leaderboard showcases the performance of various systems on the EA-MT shared task, which has been organized as part of the SemEval 2025 workshop.
- The results are still provisional and subject to change.
Task Description
The task is to translate a given input sentence from the source language (English) to the target language, where the input sentence contains named entities that may be challenging for machine translation systems to handle. The named entities may be entities that are rare, ambiguous, or unknown to the machine translation system. The task is to develop machine translation systems that can accurately translate such named entities in the input sentence to the target language.
- Learn more about the task on the EA-MT shared task page.
Scoring
The leaderboard is based on three main scores:
- M-ETA Score: A score that evaluates the translation quality of named entities in the input sentence.
- COMET Score: A score that evaluates the translation quality at the sentence level.
- Overall Score: The harmonic mean of the M-ETA and COMET scores.
Legend
- ๐ : Uses gold data, i.e., the gold Wikidata ID or information derived from it, at test time.
- ๐: Uses RAG (Retrieval-Augmented Generation) for named entity translation.
- ๐ค: Uses an LLM (Large Language Model) for named entity translation.
- ๐: The system (LLM and/or MT model) is finetuned on additional data.
Filters and Controls
Use the dropdowns and checkboxes to filter the leaderboard scores.
Leaderboard Scores
You can view the leaderboard scores for each system based on the following metrics:
- M-ETA Score: A score that evaluates the translation quality of named entities in the input sentence.
- COMET Score: A score that evaluates the translation quality at the sentence level.
- Overall Score: The harmonic mean of the M-ETA and COMET scores. Switch between the tabs to view the scores for each metric.
Note: You can sort the leaderboard by clicking on the column headers. For example, click on the "it_IT" column to sort by the Italian language scores.
Overall Score Leaderboard
Rank | Team | System | Uses Gold | Uses RAG | Uses LLM | LLM Name | Finetuned | ar_AE | de_DE | es_ES | fr_FR | it_IT | ja_JP | ko_KR | th_TH | tr_TR | zh_TW | overall |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
10 | The Five Forbidden Entities | LoRA-nllb-distilled-200-distilled-600M | ๐ | ๐ | ๐ค | Llama-3.3-70B-Instruct + DeepSeek-R1 | ๐ | 92.68 | 90.03 | 92.54 | 92.92 | 94.39 | 93.34 | 92.77 | 92.35 | 89.54 | 87.36 | 91.79 |
M-ETA Score Leaderboard
Rank | Team | System | Uses Gold | Uses RAG | Uses LLM | LLM Name | Finetuned | ar_AE | de_DE | es_ES | fr_FR | it_IT | ja_JP | ko_KR | th_TH | tr_TR | zh_TW | overall |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
11 | The Five Forbidden Entities | LoRA-nllb-distilled-200-distilled-600M | ๐ | ๐ | ๐ค | Llama-3.3-70B-Instruct + DeepSeek-R1 | ๐ | 88.78 | 82.59 | 88.42 | 86.59 | 89.88 | 87.74 | 88.17 | 86.51 | 79.28 | 80.64 | 85.86 |
Rank | Team | System | Uses Gold | Uses RAG | Uses LLM | LLM Name | Finetuned | ar_AE | de_DE | es_ES | fr_FR | it_IT | ja_JP | ko_KR | th_TH | tr_TR | zh_TW | overall |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
11 | YNU-HPCC | LLaMA + MT | ๐ | ๐ค | Llama-3.3-70B-Instruct | 88.78 | 82.59 | 88.42 | 86.59 | 89.88 | 87.74 | 88.17 | 86.51 | 79.28 | 80.64 | 85.86 | ||
20 | Lunar | LLaMA-RAFT-Plus | ๐ | ๐ค | Llama-3.1-8B-Instruct | ๐ | 67.54 | 59.55 | 66.58 | 67.15 | 73.58 | 55.66 | 61.49 | 67.22 | 71.5 | 38.79 | 62.9 | |
34 | sakura | Rakuten7b-PO10 | ๐ค | Rakuten/RakutenAI-7B-chat | ๐ | 29.5 | ||||||||||||
34 | SHEF | Llama-Wiki-DeepSeek | ๐ค | Llama-3.3-70B-Instruct + DeepSeek-R1 | ๐ | 85.57 | 90.5 | 90.05 | 93.02 | |||||||||
11 | YNU-HPCC | Qwen2.5-32B | ๐ | ๐ค | Qwen2.5-32B | ๐ | 88.78 | 82.59 | 88.42 | 86.59 | 89.88 | 87.74 | 88.17 | 86.51 | 79.28 | 80.64 | 85.86 | |
8 | UAlberta | WikiGPT4o | ๐ | ๐ | ๐ค | GPT-4o | 91.66 | 85.06 | 89.34 | 89.68 | 91.74 | 90.33 | 90.34 | 89.96 | 81.65 | 81.22 | 88.1 | |
18 | SALT ๐ง | Salt-MT-Pipeline | ๐ | N/A | ๐ | 81.72 | 73.77 | 74.58 | 74.77 | 77.62 | 72.2 | 74.24 | 65.59 | 76.86 | 45.27 | 71.66 | ||
34 | Transcreate | Chatgpt-4o-mini-llm | ๐ค | gpt-4o-mini-2024-07-18 | 29.14 | |||||||||||||
34 | silp_nlp | NER-M2M100 | ๐ | ๐ค | M2M100 | ๐ | 7.18 | |||||||||||
31 | Muhandro_HSE | NER-LLM | ๐ค | Llama-3.1-8B, Llama-3.1-70B | ๐ | 32.35 | 37.85 | 41.21 | 35.77 | 33.03 | 23.71 | 20.98 | 9.78 | 39.86 | 6.23 | 28.08 | ||
34 | CUET_DeepLearners | Spacy-NLLB | N/A | |||||||||||||||
34 | Transcreate | Claude-haiku-llm | ๐ค | claude-3-5-haiku-20241022 | 28.49 | |||||||||||||
9 | RAGthoven | GPT-4o + Wikidata | ๐ | ๐ค | GPT-4o | 91.88 | 85.07 | 89.83 | 91.02 | 92.64 | 87.57 | 88.08 | 89.21 | 79.61 | 81.46 | 87.64 | ||
34 | Transcreate | Gemini-pro-llm | ๐ค | gemini-1.5-pro | 48.33 | |||||||||||||
21 | YNU-HPCC | Qwen2.5 + M2M | ๐ค | Qwen2.5-32B | 64.09 | 60.16 | 67.27 | 65.31 | 65.4 | 64.19 | 59.92 | 56.37 | 57.84 | 59.17 | 61.97 | |||
19 | FII-UAIC-SAI | Qwen2.5-Wiki-MT | ๐ค | Qwen2.5-72B-Instruct-AWQ | 66.42 | 66.98 | 72.35 | 72.46 | 75.79 | 67.03 | 66.02 | 65.25 | 67.56 | 62.5 | 68.24 | |||
33 | silp_nlp | GPT-4o-mini | ๐ค | GPT-4o-mini | 27.64 | 0.6 | 2.21 | 1.67 | 34.48 | 30.07 | 26.37 | 0.12 | 8.52 | 0.12 | 13.18 | |||
22 | FII the Best | mBERT-WikiNEuRal | ๐ค | Gemini 1.0 Pro | 68.11 | 62.63 | 69.91 | 68.11 | 67.67 | 66.68 | 64.11 | 55.41 | 56.9 | 26.46 | 60.6 | |||
34 | Transcreate | Chatgpt-o1-llm | ๐ค | o1-2024-12-17 | 37.52 | |||||||||||||
30 | HausaNLP | Gemini-few-shot | ๐ | ๐ค | gemini-1.5-flash | 34.18 | 38.14 | 48.3 | 35.32 | 39.39 | 34.93 | 33.75 | 18.62 | 41.54 | 8.09 | 33.22 | ||
28 | Zero | FineTuned-MT | N/A | ๐ | 37.5 | 40.32 | 46.46 | 33.16 | 39.37 | 35.28 | 35.97 | 13.75 | 46.5 | 8.41 | 33.67 | |||
34 | VerbaNexAI Lab | TransNER-SpEn | ๐ | N/A | ๐ | 24.62 | ||||||||||||
34 | Transcreate | Chatgpt-4o-llm | ๐ค | gpt-4o-2024-08-06 | 39.51 | |||||||||||||
27 | The Five Forbidden Entities | Embedded Entities | ๐ค | MBart | ๐ | 53.24 | 51.79 | 53.24 | 48.14 | 54.69 | 61.65 | 39.2 | 13.64 | 48.65 | 18.93 | 44.32 | ||
34 | Transcreate | Chatgpt-o1-mini-llm | ๐ค | o1-mini-2024-09-12 | 33.06 | |||||||||||||
14 | Lunar | LLaMA-RAFT-Gold | ๐ | ๐ | ๐ค | Llama-3.1-8B-Instruct | ๐ | 86.5 | 81.72 | 87.47 | 73.19 | 90.04 | 88.27 | 88.51 | 87.03 | 80.51 | 57.99 | 82.12 |
2 | pingan_team | Qwen2.5-72B-LoRA + zhconv | ๐ | ๐ค | Qwen2.5-72B | ๐ | 91.47 | 85.72 | 90.13 | 91.44 | 93.19 | 90.92 | 90.24 | 91.18 | 84.08 | 81.09 | 88.95 | |
34 | Transcreate | Llama-llm | ๐ค | Llama-3.1-8B-Instruct | 5.63 | |||||||||||||
34 | Transcreate | Gemini-flash-llm | ๐ค | gemini-1.5-flash | 33.16 | |||||||||||||
15 | SALT ๐ง | Salt-Full-Pipeline + Gold | ๐ | ๐ | N/A | ๐ | 88.19 | 82.56 | 82.99 | 84.04 | 88.68 | 83.34 | 82.62 | 72.93 | 83.39 | 51 | 79.98 | |
1 | pingan_team | Qwen2.5-72B-LoRA | ๐ | ๐ค | Qwen2.5-72B | ๐ | 91.73 | 86.35 | 90.13 | 91.56 | 93.02 | 91.41 | 90.24 | 91.18 | 84.13 | 81.26 | 89.1 | |
17 | SALT ๐ง | Salt-Full-Pipeline | ๐ | ๐ค | GPT-4o-mini | ๐ | 83.15 | 76.07 | 82.47 | 79.82 | 80.74 | 79.91 | 80.48 | 76.18 | 77.93 | 54.59 | 77.13 | |
34 | JNLP | Multi-task-mT5 | N/A | ๐ | 13.07 | 12.01 | 11.93 | |||||||||||
34 | silp_nlp | T5-MT-Instruct | ๐ | ๐ค | T5-base | ๐ | 0 | 0 | 0 | 0.07 | 5.63 | 0 | ||||||
29 | HausaNLP | Gemini-0shot | ๐ค | gemini-1.5-flash | 32.66 | 38.16 | 47.92 | 38.77 | 40.31 | 35.1 | 34.67 | 18.8 | 40.82 | 8.53 | 33.57 | |||
13 | arancini | WikiGemmaMT | ๐ | ๐ค | gemma-2-9b-it | 90.15 | 84.8 | 89.58 | 90.7 | 92.43 | 90.74 | 90.73 | 90.8 | 82.29 | 50.79 | 85.3 | ||
34 | AMM_CUET | EA-MT-GPT4o-FR-IT-NER | ๐ค | GPT-4o | ๐ | |||||||||||||
6 | UAlberta | WikiEnsemble | ๐ | ๐ | ๐ค | GPT-4o | 91.69 | 85.23 | 89.36 | 89.62 | 91.74 | 90.39 | 90.44 | 90.02 | 83.21 | 81.15 | 88.28 | |
24 | UAlberta | PromptGPT | ๐ค | GPT4o | 43.99 | 49.57 | 57.03 | 50.39 | 51.98 | 52.37 | 48.86 | 25.27 | 48.2 | 39.04 | 46.67 | |||
34 | GinGer | LoRA-nllb-distilled-200-distilled-600M | N/A | ๐ | 18.41 | 25.52 | 0 | |||||||||||
4 | RAGthoven | GPT-4o + WikiData + RAG | ๐ | ๐ | ๐ค | GPT-4o | 91.88 | 85.07 | 89.88 | 91.02 | 92.78 | 89.57 | 90.22 | 90.48 | 82.76 | 81.47 | 88.51 | |
23 | Lunar | LLaMA-RAFT | ๐ | ๐ค | Llama-3.1-8B-Instruct | ๐ | 62.66 | 56.74 | 63.15 | 53.67 | 70.93 | 52.94 | 58.17 | 55.15 | 64.81 | 26.57 | 56.48 | |
3 | Deerlu | Qwen2.5-Max-Wiki | ๐ | ๐ | ๐ค | Qwen2.5-Max | 91.53 | 85.94 | 90.26 | 91.07 | 92.84 | 91.5 | 90.69 | 90.98 | 83.88 | 80.76 | 88.95 | |
34 | Transcreate | Claude-sonnet-llm | ๐ค | claude-3-5-sonnet-20241022 | 39.69 | |||||||||||||
16 | Howard University-AI4PC | DoubleGPT | ๐ | ๐ | ๐ค | gpt-4o-2024-08-06 | 85.18 | 77.18 | 85.16 | 79.01 | 81.82 | 85.1 | 86.09 | 84.22 | 73.33 | 42.18 | 77.93 | |
25 | The Five Forbidden Entities | MBart-KnowledgeAware | ๐ค | MBart | ๐ | 58.08 | 54.29 | 52.92 | 49.42 | 55.34 | 62.61 | 56.63 | 16.3 | 50.77 | 26.23 | 48.26 | ||
32 | silp_nlp | GPT-4o | ๐ค | GPT-4o | 28.24 | 0.92 | 2.21 | 1.5 | 32.97 | 31.15 | 28.85 | 0.09 | 9.21 | 0.12 | 13.52 | |||
26 | RAGthoven | GPT-4o + RAG | ๐ | ๐ค | GPT-4o | 43.57 | 45.93 | 52.55 | 46.29 | 48.71 | 46.61 | 46.48 | 30.81 | 50.1 | 41.78 | 45.28 | ||
34 | ASL_CUET | GPT-4o-EntityAware-FR-IT | ๐ค | GPT-4o | ๐ | |||||||||||||
10 | Lunar | LLaMA-RAFT-Plus-Gold | ๐ | ๐ | ๐ค | Llama-3.1-8B-Instruct | ๐ | 88.81 | 85.96 | 89.83 | 90.12 | 92.64 | 91.44 | 90.73 | 91.24 | 83.9 | 68.33 | 87.3 |
5 | pingan_team | Phi4-FullFT | ๐ | ๐ค | Phi-4 | ๐ | 91.22 | 85.5 | 90.09 | 91.27 | 92.86 | 91.27 | 90.89 | 91.12 | 83.93 | 80.88 | 88.9 | |
34 | HausaNLP | FT-NLLB | ๐ | ๐ค | NLLB-200-600M | ๐ | 20.61 | 20.86 | 32.75 | 22.85 | 27.29 | 12.74 | ||||||
7 | CHILL | GPT4o-RAG-Refine | ๐ | ๐ | ๐ค | GPT-4o | 91.86 | 85.23 | 89.88 | 89.95 | 92.43 | 90.86 | 90.85 | 91.53 | 84.86 | 77.77 | 88.52 |
Comet Score Leaderboard
Rank | Team | System | Uses Gold | Uses RAG | Uses LLM | LLM Name | Finetuned | ar_AE | de_DE | es_ES | fr_FR | it_IT | ja_JP | ko_KR | th_TH | tr_TR | zh_TW | overall |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
11 | The Five Forbidden Entities | LoRA-nllb-distilled-200-distilled-600M | ๐ | ๐ | ๐ค | Llama-3.3-70B-Instruct + DeepSeek-R1 | ๐ | 94.33 | 94.38 | 95.28 | 93.77 | 94.96 | 95.68 | 92.79 | 93.43 | 94.09 | 94.24 | 94.51 |
Rank | Team | System | Uses Gold | Uses RAG | Uses LLM | LLM Name | Finetuned | ar_AE | de_DE | es_ES | fr_FR | it_IT | ja_JP | ko_KR | th_TH | tr_TR | zh_TW | overall |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
11 | YNU-HPCC | LLaMA + MT | ๐ | ๐ค | Llama-3.3-70B-Instruct | 94.33 | 94.38 | 95.28 | 93.77 | 94.96 | 95.68 | 94.9 | 93.43 | 94.09 | 94.24 | 94.51 | ||
20 | Lunar | LLaMA-RAFT-Plus | ๐ | ๐ค | Llama-3.1-8B-Instruct | ๐ | 91.45 | 91.4 | 93.02 | 91.33 | 93.31 | 92.12 | 92.79 | 90.17 | 93.63 | 88.97 | 91.82 | |
34 | sakura | Rakuten7b-PO10 | ๐ค | Rakuten/RakutenAI-7B-chat | ๐ | 90.74 | ||||||||||||
34 | SHEF | Llama-Wiki-DeepSeek | ๐ค | Llama-3.3-70B-Instruct + DeepSeek-R1 | ๐ | 92.82 | 93.91 | 91.93 | 94.68 | |||||||||
11 | YNU-HPCC | Qwen2.5-32B | ๐ | ๐ค | Qwen2.5-32B | ๐ | 94.33 | 94.38 | 95.28 | 93.77 | 94.96 | 95.68 | 94.9 | 93.43 | 94.09 | 94.24 | 94.51 | |
8 | UAlberta | WikiGPT4o | ๐ | ๐ | ๐ค | GPT-4o | 94.86 | 94.28 | 95.3 | 94.26 | 95.92 | 95.79 | 95.6 | 94.12 | 95.82 | 94.28 | 95.02 | |
18 | SALT ๐ง | Salt-MT-Pipeline | ๐ | N/A | ๐ | 93.2 | 92.34 | 93.6 | 91.84 | 93.36 | 93.02 | 92.97 | 90.64 | 94.47 | 89.75 | 92.52 | ||
34 | Transcreate | Chatgpt-4o-mini-llm | ๐ค | gpt-4o-mini-2024-07-18 | 90.46 | |||||||||||||
34 | silp_nlp | NER-M2M100 | ๐ | ๐ค | M2M100 | ๐ | 86.55 | |||||||||||
31 | Muhandro_HSE | NER-LLM | ๐ค | Llama-3.1-8B, Llama-3.1-70B | ๐ | 87.07 | 90.46 | 91.65 | 88.82 | 89.45 | 90 | 89.53 | 77.58 | 92.37 | 84.99 | 88.19 | ||
34 | CUET_DeepLearners | Spacy-NLLB | N/A | |||||||||||||||
34 | Transcreate | Claude-haiku-llm | ๐ค | claude-3-5-haiku-20241022 | 80.56 | |||||||||||||
9 | RAGthoven | GPT-4o + Wikidata | ๐ | ๐ค | GPT-4o | 94.63 | 94.33 | 95.15 | 94.03 | 95.87 | 95.54 | 95.19 | 93.67 | 95.29 | 94.08 | 94.78 | ||
34 | Transcreate | Gemini-pro-llm | ๐ค | gemini-1.5-pro | 90.94 | |||||||||||||
21 | YNU-HPCC | Qwen2.5 + M2M | ๐ค | Qwen2.5-32B | 91.54 | 91.65 | 92.77 | 90.57 | 91.97 | 93.82 | 93.08 | 89.14 | 91.49 | 92.02 | 91.8 | |||
19 | FII-UAIC-SAI | Qwen2.5-Wiki-MT | ๐ค | Qwen2.5-72B-Instruct-AWQ | 91.35 | 91.3 | 92.58 | 90.59 | 92.71 | 93.56 | 92.78 | 88.62 | 91.63 | 91.25 | 91.64 | |||
33 | silp_nlp | GPT-4o-mini | ๐ค | GPT-4o-mini | 88.6 | 63.19 | 79.21 | 74.02 | 89.93 | 92.61 | 91.32 | 60.92 | 67.5 | 69.37 | 77.67 | |||
22 | FII the Best | mBERT-WikiNEuRal | ๐ค | Gemini 1.0 Pro | 90.01 | 89.13 | 91.06 | 89.89 | 88.5 | 91.82 | 90.72 | 85.2 | 90.19 | 88.29 | 89.48 | |||
34 | Transcreate | Chatgpt-o1-llm | ๐ค | o1-2024-12-17 | 91.96 | |||||||||||||
30 | HausaNLP | Gemini-few-shot | ๐ | ๐ค | gemini-1.5-flash | 89.59 | 89.86 | 92.5 | 88.95 | 90.64 | 92.31 | 91.34 | 83.97 | 92.85 | 88.66 | 90.07 | ||
28 | Zero | FineTuned-MT | N/A | ๐ | 90.82 | 90.62 | 92.38 | 89.06 | 90.78 | 92.57 | 91.78 | 82.61 | 93.83 | 88.98 | 90.34 | |||
34 | VerbaNexAI Lab | TransNER-SpEn | ๐ | N/A | ๐ | 87.09 | ||||||||||||
34 | Transcreate | Chatgpt-4o-llm | ๐ค | gpt-4o-2024-08-06 | 90.87 | |||||||||||||
27 | The Five Forbidden Entities | Embedded Entities | ๐ค | MBart | ๐ | 86.66 | 88.76 | 89.64 | 86.69 | 89.01 | 91.29 | 79.8 | 62.4 | 80.11 | 80.29 | 83.46 | ||
34 | Transcreate | Chatgpt-o1-mini-llm | ๐ค | o1-mini-2024-09-12 | 92.02 | |||||||||||||
14 | Lunar | LLaMA-RAFT-Gold | ๐ | ๐ | ๐ค | Llama-3.1-8B-Instruct | ๐ | 91.35 | 92.61 | 93.83 | 92.44 | 94.42 | 94.57 | 92.84 | 89.17 | 93.78 | 91.03 | 92.6 |
2 | pingan_team | Qwen2.5-72B-LoRA + zhconv | ๐ | ๐ค | Qwen2.5-72B | ๐ | 94.18 | 94.03 | 95.09 | 94.41 | 95.89 | 95.7 | 95.44 | 93.55 | 95.51 | 94.26 | 94.81 | |
34 | Transcreate | Llama-llm | ๐ค | Llama-3.1-8B-Instruct | 55.29 | |||||||||||||
34 | Transcreate | Gemini-flash-llm | ๐ค | gemini-1.5-flash | 90.81 | |||||||||||||
15 | SALT ๐ง | Salt-Full-Pipeline + Gold | ๐ | ๐ | N/A | ๐ | 93.62 | 93.21 | 94.26 | 92.61 | 94.58 | 94.17 | 93.69 | 91.55 | 95 | 90.72 | 93.34 | |
1 | pingan_team | Qwen2.5-72B-LoRA | ๐ | ๐ค | Qwen2.5-72B | ๐ | 93.64 | 94.05 | 95.09 | 94.31 | 95.8 | 95.36 | 95.44 | 93.55 | 95.7 | 94.44 | 94.74 | |
17 | SALT ๐ง | Salt-Full-Pipeline | ๐ | ๐ค | GPT-4o-mini | ๐ | 91.86 | 91.42 | 93.16 | 91.15 | 92.32 | 92.54 | 92.26 | 90.17 | 93.74 | 89.49 | 91.81 | |
34 | JNLP | Multi-task-mT5 | N/A | ๐ | 78.87 | 71.63 | 79.6 | |||||||||||
34 | silp_nlp | T5-MT-Instruct | ๐ | ๐ค | T5-base | ๐ | 38.87 | 45.63 | 27.93 | 45.21 | 69.13 | 38.78 | ||||||
29 | HausaNLP | Gemini-0shot | ๐ค | gemini-1.5-flash | 88.56 | 89.3 | 91.71 | 88.35 | 89.99 | 91.06 | 90.71 | 83.41 | 92.42 | 87.85 | 89.33 | |||
13 | arancini | WikiGemmaMT | ๐ | ๐ค | gemma-2-9b-it | 92.62 | 93.68 | 94.97 | 93.41 | 95.35 | 93.93 | 94.26 | 92.18 | 94.38 | 91.44 | 93.62 | ||
34 | AMM_CUET | EA-MT-GPT4o-FR-IT-NER | ๐ค | GPT-4o | ๐ | |||||||||||||
6 | UAlberta | WikiEnsemble | ๐ | ๐ | ๐ค | GPT-4o | 94.86 | 94.28 | 95.31 | 94.26 | 95.92 | 95.79 | 95.62 | 94.11 | 95.93 | 94.28 | 95.04 | |
24 | UAlberta | PromptGPT | ๐ค | GPT4o | 91.68 | 91.58 | 93.22 | 90.61 | 92.22 | 93.52 | 93.45 | 87.32 | 93.27 | 92.09 | 91.9 | |||
34 | GinGer | LoRA-nllb-distilled-200-distilled-600M | N/A | ๐ | 87.1 | 89.28 | 0 | |||||||||||
4 | RAGthoven | GPT-4o + WikiData + RAG | ๐ | ๐ | ๐ค | GPT-4o | 94.63 | 94.33 | 95.1 | 94.03 | 95.93 | 95.75 | 95.79 | 94.53 | 95.83 | 94.5 | 95.04 | |
23 | Lunar | LLaMA-RAFT | ๐ | ๐ค | Llama-3.1-8B-Instruct | ๐ | 89.67 | 90.49 | 92.29 | 90.39 | 92.59 | 92.09 | 90.64 | 85.2 | 92.57 | 88.03 | 90.4 | |
3 | Deerlu | Qwen2.5-Max-Wiki | ๐ | ๐ | ๐ค | Qwen2.5-Max | 94.29 | 94.29 | 95.13 | 93.82 | 95.81 | 95.7 | 95.18 | 93.61 | 95.23 | 94.53 | 94.76 | |
34 | Transcreate | Claude-sonnet-llm | ๐ค | claude-3-5-sonnet-20241022 | 83.84 | |||||||||||||
16 | Howard University-AI4PC | DoubleGPT | ๐ | ๐ | ๐ค | gpt-4o-2024-08-06 | 93.83 | 93.49 | 94.81 | 92.64 | 93.46 | 95.27 | 94.61 | 92.68 | 93.51 | 91.96 | 93.63 | |
25 | The Five Forbidden Entities | MBart-KnowledgeAware | ๐ค | MBart | ๐ | 86.84 | 88.03 | 88.82 | 86.06 | 87.76 | 92.31 | 83.14 | 62.49 | 84.86 | 82.84 | 84.32 | ||
32 | silp_nlp | GPT-4o | ๐ค | GPT-4o | 88.5 | 62.97 | 78.62 | 74.15 | 89.64 | 92.73 | 91.49 | 60.81 | 67.77 | 69.3 | 77.6 | |||
26 | RAGthoven | GPT-4o + RAG | ๐ | ๐ค | GPT-4o | 91.02 | 90.94 | 92.88 | 89.98 | 91.6 | 93.31 | 93.26 | 88.36 | 93.73 | 92.24 | 91.73 | ||
34 | ASL_CUET | GPT-4o-EntityAware-FR-IT | ๐ค | GPT-4o | ๐ | |||||||||||||
10 | Lunar | LLaMA-RAFT-Plus-Gold | ๐ | ๐ | ๐ค | Llama-3.1-8B-Instruct | ๐ | 94.11 | 94.33 | 95.13 | 94.23 | 95.75 | 95.68 | 95.46 | 93.49 | 95.21 | 93.61 | 94.7 |
5 | pingan_team | Phi4-FullFT | ๐ | ๐ค | Phi-4 | ๐ | 92.98 | 94.13 | 95.04 | 94.21 | 95.81 | 95.91 | 95.18 | 91.41 | 95.52 | 94.18 | 94.44 | |
34 | HausaNLP | FT-NLLB | ๐ | ๐ค | NLLB-200-600M | ๐ | 87.98 | 88.42 | 91.26 | 85.07 | 89.52 | 88.86 | ||||||
7 | CHILL | GPT4o-RAG-Refine | ๐ | ๐ | ๐ค | GPT-4o | 94.23 | 94.08 | 95 | 93.54 | 95.65 | 95.61 | 95.21 | 94.26 | 95.63 | 93.86 | 94.71 |