| | --- |
| | language: |
| | - fr |
| | - es |
| | - zh |
| | size_categories: |
| | - 1K<n<10K |
| | configs: |
| | - config_name: default |
| | data_files: |
| | - split: test |
| | path: "test/data-00000-of-00001.arrow" |
| | --- |
| | |
| | # MultiNRC: Multilingual Native Reasoning Challenge |
| |
|
| | MultiNRC is a challenging evaluation benchmark for large language models, designed to assess multilingual reasoning ability in French, Spanish, and Chinese. Unlike existing benchmarks that simply translate English-centric content, MultiNRC consists of over 1,000 native-authored reasoning questions, crafted by native speakers to capture linguistic and cultural nuances. |
| |
|
| | ## Features |
| |
|
| | - **Languages:** French, Spanish, Chinese |
| | - **Categories:** |
| | - Language-specific Linguistic Reasoning |
| | - Wordplay & Riddles |
| | - Cultural Reasoning & Traditions |
| | - Math Reasoning with Cultural Relevance |
| | - **English Equivalents:** For Cultural/Tradition and Math, human-translated English versions are provided for direct comparison. |
| | - **Ground Truth Final Answers:** Short, objective answers accompany each prompt for automatic evaluation. |
| |
|
| |
|
| | ## Dataset Structure |
| |
|
| | Each entry includes: |
| | - A native-language prompt and answer (`i18n_prompt`, `i18n_gtfa`) |
| | - (For Math Reasoning and Cultural Reasoning category tasks) An English-equivalent prompt and answer (`english_prompt`, `english_gtfa`) |
| | - Metadata: `task_id`, `language`, `category` |
| |
|
| | ## Citation |
| |
|
| | If you use MultiNRC in your research, please cite: |
| |
|
| | ```bibtex |
| | @article{fabbri2025multinrc, |
| | title = {MultiNRC: A Challenging Native Multilingual Reasoning Evaluation Benchmark for LLMs}, |
| | author = {Fabbri, Alexander R. and Mares, Diego and Flores, Jorge and Mankikar, Meher and Hernandez, Ernesto and Lee, Dean and Liu, Bing and Xing, Chen}, |
| | year = {2025}, |
| | note = {arXiv preprint, arXiv:XXXX.XXXXX} |
| | } |