Eka-Eval : a comprehensive evaluation framework for large language models in Indian languages

Show simple item record

dc.contributor.author Sinha, Samridhi Raj
dc.contributor.author Sheth, Rajvee
dc.contributor.author Upperwal, Abhishek
dc.contributor.author Singh, Mayank
dc.coverage.spatial United States of America
dc.date.accessioned 2025-07-11T08:30:50Z
dc.date.available 2025-07-11T08:30:50Z
dc.date.issued 2025-07
dc.identifier.citation Sinha, Samridhi Raj; Sheth, Rajvee; Upperwal, Abhishek and Singh, Mayank, "Eka-Eval : a comprehensive evaluation framework for large language models in Indian languages", arXiv, Cornell University Library, DOI: arXiv:2507.01853, Jul. 2025.
dc.identifier.uri https://doi.org/10.48550/arXiv.2507.01853
dc.identifier.uri https://repository.iitgn.ac.in/handle/123456789/11625
dc.description.abstract The rapid advancement of Large Language Models (LLMs) has intensified the need for evaluation frameworks that address the requirements of linguistically diverse regions, such as India, and go beyond English-centric benchmarks. We introduce EKA-EVAL, a unified evaluation framework that integrates over 35+ benchmarks (including 10 Indic benchmarks) across nine major evaluation categories. The framework provides broader coverage than existing Indian language evaluation tools, offering 11 core capabilities through a modular architecture, seamless integration with Hugging Face and proprietary models, and plug-and-play usability. As the first end-to-end suite for scalable, multilingual LLM benchmarking, the framework combines extensive benchmarks, modular workflows, and dedicated support for low-resource Indian languages to enable inclusive assessment of LLM capabilities across diverse domains. We conducted extensive comparisons against five existing baselines, demonstrating that EKA-EVAL achieves the highest participant ratings in four out of five categories.
dc.description.statementofresponsibility by Samridhi Raj Sinha, Rajvee Sheth, Abhishek Upperwal and Mayank Singh
dc.language.iso en_US
dc.publisher Cornell University Library
dc.title Eka-Eval : a comprehensive evaluation framework for large language models in Indian languages
dc.type Article
dc.relation.journal arXiv


Files in this item

Files Size Format View

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record

Search Digital Repository


Browse

My Account