Can AI Pass an Italian High School Exam? We Put LLMs to the Test
Disce aut Deficere: Evaluating LLMs Proficiency on the INVALSI Italian Benchmark
By
Fabio Mercorio|Mario Mezzanzanica|Daniele Potertì|Antonio Serino|Andrea Seveso

https://arxiv.org/abs/2406.17535v1
Summary
Imagine an AI taking your Italian high school exams. How would it fare against real students? That’s precisely what researchers explored in a new study using the INVALSI tests, Italy’s standardized national assessments, as a benchmark to evaluate the proficiency of Large Language Models (LLMs). These AI systems, trained on massive amounts of text data, have shown impressive language skills, but how do those skills measure up against real-world academic challenges? The study adapted INVALSI tests, covering reading comprehension, grammar, and vocabulary, into a format suitable for automated LLM evaluation. They tested a diverse range of models, from well-known names like OpenAI's GPT models and Google's Gemini to open-source and Italian-specific models. The results? While AI excelled in some areas, particularly reconstructing the meaning of text, it struggled in others. For instance, accurately placing the letter 'h' in Italian words, a seemingly simple task for humans, proved difficult. The study highlighted a common issue: AI models often struggle with tasks that require precise, rule-based understanding of grammar. Interestingly, larger models generally performed better, but even the most advanced couldn’t match average student performance on higher-grade tests. This suggests that while LLMs are powerful tools, they still lag behind humans in complex reasoning and nuanced language understanding, especially in academic contexts. One intriguing aspect of the research is the comparison of human and model performance. The researchers analyzed human accuracy data from several grades and found that while AI tends to perform worse in higher grades, there was no correlation between grade and human performance. The best human scores were recorded in 8th grade. However, this study is just the beginning. Researchers plan to expand the benchmark to include math and multimodal questions, allowing AI to tackle geometry problems and other visual information. The ongoing project will also include a public leaderboard, creating a competitive arena for researchers to improve their LLMs' language proficiency. The study raises interesting questions about the nature of intelligence and learning. While AI may excel in some areas, it still has some way to go before it can truly grasp the nuances of human language and reasoning. The study underscores the fact that while AI is rapidly evolving, there is still a gap between its capabilities and those of human learners, especially in the complex settings of high-stakes academic exams.
🍰 Interesting in building your own agents?
PromptLayer provides the tools to manage and monitor prompts with your whole team.
Get started for free.Question & Answers
How did researchers adapt the INVALSI tests to evaluate LLMs' language capabilities?
The researchers converted traditional INVALSI tests into an LLM-compatible format focusing on reading comprehension, grammar, and vocabulary assessment. The adaptation process involved transforming standard test questions into structured prompts that could be processed by AI models. They tested multiple LLM variants, including OpenAI's GPT models, Google's Gemini, and Italian-specific models, creating a comprehensive evaluation framework. The implementation allowed for automated assessment of AI responses across different linguistic competencies, particularly in areas like text meaning reconstruction and grammatical rule application. This methodology enabled direct comparison between human and AI performance across various grade levels.
What are the main challenges AI faces in language learning compared to humans?
AI systems face several distinct challenges in language learning that humans naturally overcome. The primary difference lies in rule-based understanding and contextual comprehension. While AI can process vast amounts of text data, it struggles with precise grammatical rules (like Italian 'h' placement) and nuanced language understanding. Humans develop these skills through natural learning and practical application, while AI relies on pattern recognition from training data. This has practical implications in education, business communication, and content creation, where AI tools can assist but may not fully replace human language expertise.
How can AI assessment tools improve educational testing and evaluation?
AI assessment tools offer several advantages in educational testing and evaluation. They can provide instant feedback, consistent grading, and personalized learning recommendations based on student performance patterns. The technology can handle large-scale assessments efficiently, reducing administrative burden on teachers while maintaining objectivity. In practical applications, these tools can help identify learning gaps, adapt teaching methods, and provide data-driven insights for curriculum improvement. However, as shown in the INVALSI study, they work best as complementary tools alongside traditional human evaluation rather than complete replacements.
.png)
PromptLayer Features
- Testing & Evaluation
- The study's systematic evaluation of LLMs on standardized tests aligns with PromptLayer's testing capabilities for assessing model performance
Implementation Details
Create standardized test suites in PromptLayer to evaluate LLM responses against known correct answers, implement scoring metrics, and track performance across model versions
Key Benefits
• Automated assessment of model performance across different linguistic tasks
• Consistent evaluation methodology for comparing multiple LLM versions
• Historical performance tracking and regression testing
Potential Improvements
• Add support for language-specific evaluation metrics
• Implement automated grading rubrics
• Develop comparative analysis tools for human vs AI performance
Business Value
.svg)
Efficiency Gains
Reduces manual evaluation time by 80% through automated testing
.svg)
Cost Savings
Minimizes resources needed for quality assurance and model validation
.svg)
Quality Improvement
Ensures consistent and objective evaluation of LLM performance
- Analytics
- Analytics Integration
- The paper's analysis of performance patterns across different tasks and grades matches PromptLayer's analytics capabilities
Implementation Details
Configure performance monitoring dashboards, set up error tracking for specific linguistic tasks, and implement detailed response analysis
Key Benefits
• Real-time visibility into model performance trends
• Detailed error analysis by task type
• Data-driven insights for model improvement
Potential Improvements
• Add specialized linguistics metrics
• Implement cross-model comparison visualizations
• Develop predictive performance analytics
Business Value
.svg)
Efficiency Gains
Reduces analysis time by providing automated performance insights
.svg)
Cost Savings
Optimizes model selection and usage based on performance data
.svg)
Quality Improvement
Enables data-driven decisions for model enhancement