Noticias
Noticias
Evaluation Agent: Efficient and Promptable Evaluation Framework for Visual Generative Models
arXiv:2412.09645v3 Announce Type: replace-cross Abstract: Recent advancements in visual generative models have enabled high-quality image...
Evaluating Speech-to-Text x LLM x Text-to-Speech Combinations for AI Interview Systems
arXiv:2507.16835v1 Announce Type: cross Abstract: Voice-based conversational AI systems increasingly rely on cascaded architectures combining...
Evaluating Rare Disease Diagnostic Performance in Symptom Checkers: A Synthetic Vignette Simulation Approach
arXiv:2506.19750v2 Announce Type: replace Abstract: Symptom Checkers (SCs) provide users with personalized medical information. To...
Evaluating LLMs on Real-World Forecasting Against Expert Forecasters
arXiv:2507.04562v3 Announce Type: replace-cross Abstract: Large language models (LLMs) have demonstrated remarkable capabilities across diverse...
Evaluating Large Language Models for Anxiety, Depression, and Stress Detection: Insights into Prompting Strategies and Synthetic Data
arXiv:2511.07044v2 Announce Type: replace Abstract: Mental health disorders affect over one-fifth of adults globally, yet...
Evaluating Evidence Grounding Under User Pressure in Instruction-Tuned Language Models
arXiv:2603.20162v1 Announce Type: new Abstract: In contested domains, instruction-tuned language models must balance user-alignment pressures...
Evaluating Creative Short Story Generation in Humans and Large Language Models
arXiv:2411.02316v5 Announce Type: replace Abstract: Story-writing is a fundamental aspect of human imagination, relying heavily...
Evaluating Autoformalization Robustness via Semantically Similar Paraphrasing
arXiv:2511.12784v1 Announce Type: new Abstract: Large Language Models (LLMs) have recently emerged as powerful tools...
Evaluating and Improving Robustness in Large Language Models: A Survey and Future Directions
arXiv:2506.11111v1 Announce Type: new Abstract: Large Language Models (LLMs) have gained enormous attention in recent...
Evaluating $n$-Gram Novelty of Language Models Using Rusty-DAWG
arXiv:2406.13069v4 Announce Type: replace Abstract: How novel are texts generated by language models (LMs) relative...
EvalTree: Profiling Language Model Weaknesses via Hierarchical Capability Trees
arXiv:2503.08893v2 Announce Type: replace Abstract: An ideal model evaluation should achieve two goals: identifying where...
EVALOOOP: A Self-Consistency-Centered Framework for Assessing Large Language Model Robustness in Programming
arXiv:2505.12185v5 Announce Type: replace-cross Abstract: Evaluating the programming robustness of large language models (LLMs) is...