publications
publications by categories in reversed chronological order. generated by jekyll-scholar.
2025
- Attributing Culture-Conditioned Generations to Pretraining CorporaHuihan Li*, Arnav Goel*, Keyu He, and 1 more authorICLR, 2025
In open-ended generative tasks like narrative writing or dialogue, large language models often exhibit cultural biases, showing limited knowledge and generating templated outputs for less prevalent cultures. Recent works show that these biases may stem from uneven cultural representation in pretraining corpora. This work investigates how pretraining leads to biased culture-conditioned generations by analyzing how models associate entities with cultures based on pretraining data patterns. We propose the MEMOed framework (MEMOrization from pretraining document) to determine whether a generation for a culture arises from memorization. Using MEMOed on culture-conditioned generations about food and clothing for 110 cultures, we find that high-frequency cultures in pretraining data yield more generations with memorized symbols, while some low-frequency cultures produce none. Additionally, the model favors generating entities with extraordinarily high frequency regardless of the conditioned culture, reflecting biases toward frequent pretraining terms irrespective of relevance. We hope that the MEMOed framework and our insights will inspire more works on attributing model performance on pretraining data.
- ELI-Why: Evaluating the Pedagogical Utility of LLM ExplanationsBrihi Joshi*, Keyu He*, Sahana Ramnath, and 5 more authorsSubmitted to ACL, Under review, 2025
Language models today are widely used in education, yet their ability to tailor responses for learners with varied informational needs and knowledge backgrounds remains under-explored. To this end, we introduce ELI-WHY, a benchmark of 13.4K "Why" questions to assess the pedagogical capabilities of LLMs. We then conduct two extensive human studies to assess the utility of LLM-generated explanatory answers (explanations) on our benchmark, tailored to three distinct educational grades: elementary, high-school, and graduate school. In our first study, human raters assume the role of an "educator" to assess model explanations’ fit to different educational grades. We find that GPT-4-generated explanations match their intended educational background only 50% of the time, compared to 79% for human-curated explanations. In our second study, human raters assume the role of a learner to assess if an explanation fits their own informational needs. Results show that users deemed GPT-4-generated explanations relatively 20% less suited to their informational needs, particularly for advanced learners. Additionally, automated evaluation metrics reveal that GPT-4 explanations for different informational needs remain indistinguishable in their grade-level, limiting their pedagogical effectiveness. These findings suggest that LLMs’ ability to follow inference-time instructions alone is insufficient for producing high-utility explanations tailored to users’ informational needs.
- Beyond the Text: How Explanation Qualities Influence User Trust in Visual Language ModelsKeyu He, Brihi Joshi, Tejas Srinivasan, and 1 more authorUnder preparation for NeurIPS, 2025
Visual Language Models (VLMs) are deployed in scenarios where users lack direct access to visual stimuli, such as remote sensing, robotics, and assistance for people with visual impairments. Despite their utility, these models can produce hallucinated outputs that may mislead users. In this work, we investigate the role of explanation quality in calibrating user trust and reliance on VLM outputs. We propose new qualities, Visual Fidelity and Contrastiveness, to complement traditional text-only measures. Through quantitative evaluations on A-OKVQA and VizWiz datasets and a user study, our results indicate that explanations enriched with quality signals lead to a lower unsure rate and improved prediction accuracy and utility in AI-assisted decision-making. We also highlight limitations and future directions to further enhance the interpretability and reliability of VLM-generated rationales.