


default search action
63rd ACL 2025: Vienna, Austria - Short Papers
- Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar:
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), ACL 2025, Vienna, Austria, July 27 - August 1, 2025. Association for Computational Linguistics 2025, ISBN 979-8-89176-252-7 - Frontmatter.
- Junlin Li, Bo Peng, Yu-Yin Hsu:
Towards LLM-powered Attentive Listener: A Pragmatic Approach through Quantity Self-Repair. 1-13 - Yin Cai, Zhouhong Gu, Zhaohan Du, Zheyu Ye, Shaosheng Cao, Yiqian Xu, Hongwei Feng, Ping Chen:
MIRAGE: Exploring How Large Language Models Perform in Complex Social Interactive Environments. 14-40 - Gyutae Park, Ingeol Baek, Byeongjeong Kim, Joongbo Shin, Hwanhee Lee:
Dynamic Label Name Refinement for Few-Shot Dialogue Intent Classification. 41-52 - Yungi Kim, Hyunsoo Ha, Sukyung Lee, Jihoo Kim, Seonghoon Yang, Chanjun Park:
Rethinking KenLM: Good and Bad Model Ensembles for Efficient Text Quality Filtering in Large Web Corpora. 53-58 - Junda Wu, Yu Xia, Tong Yu, Xiang Chen, Sai Sree Harsha, Akash V. Maharaj, Ruiyi Zhang, Victor S. Bursztyn, Sungchul Kim, Ryan A. Rossi, Julian J. McAuley, Yunyao Li, Ritwik Sinha:
Doc-React: Multi-page Heterogeneous Document Question-answering. 67-78 - Mikolaj Pokrywka, Wojciech Kusa, Mieszko Rutkowski, Mikolaj Koszowski:
ConECT Dataset: Overcoming Data Scarcity in Context-Aware E-Commerce MT. 79-86 - Pius von Däniken, Jan Milan Deriu, Mark Cieliebak:
A Measure of the System Dependence of Automated Metrics. 87-99 - Hyeonseok Moon, Jaehyung Seo, Heuiseok Lim:
Call for Rigor in Reporting Quality of Instruction Tuning Data. 100-109 - Shintaro Ozaki, Kazuki Hayashi, Miyu Oba, Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe:
BQA: Body Language Question Answering Dataset for Video Large Language Models. 110-123 - Miles Shelton, Nate Wingerd, Kritim K. Rijal, Ayush Garg, Adelina Gutic, Brett Barnes, Catherine Finegan-Dollak:
Grounded, or a Good Guesser? A Per-Question Balanced Dataset to Separate Blind from Grounded Models for Embodied Question Answering. 124-135 - Xingjian Diao, Tianzhen Yang, Chunhui Zhang, Weiyi Wu, Ming Cheng, Jiang Gui:
Learning Sparsity for Effective and Efficient Music Performance Question Answering. 136-146 - Chen Zhang, Zhiyuan Liao, Yansong Feng:
Cross-Lingual Transfer of Cultural Knowledge: An Asymmetric Phenomenon. 147-157 - Suet-Ying Lam, Qingcheng Zeng, Jingyi Wu, Rob Voigt:
Leveraging Human Production-Interpretation Asymmetries to Test LLM Cognitive Plausibility. 158-171 - Lorenzo Jaime Yu Flores, Ori Ernst, Jackie CK Cheung:
Improving the Calibration of Confidence Scores in Text Generation Using the Output Distribution's Characteristics. 172-182 - Tianshi Zheng, Weihan Li, Jiaxin Bai, Weiqi Wang, Yangqiu Song:
KnowShiftQA: How Robust are RAG Systems when Textbook Knowledge Shifts in K-12 Education? 183-195 - Shu Okabe, Katharina Hämmerl, Alexander Fraser:
Improving Parallel Sentence Mining for Low-Resource and Endangered Languages. 196-205 - Jiayu Liu, Qing Zong, Weiqi Wang, Yangqiu Song:
Revisiting Epistemic Markers in Confidence Estimation: Can Markers Accurately Reflect Large Language Models' Uncertainty? 206-221 - Marcell Fekete, Nathaniel Romney Robinson, Ernests Lavrinovics, E. Djeride Jean-Baptiste, Raj Dabre, Johannes Bjerva, Heather C. Lent:
Limited-Resource Adapters Are Regularizers, Not Linguists. 222-237 - Anna Bavaresco, Raffaella Bernardi, Leonardo Bertolazzi, Desmond Elliott, Raquel Fernández, Albert Gatt, Esam Ghaleb, Mario Giulianelli, Michael Hanna, Alexander Koller, André F. T. Martins, Philipp Mondorf, Vera Neplenbroek, Sandro Pezzelle, Barbara Plank, David Schlangen, Alessandro Suglia, Aditya K. Surikuchi, Ece Takmaz, Alberto Testoni:
LLMs instead of Human Judges? A Large Scale Empirical Study across 20 NLP Evaluation Tasks. 238-255 - Tong Liu, Xiao Yu, Wenxuan Zhou, Jindong Gu, Volker Tresp:
FocalPO: Enhancing Preference Optimizing by Focusing on Correct Preference Rankings. 256-267 - Megh Thakkar, Quentin Fournier, Matthew Riemer, Pin-Yu Chen, Amal Zouaq, Payel Das, Sarath Chandar:
Combining Domain and Alignment Vectors Provides Better Knowledge-Safety Trade-offs in LLMs. 268-277 - Shira Wein:
Can Uniform Meaning Representation Help GPT-4 Translate from Indigenous Languages? 278-285 - Bastian Bunzeck, Sina Zarrieß:
Subword models struggle with word learning, but surprisal hides it. 286-300 - Christophe Ye, Cassie S. Mitchell:
LLM as Entity Disambiguator for Biomedical Entity-Linking. 301-312 - Piyawat Lertvittayakumjorn, David Kinney, Vinodkumar Prabhakaran, Donald Martin Jr., Sunipa Dev:
Towards Geo-Culturally Grounded LLM Generations. 313-330 - Tharindu Ranasinghe, Hansi Hettiarachchi, Constantin Orasan, Ruslan Mitkov:
MUSTS: MUltilingual Semantic Textual Similarity Benchmark. 331-353 - Davis Bartels, Deepak Gupta, Dina Demner-Fushman:
Can Large Language Models Accurately Generate Answer Keys for Health-related Questions? 354-368 - Katherine Thai, Mohit Iyyer:
Literary Evidence Retrieval via Long-Context Language Models. 369-380 - Dhananjay Ashok, Jonathan May:
A Little Human Data Goes A Long Way. 381-413 - Guangzhen Zhao, Yu Yao, Dechang Kong, Zhenjiang Dong:
Seeking Rational Demonstrations for Large Language Models: A Domain Generalization Approach to Unsupervised Cross-Domain Keyphrase Generation. 414-424 - T. Y. S. S. Santosh, Elvin Quero Hernandez:
LexKeyPlan: Planning with Keyphrases and Retrieval Augmentation for Legal Text Generation: A Case Study on European Court of Human Rights Cases. 425-436 - Runnan Fang, Xiaobin Wang, Yuan Liang, Shuofei Qiao, Jialong Wu, Zekun Xi, Ningyu Zhang, Yong Jiang, Pengjun Xie, Fei Huang, Huajun Chen:
SynWorld: Virtual Scenario Synthesis for Agentic Action Knowledge Refinement. 437-448 - Felix Faltings, Wei Wei, Yujia Bao:
Enhancing Retrieval Systems with Inference-Time Logical Reasoning. 449-463 - Zhipang Wang, Yu Hong, Weihao Sun, Guodong Zhou:
Using Subtext to Enhance Generative IDRR. 464-473 - Wonjun Kang, Kevin Galim, Yuchen Zeng, Minjae Lee, Hyung Il Koo, Nam Ik Cho:
State-offset Tuning: State-based Parameter-Efficient Fine-Tuning for State Space Models. 474-487 - Yu Zhang:
Internal and External Impacts of Natural Language Processing Papers. 488-494 - Xuemei Tang, Jun Wang, Qi Su, Chu-Ren Huang, Jinghang Gu:
An Effective Incorporating Heterogeneous Knowledge Curriculum Learning for Sequence Labeling. 495-503 - Zhenyu Zhang, JiuDong Yang, Zhaowen Tao, Meng Chen:
Accelerating Dense LLMs via L0-regularized Mixture-of-Experts. 504-513 - Noriki Nishida, Koji Inoue, Hideki Nakayama, Mayumi Bono, Katsuya Takanashi:
Do Multimodal Large Language Models Truly See What We Point At? Investigating Indexical, Iconic, and Symbolic Gesture Comprehension. 514-524 - Songtao Jiang, Chenyi Zhou, Yan Zhang, Yeying Jin, Zuozhu Liu:
Fast or Slow? Integrating Fast Intuition and Deliberate Thinking for Enhancing Visual Question Answering. 525-534 - Nadav Borenstein, Greta Warren, Desmond Elliott, Isabelle Augenstein:
Can Community Notes Replace Professional Fact-Checkers? 535-552 - Sihan Tan, Taro Miyazaki, Kazuhiro Nakadai:
Multilingual Gloss-free Sign Language Translation: Towards Building a Sign Language Foundation Model. 553-561 - Xiang Fei, Jinghui Lu, Qi Sun, Hao Feng, Yanjie Wang, Wei Shi, An-Lan Wang, Jingqun Tang, Can Huang:
Advancing Sequential Numerical Prediction in Autoregressive Models. 562-574 - Hyein Seo, Taewook Hwang, Yohan Lee, Sangkeun Jung:
FEAT: A Preference Feedback Dataset through a Cost-Effective Auto-Generation and Labeling Framework for English AI Tutoring. 575-589 - Duygu Sezen Islakoglu, Jan-Christoph Kalo:
ChronoSense: Exploring Temporal Understanding in Large Language Models with Time Intervals of Events. 590-602 - Tanguy Cazalets, Ruben Janssens, Tony Belpaeme, Joni Dambre:
Human Alignment: How Much Do We Adapt to LLMs? 603-613 - Yonghyun Jun, Hwanhee Lee:
Dynamic Order Template Prediction for Generative Aspect-Based Sentiment Analysis. 614-626 - Éric Le Ferrand, Bo Jiang, Joshua K. Hartshorne, Emily Prud'hommeaux:
That doesn't sound right: Evaluating speech transcription quality in field linguistics corpora. 627-635 - William Jurayj, Jeffrey Cheng, Benjamin Van Durme:
Is That Your Final Answer? Test-Time Scaling Improves Selective Question Answering. 636-644 - Álvaro Vega-Hidalgo, Artem Abzaliev, Thore Bergman, Rada Mihalcea:
Acoustic Individual Identification of White-Faced Capuchin Monkeys Using Joint Multi-Species Embeddings. 645-659 - Danush Khanna, Pratinav Seth, Sidhaarth Sredharan Murali, Aditya Kumar Guru, Siddharth Shukla, Tanuj Tyagi, Sandeep Chaurasia, Kripabandhu Ghosh:
SELF-PERCEPT: Introspection Improves Large Language Models' Detection of Multi-Person Mental Manipulation in Conversations. 660-675 - Samuel Mensah, Elena Kochkina, Jabez Magomere, Joy Prakash Sain, Simerjot Kaur, Charese Smiley:
A Variational Approach for Mitigating Entity Bias in Relation Extraction. 676-684 - Mohammadtaha Bagherifard, Sahar Rajabi, Ali Edalat, Yadollah Yaghoobzadeh:
GenKnowSub: Improving Modularity and Reusability of LLMs through General Knowledge Subtraction. 685-694 - Zachary Nicholas Houghton, Kenji Sagae, Emily Morgan:
The Role of Abstract Representations and Observed Preferences in the Ordering of Binomials in Large Language Models. 695-702 - Payal Mohapatra, Akash Pandey, Xiaoyuan Zhang, Qi Zhu:
Can LLMs Understand Unvoiced Speech? Exploring EMG-to-Text Conversion with LLMs. 703-712 - Dan Qiao, Yuan Gao, Zheming Yang, Di Yang, Ziheng Wu, Pengcheng Lu, Minghui Qiu, Juntao Li, Min Zhang:
Decoder-Only LLMs can be Masked Auto-Encoders. 713-723 - Zikai Xiao, Ziyang Wang, Wen Ma, Yan Zhang, Wei Shen, WangYan WangYan, Luqi Gong, Zuozhu Liu:
Mitigating Posterior Salience Attenuation in Long-Context LLMs with Positional Contrastive Decoding. 724-733 - Xuan Zhang, Cunxiao Du, Sicheng Yu, Jiawei Wu, Fengzhuo Zhang, Wei Gao, Qian Liu:
Sparse-to-Dense: A Free Lunch for Lossless Acceleration of Video Understanding in LLMs. 734-742 - Andrea Santilli, Adam Golinski, Michael Kirchhof, Federico Danieli, Arno Blaas, Miao Xiong, Luca Zappella, Sinead Williamson:
Revisiting Uncertainty Quantification Evaluation in Language Models: Spurious Interactions with Response Length Bias Results. 743-759 - Verna Dankers, Vikas Raunak:
Memorization Inheritance in Sequence-Level Knowledge Distillation for Neural Machine Translation. 760-774 - Fabio Fehr, Prabhu Teja Sivaprasad, Luca Franceschi, Giovanni Zappella:
CoRet: Improved Retriever for Code Editing. 775-789 - Lorenzo Proietti, Stefano Perrella, Roberto Navigli:
Has Machine Translation Evaluation Achieved Human Parity? The Human Reference and the Limits of Progress. 790-813 - Quan Nguyen-Tri, Cong Dao Tran, Hoang Thanh-Tung:
Diffusion Directed Acyclic Transformer for Non-Autoregressive Machine Translation. 814-828 - Akshat Gupta, Maochuan Lu, Thomas Hartvigsen, Gopala Anumanchipalli:
Efficient Knowledge Editing via Minimal Precomputation. 829-840 - Dallas Card:
Meaning Variation and Data Quality in the Corpus of Founding Era American English. 841-856 - Ye Wang, Xinrun Xu, Zhiming Ding:
MindRef: Mimicking Human Memory for Hierarchical Reference Retrieval with Fine-Grained Location Awareness. 857-872 - Florian Kandra, Vera Demberg, Alexander Koller:
LLMs syntactically adapt their language use to their conversational partner. 873-886 - Nishat Raihan, Marcos Zampieri:
TigerLLM - A Family of Bangla Large Language Models. 887-896 - Ronja Stern, Ken Kawamura, Matthias Stürmer, Ilias Chalkidis, Joel Niklaus:
From Citations to Criticality: Predicting Legal Decision Influence in the Multilingual Swiss Jurisprudence. 897-905 - Junwoo Park, Hyuck Lee, Dohyun Lee, Daehoon Gwak, Jaegul Choo:
Revisiting LLMs as Zero-Shot Time Series Forecasters: Small Noise Can Break Large Models. 906-922 - Chen-An Li, Tzu-Han Lin, Yun-Nung Chen, Hung-yi Lee:
Transferring Textual Preferences to Vision-Language Understanding through Model Merging. 923-943 - Xiaoshuai Song, Yanan Wu, Weixun Wang, Jiaheng Liu, Wenbo Su, Bo Zheng:
ProgCo: Program Helps Self-Correction of Large Language Models. 944-959 - Ananth Muppidi, Abhilash Nandy, Sambaran Bandyopadhyay:
Leveraging Self-Attention for Input-Dependent Soft Prompting in LLMs. 960-969 - Andrew Gambardella, Takeshi Kojima, Yusuke Iwasawa, Yutaka Matsuo:
Inconsistent Tokenizations Cause Language Models to be Perplexed by Japanese Grammar. 970-976 - Selim Jerad, Anej Svete, Jiaoda Li, Ryan Cotterell:
Unique Hard Attention: A Tale of Two Sides. 977-996 - Keqin Peng, Liang Ding, Yuanxin Ouyang, Meng Fang, Yancheng Yuan, Dacheng Tao:
Enhancing Input-Label Mapping in In-Context Learning with Contrastive Decoding. 997-1004 - Dennis Fucci, Marco Gaido, Matteo Negri, Luisa Bentivogli, André F. T. Martins, Giuseppe Attanasio:
Different Speech Translation Models Encode and Translate Speaker Gender Differently. 1005-1019 - Kaikai An, Shuzheng Si, Helan Hu, Haozhe Zhao, Yuchi Wang, Qingyan Guo, Baobao Chang:
Rethinking Semantic Parsing for Large Language Models: Enhancing LLM Performance with Semantic Hints. 1020-1029 - Pegah Alipoormolabashi, Ajay Patel, Niranjan Balasubramanian:
Quantifying Misattribution Unfairness in Authorship Attribution. 1030-1041 - Thi Vu, Linh The Nguyen, Dat Quoc Nguyen:
Zero-Shot Text-to-Speech for Vietnamese. 1042-1049 - Zheyuan Yang, Zexi Kuang, Xue Xia, Yilun Zhao:
Can LLMs Generate High-Quality Test Cases for Algorithm Problems? TestCase-Eval: A Systematic Evaluation of Fault Coverage and Exposure. 1050-1063 - Juan Wisznia, Cecilia Bolaños, Juan Tollo, Giovanni Marraffini, Agustín Gianolini, Noe Hsueh, Luciano Del Corro:
Are Optimal Algorithms Still Optimal? Rethinking Sorting in LLM-Based Pairwise Ranking with Batching and Caching. 1064-1072 - Jialin Ouyang:
TreeCut: A Synthetic Unanswerable Math Word Problem Dataset for LLM Hallucination Evaluation. 1073-1085 - Zheng Hui, Yinheng Li, Dan Zhao, Colby R. Banbury, Tianyi Chen, Kazuhito Koishida:
WinSpot: GUI Grounding Benchmark with Multimodal Large Language Models. 1086-1096 - Fardin Ahsan Sakib, Ziwei Zhu, Karen Trister Grace, Meliha Yetisgen, Özlem Uzuner:
Spurious Correlations and Beyond: Understanding and Mitigating Shortcut Learning in SDOH Extraction with Large Language Models. 1097-1106 - Taku Oi, Makoto Miwa:
Enhancing NER by Harnessing Multiple Datasets with Conditional Variational Autoencoders. 1107-1117 - Phan Anh Duong, Cat Luong, Divyesh Bommana, Tianyu Jiang:
CHEER-Ekman: Fine-grained Embodied Emotion Classification. 1118-1131 - Ekta Sood, Prajit Dhar, Enrica Troiano, Rosy Southwell, Sidney K. D'Mello:
ScanEZ: Integrating Cognitive Models with Self-Supervised Learning for Spatiotemporal Scanpath Prediction. 1132-1142 - Haoyuan Li, Rui Zhang, Snigdha Chaturvedi:
Improving Fairness of Large Language Models in Multi-document Summarization. 1143-1154 - Yue Wu, Yangmin Huang, Qianyun Du, Lixian Lai, Zhiyang He, Jiaxue Hu, Xiaodong Tao:
Should I Believe in What Medical AI Says? A Chinese Benchmark for Medication Based on Knowledge and Reasoning. 1155-1164 - Takumi Goto, Yusuke Sakai, Taro Watanabe:
Rethinking Evaluation Metrics for Grammatical Error Correction: Why Use a Different Evaluation Process than Human? 1165-1172 - Chengwei Qin, Wenxuan Zhou, Karthik Abinav Sankararaman, Nanshu Wang, Tengyu Xu, Alexander Radovic, Eryk Helenowski, Arya Talebzadeh, Aditya Tayade, Sinong Wang, Shafiq Joty, Han Fang, Hao Ma:
Learning Auxiliary Tasks Improves Reference-Free Hallucination Detection in Open-Domain Long-Form Generation. 1173-1182 - Ahmed Elhady, Eneko Agirre, Mikel Artetxe:
WiCkeD: A Simple Method to Make Multiple Choice Benchmarks More Challenging. 1183-1192 - Nathaniel Krasner, Nicholas Lanuzo, Antonios Anastasopoulos:
Cross-Lingual Representation Alignment Through Contrastive Image-Caption Tuning. 1193-1199 - Zhifan Ye, Zheng Wang, Kejing Xia, Jihoon Hong, Leshu Li, Lexington Whalen, Cheng Wan, Yonggan Fu, Yingyan Celine Lin, Souvik Kundu:
LAMB: A Training-Free Method to Enhance the Long-Context Understanding of SSMs via Attention-Guided Token Filtering. 1200-1209 - Jongho Kim, Seung-won Hwang:
Counterfactual-Consistency Prompting for Relative Temporal Understanding in Large Language Models. 1210-1225

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.