Publications | Natural Language Processing
Preprints
-
ChartGemma: Visual Instruction-tuning for Chart Reasoning in the Wild
Ahmed Masry*, Megh Thakkar*, Aayush Bajaj, Aaryaman Kartha, Enamul Hoque, and Shafiq Joty
In arXiv, 2024.
#NLP
[arXiv], [code] -
Exploring the Plasticity of Neural Network for NLP Tasks in Continual Learning
Maryam Hashemzadeh, Pranshu Malviya*, Darshan Patil*, and Sarath Chandar
Conference on Lifelong Learning Agents (CoLLAs) workshop, 2024.
#DL, #NLP
-
Interpretability Needs a New Paradigm
Andreas Madsen, Himabindu Lakkaraju, Siva Reddy, and Sarath Chandar
In ArXiv, 2024.
#NLP, #DL, #Other
[arXiv]
Conference and Journal Papers
2024
-
Exploring Quantization for Efficient Pre-Training of Transformer Language Models
Kamran Chitsaz, Quentin Fournier, Gonçalo Mordido, and Sarath Chandar
Findings of the Association for Computational Linguistics (EMNLP), 2024.
#NLP, #DL
[arXiv] -
Context-Aware Assistant Selection for Improved Inference Acceleration with Large Language Models
Jerry Huang, Prasanna Parthasarathi, Mehdi Rezagholizadeh, and Sarath Chandar
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2024.
#NLP
[arXiv] -
Do Large Language Models Know How Much They Know?
Gabriele Prato, Jerry Huang, Prasanna Parthasarathi, Shagun Sodhani, and Sarath Chandar
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2024.
#NLP
-
WorkArena++: Towards Compositional Planning and Reasoning-based Common Knowledge Work Tasks
Leo Boisvert*, Megh Thakkar*, Maxime Gasse, Massimo Caccia, Thibault Le Sellier De Chezelles, Quentin Cappart, Nicolas Chapados, Alexandre Lacoste, and Alexandre Drouin
Neural Information Processing Systems (NeurIPS), Datasets and Benchmarks, 2024.
#NLP
[arXiv], [code] -
Should We Attend More or Less? Modulating Attention for Fairness
Abdelrahman Zayed, Gonçalo Mordido, Samira Shabanian, and Sarath Chandar
Conference on Language Modeling (COLM), 2024.
#NLP
[arXiv] -
Are self-explanations from Large Language Models faithful?
Andreas Madsen, Sarath Chandar, and Siva Reddy
Findings of the Association for Computational Linguistics (ACL), 2024.
#NLP
[arXiv], [code] -
A deep-dive into the tradeoffs of preference alignment with PEFT
Megh Thakkar, Quentin Fournier, Matthew Riemer, Pin-Yu Chen, Amal Zouaq, Payel Das, and Sarath Chandar
Annual Meeting of the Association for Computational Linguistics (ACL), 2024.
#NLP
[arXiv] -
Why Don’t Prompt-Based Fairness Metrics Correlate?
Abdelrahman Zayed, Gonçalo Mordido, Ioana Baldini, and Sarath Chandar
Annual Meeting of the Association for Computational Linguistics (ACL), 2024.
#NLP
[arXiv], [YouTube] -
Sub-goal Distillation: A Method to Improve Small Language Agents
Maryam Hashemzadeh, Elias Stengel-Eskin, Sarath Chandar, and Marc-Alexandre Cote
Conference on Lifelong Learning Agents (CoLLAs), 2024. [Oral presentation.]
#RL, #NLP
[arXiv] -
Faithfulness Measurable Masked Language Models
Andreas Madsen, Siva Reddy, and Sarath Chandar
International Conference on Machine Learning (ICML), 2024. [Spotlight award - top 3.5%]
#NLP
[arXiv], [code], [YouTube], [blogpost] -
MVP: Minimal Viable Phrase for Long Text Understanding
Louis Clouâtre, Amal Zouaq, and Sarath Chandar
Joint International Conference on Computational Linguistics, Language, Resources and Evaluation (LREC-COLING), 2024.
#NLP
-
Fairness-Aware Structured Pruning in Transformers
Abdelrahman Zayed, Gonçalo Mordido, Samira Shabanian, Ioana Baldini, and Sarath Chandar
AAAI Conference on Artificial Intelligence (AAAI), 2024.
#NLP
[arXiv], [YouTube]
2023
-
Self-Influence Guided Data Reweighting for Language Model Pre-training
Megh Thakkar, Tolga Bolukbasi, Sriram Ganapathy, Shikhar Vashishth, Sarath Chandar, and Partha Talukdar
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023.
#NLP
[arXiv] -
EpiK-Eval: Evaluation for Language Models as Epistemic Models
Gabriele Prato, Jerry Huang, Prasanna Parthasarathi, Shagun Sodhani, and Sarath Chandar
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023.
#NLP
[arXiv], [code] -
Measuring the Knowledge Acquisition-Utilization Gap in Pretrained Language Models
Amirhossein Kazemnejad, Mehdi Rezagholizadeh, Prasanna Parthasarathi, and Sarath Chandar
Findings of the Association for Computational Linguistics (EMNLP), 2023.
#NLP
[arXiv] -
Deep Learning on a Healthy Data Diet: Finding Important Examples for Fairness
Abdelrahman Zayed, Prasanna Parthasarathi, Gonçalo Mordido, Hamid Palangi, Samira Shabanian, and Sarath Chandar
AAAI Conference on Artificial Intelligence (AAAI), 2023.
#NLP
[arXiv], [YouTube]
2022
-
Evaluating the Faithfulness of Importance Measures in NLP by Recursively Masking Allegedly Important Tokens and Retraining
Andreas Madsen, Nicholas Meade, Vaibhav Adlakha, and Siva Reddy
Findings of the Association for Computational Linguistics (EMNLP), 2022.
[BlackboxNLP Workshop, 2022]
#NLP
[arXiv], [code] -
Detecting Languages Unintelligible to Multilingual Models through Local Structure Probes
Louis Clouâtre, Prasanna Parthasarathi, Amal Zouaq, and Sarath Chandar
Findings of the Association for Computational Linguistics (EMNLP), 2022.
#NLP
-
Local Structure Matters Most in Most Languages
Louis Clouâtre, Prasanna Parthasarathi, Amal Zouaq, and Sarath Chandar
Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the International Joint Conference on Natural Language Processing (AACL-IJCNLP), 2022.
#NLP
-
Post-hoc Interpretability for Neural NLP: A Survey
Andreas Madsen, Siva Reddy, and Sarath Chandar
ACM Computing Surveys, 2022.
#NLP
[arXiv] -
Local Structure Matters Most: Perturbation Study in NLU
Louis Clouâtre, Prasanna Parthasarathi, Amal Zouaq, and Sarath Chandar
Findings of the Association for Computational Linguistics (ACL), 2022.
#NLP
[arXiv]
2021
-
MLMLM: Link Prediction with Mean Likelihood Masked Language Model
Louis Clouâtre, Philippe Trempe, Amal Zouaq, and Sarath Chandar
Findings of the Association for Computational Linguistics (ACL-IJCNLP), 2021.
#NLP
[arXiv] -
Benchmarking Bias Mitigation Algorithms in Representation Learning through Fairness Metrics
Charan Reddy, Deepak Sharma, Soroush Mehri, Adriana Romero, Samira Shabanian, and Sina Honari
Proceedings of the Neural Information Processing Systems (NeurIPS) Track on Datasets and Benchmarks, 2021.
#NLP
[openreview], [code] -
A Brief Study on the Effects of Training Generative Dialogue Models with a Semantic loss
Prasanna Parthasarathi, Mohamed Abdelsalam, Joelle Pineau, and Sarath Chandar
Proceedings of the 22nd Annual SIGdial Meeting on Discourse and Dialogue, 2021.
#NLP
-
Do Encoder Representations of Generative Dialogue Models Encode Sufficient Information about the Task ?
Prasanna Parthasarathi, Sarath Chandar, and Joelle Pineau
Proceedings of the 22nd Annual SIGdial Meeting on Discourse and Dialogue, 2021.
#NLP
-
A Survey of Data Augmentation Approaches for NLP
Steven Y. Feng, Varun Gangal, Jason Wei, Sarath Chandar, Soroush Vosoughi, Teruko Mitamura, and Eduard Hovy
Findings of the Association for Computational Linguistics (ACL-IJCNLP), 2021.
#NLP
[arXiv]