is a group of researchers who study, analyze, and build various aspects of AI (including social) systems. Our work spans several areas - Applied Machine Learning, Responsible and Safe AI, Natural Language Processing, and Social Network Analysis. By understanding and measuring AI systems, we aim to develop solutions that contribute to the greater good of society.
As Language Model (LM) capabilities advance, evaluating and supervising them at scale is getting harder for humans. There is hope that other language models can automate both these tasks, which we refer to as "AI Oversight". We study how model similarity affects both aspects of AI oversight by proposing a probabilistic metric for LM similarity based on overlap in model mistakes. Using this metric, we first show that LLM-as-a-judge scores favor models similar to the judge, generalizing recent self-preference results. Then, we study training on LM annotations, and find complementary knowledge between the weak supervisor and strong student model plays a crucial role in gains from "weak-to-strong generalization". As model capabilities increase, it becomes harder to find their mistakes, and we might defer more to AI oversight. However, we observe a concerning trend – model mistakes are becoming more similar with increasing capabilities, pointing to risks from correlated failures. Our work underscores the importance of reporting and correcting for model similarity, especially in the emerging paradigm of AI oversight.
@article{goel2025greatmodelsthinkalike,title={Great Models Think Alike and this Undermines AI Oversight},author={Goel, Shashwat and Struber, Joschka and Auzina, Ilze Amanda and Chandra, Karuna K and Kumaraguru, Ponnurangam and Kiela, Douwe and Prabhu, Ameya and Bethge, Matthias and Geiping, Jonas},year={2025},journal={Forty Second International Conference on Machine Learning},}
ICML
A Cognac shot to forget bad memories: Corrective Unlearning in GNNs
Graph Neural Networks (GNNs) are increasingly being used for a variety of ML applications on graph data. Because graph data does not follow the independently and identically distributed (i.i.d.) assumption, adversarial manipulations or incorrect data can propagate to other data points through message passing, which deteriorates the model’s performance. To allow model developers to remove the adverse effects of manipulated entities from a trained GNN, we study the recently formulated problem of Corrective Unlearning. We find that current graph unlearning methods fail to unlearn the effect of manipulations even when the whole manipulated set is known. We introduce a new graph unlearning method, Cognac, which can unlearn the effect of the manipulation set even when only 5% of it is identified.
@article{kolipaka2024cognacshotforgetbad,title={A Cognac shot to forget bad memories: Corrective Unlearning in GNNs},author={Kolipaka, Varshita and Sinha, Akshit and Mishra, Debangan and Kumar, Sumit and Arun, Arvindh and Goel, Shashwat and Kumaraguru, Ponnurangam},journal={Forty-Second International Conference on Machine Learning},year={2025},}
EASE
Small Models, Big Tasks: An Exploratory Empirical Study on Small Language Models for Function Calling
Ishan
Kavathekar, Raghav
Donakanti, Ponnurangam
Kumaraguru, and Karthik
Vaidhyanathan
International Conference on Evaluation and Assessment in Software Engineering (EASE), 2025, 2025
Function calling is a complex task with widespread applications in domains such as information retrieval, software engineering and automation. For example, a query to book the shortest flight from New York to London on January 15 requires identifying the correct parameters to generate accurate function calls. Large Language Models (LLMs) can automate this process but are computationally expensive and impractical in resource-constrained settings. In contrast, Small Language Models (SLMs) can operate efficiently, offering faster response times, and lower computational demands, making them potential candidates for function calling on edge devices. In this exploratory empirical study, we evaluate the efficacy of SLMs in generating function calls across diverse domains using zero-shot, few-shot, and fine-tuning approaches, both with and without prompt injection, while also providing the finetuned models to facilitate future applications. Furthermore, we analyze the model responses across a range of metrics, capturing various aspects of function call generation. Additionally, we perform experiments on an edge device to evaluate their performance in terms of latency and memory usage, providing useful insights into their practical applicability. Our findings show that while SLMs improve from zero-shot to few-shot and perform best with fine-tuning, they struggle significantly with adhering to the given output format. Prompt injection experiments further indicate that the models are generally robust and exhibit only a slight decline in performance. While SLMs demonstrate potential for the function call generation task, our results also highlight areas that need further refinement for real-time functioning.
@article{kavathekar2025smallmodelsbigtasks,title={Small Models, Big Tasks: An Exploratory Empirical Study on Small Language Models for Function Calling},author={Kavathekar, Ishan and Donakanti, Raghav and Kumaraguru, Ponnurangam and Vaidhyanathan, Karthik},year={2025},journal={International Conference on Evaluation and Assessment in Software Engineering (EASE), 2025},eprint={2504.19277},archiveprefix={arXiv},primaryclass={cs.AI},url={https://arxiv.org/abs/2504.19277}}
WebSci
COBIAS: Assessing the Contextual Reliability of Bias Benchmarks for Language Models
Priyanshul
Govil, Hemang
Jain, Vamshi
Bonagiri, Aman
Chadha, Ponnurangam
Kumaraguru, Manas
Gaur, and Sanorita
Dey
In Proceedings of the 17th ACM Web Science Conference 2025, 2025
Large Language Models (LLMs) often inherit biases from the web data they are trained on, which contains stereotypes and prejudices. Current methods for evaluating and mitigating these biases rely on bias-benchmark datasets. These benchmarks measure bias by observing an LLM’s behavior on biased statements. However, these statements lack contextual considerations of the situations they try to present. To address this, we introduce a contextual reliability framework, which evaluates model robustness to biased statements by considering the various contexts in which they may appear. We develop the Context-Oriented Bias Indicator and Assessment Score (COBIAS) to measure a biased statement’s reliability in detecting bias based on the variance in model behavior across different contexts. To evaluate the metric, we augment 2,291 stereotyped statements from two existing benchmark datasets by adding contextual information. We show that COBIAS aligns with human judgment on the contextual reliability of biased statements and can be used to create reliable datasets, which would assist bias mitigation works.
@inproceedings{govil2025cobias,title={COBIAS: Assessing the Contextual Reliability of Bias Benchmarks for Language Models},author={Govil, Priyanshul and Jain, Hemang and Bonagiri, Vamshi and Chadha, Aman and Kumaraguru, Ponnurangam and Gaur, Manas and Dey, Sanorita},year={2025},booktitle={Proceedings of the 17th ACM Web Science Conference 2025},}
WebSci
Framing the Fray: Conflict Framing in Indian Election News Coverage
In covering elections, journalists often use conflict frames which depict events and issues as adversarial, often highlighting confrontations between opposing parties. Although conflict frames result in more citizen engagement, they may distract from substantive policy discussion. In this work, we analyze the use of conflict frames in online English-language news articles by seven major news outlets in the 2014 and 2019 Indian general elections. We find that the use of conflict frames is not linked to the news outlets’ ideological biases but is associated with TV-based (rather than print-based) media. Further, the majority of news outlets do not exhibit ideological biases in portraying parties as aggressors or targets in articles with conflict frames. Finally, comparing news articles reporting on political speeches to their original speech transcripts, we find that, on average, news outlets tend to consistently report on attacks on the opposition party in the speeches but under-report on more substantive electoral issues covered in the speeches such as farmers’ issues and infrastructure.
@inproceedings{chebroluframing,title={Framing the Fray: Conflict Framing in Indian Election News Coverage},author={Chebrolu, Tejasvi and Modepalle, Rohan and Vardhan, Harsha and Rajadesingan, Ashwin and Kumaraguru, Ponnurangam},year={2025},booktitle={Proceedings of the 17th ACM Conference on Web Science},}
WebSci
Personal Narratives Empower Politically Disinclined Individuals to Engage in Political Discussions
Tejasvi
Chebrolu, Ashwin
Rajadesingan, and Ponnurangam
Kumaraguru
In Proceedings of the 17th ACM Conference on Web Science, 2025
Engaging in political discussions is crucial in democratic societies, yet many individuals remain politically disinclined due to various factors such as perceived knowledge gaps, conflict avoidance, or a sense of disconnection from the political system. In this paper, we explore the potential of personal narratives—short, first-person accounts emphasizing personal experiences—as a means to empower these individuals to participate in online political discussions. Using a text classifier that identifies personal narratives, we conducted a large-scale computational analysis to evaluate the relationship between the use of personal narratives and participation in political discussions on Reddit. We find that politically disinclined individuals (PDIs) are more likely to use personal narratives than more politically active users. Personal narratives are more likely to attract and retain politically disinclined individuals in political discussions than other comments. Importantly, personal narratives posted by politically disinclined individuals are received more positively than their other comments in political communities. These results emphasize the value of personal narratives in promoting inclusive political discourse.
@inproceedings{chebrolunarrative,title={Personal Narratives Empower Politically Disinclined Individuals to Engage in Political Discussions},author={Chebrolu, Tejasvi and Rajadesingan, Ashwin and Kumaraguru, Ponnurangam},year={2025},booktitle={Proceedings of the 17th ACM Conference on Web Science},}
Thank you for your interest in joining our team! We are always looking for talented and motivated individuals to join our team. If you are interested in working with us, please
apply here.