is a group of researchers who study, analyze, and build various aspects of AI (including social) systems. Our work spans several areas - Applied Machine Learning, Responsible and Safe AI, Natural Language Processing, and Social Network Analysis. By understanding and measuring AI systems, we aim to develop solutions that contribute to the greater good of society.
Large Language Models (LLMs) often inherit biases from the web data they are trained on, which contains stereotypes and prejudices. Current methods for evaluating and mitigating these biases rely on bias-benchmark datasets. These benchmarks measure bias by observing an LLM’s behavior on biased statements. However, these statements lack contextual considerations of the situations they try to present. To address this, we introduce a contextual reliability framework, which evaluates model robustness to biased statements by considering the various contexts in which they may appear. We develop the Context-Oriented Bias Indicator and Assessment Score (COBIAS) to measure a biased statement’s reliability in detecting bias based on the variance in model behavior across different contexts. To evaluate the metric, we augment 2,291 stereotyped statements from two existing benchmark datasets by adding contextual information. We show that COBIAS aligns with human judgment on the contextual reliability of biased statements and can be used to create reliable datasets, which would assist bias mitigation works.
@inproceedings{govil2025cobias,title={COBIAS: Assessing the Contextual Reliability of Bias Benchmarks for Language Models},author={Govil, Priyanshul and Jain, Hemang and Bonagiri, Vamshi and Chadha, Aman and Kumaraguru, Ponnurangam and Gaur, Manas and Dey, Sanorita},year={2025},booktitle={Proceedings of the 17th ACM Web Science Conference 2025},}
WebSci
Framing the Fray: Conflict Framing in Indian Election News Coverage
In covering elections, journalists often use conflict frames which depict events and issues as adversarial, often highlighting confrontations between opposing parties. Although conflict frames result in more citizen engagement, they may distract from substantive policy discussion. In this work, we analyze the use of conflict frames in online English-language news articles by seven major news outlets in the 2014 and 2019 Indian general elections. We find that the use of conflict frames is not linked to the news outlets’ ideological biases but is associated with TV-based (rather than print-based) media. Further, the majority of news outlets do not exhibit ideological biases in portraying parties as aggressors or targets in articles with conflict frames. Finally, comparing news articles reporting on political speeches to their original speech transcripts, we find that, on average, news outlets tend to consistently report on attacks on the opposition party in the speeches but under-report on more substantive electoral issues covered in the speeches such as farmers’ issues and infrastructure.
@inproceedings{chebroluframing,title={Framing the Fray: Conflict Framing in Indian Election News Coverage},author={Chebrolu, Tejasvi and Modepalle, Rohan and Vardhan, Harsha and Rajadesingan, Ashwin and Kumaraguru, Ponnurangam},year={2025},booktitle={Proceedings of the 17th ACM Conference on Web Science},}
WebSci
Personal Narratives Empower Politically Disinclined Individuals to Engage in Political Discussions
Tejasvi
Chebrolu, Ashwin
Rajadesingan, and Ponnurangam
Kumaraguru
In Proceedings of the 17th ACM Conference on Web Science, 2025
Engaging in political discussions is crucial in democratic societies, yet many individuals remain politically disinclined due to various factors such as perceived knowledge gaps, conflict avoidance, or a sense of disconnection from the political system. In this paper, we explore the potential of personal narratives—short, first-person accounts emphasizing personal experiences—as a means to empower these individuals to participate in online political discussions. Using a text classifier that identifies personal narratives, we conducted a large-scale computational analysis to evaluate the relationship between the use of personal narratives and participation in political discussions on Reddit. We find that politically disinclined individuals (PDIs) are more likely to use personal narratives than more politically active users. Personal narratives are more likely to attract and retain politically disinclined individuals in political discussions than other comments. Importantly, personal narratives posted by politically disinclined individuals are received more positively than their other comments in political communities. These results emphasize the value of personal narratives in promoting inclusive political discourse.
@inproceedings{chebrolunarrative,title={Personal Narratives Empower Politically Disinclined Individuals to Engage in Political Discussions},author={Chebrolu, Tejasvi and Rajadesingan, Ashwin and Kumaraguru, Ponnurangam},year={2025},booktitle={Proceedings of the 17th ACM Conference on Web Science},}
SIFM@ICLR
Great Models Think Alike and this Undermines AI Oversight
Shashwat
Goel, Joschka
Struber, Ilze Amanda
Auzina, Karuna K
Chandra, Ponnurangam
Kumaraguru, Douwe
Kiela, Ameya
Prabhu, Matthias
Bethge, and Jonas
Geiping
In ICLR Workshop on Self-Improving Foundation Models, 2025
As Language Model (LM) capabilities advance, evaluating and supervising them at scale is getting harder for humans. There is hope that other language models can automate both these tasks, which we refer to as "AI Oversight". We study how model similarity affects both aspects of AI oversight by proposing a probabilistic metric for LM similarity based on overlap in model mistakes. Using this metric, we first show that LLM-as-a-judge scores favor models similar to the judge, generalizing recent self-preference results. Then, we study training on LM annotations, and find complementary knowledge between the weak supervisor and strong student model plays a crucial role in gains from "weak-to-strong generalization". As model capabilities increase, it becomes harder to find their mistakes, and we might defer more to AI oversight. However, we observe a concerning trend – model mistakes are becoming more similar with increasing capabilities, pointing to risks from correlated failures. Our work underscores the importance of reporting and correcting for model similarity, especially in the emerging paradigm of AI oversight.
@inproceedings{goel2025greatmodelsthinkalike,title={Great Models Think Alike and this Undermines AI Oversight},author={Goel, Shashwat and Struber, Joschka and Auzina, Ilze Amanda and Chandra, Karuna K and Kumaraguru, Ponnurangam and Kiela, Douwe and Prabhu, Ameya and Bethge, Matthias and Geiping, Jonas},year={2025},booktitle={ICLR Workshop on Self-Improving Foundation Models},}
Thank you for your interest in joining our team! We are always looking for talented and motivated individuals to join our team. If you are interested in working with us, please
apply here.