ICLR 2025 Workshop
Building Trust in LLMs and LLM Applications:
From Guardrails to Explainability to Regulation

Welcome!

As Large Language Models (LLMs) are rapidly adopted across diverse industries, concerns around their trustworthiness, safety, and ethical implications increasingly motivate academic research, industrial development, and legal innovation. LLMs are increasingly integrated into complex applications, where they must navigate challenges related to data privacy, regulatory compliance, and dynamic user interactions. These complex applications amplify the potential of LLMs to violate the trust of humans. Ensuring the trustworthiness of LLMs is paramount as they transition from standalone tools to integral components of real-world applications used by millions. This workshop addresses the unique challenges posed by the deployment of LLMs, ranging from guardrails to explainability to regulation and beyond. The proposed workshop will bring together researchers and practitioners from academia and industry to explore cutting-edge solutions for improving the trustworthiness of LLMs and LLM-driven applications. The workshop will feature invited talks, a panel discussion, interactive breakout discussion sessions, and poster presentations, fostering rich dialogue and knowledge exchange. We aim to bridge the gap between foundational research and the practical challenges of deploying LLMs in trustworthy, use-centric systems.

Workshop Scope:

This workshop has a broad focus, including but not limited to:

1. Metrics, benchmarks, and evaluation of trustworthy LLMs

2. Improving reliability and truthfulness of LLMs

3. Explainability and interpretability of language model responses

4. Robustness of LLMs

5. Unlearning for LLMs

6. Fairness of LLMs

7. Guardrails and regulations for LLMs

8. Error detection and correction

Organizers

Micah Goldblum
Micah Goldblum
Columbia
Ramasuri Narayanam
Ramasuri Narayanam
Adobe
Bang An
Bang An
UMD
Soumyabrata Pal
Soumyabrata Pal
Adobe
Martin Pawelczyk
Martin Pawelczyk
Harvard
HimaBindu Lakkaraju
HimaBindu Lakkaraju
Harvard
Shiv Kumar Saini
Shiv Kumar Saini
Adobe

Contact

To mail the organizers, please send an email to iclr-workshop-building-trust@googlegroups.com, and please CC the following contacts:

Correspondence:
Ramasuri Narayanam: rnarayanam@adobe.com,
Martin Pawelczyk: martin.pawelczyk.1@gmail.com