Welcome!
As Large Language Models (LLMs) are rapidly adopted across diverse industries, concerns around their trustworthiness, safety, and ethical implications increasingly motivate academic research, industrial development, and legal innovation. LLMs are increasingly integrated into complex applications, where they must navigate challenges related to data privacy, regulatory compliance, and dynamic user interactions. These complex applications amplify the potential of LLMs to violate the trust of humans. Ensuring the trustworthiness of LLMs is paramount as they transition from standalone tools to integral components of real-world applications used by millions. This workshop addresses the unique challenges posed by the deployment of LLMs, ranging from guardrails to explainability to regulation and beyond. The proposed workshop will bring together researchers and practitioners from academia and industry to explore cutting-edge solutions for improving the trustworthiness of LLMs and LLM-driven applications. The workshop will feature invited talks, a panel discussion, interactive breakout discussion sessions, and poster presentations, fostering rich dialogue and knowledge exchange. We aim to bridge the gap between foundational research and the practical challenges of deploying LLMs in trustworthy, use-centric systems.
Workshop Scope:
This workshop has a broad focus, including but not limited to:
1. Metrics, benchmarks, and evaluation of trustworthy LLMs
2. Improving reliability and truthfulness of LLMs
3. Explainability and interpretability of language model responses
4. Robustness of LLMs
5. Unlearning for LLMs
6. Fairness of LLMs
7. Guardrails and regulations for LLMs
8. Error detection and correction
Organizers

Columbia

Adobe

UMD

Adobe

Harvard

Harvard

Adobe
Contact
To mail the organizers, please send an email to iclr-workshop-building-trust@googlegroups.com, and please CC the following contacts:
Correspondence:
Ramasuri Narayanam:
rnarayanam@adobe.com,
Martin Pawelczyk:
martin.pawelczyk.1@gmail.com