TruthTorchLM: A Comprehensive Library for Predicting Truthfulness in LLM Outputs
Duygu Nur Yaldiz, Yavuz Faruk Bakman, Sungmin Kang, Alperen Ozis, Hayrettin Eren Yildiz, Mitash Ashish Shah, Zhiqi Huang, Anoop Kumar, Alfy Samuel, Daben Liu, Sai Praneeth Karimireddy, Salman Avestimehr
November, 2025
Abstract
TruthTorchLM is an open-source library for estimating the truthfulness of LLM outputs. It unifies over 30 truthfulness methods spanning uncertainty estimation, verification, and supervised detectors, with different trade-offs in compute, model access, and grounding requirements. The toolkit integrates with HuggingFace and LiteLLM and provides interfaces for generation, calibration, evaluation, and long-form truthfulness analysis. Benchmarking on TriviaQA, GSM8K, and FactScore-Bio demonstrates practical utility for reliable deployment workflows.
Publication
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: System Demonstrations

PhD Student in Computer Science Capital One Responsible AI Fellow
My research interests include Trustworthy LLM, Continual Learning and Federated Learning.