Mind the Value-Action Gap: Do LLMs Act in Alignment with Their Values?
EMNLP Main Conference 🏆 Outstanding Paper Award, 2025
Humans often act in ways that are inconsistent with their values. We hypothesize that this is also true for LLMs.
EMNLP Main Conference 🏆 Outstanding Paper Award, 2025
Humans often act in ways that are inconsistent with their values. We hypothesize that this is also true for LLMs.
COLM, 2025
A framework for understanding how LLMs can better align their knowledge with users.
AAAI Spring Symposium, 2024
Do social media users realize that they are interacting with bots? Short answer: probably not.
IEEE Access, 2025
A method for quantifying uncertainty in LLM classifications using ensemble techniques.
Computers & Security, 2024
Which web browser has the best default privacy settings?