Post
π Today's pick in Interpretability & Analysis of LMs: LLMCheckup: Conversational Examination of Large Language Models via Interpretability Tools by
@qiaw99
@tanikina
@nfel
et al.
Authors introduce LLMCheckup, a conversational interface connecting an LLM to several interpretability tools (feature attribution methods, similarity, counterfactual/rationale generation) allowing users to inquire about LLM predictions using natural language. The interface consolidates several interpretability methods in a unified chat interface, simplifying future investigations into natural language explanations.
π Paper: LLMCheckup: Conversational Examination of Large Language Models via Interpretability Tools (2401.12576)
π» Code: https://github.com/DFKI-NLP/LLMCheckup
π₯ Demo video: https://www.youtube.com/watch?v=ZwN8ZQSXoOU
Authors introduce LLMCheckup, a conversational interface connecting an LLM to several interpretability tools (feature attribution methods, similarity, counterfactual/rationale generation) allowing users to inquire about LLM predictions using natural language. The interface consolidates several interpretability methods in a unified chat interface, simplifying future investigations into natural language explanations.
π Paper: LLMCheckup: Conversational Examination of Large Language Models via Interpretability Tools (2401.12576)
π» Code: https://github.com/DFKI-NLP/LLMCheckup
π₯ Demo video: https://www.youtube.com/watch?v=ZwN8ZQSXoOU