Patchscope
Unified framework for probing hidden representations in language models
CommonProductProgrammingLanguage ModelsInterpretability
Patchscope is a unified framework for probing the hidden representations of large language models (LLMs). It enables the interpretation of model behavior and the validation of its alignment with human values. By leveraging the model's own capacity to generate human-understandable text, we propose utilizing the model itself to explain its internal natural language representations. We demonstrate how the Patchscope framework can be used to answer a wide range of research questions about LLM computation. We show that prior interpretability methods based on projecting representations into the vocabulary space and intervening with LLM computation can be viewed as special instances of this framework. Furthermore, Patchscope opens new possibilities, such as using more powerful models to interpret the representations of smaller models and unlocking novel applications like self-correction and multi-hop reasoning.
Patchscope Visit Over Time
Monthly Visits
17788201
Bounce Rate
44.87%
Page per Visit
5.4
Visit Duration
00:05:32