|
|
--- |
|
|
license: cc-by-nc-4.0 |
|
|
datasets: |
|
|
- togethercomputer/RedPajama-Data-V2 |
|
|
base_model: |
|
|
- meta-llama/Llama-3.1-8B-Instruct |
|
|
--- |
|
|
|
|
|
# TopK Transcoder Based on Llama 3.1 8B Instruct |
|
|
|
|
|
This repository provides the TopK transcoder checkpoints used in the paper [**“Verifying Chain-of-Thought Reasoning via Its Computational Graph”**](https://arxiv.org/abs/2510.09312). |
|
|
The model is based on **Llama 3.1 8B Instruct** and trained with the TopK transcoder method described in the paper. |
|
|
|
|
|
## Installation |
|
|
|
|
|
To run the model, you need the Circuit Tracer library. |
|
|
It can be installed from the project page: |
|
|
|
|
|
https://github.com/zsquaredz/circuit-tracer |
|
|
|
|
|
Note that this is a fork of the original library as they don't yet support TopK transcoder. |
|
|
|
|
|
After installing the library, you can load and run the transcoder as shown below. |
|
|
|
|
|
## Minimal Usage Example |
|
|
|
|
|
```python |
|
|
from circuit_tracer import ReplacementModel |
|
|
import torch |
|
|
|
|
|
# Load transcoders into a ReplacementModel |
|
|
model = ReplacementModel.from_pretrained("meta-llama/Llama-3.1-8B-Instruct", "facebook/crv-8b-instruct-transcoders", dtype=torch.bfloat16) |
|
|
``` |
|
|
Once you have loaded the model, you can perform attribution or intervention as shown in [this demo](https://github.com/safety-research/circuit-tracer/blob/main/demos/llama_demo.ipynb). |
|
|
|
|
|
## Citation |
|
|
|
|
|
If you use this model, please cite our paper: |
|
|
|
|
|
```bibtex |
|
|
@article{zhao2025verifying, |
|
|
title={Verifying Chain-of-Thought Reasoning via Its Computational Graph}, |
|
|
author={Zheng Zhao and Yeskendir Koishekenov and Xianjun Yang and Naila Murray and Nicola Cancedda}, |
|
|
year={2025}, |
|
|
eprint={2510.09312}, |
|
|
archivePrefix={arXiv}, |
|
|
primaryClass={cs.CL}, |
|
|
url={https://arxiv.org/abs/2510.09312}, |
|
|
} |
|
|
``` |