zsquaredz's picture
Update README.md
2cc7ffe verified
|
raw
history blame
1.73 kB
metadata
license: cc-by-nc-4.0
datasets:
  - togethercomputer/RedPajama-Data-V2
base_model:
  - meta-llama/Llama-3.1-8B-Instruct

TopK Transcoder Based on Llama 3.1 8B Instruct

This repository provides the TopK transcoder checkpoints used in the paper “Verifying Chain-of-Thought Reasoning via Its Computational Graph”.
The model is based on Llama 3.1 8B Instruct and trained with the TopK transcoder method described in the paper.

Installation

To run the model, you need the Circuit Tracer library.
It can be installed from the project page:

https://github.com/zsquaredz/circuit-tracer

Note that this is a fork of the original library as they don't yet support TopK transcoder.

After installing the library, you can load and run the transcoder as shown below.

Minimal Usage Example

from circuit_tracer import ReplacementModel
import torch

# Load transcoders into a ReplacementModel
model = ReplacementModel.from_pretrained("meta-llama/Llama-3.1-8B-Instruct", "facebook/crv-8b-instruct-transcoders", dtype=torch.bfloat16)

Once you have loaded the model, you can perform attribution or intervention as shown in this demo.

Citation

If you use this model, please cite our paper:

@article{zhao2025verifying,
      title={Verifying Chain-of-Thought Reasoning via Its Computational Graph},
      author={Zheng Zhao and Yeskendir Koishekenov and Xianjun Yang and Naila Murray and Nicola Cancedda},
      year={2025},
      eprint={2510.09312},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2510.09312},
}