© Neuronpedia 2026
    Privacy & TermsBlogGitHubSlackTwitterContact
    Neuronpedia logo - a computer chip with a rounded viewfinder border around it

    Neuronpedia

    Natural Language
    Autoencoders
    NEW
    Assistant AxisNEWCircuit TracerUPDATESteerSAE EvalsExportsAPI Community BlogPrivacy & TermsContact
    1. Home
    2. Circuit Tracing with Interpretable Attention
    Circuit Tracing with Interpretable Attention
    OpenMOSS Team, Fudan University
    ·huggingface.co ↗
    llama-scope-2

    OpenMOSS extended Anthropic's circuit tracing work to add interpretable attention in addition to MLP transcoders, calling themComplete Replacement Models (CRMs). Neuronpedia now supports generating CRM graphs on Qwen3-1.7B.

    CRM graphs have a new node type to represent attention calledLORSA (Low-Rank Sparse Attention), which are displayed as triangle ▲ nodes to visually distinguish them from transcoder circle ⏺ nodes.

    Since CRM graphs incorporate both transcoders and LORSA, they refer to two sets of dashboards. When selecting LORSA (triangle) nodes, you'll see the LORSA dashboard, which shows attention Z patterns when hovering over top activation tokens.

    Additionally, LORSA nodes show QK tracing results under the Node Connections panel — including top marginal and pairwise (query-feature, key-feature) contributors. These tell us why a LORSA feature attends from one position to another.

    Jump To

    Jump to Source/SAE
    Jump to Feature
    INDEX
    Random Feature

    Search Explanations

    Browse

    Features in QWEN3-1.7B@0-llamascope-2-lorsa-16k-k64
    1. Hover over a feature on the left to preview its details.
    2. Click a feature to lock it and interact with it.