Popular repositories Loading
-
-
-
Analyzing-Attention-Head-Specialization-in-Transformer-Language-Models
Analyzing-Attention-Head-Specialization-in-Transformer-Language-Models PublicPerformed head-level interpretability analysis on Transformer models using masking experiments. Evaluated attention head contribution through accuracy and logit-based metrics (91% baseline accuracy).
Jupyter Notebook
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.