📄️ Attention Manipulation (AtMan)
AtMan is our method to manipulate the attention of an input sequence (this can be a token, a word, or a sentence) to steer the model's prediction in a different contextual direction.
📄️ Explainability
In the previous section, we explained how you can steer the attention of our models and either suppress or amplify parts of the input sequences.