An awesome repository & A comprehensive survey on interpretability of LLM attention heads.
-
Updated
Mar 2, 2025 - TeX
8000
An awesome repository & A comprehensive survey on interpretability of LLM attention heads.
IT-Based textgeneration with the use of NLP methods. A text summarization task is conducted with the amazon fine food review dataset from Kaggle. This task is done by attention and lstm neural networks.
Master Thesis: Energy-based Multi-Modal Attention (EMMA). A novel method improving the robustness of multi-modal deep learning.
Add a description, image, and links to the attention-mechanism topic page so that developers can more easily learn about it.
To associate your repository with the attention-mechanism topic, visit your repo's landing page and select "manage topics."