Despite advancements in Deep Learning, adversarial attacks remain problematic, including NLP and transformer models. This project assesses various self-attention mechanisms to enhance transformer robustness against adversarial attacks in NLP.
Comparing sequential models like RNN and LSTM with the parallel processing of transformers.
Explanation of NLP-focused adversarial attacks and their impact.
We are utilizing Yelp-polarity sentiment analysis and TextAttack for robustness evaluation.
Brief overview of Additive Attention, Paas, Linformer, SimA, SOFT, CosFormer, and TransNormer.
Exploration of Diag attention and its impact on model robustness.
Comparison of custom, GloVe, and Counter-fitting word embeddings.
Examining the influence of head number on transformer model robustness.
Introduction of ReLU Value attention and ReLU Value CosFormer for enhanced robustness.
Results of adversarial training using Textfooler method.
Discussion on the scalability of attention mechanisms in larger transformer models.