Ensuring Stability in Large Language Models: Countering Adversarial Attacks and Input Perturbations

Authors

  • Abuelgasim Saadeldin Author

Keywords:

Large Language Models, Robustness, Adversarial Attacks, Input Perturbations, Adversarial Training, Robust Optimization, Input Preprocessing, Vulnerabilities

Abstract

In recent years, large language models (LLMs) have demonstrated remarkable capabilities across various natural language processing tasks. However, their susceptibility to adversarial attacks and input perturbations poses significant challenges to their robustness and reliability. This paper presents an investigation into methods aimed at ensuring stability in LLMs by countering adversarial attacks and input perturbations. By addressing the critical issue of stability, this research contributes to the advancement of dependable and trustworthy large language models in real-world applications.

Downloads

Published

01-05-2024

How to Cite

Ensuring Stability in Large Language Models: Countering Adversarial Attacks and Input Perturbations. (2024). Asian American Research Letters Journal, 1(1). https://aarlj.com/index.php/AARLJ/article/view/11

Similar Articles

1-10 of 29

You may also start an advanced similarity search for this article.