Ensuring Stability in Large Language Models: Countering Adversarial Attacks and Input Perturbations
Keywords:
Large Language Models, Robustness, Adversarial Attacks, Input Perturbations, Adversarial Training, Robust Optimization, Input Preprocessing, VulnerabilitiesAbstract
In recent years, large language models (LLMs) have demonstrated remarkable capabilities across various natural language processing tasks. However, their susceptibility to adversarial attacks and input perturbations poses significant challenges to their robustness and reliability. This paper presents an investigation into methods aimed at ensuring stability in LLMs by countering adversarial attacks and input perturbations. By addressing the critical issue of stability, this research contributes to the advancement of dependable and trustworthy large language models in real-world applications.
Downloads
Published
01-05-2024
Issue
Section
Articles
How to Cite
Ensuring Stability in Large Language Models: Countering Adversarial Attacks and Input Perturbations. (2024). Asian American Research Letters Journal, 1(1). https://aarlj.com/index.php/AARLJ/article/view/11