Guarding the Integrity of Large Language Models

Authors

  • Abuelgasim Saadeldin Author

Keywords:

Large Language Models, Robustness, Adversarial Attacks, Input Perturbations, Adversarial Training, Robust Optimization, Input Preprocessing, Vulnerabilities

Abstract

Large language models (LLMs) have exhibited remarkable capabilities in various natural language processing tasks. However, their integrity is threatened by adversarial attacks and variations in input data. This paper explores methods to guard the integrity of LLMs by countering adversarial threats and addressing input variations. By focusing on guarding LLM integrity, this research contributes to the development of robust and dependable large language models suitable for real-world applications. This paper explores strategies aimed at preserving the integrity of LLMs by countering adversarial threats and accommodating input variations. We survey existing techniques for enhancing LLM robustness and propose novel methods to mitigate adversarial threats and address input variations.

Downloads

Published

01-05-2024

How to Cite

Guarding the Integrity of Large Language Models. (2024). Asian American Research Letters Journal, 1(1). https://aarlj.com/index.php/AARLJ/article/view/8

Similar Articles

1-10 of 29

You may also start an advanced similarity search for this article.