Adversarial Testing of LLMs Across Multiple Languages

Y. Kumar, C. Paredes, G. Yang, J. J. Li, P. Morreale

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

2 Scopus citations

Abstract

This study builds on prior research in the field of jailbreaking, focusing on the vulnerabilities of state-of-the-art Large Language Models (LLMs) and their associated chatbots. The primary objective is to evaluate these vulnerabilities through a multilingual lens, extending the scope of previous monolingual studies. The main chatbots under examination include ChatGPT (both legacy ChatGPT -3.5 and the latest ChatGPT -40 model), Gemini, Microsoft Copilot, and Perplexity. Researchers conducted multilingual adversarial attacks, facilitating cross-language and cross-model comparisons, and explored different modalities, such as text versus speech inputs. The findings reveal significant weaknesses in these major models, particularly in their susceptibility to adversarial attacks conducted in languages such as Spanish, Russian, and Traditional Chinese. Given the global proliferation and accessibility of these models, it is imperative to rigorously assess the robustness of LLMs against adversarial inputs across multiple languages and modalities.

Original languageEnglish
Title of host publication2024 International Symposium on Networks, Computers and Communications, ISNCC 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798350364910
DOIs
StatePublished - 2024
Event2024 International Symposium on Networks, Computers and Communications, ISNCC 2024 - Washington, United States
Duration: 22 Oct 202425 Oct 2024

Publication series

Name2024 International Symposium on Networks, Computers and Communications, ISNCC 2024

Conference

Conference2024 International Symposium on Networks, Computers and Communications, ISNCC 2024
Country/TerritoryUnited States
CityWashington
Period22/10/2425/10/24

Keywords

  • Large Language Models (LLMs)
  • Multi-Language Jailbreaking
  • Multimodal Jailbreaking
  • Robust Testing

Fingerprint

Dive into the research topics of 'Adversarial Testing of LLMs Across Multiple Languages'. Together they form a unique fingerprint.

Cite this