TY - GEN
T1 - Adversarial Testing of LLMs Across Multiple Languages
AU - Kumar, Y.
AU - Paredes, C.
AU - Yang, G.
AU - Li, J. J.
AU - Morreale, P.
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - This study builds on prior research in the field of jailbreaking, focusing on the vulnerabilities of state-of-the-art Large Language Models (LLMs) and their associated chatbots. The primary objective is to evaluate these vulnerabilities through a multilingual lens, extending the scope of previous monolingual studies. The main chatbots under examination include ChatGPT (both legacy ChatGPT -3.5 and the latest ChatGPT -40 model), Gemini, Microsoft Copilot, and Perplexity. Researchers conducted multilingual adversarial attacks, facilitating cross-language and cross-model comparisons, and explored different modalities, such as text versus speech inputs. The findings reveal significant weaknesses in these major models, particularly in their susceptibility to adversarial attacks conducted in languages such as Spanish, Russian, and Traditional Chinese. Given the global proliferation and accessibility of these models, it is imperative to rigorously assess the robustness of LLMs against adversarial inputs across multiple languages and modalities.
AB - This study builds on prior research in the field of jailbreaking, focusing on the vulnerabilities of state-of-the-art Large Language Models (LLMs) and their associated chatbots. The primary objective is to evaluate these vulnerabilities through a multilingual lens, extending the scope of previous monolingual studies. The main chatbots under examination include ChatGPT (both legacy ChatGPT -3.5 and the latest ChatGPT -40 model), Gemini, Microsoft Copilot, and Perplexity. Researchers conducted multilingual adversarial attacks, facilitating cross-language and cross-model comparisons, and explored different modalities, such as text versus speech inputs. The findings reveal significant weaknesses in these major models, particularly in their susceptibility to adversarial attacks conducted in languages such as Spanish, Russian, and Traditional Chinese. Given the global proliferation and accessibility of these models, it is imperative to rigorously assess the robustness of LLMs against adversarial inputs across multiple languages and modalities.
KW - Large Language Models (LLMs)
KW - Multi-Language Jailbreaking
KW - Multimodal Jailbreaking
KW - Robust Testing
UR - http://www.scopus.com/inward/record.url?scp=85203633819&partnerID=8YFLogxK
U2 - 10.1109/ISNCC62547.2024.10758949
DO - 10.1109/ISNCC62547.2024.10758949
M3 - Conference contribution
AN - SCOPUS:85203633819
T3 - 2024 International Symposium on Networks, Computers and Communications, ISNCC 2024
BT - 2024 International Symposium on Networks, Computers and Communications, ISNCC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2024 International Symposium on Networks, Computers and Communications, ISNCC 2024
Y2 - 22 October 2024 through 25 October 2024
ER -