
<Photo 1. (From left) Ph.D. candidates Mingyoo Song and Jaehan Kim, Professor Sooel Son, (Top right) Professor Seungwon Shin, Lead Researcher Seung Ho Na>
Most major commercial Large Language Models (LLMs), such as Google’s Gemini, utilize a Mixture-of-Experts (MoE) structure. This architecture enhances efficiency by dynamically selecting and using multiple "small AI models (Expert AIs)" depending on input queries . However, KAIST research team has revealed for the first time in the world that this very structure can actually become a new security threat.
A joint research team led by Professor Seungwon Shin (School of Electrical Engineering) and Professor Sooel Son (School of Computing) announced on December 26th that they have identified an attack technique that can seriously compromise the safety of LLMs by exploiting the MoE structure. For this research, they received the Distinguished Paper Award at ACSAC 2025, one of the most prestigious international conferences in the field of information security.
ACSAC (Annual Computer Security Applications Conference) is among the most influential international academic conferences in security. This year, only two papers out of all submissions were selected as Distinguished Papers. It is highly unusual for a domestic Korean research team to achieve such a feat in the field of AI security.
In this study, the team systematically analyzed the fundamental security vulnerabilities of the MoE structure. In particular, they demonstrated that even if an attacker does not have direct access to the internal structure of a commercial LLM, the entire model can be induced to generate dangerous responses if just one maliciously manipulated "Expert Model" is distributed through open-source channels and integrated into the system.

<Figure 1. Conceptual diagram of the attack technology proposed by the research team.>
To put it simply: even if there is only one "malicious expert" mixed among normal AI experts, that specific expert may be repeatedly selected for processing harmful queries, causing the overall safety of the AI to collapse. A particularly dangerous factor highlighted was that this process causes almost no degradation in model performance, making the problem extremely difficult to detect in advance.
Experimental results showed that the attack technique proposed by the research team could increase the harmful response rate from 0% to up to 80%. They confirmed that the safety of the entire model significantly deteriorates even if only one out of many experts is "infected."
This research is highly significant as it presents the first new security threat that can occur in the rapidly expanding global open-source-based LLM development environment. Simultaneously, it suggests that verifying the "source and safety of individual expert models" is now essential—not just performance—during the AI model development process.
Professors Seungwon Shin and Sooel Son stated, "Through this study, we have empirically confirmed that the MoE structure, which is spreading rapidly for the sake of efficiency, can become a new security threat. This award is a meaningful achievement that recognizes the importance of AI security on an international level."
The study involved Ph.D. candidates Jaehan Kim and Mingyoo Song, Dr. Seung Ho Na (currently at Samsung Electronics), Professor Seungwon Shin, and Professor Sooel Son. The results were presented at ACSAC in Hawaii, USA, on December 12, 2025.

<Figure 2. Photo of the Distinguished Paper Award certificate>
This research was supported by the Korea Internet & Security Agency (KISA) and the Institute of Information & Communications Technology Planning & Evaluation (IITP) under the Ministry of Science and ICT.
< Professor Youngjin Kwon > Modern CPUs have complex structures, and in the process of handling multiple tasks simultaneously, an order-scrambling error known as a 'concurrency bug' can occur. Although this can lead to security issues, these bugs were extremely difficult to detect using conventional methods. Our university's research team has developed a world-first-level technology to automatically detect these bugs by precisely reproducing the internal operation of the CPU in a virt
2025-11-21Breakthrough Discovery Reveals How Attackers Can Remotely Manipulate User Data Without Physical Proximity DAEJEON, South Korea — In an era when recent cyberattacks on major telecommunications providers have highlighted the fragility of mobile security, researchers at the Korea Advanced Institute of Science and Technology have identified a class of previously unknown vulnerabilities that could allow remote attackers to compromise cellular networks serving billions of users worldwide. Th
2025-11-03<(From Left) Ph.D candidate Jinseo Lee, Hobin Kim, Professor Min Suk Kang> KAIST research team has made a new milestone in global security research, becoming the first Korean research team to identify a security vulnerability in Tor, the world's largest anonymous network, and propose a solution. On September 12, our university's Professor Min Suk Kang's research team from the School of Computing announced that they had received an Honorable Mention Award at the USENIX Security 2025 c
2025-09-12KAIST (President Kwang Hyung Lee) announced on the 27th of August that a research team led by Professor Hae-Won Jeon of the Graduate School of Green Growth and Sustainable Development has signed a memorandum of understanding (MOU) with the Andlinger Center for Energy and the Environment at Princeton University in the United States to promote joint research on carbon neutrality, officially launching the Net-Zero Korea (NZK) project. This project was unveiled at the World Climate Industry EXPO (
2025-08-28KAIST (President Kwang Hyung Lee) announced on the 22nd of August that, together with the National Intelligence Service, it held the “4th University Research Security Education Council Workshop” at the Academic Cultural Complex on KAIST’s main campus in Daejeon on the 21st. This 4th workshop was held under the theme of “Global Research Security,” based on the recognition that technology leakage can have serious impacts not only on the national economy and securit
2025-08-22