
<(From Left) Ph.D candidate Seongryong Oh, Ph.D candidate Yoonsung Kim, Ph.D candidate Wonung Kim, Ph.D candidate Yubin Lee, M.S candidate Jiyong Jung, Professor Jongse Park, Professor Divya Mahajan, Professor Chang Hyun Park>
As recent Artificial Intelligence (AI) models’ capacity to understand and process long, complex sentences grows, the necessity for new semiconductor technologies that can simultaneously boost computation speed and memory efficiency is increasing. Amidst this, a joint research team featuring KAIST researchers and international collaborators has successfully developed a core AI semiconductor 'brain' technology based on a hybrid Transformer and Mamba structure, which was implemented for the first time in the world in a form capable of direct computation inside the memory, resulting in a four-fold increase in the inference speed of Large Language Models (LLMs) and a 2.2-fold reduction in power consumption.
KAIST (President Kwang Hyung Lee) announced on the 17th of October that the research team led by Professor Jongse Park from KAIST School of Computing, in collaboration with Georgia Institute of Technology in the United States and Uppsala University in Sweden, developed 'PIMBA,' a core technology based on 'AI Memory Semiconductor (PIM, Processing-in-Memory),' which acts as the brain for next-generation AI models.
Currently, LLMs such as ChatGPT, GPT-4, Claude, Gemini, and Llama operate based on the 'Transformer' brain structure, which sees all of the words simultaneously. Consequently, as the AI model grows and the processed sentences become longer, the computational load and memory requirements surge, leading to speed reductions and high energy consumption as major issues.
To overcome these problems with Transformer, the recently proposed sequential memory-based 'Mamba' structure introduced a method for processing information over time, increasing efficiency. However, memory bottlenecks and power consumption limits still remained.
Professor Park Jongse's research team designed 'PIMBA,' a new semiconductor structure that directly performs computations inside the memory in order to maximize the performance of the 'Transformer–Mamba Hybrid Model,' which combines the advantages of both Transformer and Mamba.
While existing GPU-based systems move data out of the memory to perform computations, PIMBA performs calculations directly within the storage device without moving the data. This minimizes data movement time and significantly reduces power consumption.

<Analysis of Post-Transformer Models and Proposal of a Problem-Solving Acceleration System>
As a result, PIMBA showed up to a 4.1-fold improvement in processing performance and an average 2.2-fold decrease in energy consumption compared to existing GPU systems.
The research outcome is scheduled to be presented on October 20th at the '58th International Symposium on Microarchitecture (MICRO 2025),' a globally renowned computer architecture conference that will be held in Seoul. It was previously recognized for its excellence by winning the Gold Prize at the '31st Samsung Humantech Paper Award.' ※Paper Title: Pimba: A Processing-in-Memory Acceleration for Post-Transformer Large Language Model Serving, DOI: 10.1145/3725843.3756121
This research was supported by the Institute for Information & Communications Technology Planning & Evaluation (IITP), the AI Semiconductor Graduate School Support Project, and the ICT R&D Program of the Ministry of Science and ICT and the IITP, with assistance from the Electronics and Telecommunications Research Institute (ETRI). The EDA tools were supported by IDEC (the IC Design Education Center).
< Professor Yiyun Kang (Photo Credit: Ryan Lash / TED) > KAIST announced on April 17th that Professor Yiyun Kang of the Department of Industrial Design has been selected as a speaker for the Main Stage at TED 2026, the world-renowned knowledge conference. Founded in 1984 under the motto "Ideas Worth Spreading," TED is an American non-profit knowledge platform where scholars, innovators, and artists from around the globe gather annually to lead global discourse. Previous Korean speakers
2026-04-18< (From left) Undergraduate researcher Taewon Kim and Professor Sangsik Kim > A new technology has been developed that allows light to be "designed" into desired forms, potentially making Artificial Intelligence (AI) and communication technologies faster and more accurate. A KAIST research team has developed an "integrated photonic resonator"—a core component of next-generation optical integrated circuits that process data using light. The research is particularly significant as i
2026-04-16<(From left) Photos of the KAIST Science Festival exhibition hall and booths from the previous year> KAIST announced on April 10th that KAIST will participate in the ‘2026 Korea Science and Technology Festival,’ the largest science festival in the country, to mark Science Month in April. KAIST will operate ‘KAIST Play World,’ an interactive exhibition hall showcasing the pinnacle of AI and robotics. This year’s festival will be held in two parts: ‘20
2026-04-13< (From left) Professor Gyu Rie Lee, Professor David Baker > Under the foundation of research cooperation established through the Ministry of Science and ICT's InnoCORE (InnoCORE) project, KAIST InnoCORE researchers have derived meaningful research results. Following a visit by Professor David Baker (University of Washington, USA), the 2024 Nobel Laureate in Chemistry, KAIST has revealed research findings on designing proteins that accurately recognize desired compounds using AI through
2026-04-09<(From Left) Prof. Dong-Soo Han, Dr. Kyuho Son, Dr. Byeongcheol Moon, Dr. Sumin Ahn, Ph.D candidate Seungwoo Chae> A Korean research team has developed a technology that enables precise indoor positioning using only a smartphone. Developed over eight years by KAIST researchers, this technology is expected to help secure critical time in missing-person searches and is being recognized as a “location sovereignty” solution that could reshape the current location service ecosyst
2026-04-03