
<(From Left) Ph.D candidate Seongryong Oh, Ph.D candidate Yoonsung Kim, Ph.D candidate Wonung Kim, Ph.D candidate Yubin Lee, M.S candidate Jiyong Jung, Professor Jongse Park, Professor Divya Mahajan, Professor Chang Hyun Park>
As recent Artificial Intelligence (AI) models’ capacity to understand and process long, complex sentences grows, the necessity for new semiconductor technologies that can simultaneously boost computation speed and memory efficiency is increasing. Amidst this, a joint research team featuring KAIST researchers and international collaborators has successfully developed a core AI semiconductor 'brain' technology based on a hybrid Transformer and Mamba structure, which was implemented for the first time in the world in a form capable of direct computation inside the memory, resulting in a four-fold increase in the inference speed of Large Language Models (LLMs) and a 2.2-fold reduction in power consumption.
KAIST (President Kwang Hyung Lee) announced on the 17th of October that the research team led by Professor Jongse Park from KAIST School of Computing, in collaboration with Georgia Institute of Technology in the United States and Uppsala University in Sweden, developed 'PIMBA,' a core technology based on 'AI Memory Semiconductor (PIM, Processing-in-Memory),' which acts as the brain for next-generation AI models.
Currently, LLMs such as ChatGPT, GPT-4, Claude, Gemini, and Llama operate based on the 'Transformer' brain structure, which sees all of the words simultaneously. Consequently, as the AI model grows and the processed sentences become longer, the computational load and memory requirements surge, leading to speed reductions and high energy consumption as major issues.
To overcome these problems with Transformer, the recently proposed sequential memory-based 'Mamba' structure introduced a method for processing information over time, increasing efficiency. However, memory bottlenecks and power consumption limits still remained.
Professor Park Jongse's research team designed 'PIMBA,' a new semiconductor structure that directly performs computations inside the memory in order to maximize the performance of the 'Transformer–Mamba Hybrid Model,' which combines the advantages of both Transformer and Mamba.
While existing GPU-based systems move data out of the memory to perform computations, PIMBA performs calculations directly within the storage device without moving the data. This minimizes data movement time and significantly reduces power consumption.

<Analysis of Post-Transformer Models and Proposal of a Problem-Solving Acceleration System>
As a result, PIMBA showed up to a 4.1-fold improvement in processing performance and an average 2.2-fold decrease in energy consumption compared to existing GPU systems.
The research outcome is scheduled to be presented on October 20th at the '58th International Symposium on Microarchitecture (MICRO 2025),' a globally renowned computer architecture conference that will be held in Seoul. It was previously recognized for its excellence by winning the Gold Prize at the '31st Samsung Humantech Paper Award.' ※Paper Title: Pimba: A Processing-in-Memory Acceleration for Post-Transformer Large Language Model Serving, DOI: 10.1145/3725843.3756121
This research was supported by the Institute for Information & Communications Technology Planning & Evaluation (IITP), the AI Semiconductor Graduate School Support Project, and the ICT R&D Program of the Ministry of Science and ICT and the IITP, with assistance from the Electronics and Telecommunications Research Institute (ETRI). The EDA tools were supported by IDEC (the IC Design Education Center).
< Group photo of the KAIST-MIT Quantum Information Winter School > “Through the KAIST-MIT Quantum Information Winter School, I was able to view research from a broader perspective. The experience of collaborating with students from various universities and majors to complete a project was very refreshing,” said Jun-hyeong Cho, a student at the KAIST School of Electrical Engineering. KAIST announced on the 16th that the Graduate School of Quantum Science and Technology suc
2026-01-16<Jae-Chul Kim, Honorary Chairman of Dongwon Group> "In the era of AI, a new future lies within the sea of data. I ask that KAIST leaps forward to become the world's No. 1 AI research group." — Jae-Chul Kim, Honorary Chairman of Dongwon Group KAIST announced on January 16th that Honorary Chairman Jae-Chul Kim of Dongwon Group has pledged an additional 5.9 billion KRW in development funds to foster Artificial Intelligence (AI) talent and strengthen research infrastructure,
2026-01-16<(Front rwo, from left) KAIST co-first author Changhyun Joo, co-first author Seongbeom Yeon, (Back row, from left) Jaeyoung Ha, Professor Himchan Cho, Jaedong Jang> Light-emitting semiconductors are used throughout everyday life in TVs, smartphones, and lighting. However, many technical barriers remain in developing environmentally friendly semiconductor materials. In particular, nanoscale semiconductors that are tens of thousands of times smaller than the width of a human hair (about 1
2026-01-14<(From Left) Distinguisehd Professor Sang Yup Lee, Dr. Gi Bae Kim, Professor Bernhard O. Palsson> “We know the genes, but not their functions.” To resolve this long-standing bottleneck in microbial research, a joint research team has proposed a cutting-edge research strategy that leverages Artificial Intelligence (AI) to drastically accelerate the discovery of microbial gene functions. KAIST announced on January 12th that a research team led by Distinguished Professor Sang
2026-01-12<Dr. Jung Won Park, (Upper Right) Professor Jeong Ho Lee, Professor Seok-Gu Kang> IDH-mutant glioma, caused by abnormalities in a specific gene (IDH), is the most common malignant brain tumor among young adults under the age of 50. It is a refractory brain cancer that is difficult to treat due to its high recurrence rate. Until now, treatment has focused primarily on removing the visible tumor mass. However, a Korean research team has discovered for the first time that normal brain cell
2026-01-09