
<(From Left) Ph.D candidate Seongryong Oh, Ph.D candidate Yoonsung Kim, Ph.D candidate Wonung Kim, Ph.D candidate Yubin Lee, M.S candidate Jiyong Jung, Professor Jongse Park, Professor Divya Mahajan, Professor Chang Hyun Park>
As recent Artificial Intelligence (AI) models’ capacity to understand and process long, complex sentences grows, the necessity for new semiconductor technologies that can simultaneously boost computation speed and memory efficiency is increasing. Amidst this, a joint research team featuring KAIST researchers and international collaborators has successfully developed a core AI semiconductor 'brain' technology based on a hybrid Transformer and Mamba structure, which was implemented for the first time in the world in a form capable of direct computation inside the memory, resulting in a four-fold increase in the inference speed of Large Language Models (LLMs) and a 2.2-fold reduction in power consumption.
KAIST (President Kwang Hyung Lee) announced on the 17th of October that the research team led by Professor Jongse Park from KAIST School of Computing, in collaboration with Georgia Institute of Technology in the United States and Uppsala University in Sweden, developed 'PIMBA,' a core technology based on 'AI Memory Semiconductor (PIM, Processing-in-Memory),' which acts as the brain for next-generation AI models.
Currently, LLMs such as ChatGPT, GPT-4, Claude, Gemini, and Llama operate based on the 'Transformer' brain structure, which sees all of the words simultaneously. Consequently, as the AI model grows and the processed sentences become longer, the computational load and memory requirements surge, leading to speed reductions and high energy consumption as major issues.
To overcome these problems with Transformer, the recently proposed sequential memory-based 'Mamba' structure introduced a method for processing information over time, increasing efficiency. However, memory bottlenecks and power consumption limits still remained.
Professor Park Jongse's research team designed 'PIMBA,' a new semiconductor structure that directly performs computations inside the memory in order to maximize the performance of the 'Transformer–Mamba Hybrid Model,' which combines the advantages of both Transformer and Mamba.
While existing GPU-based systems move data out of the memory to perform computations, PIMBA performs calculations directly within the storage device without moving the data. This minimizes data movement time and significantly reduces power consumption.

<Analysis of Post-Transformer Models and Proposal of a Problem-Solving Acceleration System>
As a result, PIMBA showed up to a 4.1-fold improvement in processing performance and an average 2.2-fold decrease in energy consumption compared to existing GPU systems.
The research outcome is scheduled to be presented on October 20th at the '58th International Symposium on Microarchitecture (MICRO 2025),' a globally renowned computer architecture conference that will be held in Seoul. It was previously recognized for its excellence by winning the Gold Prize at the '31st Samsung Humantech Paper Award.' ※Paper Title: Pimba: A Processing-in-Memory Acceleration for Post-Transformer Large Language Model Serving, DOI: 10.1145/3725843.3756121
This research was supported by the Institute for Information & Communications Technology Planning & Evaluation (IITP), the AI Semiconductor Graduate School Support Project, and the ICT R&D Program of the Ministry of Science and ICT and the IITP, with assistance from the Electronics and Telecommunications Research Institute (ETRI). The EDA tools were supported by IDEC (the IC Design Education Center).
<(From left) Photos of the KAIST Science Festival exhibition hall and booths from the previous year> KAIST announced on April 10th that KAIST will participate in the ‘2026 Korea Science and Technology Festival,’ the largest science festival in the country, to mark Science Month in April. KAIST will operate ‘KAIST Play World,’ an interactive exhibition hall showcasing the pinnacle of AI and robotics. This year’s festival will be held in two parts: ‘20
2026-04-13< (From left) Professor Gyu Rie Lee, Professor David Baker > Under the foundation of research cooperation established through the Ministry of Science and ICT's InnoCORE (InnoCORE) project, KAIST InnoCORE researchers have derived meaningful research results. Following a visit by Professor David Baker (University of Washington, USA), the 2024 Nobel Laureate in Chemistry, KAIST has revealed research findings on designing proteins that accurately recognize desired compounds using AI through
2026-04-09<(From Left) Prof. Dong-Soo Han, Dr. Kyuho Son, Dr. Byeongcheol Moon, Dr. Sumin Ahn, Ph.D candidate Seungwoo Chae> A Korean research team has developed a technology that enables precise indoor positioning using only a smartphone. Developed over eight years by KAIST researchers, this technology is expected to help secure critical time in missing-person searches and is being recognized as a “location sovereignty” solution that could reshape the current location service ecosyst
2026-04-03<(From Left) Professor Sang Woo Han, Researcher Jin Wook Baek> In chemical processes for producing pharmaceuticals, catalysts are the key to determine production speed and cost. However, until now, there has been a trade-off between “precise but disposable catalysts” and “reusable catalysts.” A KAIST research team has developed an eco-friendly catalytic technology that combines these two types, operating with light and air. This opens a pathway to producing pharm
2026-03-30< Integrated Operation of Heterogeneous Logistics Robot Systems > KAIST announced on March 23rd that Professor Young Jae Jang's team from the Department of Industrial and Systems Engineering has constructed ‘KAIROS’ (KAIST AI Robot Orchestration Systems), a physical AI testbed that integrates and controls heterogeneous robots, sensors, facilities, and digital twins into a single system. KAIROS is a 100% unmanned factory platform based on physical AI and is the first integr
2026-03-23