3D Worlds from Just a Few Phone Photos
<(From Left) Ph.D candidate Jumin Lee, Ph.D candidate Woo Jae Kim, Ph.D candidate Youngju Na, Ph.D candidate Kyu Beom Han, Professor Sung-eui Yoon>
Existing 3D scene reconstructions require a cumbersome process of precisely measuring physical spaces with LiDAR or 3D scanners, or correcting thousands of photos along with camera pose information. The research team at KAIST has overcome these limitations and introduced a technology enabling the reconstruction of 3D —from tabletop objects to outdoor scenes—with just two to three ordinary photographs. The breakthrough suggests a new paradigm in which spaces captured by camera can be immediately transformed into virtual environments.
KAIST announced on November 6 that the research team led by Professor Sung-Eui Yoon from the School of Computing has developed a new technology called SHARE (Shape-Ray Estimation), which can reconstruct high-quality 3D scenes using only ordinary images, without precise camera pose information.
Existing 3D reconstruction technology has been limited by the requirement of precise camera position and orientation information at the time of shooting to reproduce 3D scenes from a small number of images. This has necessitated specialized equipment or complex calibration processes, making real-world applications difficult and slowing widespread adoption.
To solve these problems, the research team developed a technology that constructs accurate 3D models by simultaneously estimating the 3D scene and the camera orientation using just two to three standard photographs. The technology has been recognized for its high efficiency and versatility, enabling rapid and precise reconstruction in real-world environments without additional training or complex calibration processes.
While existing methods calculate 3D structures from known camera poses, SHARE autonomously extracts spatial information from images themselves and infers both camera pose and scene structure. This enables stable 3D reconstruction without shape distortion by aligning multiple images taken from different positions into a single unified space.
<Representative Image of SHARE Technology>
"The SHARE technology is a breakthrough that dramatically lowers the barrier to entry for 3D reconstruction,” said Professor Sung-Eui Yoon. “It will enable the creation of high-quality content in various industries such as construction, media, and gaming using only a smartphone camera. It also has diverse application possibilities, such as building low-cost simulation environments in the fields of robotics and autonomous driving."
<SHARE Technology, Precise Camera Information and 3D Scene Prediction Technology>
Ph.D. Candidate Youngju Na and M.S candidate Taeyeon Kim participated as co-first authors on the research. The results were presented on September 17th at the IEEE International Conference on Image Processing (ICIP 2025), where the paper received the Best Student Paper Award.
The award, given to only one paper among 643 accepted papers this year—a selection rate of 0.16 percent—once again underscores the excellent research capabilities of the KAIST research team.
Paper Title: Pose-free 3D Gaussian Splatting via Shape-Ray Estimation, DOI: https://arxiv.org/abs/2505.22978
Award Information: https://www.linkedin.com/posts/ieeeicip_congratulations-to-the-icip-2025-best-activity-7374146976449335297-6hXz
This achievement was carried out with support from the Ministry of Science and ICT's SW Star Lab Project under the task 'Development of Perception, Action, and Interaction Algorithms for Unspecified Environments for Open World Robot Services.'
'Fingerprint' Machine Learning Technique Identifies Different Bacteria in Seconds
A synergistic combination of surface-enhanced Raman spectroscopy and deep learning serves as an effective platform for separation-free detection of bacteria in arbitrary media
Bacterial identification can take hours and often longer, precious time when diagnosing infections and selecting appropriate treatments. There may be a quicker, more accurate process according to researchers at KAIST. By teaching a deep learning algorithm to identify the “fingerprint” spectra of the molecular components of various bacteria, the researchers could classify various bacteria in different media with accuracies of up to 98%.
Their results were made available online on Jan. 18 in Biosensors and Bioelectronics, ahead of publication in the journal’s April issue.
Bacteria-induced illnesses, those caused by direct bacterial infection or by exposure to bacterial toxins, can induce painful symptoms and even lead to death, so the rapid detection of bacteria is crucial to prevent the intake of contaminated foods and to diagnose infections from clinical samples, such as urine. “By using surface-enhanced Raman spectroscopy (SERS) analysis boosted with a newly proposed deep learning model, we demonstrated a markedly simple, fast, and effective route to classify the signals of two common bacteria and their resident media without any separation procedures,” said Professor Sungho Jo from the School of Computing.
Raman spectroscopy sends light through a sample to see how it scatters. The results reveal structural information about the sample — the spectral fingerprint — allowing researchers to identify its molecules. The surface-enhanced version places sample cells on noble metal nanostructures that help amplify the sample’s signals.
However, it is challenging to obtain consistent and clear spectra of bacteria due to numerous overlapping peak sources, such as proteins in cell walls. “Moreover, strong signals of surrounding media are also enhanced to overwhelm target signals, requiring time-consuming and tedious bacterial separation steps,” said Professor Yeon Sik Jung from the Department of Materials Science and Engineering.
To parse through the noisy signals, the researchers implemented an artificial intelligence method called deep learning that can hierarchically extract certain features of the spectral information to classify data. They specifically designed their model, named the dual-branch wide-kernel network (DualWKNet), to efficiently learn the correlation between spectral features. Such an ability is critical for analyzing one-dimensional spectral data, according to Professor Jo.
“Despite having interfering signals or noise from the media, which make the general shapes of different bacterial spectra and their residing media signals look similar, high classification accuracies of bacterial types and their media were achieved,” Professor Jo said, explaining that DualWKNet allowed the team to identify key peaks in each class that were almost indiscernible in individual spectra, enhancing the classification accuracies. “Ultimately, with the use of DualWKNet replacing the bacteria and media separation steps, our method dramatically reduces analysis time.”
The researchers plan to use their platform to study more bacteria and media types, using the information to build a training data library of various bacterial types in additional media to reduce the collection and detection times for new samples.
“We developed a meaningful universal platform for rapid bacterial detection with the collaboration between SERS and deep learning,” Professor Jo said. “We hope to extend the use of our deep learning-based SERS analysis platform to detect numerous types of bacteria in additional media that are important for food or clinical analysis, such as blood.”
The National R&D Program, through a National Research Foundation of Korea grant funded by the Ministry of Science and ICT, supported this research.
-PublicationEojin Rho, Minjoon Kim, Seunghee H. Cho, Bongjae Choi, Hyungjoon Park, Hanhwi Jang, Yeon Sik Jung, Sungho Jo, “Separation-free bacterial identification in arbitrary media via deepneural network-based SERS analysis,” Biosensors and Bioelectronics online January 18, 2022 (doi.org/10.1016/j.bios.2022.113991)
-ProfileProfessor Yeon Sik JungDepartment of Materials Science and EngineeringKAIST
Professor Sungho JoSchool of ComputingKAIST
Newdin Contents Donates 'Strikezon'
Newdin Contents, an online and mobile game maker, made a gift of ‘Strikezon' to KAIST on April 19. The screen game valued at 100 million KRW will be placed in the lobby of the School of Computing, enriching the diverse physical activity options for the KAIST community. The donation was made at a ceremony attended by KAIST President Sung-Chul Shin, the CEO of the Newdin, Hyo-Kyum Kim, and Head of the School of Computing Professor Myoung Ho Kim.
At the Strikezon, students can enjoy mini baseball games indoors including a batting challenge and a pitching mode indoors for free.
President Shin thanked Mr. Kim of Newdin Contents, saying the donation will be a stepping stone for possible mutual collaborations which will play a synergistic role for technological development. Mr. Kim noted, “We are very pleased to donate the program to KAIST, which is the alma mater of Joon-Mo Hwang, the developer of Strikezon.” He added that Newdin Contents will make every effort to produce advanced game products with state of the art technology.
(Photo caption:President Sung-Chul Shin hits the ball at the Strikezon on April 19.)