Receive KAIST news by email!
Type your e-mail address here.
by recently order
by view order
Professor Juho Kim’s Team Wins Best Paper Award at ACM CHI 2022
The research team led by Professor Juho Kim from the KAIST School of Computing won a Best Paper Award and an Honorable Mention Award at the Association for Computing Machinery Conference on Human Factors in Computing Systems (ACM CHI) held between April 30 and May 6. ACM CHI is the world’s most recognized conference in the field of human computer interactions (HCI), and is ranked number one out of all HCI-related journals and conferences based on Google Scholar’s h-5 index. Best paper awards are given to works that rank in the top one percent, and honorable mention awards are given to the top five percent of the papers accepted by the conference. Professor Juho Kim presented a total of seven papers at ACM CHI 2022, and tied for the largest number of papers. A total of 19 papers were affiliated with KAIST, putting it fifth out of all participating institutes and thereby proving KAIST’s competence in research. One of Professor Kim’s research teams composed of Jeongyeon Kim (first author, MS graduate) from the School of Computing, MS candidate Yubin Choi from the School of Electrical Engineering, and Dr. Meng Xia (post-doctoral associate in the School of Computing, currently a post-doctoral associate at Carnegie Mellon University) received a best paper award for their paper, “Mobile-Friendly Content Design for MOOCs: Challenges, Requirements, and Design Opportunities”. The study analyzed the difficulties experienced by learners watching video-based educational content in a mobile environment and suggests guidelines for solutions. The research team analyzed 134 survey responses and 21 interviews, and revealed that texts that are too small or overcrowded are what mainly brings down the legibility of video contents. Additionally, lighting, noise, and surrounding environments that change frequently are also important factors that may disturb a learning experience. Based on these findings, the team analyzed the aptness of 41,722 frames from 101 video lectures for mobile environments, and confirmed that they generally show low levels of adequacy. For instance, in the case of text sizes, only 24.5% of the frames were shown to be adequate for learning in mobile environments. To overcome this issue, the research team suggested a guideline that may improve the legibility of video contents and help overcome the difficulties arising from mobile learning environments. The importance of and dependency on video-based learning continue to rise, especially in the wake of the pandemic, and it is meaningful that this research suggested a means to analyze and tackle the difficulties of users that learn from the small screens of mobile devices. Furthermore, the paper also suggested technology that can solve problems related to video-based learning through human-AI collaborations, enhancing existing video lectures and improving learning experiences. This technology can be applied to various video-based platforms and content creation. Meanwhile, a research team composed of Ph.D. candidate Tae Soo Kim (first author), MS candidate DaEun Choi, and Ph.D. candidate Yoonseo Choi from the School of Computing received an honorable mention award for their paper, “Stylette: styling the Web with Natural Language”. The research team developed a novel interface technology that allows nonexperts who are unfamiliar with technical jargon to edit website features through speech. People often find it difficult to use or find the information they need from various websites due to accessibility issues, device-related constraints, inconvenient design, style preferences, etc. However, it is not easy for laymen to edit website features without expertise in programming or design, and most end up just putting up with the inconveniences. But what if the system could read the intentions of its users from their everyday language like “emphasize this part a little more”, or “I want a more modern design”, and edit the features automatically? Based on this question, Professor Kim’s research team developed ‘Stylette’, a system in which AI analyses its users’ speech expressed in their natural language and automatically recommends a new style that best fits their intentions. The research team created a new system by putting together language AI, visual AI, and user interface technologies. On the linguistic side, a large-scale language model AI converts the intentions of the users expressed through their everyday language into adequate style elements. On the visual side, computer vision AI compares 1.7 million existing web design features and recommends a style adequate for the current website. In an experiment where 40 nonexperts were asked to edit a website design, the subjects that used this system showed double the success rate in a time span that was 35% shorter compared to the control group. It is meaningful that this research proposed a practical case in which AI technology constructs intuitive interactions with users. The developed technology can be applied to existing design applications and web browsers in a plug-in format, and can be utilized to improve websites or for advertisements by collecting the natural intention data of users on a large scale.
KAIST's Patina Engraving System Awarded at ACM CHI
Professor Tek-Jin Nam’s research team of the Industrial Design Department of KAIST received the Best Paper Award in the 2015 Association for Computing Machinery’s (ACM) Conference on Human Factors in Computing Systems (CHI) which was held from April 18 to 23, 2015. The team consisted of two KAIST students: Moon-Hwan Lee, a Ph.D. candidate, and Sejin Cha, a master's student. The team was the first in Asia to receive the award. The ACM CHI represents the premier conference in the field of Human-Computer Interaction (HCI). This year’s event, held in Seoul, South Korea, was the first conference that the ACM had held in Asia in its thirty-three year history. The KAIST team’s paper, entitled “Patina Engraver: Visualizing Activity Logs as Patina in Fashionable Trackers,” ranked in the top 1% of 2,000 submitted papers. The team developed Patina Engraver, an activity tracker, which monitors and tracks fitness-related metrics such as distances walked or run, calorie consumption, heartbeat, sleep quality, and blood pressure. The device wirelessly connects to a computer or smartphone so that it can store and utilize long-term tracking data. However, what makes Patina Engraver, a smart wristband, different from other health trackers is its ability to display different design patterns based on users’ activity on the surface of the wristband. The research team was inspired to build this system from the fact that wearable electronics including activity trackers can be used not only as health care devices, but also as fashion items to express emotions and personalities. Equipped with an engraving feature, the charging pad or holder for Patina Engraver draws individualized patterns to reflect the user’s activities, such as walking or running, while the device is being charged. The pattern display syncs with the frequency of usage, therefore, the more the tracker is used, the greater the number of patterns will show up. According to the team, since Patina Engraver provides users with a personalized illustration of their activity on the tracker, users are more motivated to put on the tracker and exercise. Professor Nam said, “This research can be applied in producing other wearable devices to enhance users’ emotional satisfaction. When wearable technology is combined with design and emotion, the industry market will quickly expand.” Figure 1: Patina engraving system developed by KAIST research team Figure 2: The process of engraving illustrations of the activity records onto the tracker Figure 3: Personalized activity trackers based on activity records
Remote Follows Your Thumb by Discovery News, May 19, 2011
The ACM CHI Conference on Human Factors in Computing Systems, an international conference of human-computer interaction, was held on May 7-12, 2011 in Vancouver, Canada. At the conference, KAIST’s research team presented a paper on the development of prototype, called "remote touch system," for manipulating a LED screen by putting user’s thumb’s shadow on a television or smart phone screen. Discovery News posted an online article on the technology, dated May 19, 2011. For the article, please copy and paste the following link in the address bar of Internet Explorer: http://news.discovery.com/tech/shadow-remote-touchscreen-110519.html?print=true
마지막 페이지 1
KAIST, 291 Daehak-ro, Yuseong-gu, Daejeon 34141, Republic of Korea
Copyright(C) 2020, Korea Advanced Institute of Science and Technology,
All Rights Reserved.