Wouldn't it be great if music creators had someone to brainstorm with, help them when they're stuck, and explore different musical directions together? Researchers of KAIST and Carnegie Mellon University (CMU) have developed AI technology similar to a fellow songwriter who helps create music.
KAIST (President Kwang-Hyung Lee) has developed an AI-based music creation support system, Amuse, by a research team led by Professor Sung-Ju Lee of the School of Electrical Engineering in collaboration with CMU. The research was presented at the ACM Conference on Human Factors in Computing Systems (CHI), one of the world’s top conferences in human-computer interaction, held in Yokohama, Japan from April 26 to May 1. It received the Best Paper Award, given to only the top 1% of all submissions.
< (From left) Professor Chris Donahue of Carnegie Mellon University, Ph.D. Student Yewon Kim and Professor Sung-Ju Lee of the School of Electrical Engineering >
The system developed by Professor Sung-Ju Lee’s research team, Amuse, is an AI-based system that converts various forms of inspiration such as text, images, and audio into harmonic structures (chord progressions) to support composition.
For example, if a user inputs a phrase, image, or sound clip such as “memories of a warm summer beach”, Amuse automatically generates and suggests chord progressions that match the inspiration.
Unlike existing generative AI, Amuse is differentiated in that it respects the user's creative flow and naturally induces creative exploration through an interactive method that allows flexible integration and modification of AI suggestions.
The core technology of the Amuse system is a generation method that blends two approaches: a large language model creates music code based on the user's prompt and inspiration, while another AI model, trained on real music data, filters out awkward or unnatural results using rejection sampling.
< Figure 1. Amuse system configuration. After extracting music keywords from user input, a large language model-based code progression is generated and refined through rejection sampling (left). Code extraction from audio input is also possible (right). The bottom is an example visualizing the chord structure of the generated code. >
The research team conducted a user study targeting actual musicians and evaluated that Amuse has high potential as a creative companion, or a Co-Creative AI, a concept in which people and AI collaborate, rather than having a generative AI simply put together a song.
The paper, in which a Ph.D. student Yewon Kim and Professor Sung-Ju Lee of KAIST School of Electrical and Electronic Engineering and Carnegie Mellon University Professor Chris Donahue participated, demonstrated the potential of creative AI system design in both academia and industry.
※ Paper title: Amuse: Human-AI Collaborative Songwriting with Multimodal Inspirations
DOI: https://doi.org/10.1145/3706598.3713818
※ Research demo video: https://youtu.be/udilkRSnftI?si=FNXccC9EjxHOCrm1
※ Research homepage: https://nmsl.kaist.ac.kr/projects/amuse/
Professor Sung-Ju Lee said, “Recent generative AI technology has raised concerns in that it directly imitates copyrighted content, thereby violating the copyright of the creator, or generating results one-way regardless of the creator’s intention. Accordingly, the research team was aware of this trend, paid attention to what the creator actually needs, and focused on designing an AI system centered on the creator.”
He continued, “Amuse is an attempt to explore the possibility of collaboration with AI while maintaining the initiative of the creator, and is expected to be a starting point for suggesting a more creator-friendly direction in the development of music creation tools and generative AI systems in the future.”
This research was conducted with the support of the National Research Foundation of Korea with funding from the government (Ministry of Science and ICT). (RS-2024-00337007)
Music, often referred to as the universal language, is known to be a common component in all cultures. Then, could ‘musical instinct’ be something that is shared to some degree despite the extensive environmental differences amongst cultures? On January 16, a KAIST research team led by Professor Hawoong Jung from the Department of Physics announced to have identified the principle by which musical instincts emerge from the human brain without special learning using an artificial neu
2024-01-23< Photo 1. Conductor and Pianist João Carlos Martins before the Recital at the Carnegie Hall preparing with his bionic gloves > KAIST’s neuroscientist and professor, Dr. Daesoo Kim attended the “Conference for Musicians with Dystonia” supported by the World Health Organization (WHO) and the Carnegie Hall concert of legendary pianist João Carlos Martins, who is also a dystonia patient, to announce his team’s recent advancements toward finding a cure
2022-12-27Decline in human mobility has stunning consequences for content streaming The Covid-19 pandemic and lockdowns significantly reduced the consumption of audio music streaming in many countries as people turned to video platforms. On average, audio music consumption decreased by 12.5% after the World Health Organization’s (WHO) pandemic declaration in March 2020. Music streaming services were an unlikely area hit hard by the Covid-19 pandemic. New research in Marketing Science found that
2022-02-15In commemoration of the 6th anniversary of the establishment of the Graduate School of Cultural Technology, KAIST organized an English musical show on space at the Auditorium on the 29th and 30th of September. The name of the musical was NARO. The musical was funded by the ‘NaDa Center’ operated by KAIST’s Graduate School of Cultural Technology. The musical was created with participation from adolescents, which told a tale about a genius boy Naro’s journey in space
2011-10-10Renowned musicians in five international locations perform new contemporary music works for peace through a real-time performance on the internet. Local audiences in Seoul, Banff, New York, San Diego and Belfast will also have a chance to hear a program. In Seoul, the "International Telematic Music Concert for Peace" will be held at the LeeHaeRang Art Theater, Dongguk University, in Seoul on Nov. 20 at 9:30 a.m., under the presentation of KAIST"s Graduate School of Culture Technology and MART
2009-11-19