The AI model developed by the team demonstrates the ability to detect and differentiate various sounds in indoor settings
With the use of artificial intelligence (AI), acoustic recognition techniques can excel in listening and interpreting various sounds in different contexts. Now, a research team from the Gwangju Institute of Science and Technology (GIST) has made significant strides in enhancing the performance of acoustic recognition to achieve diverse audio intelligence. Their groundbreaking work bagged the top spot in the indoor acoustic event detection category at “DCASE Challenge 2023” organized by IEEE.
In our daily lives, sounds carry a large volume of information about the environment and the events that take place in the space around us. Humans can perceive the surrounding sound environment (for examples, sounds on busy streets, offices, or at home), and recognize the source of these sounds. With the advent of artificial intelligence (AI), the performance of acoustic recognition techniques can be further improved, enabling machines to listen to and interpret various sounds in different contexts.
This research has tremendous potential in several applications in multimedia search based on its audio content, context-aware mobile devices, robots, cars, intelligent surveillance and monitoring systems. However, the challenge in achieving this primarily lies in the recognition of sound scenes and individual sound sources in realistic soundscapes, where multiple sounds are simultaneously present.
To this end, a collaborative research team from GIST (with Acting President Rae-gil Park), have now excelled in developing an excellent acoustic recognition technique using AI. Their work has now bagged the top spot in the International Challenge on Detection and Classification of
“In the category of Indoor Acoustic Event Detection, the performance of AI technology was evaluated based on its ability to detect and differentiate between 10 different sounds commonly found in indoor environments, including vacuum cleaners, dishes, barking dogs, and running water,”
says Prof. Kim, while speaking about their achievement.
The GIST-Hanwha Vision team made significant strides in enhancing the performance of acoustic recognition by integrating a range of AI technologies capable of achieving diverse audio intelligence. This included semi-learning technology, which leverages answers inferred from AI; fusion technology, which combines the inference results of pre-learning training models with those of existing models; data refinement techniques, which optimizes performance; and ensemble technologies, which integrates multiple techniques to enhance AI performance.
“The experience and knowledge gained from this competition will be applied for the detection of acoustic events in the CCTV systems developed by Hanwha Vision. Furthermore, efforts will be made to develop more efficient and user-friendly services, such as a technology for detecting speech segments and acoustic events in social media content,”
remarks elated Prof. Kim.
This groundbreaking technology is expected to have a wide range of applications, including indoor surveillance and AI speakers. We definitely wish it enables the AI to understand the is happenings in our surrounding by solely analyzing the sounds, even in situations where visual observation are not possible.
The Gwangju Institute of Science and Technology (GIST) was founded in 1993 by the Korean government as a research-oriented graduate school to help ensure Korea’s continued economic growth and prosperity by developing advanced science and technology with an emphasis on collaboration with the international community. Since that time, GIST has pioneered a highly regarded undergraduate science curriculum in 2010 that has become a model for other science universities in Korea. To learn more about GIST and its exciting opportunities for researchers and students alike, please visit: gist.ac.kr/.
Prof. Hong Kook Kim received a B. S. degree in Control and Instrumentation Engineering from Seoul National University, Korea, in 1988. He then received both M.S. and Ph.D. degrees in Electrical Engineering from the Korea Advanced Institute of Science and Technology (KAIST), Korea in 1990 and 1994, respectively. He is currently an IEEE Senior Member and an affiliate member of IEEE Speech and Language Technical Committee. His current research interests include large vocabulary speech recognition, audio coding and speech/audio source separation, and embedded algorithms and solutions for speech and audio processing for handheld devices.
See more articles on: The Gwangju Institute of Science and Technology
Source: gist.ac.kr