Abstract
The human activity that takes place in meeting rooms or classrooms is reflected in a rich variety of acoustic events (AE), produced either by the human body or by objects handled by humans, so the determination of both the identity of sounds and their position in time may help to detect and describe that human activity. Indeed, speech is usually the most informative sound, but other kinds of AEs may also carry useful information, for example, clapping or laughing inside a speech, a strong yawn in the middle of a lecture, a chair moving or a door slam when the meeting has just started. Additionally, detection and classification of sounds other than speech may be useful to enhance the robustness of speech technologies like automatic speech recognition.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
G. B. D. Wang. Computational Auditory Scene Analysis: Principles, Algorithms and Applications. Wiley-IEEE Press, 2006.
R. Malkin, D. Macho, A. Temko, and C. Nadeu. First evaluation of acoustic event classification systems in the CHIL project. In Joint Workshop on Hands-Free Speech Communication and Microphone Array, HSCMA’05, March 2005.
C. Segura, A. Abad, C. Nadeu, and J. Hernando. Multispeaker localization and tracking in intelligent environments. In Multimodal Technologies for Perception of Humans, Proceedings of the International Evaluation Workshops CLEAR 2007 and RT 2007, LNCS 4625, pages 82–90, Baltimore, MD, May 8-11 2007.
R. Stiefelhagen, R. Bowers, and J. Fiscus, editors. Multimodal Technologies for Perception of Humans, Proceedings of the International Evaluation Workshops CLEAR 2007 and RT 2007. LNCS 4625. Springer, Baltimore, MD, May 8-11 2007.
R. Stiefelhagen and J. Garofolo, editors. Multimodal Technologies for Perception of Humans, First International Evaluation Workshop on Classification of Events, Activities and Relationships, CLEAR’06. LNCS 4122. Springer, Southampton, UK, Apr. 6-7 2006.
A. Temko. Acoustic Event Detection and Classification. PhD thesis, Universitat Politècnica de Catalunya, Barcelona, 2007.
A. Temko, R. Malkin, C. Zieger, D. Macho, C. Nadeu, and M. Omologo. Evaluation of acoustic event detection and classification systems. In Multimodal Technologies for Perception of Humans. First International Evaluation Workshop on Classification of Events, Activities and Relationships CLEAR 2006, LNCS 4122, pages 311–322. Springer-Verlag, Southampton, UK, Apr. 6-7 2006.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer-Verlag London Limited
About this chapter
Cite this chapter
Temko, A., Nadeu, C., Macho, D., Malkin, R., Zieger, C., Omologo, M. (2009). Acoustic Event Detection and Classification. In: Waibel, A., Stiefelhagen, R. (eds) Computers in the Human Interaction Loop. Human–Computer Interaction Series. Springer, London. https://doi.org/10.1007/978-1-84882-054-8_7
Download citation
DOI: https://doi.org/10.1007/978-1-84882-054-8_7
Publisher Name: Springer, London
Print ISBN: 978-1-84882-053-1
Online ISBN: 978-1-84882-054-8
eBook Packages: Computer ScienceComputer Science (R0)