Audio signals, such as speech and ambient sounds convey rich information pertaining to a user’s activity, mood or intent. Enabling machines to understand this contextual information is necessary to bridge the gap in human-machine interaction. This is challenging due to its subjective nature, hence, requiring sophisticated techniques. This dissertation presents a set of computational methods, that generalize well across different conditions, for speech-based applications involving emotion recognition and keyword detection, and ambient sounds-based applications such as lifelogging.
Download count: 0
- Partial requirement for: Ph.D., Arizona State University, 2015Note typethesis
- Includes bibliographical references (p. 134-144)Note typebibliography
- Field of study: Electrical engineering