Spoken Language Understanding Task
(Redirected from spoken language understanding)
Jump to navigation
Jump to search
A Spoken Language Understanding Task is a Natural Language Understanding Task that accepts spoken language.
- Context:
- It can be solved by a Spoken Language Understanding System (that implements a Spoken Language Understanding Algorithm).
- See: Spoken Dialog System.
References
2015
- (Mesnil et al., 2015) ⇒ Grégoire Mesnil, Yann Dauphin, Kaisheng Yao, Yoshua Bengio, Li Deng, Dilek Hakkani-Tur, Xiaodong He, Larry Heck, Gokhan Tur, Dong Yu, and Geoffrey Zweig. (2015). “Using Recurrent Neural Networks for Slot Filling in Spoken Language Understanding.” In: IEEE/ACM Transactions on Audio, Speech and Language Processing (TASLP) Journal, 23(3). doi:10.1109/TASLP.2014.2383614
- QUOTE: Semantic slot filling is one of the most challenging problems in spoken language understanding (SLU).
2008
- (De Mori et al., 2008) ⇒ Renato De Mori, Frédéric Bechet, Dilek Hakkani-Tur, Michael McTear, Giuseppe Riccardi, and Gokhan Tur. (2008). “Spoken Language Understanding.” IEEE Signal Processing Magazine, 25(3)
- ABSTRACT: Semantics deals with the organization of meanings and the relations between sensory signs or symbols and what they denote or mean. Computational semantics performs a conceptualization of the world using computational processes for composing a meaning representation structure from available signs and their features present, for example, in words and sentences. Spoken language understanding (SLU) is the interpretation of signs conveyed by a speech signal. SLU and natural language understanding (NLU) share the goal of obtaining a conceptual representation of natural language sentences. Specific to SLU is the fact that signs to be used for interpretation are coded into signals along with other information such as speaker identity. Furthermore, spoken sentences often do not follow the grammar of a language; they exhibit self-corrections, hesitations, repetitions, and other irregular phenomena. SLU systems contain an automatic speech recognition (ASR) component and must be robust to noise due to the spontaneous nature of spoken language and the errors introduced by ASR. Moreover, ASR components output a stream of words with no structure information like punctuation and sentence boundaries. Therefore, SLU systems cannot rely on such markers and must perform text segmentation and understanding at the same time.