L | M | X | J | V | S | D |
---|---|---|---|---|---|---|
1 | 2 | 3 | ||||
4 | 5 | 6 | 7 | 8 | 9 | 10 |
11 | 12 | 13 | 14 | 15 | 16 | 17 |
18 | 19 | 20 | 21 | 22 | 23 | 24 |
25 | 26 | 27 | 28 | 29 | 30 |
Research group Spoken and Multimodal Dialogue Systems (SISDIAL)
Dept. of Languages and Computer Systems, Faculty of Computer Science and Telecommunications, University of Granada, Spain
Spoken dialogue systems enable human-computer interaction (HCI) using spontaneous speech. Many of these systems are employed currently to automate telephone-based information services.
Mutlimodal dialogue systems enable HCI employing a number of interaction modalities, for example, speech, body gestures and facial expresions. These systems are employed nowadays to make more human-like the interaction. For example, they are used in tutoring systems, Ambient Intelligence (AmI), robots and healthcare systems.
The research group Spoken and Multimodal Dialogue Systems (SISDIAL) focuses on the analysis, design and development of these systems, paying special attention to the following issues: