Abstract
A fondamental challenge we must meet for computers to eventually process spoken language as effectively as humans is to capture the immensely rich fund of information we have in our heads that is NOT in the speech signal. This information is what gives us the ability to supply acoustic cues when these are degraded or missing, or to zero in on one speaker amid a chorus of other voices. While the powerful statistical methods currently used in speech recognition and synthesis have brought some success and useful applications, future progress will depend crucially on a deeper knowledge and greater use of this information. Some of this information is applicable to all languages, and some of it is specific to individual language types. In my discussion, special attention will be given to the processing of spoken Chinese.
Original language | English |
---|---|
Title of host publication | 2004 International Symposium on Chinese Spoken Language Processing - Proceedings |
Publication status | Published - 1 Dec 2004 |
Externally published | Yes |
Event | 2004 International Symposium on Chinese Spoken Language Processing - Hong Kong, China, Hong Kong Duration: 15 Dec 2004 → 18 Dec 2004 |
Conference
Conference | 2004 International Symposium on Chinese Spoken Language Processing |
---|---|
Country/Territory | Hong Kong |
City | Hong Kong, China |
Period | 15/12/04 → 18/12/04 |
ASJC Scopus subject areas
- General Engineering