photograph

 Alexander I. Rudnicky


I'm a Research Professor at the Computer Science Department in the School of Computer Science at Carnegie Mellon University. I'm a part of the Carnegie Mellon Speech Group, one of the oldest such groups in the country, founded by Raj Reddy. I'm also the Director of the Carnegie Mellon Speech Consortium, a partnership between Carnegie Mellon University and industry. I am part of the faculty in the Language Technologies Institute and, actually, most of my students are based there.

My current interests center on language-based communication between humans and robots and on aspects of core speech recognition, such as out-of-vocabulary (OOV) word processing. I am also interested in approaches to learning based on implicit supervision and on improvement of speech system knowledge through dialog. Check this list of publications.

Over time my research interests have revolved around speech perception and recognition, speech interfaces, dialog systems and language in general. I've headed the SpeechWear project that produced an early mobile speech system; I'm still interested in the topic of mobile speech. I also headed the Communicator project, which dealt with dialog system architectures. Olympus, a successor system, is available in Open Source. The Ravenclaw dialog manager came out of the Communicator work and is the current foundation for many of the systems that we create. RavenClaw incorporates dialog management ideas that were first introduced in the Office Manager (OM) system and further developed in successor systems, including Scheduler and AGENDA. Papers listed on the publications describe this work. Some videos from our group.

In 1996 I implemented a set of web-based tools that are used to generate a knowledge base for the open source Sphinx recognition system, including a language model and a pronouncing dictionary. It's proved to be quite popular, and I've continued to maintain the tool and expand its capabilities. Do try it out if you are building a recognition system (the model formats work for any ARPA-compliant system) .

If you'd like to know more about me by reading either a short or a long biography. You can also find me at LinkedIn and on Plaxo; these sites will have somewhat different information about me. I'm involved in several organizations focused on spoken language interaction (SIGdial) and speech technologies (AVIOS).


Some papers covering work in the past year or so; check the publications for a complete list.

Y-N Chen, W. Y. Wang, A. I. Rudnicky Unsupervised Induction and Filling of Semantic Slots for Spoken Dialogue Systems Using Frame-Semantic Parsing, Proceedings of ASRU 2013, Olomouc, CZ.

L. Qin, A. Rudnicky Learning Better Lexical Properties for Recurrent OOV Words, Proceedings of ASRU, 2013, Olomouc, CZ.

L. Qin, A. Rudnicky Finding Recurrent Out-of-Vocabulary Words. Proc. of Interspeech, 2013, Lyon, France.

A. Pappu, A. & Rudnicky, A. Predicting Tasks in Goal-Oriented Spoken Dialog Systems using Semantic Knowledge Bases, In Proceedings of the SIGDIAL 2013 Conference, 2013, Metz, France.

J. Chiu, A. Rudnicky Using Conversational Word Bursts in Spoken Term Detection, Proc. of Interspeech, 2013, Lyon, France.

Teodoro, G., Martin, N., Keshner, E., Shi, J. Y. and Rudnicky, A. Virtual clinicians for the treatment of aphasia and speech disorders Proceedings of ICVR, 2013, Philadelphia, PA US, 158-159.

A. Pappu, M. Sun, S. Sridharan, A. Rudnicky Situated Multiparty Interaction between Humans and Agents. Proceedings of HCII 2013, Las Vegas, US.

S. Sridharan, Y-N Chen, K-M Chang, A. I. Rudnicky NeuroDialog: An EEG-enabled spoken language interface. Proceedings of ICMI, 2012, Santa Monica, CA US.