James Carmichael

Learn More
Automatic speech recognition (ASR) can provide a rapid means of controlling electronic assistive technology. Off-the-shelf ASR systems function poorly for users with severe dysarthria because of the increased variability of their articulations. We have developed a limited vocabulary speaker dependent speech recognition application which has greater(More)
This paper describes a new formulation of a polynomial sequence kernel based on dynamic time warping (DTW) for support vector machine (SVM) classification of isolated words given very sparse training data. The words are uttered by dysarthric speakers who suffer from debilitating neurological conditions that make the collection of speech samples a(More)
We describe an unusual ASR application: recognition of command words from severely dysarthric speakers, who have poor control of their articulators. The goal is to allow these clients to control assistive technology by voice. While this is a small vocabulary, speaker-dependent, isolated-word application, the speech material is more variable than normal, and(More)
Automatic speech recognition (ASR) can provide a rapid means of controlling EAT. Off-the-shelf ASR systems function poorly for users with severe dysarthria because of the increased variability of their articulations compared to ‘normal’ speech. A two-pronged approach has been applied to this problem: 1. To develop a computerised training package which will(More)
This study reports on the development of a diagnostic expert system – incorporating a multilayer perceptron (MLP) – designed to identify any sub-type of dysarthria (loss of neuromuscular control over the articulators) manifested by a patient undergoing a Frenchay Dysarthria Assessment (FDA) evaluation. If sufficient information is provided describing(More)
Computer based speech training systems aim to provide the client with customised tools for improving articulation based on audio-visual stimuli and feedback. They require the integration of various components of speech technology, such as speech recognition and transcription tools, and a database management system which supports multiple on-the-fly(More)
This paper describes a multimedia multimodal information access sub-system (MIAS) for digital audio-visual documents, typically presented in streaming media format. The system is designed to provide both professional and general users with entry points into video documents that are relevant to their information needs. In this work, we focus on the(More)
This study discusses the implementation of a digital signal processing computerised application which is specifically designed to measure and classify aberrant voice quality in the context of the Frenchay Dysarthria Assessment (FDA) diagnostic test series. It has been observed that subjective evaluations of voice quality -- particularly when performed by(More)
This study discusses the findings of an evaluation study on the performance of a multimedia multimodal information access sub-system (MIAS), incorporating automatic speech recognition technology (ASR) to automatically transcribe the speech content of video soundtracks. The study’s results indicate that an information-rich but minimalist graphical interface(More)