
 
mechanical pulses at a single frequency within the 
natural range of the human voice. The transducer 
uses an a coil-magnet arrangement that vibrates 
against a diaphragm when the output of an electrical 
oscillator is applied to its winding. The device is 
pressed against the mandible and this vibrates the 
pharynx which in turn resonates the air in the vocal 
and/or nasal tract. The vibrations are formed into 
speech by the articulators of the upper vocal tract. 
Research to date has focused on the 
improvement of the quality of speech produced by 
the electro-larynx. Some significant contributions 
have been made by (Houston et al. 1999) who 
developed an electro-larynx which used digital 
signal processing to create a superior quality of 
sound. (Shoureshi et al. 2003) used neural-based 
signal processing and smart materials to improve the 
sound created. (Liu et al. 2006) and (Cole et al. 
1997) focused on removing the buzzing sound 
created by the transducer. (Uemi et al. 1994) 
developed a system that utilized measurements from 
air pressure that was obtained from a resistive 
component that was placed over the stoma to 
maintain the electrolarynx’s fundamental frequency. 
(Ma et al. 1999) used cepstral analysis of speech to 
replace the electrolarynx excitation signal with a 
normal speech excitation signal.  
Despite all the acoustic improvements these 
studies have shown, they have however, been 
performed in isolation and have been deemed to be 
difficult to implement into the existing technology. 
Therefore, the basic concept of design which was 
first introduced by Barney et al. in the 1950’s 
remains the same to this day. It has been shown that 
up to 50%-66% of all laryngectomees use some 
form of electrolarynx speech (Gray et al. 1976) and 
(Hillman et al. 1998): either as a method of 
communication for speech rehabilitation post-
surgery or as a reliable back-up in situations where 
esophageal or trcheo-esophageal speech is proving 
difficult. 
1.2  Speech Intelligibility  
When determining the intelligibility of a speech 
signal, it is important to choose a suitable linguistic 
level at which to make measurements. Is it necessary 
to measure the accuracy at with which each phonetic 
element is communicated in order to assess whether 
each word is identifiable. It is also necessary to 
investigate whether the communication of a sentence 
is clear. 
This type of linguistic dismissal can introduce an 
additional difficulty in that individual human 
listeners will ultimately differ in their capability to 
make use of these linguistic constraints. Even 
though it may be deemed necessary to assess the 
utility of a particular channel in order to convey the 
meanings of real spoken utterances, listeners will 
inevitably vary in their capacity to comprehend the 
speech, depending on their own linguistic ability. 
Many speech intelligibility tests consist of either 
phonetic unit, which are composed into: nonsense 
syllables, words which are used in isolation or in 
short sentences spoken in one breath for comfort 
(Crystal et al 1982) and (Mitchell et al. 1996). 
An issue that arises through the use of nonsense 
syllables is that many listeners could require training 
in order to be able to identify the component 
phonetic units, and they may be confused by 
phonemes which don't compare well with the 
spelling e.g. there, their, they’re. Therefore by 
limiting listener reply’s to real words thus allowing 
them to respond in ordinary spelling. This can 
however introduce other difficulties: firstly, that 
varying listeners may possess differing degrees of 
familiarity with the words that are being used; 
secondly, that some words are memorable and 
having heard a word once, some listeners may be 
biased in their usage of a particular word another 
time. 
A possible solution to these problems includes 
the formulation of multiple word lists of reasonable 
difficulty, allowing a listener to be used within a test 
more than once. Another option is to create tests 
consisting of closed response sets, making every 
listener needing to make the matching choices about 
the word which is under test. 
(Egan et al. 1948) pioneered one of the first lists 
of words for an intelligibility test in 1948. He 
created the list by using the concept of “phonetic 
balance” which meant that the relative frequency of 
the phonemes in the word lists corresponded to the 
relative frequency of phonemes in conversational 
speech. He constructed 20 lists containing 50 
monosyllabic words and his intention was to balance 
average difficulty and range of difficulty throughout 
the lists whilst ensuring that the phonetic units that 
were present were represented equally.  
2 METHODS 
2.1  Novel Actuator Design  
This design consists of a simple pager motor (which 
is typically found in a mobile phone) attached to a 
thin piece of high-density polyethylene (HDPE) by 
BIOSIGNALS 2011 - International Conference on Bio-inspired Systems and Signal Processing
266