next up previous contents
Next: What is a Up: Spoken Language Lexica Previous: Spoken Language Lexica

Introduction

Lexica for spoken language systems

Spoken language systems are becoming increasingly versatile, and a central problem in developing such a system is the creation of its lexicon. In other areas of language processing, such as computational linguistics, the lexicon is also taking on a more and more central role. The lexicon of a spoken language system may be designed for broad or narrow coverage, for specific applications, with a particular kind of organisation, and optimised for a specific strategy of lexical search. Since the construction of a lexicon is a highly labour-intensive and thus also error-prone job, a prime requirement is for automatising lexicon development as far as possible, and in transferring lexical information from previous applications to new applications.

The main object of this chapter is to provide a framework for relating such concepts to each other and for the formulation of recommendations for development and use of lexica for spoken language systems.

In this introductory section, some basic concepts connected with the use and structure of lexica in spoken language systems are outlined; in the following sections, specific dimensions of spoken language lexica are discussed in more detail. Discussion is restricted to spoken language system lexica; non-electronic lexica for human use (e.g. pronunciation dictionaries in book form) are not considered. Features common to spoken and written language lexica, such as syntactic and semantic information in lexical entries, are only mentioned in passing; see the contribution of the EAGLES Working Group on Computational Lexica on these points. The close relation between spoken language lexica and speech corpora entails a minimal amount of overlap with the contribution of the Spoken Language Corpus chapter of this handbook, which should also be consulted.

In the remainder of the introductory section of this chapter, recent results in the development of spoken language lexica are summarised. The following sections of the chapter are concerned with basic features of spoken language lexica, lexical information, lexicon structure, lexical access, and lexical knowledge acquisition for spoken language lexica.

Lexical information as properties of words

At the present time, information about lexica for spoken language systems is relatively hard to come by. One reason for this is that such information is largely contained in specifications of particular systems, technical reports and books on various aspects of speech processing. Another is that there is not a close relation between concepts and terminology in the speech processing field, and concepts and terminology in traditional lexicography, natural language processing and computational linguistics. Components such as Hidden Markov Models  for word recognition, stochastic  language models for word sequence patterns, grapheme-phoneme tables and rules, word-oriented knowledge bases for semantic interpretation or text construction are all concerned with the areas of word identity and lexical access, lexical disambiguation , lexicon architecture and lexical representation, but these relations are not immediately obvious within the context of speech technology as a whole, and stochastic word models, for instance, would not generally be regarded as providing lexical information (though in the strict sense of the term, they evidently do provide such information). For the purposes of this handbook, the broader view will be taken.

It is customary to distinguish between system lexica and lexical databases. The distinction between the two may, in specific cases, be blurred, and the unity of the two concepts may also be rather loose, if the system lexicon is highly modular or several lexical databases are used. However, the distinction is a useful one. The distinction between lexica and lexical databases will be discussed below. Since the kinds of information in both these types of lexical object overlap, the term ``spoken language lexicon'' will generally be used in this chapter to cover both types.

Applications of spoken language lexica

With the advent of organisations for coordinating the use of language resources, such as ELRA (The European Language Resources Association) and the word-wide LDC Linguistic Data Consortium, access to information on spoken language lexica will become more widely available. The following overview is temporary, and therefore necessarily partial.

Types of application for spoken language lexica

Lexica for spoken language are used in a variety of systems, including the following:

Spoken language lexical databases as a general resource

Spoken language lexica may be components of systems such as those listed above, or reusable background resources. System lexica are generally only of local interest within institutes, companies or projects. Lexical databases as reusable background resources which are intended to be more generally available raise the question of standardised storage and dissemination. In general, the same principles apply as for Spoken Language Corpora: they are stored and disseminated using a variety of media. In research and development contexts, magnetic media (disk or tape) were preferred until recently; in recent years, local magnetic storage and wider informal dissemination within projects or other relevant communities via electronic networks such as Internet using standard file transfer protocols and electronic mail has become the norm. Large lexica, and corpora on which large lexica are based, are in general stored and disseminated in the form of ISO standard CD-ROMs.

The following brief overview can do no more than list a number of examples of current work on spoken language lexicography. At this stage, no claim to exhaustiveness is made, and no valuation of cited or uncited work is intended. It is intended to include a more detailed overview in later versions of this chapter.

Lexica in selected spoken language systems

The range of existing spoken language systems is large, so that only a small selection can be outlined, concentrating on older and established systems whose lexicon requirements are fairly well known; the situation is currently in a state of flux and for this reason the most recent developments are not included. Small vocabulary systems are also excluded, as their strong points are evidently not in the area of the lexicon. The concepts referred to in the descriptions are discussed in the relevant sections below.

HARPY
is a large-vocabulary (1011 words) continuous speech recognition system. It was developed at Carnegie Mellon University. HARPY was the best performing speech recognition system developed under the five-year ARPA project launched in 1971. HARPY makes use of various knowledge sources, including a highly constrained grammar  (a finite state grammar  in BNF [Backus Naur Form] notation) and lexical knowledge in the form of a pronunciation dictionary that contains alternative pronunciations of each word. Initial attempts to derive within-word phonological variations with a set of phonological rules operating on a baseform failed. A set of juncture rules describes inter-word phonological phenomena such as /p/ deletion  at /pm/ junctures: /helpmi/ --- /helmi/. The spectral  characteristics of allophones  of a given phoneme , including their empirically determined durations , are stored in phone templates . The HARPY system compiles all knowledge into a unified directed graph representation, a transition network of 15,000 states (the so-called blackboard model). Each state in the network corresponds to a spectral template . The spectra  of the observed segments are compared with the spectral templates  in the network. The system determines which sequence of spectra , that is, which path through the network, provides the best match with the acoustic input spectral  sequence.

(cf. =1 (

; Klatt 1977) ; see also =1 (

; Lowerre Reddy 1980) )

HEARSAY-II
also uses the blackboard principle (see HARPY), where knowledge sources contribute to the recognition process via a global data base. In the recognition process, an utterance is segmented into categories of manner-of-articulation features, e.g. a stop -vowel-stop  pattern. All words with a syllable structure  corresponding to that of the input are proposed as hypotheses. However, words can also be hypothesised top-down by the syntactic component. So misses by the lexical hypothesiser, which are very likely, can be made up for by the syntactic predictor. The lexicon for word verification has the same structure as HARPY; It is defined in terms of spectral  patterns.

(cf. =1 (

; Klatt 1977) , see also =1 (

; Erman 1977) and =1 (

; Erman Lesser 1980) )

SPHINX
is currently regarded as the state of the art system. It is a large-vocabulary continuous speech recognition system for speaker-independent application. It was evaluated on the DARPA naval resource management task. The baseline SPHINX system works with Hidden Markov Models (HMMs ) where each HMM  represents a phone . The total of phones  is 45. The phone models  are concatenated to create word models, which in turn serve to create sentence models. The phonetic spelling  of a word was adopted from the ANGEL System (cf. =1 (

; Rudnicky et al. 1987) ). The SPHINX baseline system has been improved by introducing multiple codebooks and adding information to the lexical-phonological component:

The SPHINX system works with grammars  of different perplexity  (average branching factor; see the section on word models); the grammars  are of a type which can, in principle, be regarded as a specialised tabular, network-like or tree-structured lexicon with probabilistic word-class information:

In word recognition tests, the best results were obtained with the bigram grammar , the most restrictive kind of the grammars  mentioned above (96% accuracy compared with 71% for null grammars ).

The SPHINX system makes use of various levels of representation of linguistic units:

(cf. =1 (

; Lee Hon Reddy 1990) ; see also =1 (

; Alleva et al. 1992) )

EVAR
(`Erkennen --- Verstehen --- Antworten --- Rückfragen') is a large-vocabulary continuous speech recognition and dialogue system. It is designed to understand standard German sentences and to react either in form of an answer or a question referring back to what has been said, within the specific discourse domain  of enquiries concerning Intercity timetables. The EVAR lexicon has the following properties:

A lexicon administration system is being developed which uses tools for extracting words according to specified criteria, such as ``Look for nouns that express a location'' or ``Look for prepositions that express a direction''.

(cf. =1 (

; Ehrlich 1986) , =1 (

; Brietzmann et al. 1983) , =1 (

; Niemann et al. 1985) , =1 (

; Niemann et al. 1992) )

Recommendations on resources

The following recommendations should be seen in conjunction with recommendations made after the following more specialised sections of this chapter.

  1. For information on resources in specific areas of spoken language systems, consult the other chapters and appendices in this handbook.
  2. For information on resources which relate to written language, consult the other EAGLES working groups.
  3. For general information on resources consult the organisations ELRA ( European Language Resources Association and LDC Linguistic Data Consortium
  4. In preparation for decisions on the use of resources, distinguish between lexical database and the system lexicon
  5. Identify the types of information required for the lexical database and the system lexicon.
  6. Consider the relevant lexical database models and system lexicon architectures.
  7. Develop a systematic concept for the tools required in producing and accessing a lexicon or a lexical database.



next up previous contents
Next: What is a Up: Spoken Language Lexica Previous: Spoken Language Lexica



WWW Administrator
Fri May 19 11:53:36 MET DST 1995