This thesis describes a structured connectionist system for natural language generation. 'FIG', short for 'Flexible Increment Generator', is based on a single network which encodes lexical knowledge, syntactic knowledge, and world knowledge. In the initial state, some nodes representing concepts are sources of activation; this represents the input. Activation flows from these nodes to nodes representing words via the various knowledge structures of the network. When the network settles, the most highly activated word is selected and emitted, activation levels are updated to represent the new current state. This process of settle, emit, and update repeats until all of the input has been conveyed. An utterance is simply the result of successive word choices.
Syntactic knowledge is encoded with network structures representing constructions and their constituents. Constituents are linked to words, syntactic categories, relations, and other constructions. Activation flow via these links, and eventually to words, provides for constituency and subcategorization. The links to constituents are gated by 'cursors,' which are updated over time, based on feedback from the words output. This mechanism ensures that words and concepts which are syntactically appropriate become highly activated at the right time, thus causing words to appear in the right order. FIG suggests that the complexity present in most treatments of syntax is unnecessary. For example, it does not assemble or manipulate syntactic structures; constructions affect the utterance only by the activation they transmit, directly or indirectly, to words.
Unlike previous generators, FIG handles arbitrarily large inputs, since the number of nodes activated in the initial state makes no difference to its operation; it handles trade-offs among competing goals without additional mechanism, since all computation is in terms of commensurate levels of activation; and it handles interaction among choices easily, since it tends to settle into a state representing a compatible set of choices, due to links among nodes representing such choices. Abstracting from the implementation leads to several design principles for generation, including explicit representation of the current state and pervasive use of parallelism. The design is a weak cognitive model and also points the way to more natural machine translation.
Title
A Flexible, Parallel Model of Natural Language Generation
Published
1991-04-01
Full Collection Name
Electrical Engineering & Computer Sciences Technical Reports
Other Identifiers
CSD-91-629
Type
Text
Extent
176 p
Archive
The Engineering Library
Usage Statement
Researchers may make free and open use of the UC Berkeley Library’s digitized public domain materials. However, some materials in our online collections may be protected by U.S. copyright law (Title 17, U.S.C.). Use or reproduction of materials protected by copyright beyond that allowed by fair use (Title 17, U.S.C. § 107) requires permission from the copyright owners. The use or reproduction of some materials may also be restricted by terms of University of California gift or purchase agreements, privacy and publicity rights, or trademark law. Responsibility for determining rights status and permissibility of any use or reproduction rests exclusively with the researcher. To learn more or make inquiries, please see our permissions policies (https://www.lib.berkeley.edu/about/permissions-policies).