Preface
Acknowledgements
Chapter One Prologue
Chapter Two MT state of the art
2.1 MT as symbolic systems
2.2 Practical MT
2.3 Alternative technique of MT
2.3.1 Theoretical foundation
2.3.2 Translation model
2.3.3 Language model
2.4 Discussion
Chapter Three Connectionist solutions
3.1 NLP models
3.2 Representation
3.3 Phonological processing
3.4 Learning verb past tense
3.5 Part of speech tagging
3.6 Chinese collocation learning
3.7 Syntactic parsing
3.7.1 Learning active/passive transformation
3.7.2 Confluent preorder parsing
3.7.3 Parsing with fiat structures
3.7.4 Parsing embedded clauses
3.7.5 Parsing with deeper structures
3.8 Discourse analysis
3.8.1 Story gestalt and text understanding
3.8.2 Processing stories with scriptural knowledge
3.9 Machine translation
3.10 Conclusion
Chapter Four NeuroTrans design considerations
4.1 Scalability and extensibility
4.2 Transfer or inter lingual
4.3 Hybrid or fully connectionist
4.4 The use of linguistic knowledge
4.5 Translation as a two stage process
4.6 Selection of network models
4.7 Connectionist implementation
4.8 Connectionist representation issues
4.9 Conclusion
Chapter Five A neural lexicon model
5.1 Language data
5.2 Knowledge representation
5.2.1 Symbolic approach
5.2.2 The statistical approach
5.2.3 Connectionist approach
5.2.4 NeuroTrans input/output representation
5.2.5 NeuroTrans lexicon representation
5.3 Implementing the neural lexicon
5.3.1 Words in context
5.3.2 Context with weights
5.3.3 Details of algorithm
5.3.4 The Neural Lexicon Builder
5.4 Training
5.4.1 Sample preparation
5.4.2 Training results
5.4.3 Generalization test
5.5 Discussion
5.5.1 Adequacy
5.5.2 Scalability and Extensibility
5.5.3 Efficiency
5.5.4 Weaknesses
Chapter Six Implementing the language model
6.1 Overview
6.2 Design
6.2.1 Redefining the generation problem
6.2.2 Defining jumble activity
6.2.3 Language model structure
6.3 Implementation
6.3.1 Network structure Sampling Training and results
6.3.2 Generalization test
6.4 Discussion
6.4.1 Insufficient data
6.4.2 Information richness
6.4.3 Insufficient contextual information
6.4.4 Distributed language model
Chapter Seven Conclusion
Chapter Eight References
Index