Speech and Language Processing, 2nd edition

Published by Pearson (May 16, 2008) © 2009

  • Daniel Jurafsky
  • James H. Martin

eTextbook

$64.99

  • Easy-to-use search and navigation
  • Add notes and highlights
  • Search by keyword or page
$191.99

  • Hardcover, paperback or looseleaf edition
  • Affordable rental option for select titles

For undergraduate or advanced undergraduate courses in Classical Natural Language Processing, Statistical Natural Language Processing, Speech Recognition, Computational Linguistics, and Human Language Processing.

An explosion of Web-based language techniques, merging of distinct fields, availability of phone-based dialogue systems, and much more make this an exciting time in speech and language processing. The first of its kind to thoroughly cover language technology — at all levels and with all modern technologies — this text takes an empirical approach to the subject, based on applying statistical and other machine-learning algorithms to large corporations. The authors cover areas that traditionally are taught in different courses, to describe a unified vision of speech and language processing. Emphasis is on practical applications and scientific evaluation. An accompanying Website contains teaching materials for instructors, with pointers to language processing resources on the Web. The Second Edition offers a significant amount of new and extended material.

Supplements:

Click on the "Resources" tab to View Downloadable Files:

• Each chapter is built around one or more worked examples demonstrating the main idea of the chapter – Uses the examples to illustrate the relative strengths and weaknesses of various approaches.

• Methodology boxes included in each chapter – Introduces important methodological tools such as evaluation, wizard of oz techniques, etc.

• Problem sets included in each chapter.

• Integration of speech and text processing – Merges speech processing and natural language processing fields.

• Empiricist/statistical/machine learning approaches to language processing–Covers all of the new statistical approaches, while still completely covering the earlier more structured and rule-based methods.

• Modern rigorous evaluation metrics.

• Unified and comprehensive coverage of the field – Covers the fundamental algorithms of various fields, whether originally proposed for spoken or written language.

• Emphasis on Web and other practical applications – Gives students an understanding of how language-related algorithms can be applied to important real-world problems.

• Emphasis on scientific evaluation – Offers a description of how systems are evaluated with each problem domain.

• Description of widely available language processing resources.

• Text-specific website at www.cs.colorado.edu/~martin/slp2.html – Contains teaching materials for instructors and pointers to language processing resources on the Web.

Seven new chapters that extend coverage to include:

• Statistical sequence labeling – Features a comprehensive introduction to hidden Markov models (HMMs) including the forward, Viterbi, and forward-backward algorithms, maximum entropy models, and  maximum entropy Markov models (MEMMs).

• Information extraction – Includes machine learning-based approaches to named entity recognition, relation detection and temporal processing.

• Question answering and summarization, including:

– factoid question answering

– single document summarization

– generic multiple document summarization

– query-focused summarization.

• Advanced topics in speech recognition, including:

– decision-tree clustering for context-dependent phones,

– advanced decoding (including n-best lists, lattices, confusion networks, and stack decoding)

– robustness (including MLLR adaptation)

– discriminative training (MMIE).

• Speech synthesis, including:

– text normalization

– grapheme-to-phoneme conversion

– diphone and unit selection synthesis,

– PSOLA

– evaluation.

Nine heavily revised chapters that update and extend coverage of the following material:

• Language modeling, including:

– a more complete description of Good-Turing

– expanded sections on practical issues like perplexity and evaluation

– language modeling toolkits, including ARPA format

– an overview of modern methods like interpolated Kneser-Ney.

• Formal grammars, including:

– new and expanded sections covering treebanks with a focus on the Penn Treebank

– searching treebanks with tgrep and tgrep2

– heads and head-finding rules

– dependency grammars

– categorial grammar

– grammars for spoken language processing.

• Statistical parsing now covers the following:

– PCFGs

– probabilistic CKY parsing

– parent annotations

– a complete description of the Collins parser

– advanced topics such as discriminative reranking

– parsing for language modeling.

• Machine translation now includes:

– a complete description of modern statistical MT with

– phrase-based MT

– IBM model 1

– IBM model 3

– HMM alignment

– evaluation via BLEU

• Dialog processing now includes:

– more information on modern dialogue systems, including VoiceXML

– confirmation and clarification dialogues

– the information-state model

– recent Markov Decision Processes approaches to dialogue.

Foreword

Preface

About the Authors 

 

1 Introduction

1.1 Knowledge in Speech and Language Processing

1.2 Ambiguity

1.3 Models and Algorithms

1.4 Language, Thought, and Understanding

1.5 The State of the Art

1.6 Some Brief History

1.6.1 Foundational Insights: 1940s and 1950s

1.6.2 The Two Camps: 1957—1970

1.6.3 Four Paradigms: 1970—1983

1.6.4 Empiricism and Finite State Models Redux: 1983—1993

1.6.5 The Field Comes Together: 1994—1999

1.6.6 The Rise of Machine Learning: 2000—2008

1.6.7 On Multiple Discoveries

1.6.8 A Final Brief Note on Psychology

1.7 Summary

Bibliographical and Historical Notes

 

Part I Words

 

2 Regular Expressions and Automata

2.1 Regular Expressions

2.1.1 Basic Regular Expression Patterns

2.1.2 Disjunction, Grouping, and Precedence

2.1.3 A Simple Example

2.1.4 A More Complex Example

2.1.5 Advanced Operators

2.1.6 Regular Expression Substitution, Memory, and ELIZA

2.2 Finite-State Automata

2.2.1 Using an FSA to Recognize Sheeptalk

2.2.2 Formal Languages

2.2.3 Another Example

2.2.4 Non-Deterministic FSAs

2.2.5 Using an NFSA to Accept Strings

2.2.6 Recognition as Search

2.2.7 Relating Deterministic and Non-Deterministic Automata

2.3 Regular Languages and FSAs

2.4 Summary

Bibliographical and Historical Notes

Exercises

 

3 Words and Transducers

3.1 Survey of (Mostly) English Morphology

3.1.1 Inflectional Morphology

3.1.2 Derivational Morphology

3.1.3 Cliticization

3.1.4 Non-Concatenative Morphology

3.1.5 Agreement

3.2 Finite-State Morphological Parsing

3.3 Construction of a Finite-State Lexicon

3.4 Finite-State Transducers

3.4.1 Sequential Transducers and Determinism

3.5 FSTs for Morphological Parsing

3.6 Transducers and Orthographic Rules

3.7 The COmbination of an FST Lexicon and Rules

3.8 Lexicon-Free FSTs: The Porter Stemmer

3.9 Word and Sentence Tokenization

3.9.1 Segmentation in Chinese

3.10 Detection and Correction of Spelling Errors

3.11 Minimum Edit Distance

3.12 Human Morphological Processing

3.13 Summary

Bibliographical and Historical Notes

Exercises

 

4 N-grams

4.1 Word Counting in Corpora

4.2 Simple (Unsmoothed) N-grams

4.3 Training and Test Sets

4.3.1 N-gram Sensitivity to the Training Corpus

4.3.2 Unknown Words: Open Versus Closed Vocabulary Tasks

4.4 Evaluating N-grams: Perplexity

4.5 Smoothing

4.5.1 Laplace Smoothing

4.5.2 Good-Turing Discounting

4.5.3 Some Advanced Issues in Good-Turing Estimation

4.6 Interpolation

4.7 Backoff

4.7.1 Advanced: Details of Computing Katz Backoff a and P∗

4.8 Practical Issues: Toolkits and Data Formats

4.9 Advanced Issues in Language Modeling

4.9.1 Advanced Smoothing Methods: Kneser-Ney Smoothing

4.9.2 Class-Based N-grams

4.9.3 Language Model Adaptation and Web Use

4.9.4 Using Longer Distance Information: A Brief Summary

4.10 Advanced: Information Theory Background

4.10.1 Cross-Entropy for Comparing Models

4.11 Advanced: The Entropy of English and Entropy Rate Constancy

4.12 Summary

Bibliographical and Historical Notes

Exercises

 

5 Part-of-Speech Tagging

5.1 (Mostly) English Word Classes

5.2 Tagsets for English

5.3 Part-of-Speech Tagging

5.4 Rule-Based Part-of-Speech Tagging

5.5 HMM Part-of-Speech Tagging

5.5.1 Computing the Most-Likely Tag Sequence: An Example

5.5.2 Formalizing Hidden Markov Model Taggers

5.5.3 Using the Viterbi Algorithm for HMM Tagging

5.5.4 Extending the HMM Algorithm to Trigrams

5.6 Transformation-Based Tagging

5.6.1 How TBL Rules Are Applied

5.6.2 How TBL Rules Are Learned

5.7 Evaluation and Error Analysis

5.7.1 Error Analysis

5.8 Advanced Issues in Part-of-Speech Tagging

5.8.1 Practical Issues: Tag Indeterminacy and Tokenization

5.8.2 Unknown Words

5.8.3 Part-of-Speech Tagging for Other Languages

5.8.4 Tagger Combination

5.9 Advanced: The Noisy Channel Model for Spelling

5.9.1 Contextual Spelling Error Correction

5.10 Summary

Bibliographical and Historical Notes

Exercises

 

6 Hidden Markov and Maximum Entropy Models

6.1 Markov Chains

6.2 The Hidden Markov Model

6.3 Likelihood Computation: The Forward Algorithm

6.4 Decoding: The Viterbi Algorithm

6.5 HMM Training: The Forward-Backward Algorithm

6.6 Maximum Entropy Models: Background

6.6.1 Linear Regression

6.6.2 Logistic Regression

6.6.3 Logistic Regression: Classification

6.6.4 Advanced: Learning in Logistic Regression

6.7 Maximum Entropy Modeling

6.7.1 Why We Call it Maximum Entropy

6.8 Maximum Entropy Markov Models

6.8.1 Decoding and Learning in MEMMs

6.9 Summary

Bibliographical and Historical Notes

Exercises

 

Part II Speech

 

7 Phonetics

7.1 Speech Sounds and Phonetic Transcription

7.2 Articulatory Phonetics

7.2.1 The Vocal Organs

7.2.2 Consonants: Place of Articulation

7.2.3 Consonants: Manner of Articulation

7.2.4 Vowels

7.2.5 Syllables 

7.3 Phonological Categories and Pronunciation Variation

7.3.1 Phonetic Features

7.3.2 Predicting Phonetic Variation

7.3.3 Factors Influencing Phonetic Variation

7.4 Acoustic Phonetics and Signals

7.4.1 Waves

7.4.2 Speech Sound Waves

7.4.3 Frequency and Amplitude; Pitch and Loudness

7.4.4 Interpretation of Phones from a Waveform

7.4.5 Spectra and the Frequency Domain

7.4.6 The Source-Filter Model

7.5 Phonetic Resources

7.6 Advanced: Articulatory and Gestural Phonology

7.7 Summary

Bibliographical and Historical Notes

Exercises

 

8 Speech Synthesis

8.1 Text Normalization

8.1.1 Sentence Tokenization

8.1.2 Non-Standard Words

8.1.3 Homograph Disambiguation

8.2 Phonetic Analysis

8.2.1 Dictionary Lookup

8.2.2 Names

8.2.3 Grapheme-to-Phoneme Conversion

8.3 Prosodic Analysis

8.3.1 Prosodic Structure

8.3.2 Prosodic Prominence

8.3.3 Tune

8.3.4 More Sophisticated Models: ToBI

8.3.5 Computing Duration from Prosodic Labels

8.3.6 Computing F0 from Prosodic Labels

8.3.7 Final Result of Text Analysis: Internal Representation

8.4 Diphone Waveform synthesis

8.4.1 Steps for Building a Diphone Database

8.4.2 Diphone Concatenation and TD-PSOLA for Prosody

8.5 Unit Selection (Waveform) Synthesis

8.6 Evaluation

Bibliographical and Historical Notes

Exercises

 

9 Automatic Speech Recognition

9.1 Speech Recognition Architecture

9.2 Applying the Hidden Markov Model to Speech

9.3 Feature Extraction: MFCC vectors

9.3.1 Preemphasis

9.3.2 Windowing

9.3.3 Discrete Fourier Transform

9.3.4 Mel Filter Bank and Log

9.3.5 The Cepstrum: Inverse Discrete Fourier Transform

9.3.6 Deltas and Energy

9.3.7 Summary: MFCC

9.4 Acoustic Likelihood Computation 

9.4.1 Vector Quantization

9.4.2 Gaussian PDFs

9.4.3 Probabilities, Log Probabilities and Distance Functions

9.5 The Lexicon and Language Model

9.6 Search and Decoding

9.7 Embedded Training

9.8 Evaluation: Word Error Rate

9.9 Summary

Bibliographical and Historical Notes

Exercises

 

10 Speech Recognition: Advanced Topics

10.1 Multipass Decoding: N-best Lists and Lattices

10.2 A∗ (‘Stack’) Decoding

10.3 Context-Dependent Acoustic Models: Triphones

10.4 Discriminative Training

10.4.1 Maximum Mutual Information Estimation

10.4.2 Acoustic Models Based on Posterior Classifiers

10.5 Modeling Variation

10.5.1 Environmental Variation and Noise

10.5.2 Speaker Variation and Speaker Adaptation

10.5.3 Pronunciation Modeling: Variation Due to Genre

10.6 Metadata: Boundaries, Punctuation, and Disfluencies

10.7 Speech Recognition by Humans

10.8 Summary

Bibliographical and Historical Notes

Exercises

 

11 Computational Phonology

11.1 Finite-State Phonology

11.2 Advanced Finite-State Phonology

11.2.1 Harmony

11.2.2 Templatic Morphology

11.3 Computational Optimality Theory

11.3.1 Finite-State Transducer Models of Optimality Theory

11.3.2 Stochastic Models of Optimality Theory

11.4 Syllabification

11.5 Learning Phonology and Morphology

11.5.1 Learning Phonological Rules

11.5.2 Learning Morphology

11.5.3 Learning in Optimality Theory

11.6 Summary

Bibliographical and Historical Notes

Exercises

 

Part III Syntax

 

12 Formal Grammars of English

12.1 Constituency

12.2 Context-Free Grammars

12.2.1 Formal definition of Context-Free Grammar

12.3 Some Grammar Rules for English

12.3.1 Sentence-Level Constructions

12.3.2 Clauses and Sentences

12.3.3 The Noun Phrase

12.3.4 Agreement

12.3.5 The Verb Phrase and Subcategorization

12.3.6 Auxiliaries

12.3.7 Coordination

12.4 Treebanks

12.4.1 Example: The Penn Treebank Project

12.4.2 Treebanks as Grammars

12.4.3 Treebank Searching

12.4.4 Heads and Head Finding

12.5 Grammar Equivalence and Normal Form

12.6 Finite-State and Context-Free Grammars

12.7 Dependency Grammars

12.7.1 The Relationship Between Dependencies and Heads

12.7.2 Categorial Grammar

12.8 Spoken Language Syntax

12.8.1 Disfluencies and Repair

12.8.2 Treebanks for Spoken Language

12.9 Grammars and Human Processing

12.10 Summary

Bibliographical and Historical Notes

Exercises

 

13 Syntactic Parsing

13.1 Parsing as Search

13.1.1 Top-Down Parsing

13.1.2 Bottom-Up Parsing

13.1.3 Comparing Top-Down and Bottom-Up Parsing

13.2 Ambiguity

13.3 Search in the Face of Ambiguity

13.4 Dynamic Programming Parsing Methods

13.4.1 CKY Parsing

13.4.2 The Earley Algorithm

13.4.3 Chart Parsing

13.5 Partial Parsing

13.5.1 Finite-State Rule-Based Chunking

13.5.2 Machine Learning-Based Approaches to Chunking

13.5.3 Evaluating Chunking Systems

13.6 Summary

Bibliographical and Historical Notes

Exercises

 

14 Statistical Parsing

14.1 Probabilistic Context-Free Grammars

14.1.1 PCFGs for Disambiguation

14.1.2 PCFGs for Language Modeling

14.2 Probabilistic CKY Parsing of PCFGs

14.3 Learning PCFG Rule Probabilities

14.4 Problems with PCFGs

14.4.1 Independence Assumptions Miss Structural Dependencies Between Rules

14.4.2 Lack of Sensitivity to Lexical Dependencies

14.5 Improving PCFGs by Splitting Non-Terminals

14.6 Probabilistic Lexicalized CFGs

14.6.1 The Collins Parser

14.6.2 Advanced: Further Details of the Collins Parser

14.7 Evaluating Parsers

14.8 Advanced: Discriminative Reranking

14.9 Advanced: Parser-Based Language Modeling

14.10 Human Parsing

14.11 Summary

Bibliographical and Historical Notes

Exercises

 

15 Features and Unification

15.1 Feature Structures

15.2 Unification of Feature Structures

15.3 Feature Structures in the Grammar

15.3.1 Agreement

15.3.2 Head Features

15.3.3 Subcategorization

15.3.4 Long-Distance Dependencies

15.4 Implementation of Unification

15.4.1 Unification Data Structures

15.4.2 The Unification Algorithm

15.5 Parsing with Unification Constraints

15.5.1 Integration of Unification into an Earley Parser

15.5.2 Unification-Based Parsing

15.6 Types and Inheritance

15.6.1 Advanced: Extensions to Typing

15.6.2 Other Extensions to Unification

15.7 Summary

Bibliographical and Historical Notes

Exercises

 

16 Language and Complexity

16.1 The Chomsky Hierarchy

16.2 Ways to Tell if a Language Isn’t Regular

16.2.1 The Pumping Lemma

16.2.2 Proofs That Various Natural Languages Are Not Regular

16.3 Is Natural Language Context-Free?

16.4 Complexity and Human Processing

16.5 Summary

Bibliographical and Historical Notes

Exercises

 

Part IV Semantics and Pragmatics

 

17 The Representation of Meaning

17.1 Computational Desiderata for Representations

17.1.1 Verifiability

17.1.2 Unambiguous Representations

17.1.3 Canonical Form

17.1.4 Inference and Variables

17.1.5 Expressiveness

17.2 Model-Theoretic Semantics

17.3 First-Order Logic

17.3.1 Basic Elements of First-Order Logic

17.3.2 Variables and Quantifiers

17.3.3 Lambda Notation

17.3.4 The Semantics of First-Order Logic

17.3.5 Inference

17.4 Event and State Representations 

17.4.1 Representing Time

17.4.2 Aspect

17.5  Description Logics

17.6 Embodied and Situated Approaches to Meaning

17.7 Summary

Bibliographical and Historical Notes

Exercises

 

18 Computational Semantics

18.1 Syntax-Driven Semantic Analysis

18.2 Semantic Augmentations to Syntactic Rules

18.3 Quantifier Scope Ambiguity and Underspecification

18.3.1 Store and Retrieve Approaches

18.3.2 Constraint-Based Approaches

18.4 Unification-Based Approaches to Semantic Analysis

18.5 Integration of Semantics into the Earley Parser

18.6 Idioms and Compositionality

18.7 Summary

Bibliographical and Historical Notes

Exercises

 

19 Lexical Semantics

19.1 Word Senses

19.2 Relations Between Senses

19.2.1 Synonymy and Antonymy

19.2.2 Hyponymy

19.2.3 Semantic Fields

19.3 WordNet: A Database of Lexical Relations

19.4 Event Participants

19.4.1 Thematic Roles

19.4.2 Diathesis Alternations

19.4.3 Problems with Thematic Roles

19.4.4 The Proposition Bank

19.4.5 FrameNet

19.4.6 Selectional Restrictions

19.5 Primitive Decomposition

19.6 Advanced: Metaphor

19.7 Summary

Bibliographical and Historical Notes

Exercises

 

20 Computational Lexical Semantics

20.1 Word Sense Disambiguation: Overview

20.2 Supervised Word Sense Disambiguation

20.2.1 Feature Extraction for Supervised Learning

20.2.2 Naive Bayes and Decision List Classifiers

20.3 WSD Evaluation, Baselines, and Ceilings

20.4 WSD: Dictionary and Thesaurus Methods

20.4.1 The Lesk Algorithm

20.4.2 Selectional Restrictions and Selectional Preferences . . . . 684

20.5 Minimally Supervised WSD: Bootstrapping

20.6 Word Similarity: Thesaurus Methods

20.7 Word Similarity: Distributional Methods

20.7.1 Defining a Word’s Co-Occurrence Vectors

20.7.2 Measuring Association with Context

20.7.3 Defining Similarity Between Two Vectors

20.7.4 Evaluating Distributional Word Similarity

20.8 Hyponymy and Other Word Relations

20.9 Semantic Role Labeling

20.10 Advanced: Unsupervised Sense Disambiguation

20.11 Summary

Bibliographical and Historical Notes

Exercises

 

21 Computational Discourse

21.1 Discourse Segmentation

21.1.1 Unsupervised Discourse Segmentation

21.1.2 Supervised Discourse Segmentation

21.1.3 Discourse Segmentation Evaluation

21.2 Text Coherence

21.2.1 Rhetorical Structure Theory

21.2.2 Automatic Coherence Assignment

21.3 Reference Resolution

21.4 Reference Phenomena

21.4.1 Five Types of Referring Expressions

21.4.2 Information Status

21.5 Features for Pronominal Anaphora Resolution

21.6 Three Algorithms for Pronominal Anaphora Resolution

21.6.1 Pronominal Anaphora Baseline: The Hobbs Algorithm

21.6.2 A Centering Algorithm for Anaphora Resolution

21.6.3 A Log-Linear Model for Pronominal Anaphora Resoluton

21.6.4 Features for Pronominal Anaphora Resoluton

21.7 Coreference Resolution

21.8 Evaluation of Coreference Resolution

21.9 Advanced: Inference-Based Coherence Resolution

21.10 Psycholinguistic Studies of Reference

21.11 Summary

Bibliographical and Historical Notes

Exercises

 

Part V Applications

 

22 Information Extraction

22.1 Named Entity Recognition

22.1.1 Ambiguity in Named Entity Recognition

22.1.2 NER as Sequence Labeling

22.1.3 Evaluation of Named Entity Recognition

22.1.4 Practical NER Architectures

22.2 Relation Detection and Classification

22.2.1 Supervised Learning Approaches to Relation Analysis

22.2.2 Lightly Supervised Approaches to Relation Analysis

22.2.3 Evaluation of Relation Analysis Systems

22.3 Temporal and Event Processing

22.3.1 Temporal Expression Recognition

22.3.2 Temporal Normalization

22.3.3 Event Detection and Analysis

22.3.4 TimeBank

22.4 Template-Filling

22.4.1 Statistical Approaches to Template-Filling

22.4.2 Finite-State Template-Filling Systems

22.5 Advanced: Biomedical Information Extraction

22.5.1 Biological Named Entity Recognition

22.5.2 Gene Normalization

22.5.3 Biological Roles and Relations

22.6 Summary

Bibliographical and Historical Notes

Exercises

 

23 Question Answering and Summarization

23.1 Information Retrieval

23.1.1 The Vector Space Model

23.1.2 Term Weighting

23.1.3 Term Selection and Creation

23.1.4 Evaluation of Information-Retrieval Systems

23.1.5 Homonymy, Polysemy, and Synonymy

23.1.6 Ways to Improve User Queries

23.2 Factoid Question Answering

23.2.1 Question Processing

23.2.2 Passage Retrieval

23.2.3 Answer Processing

23.2.4 Evaluation of Factoid Answers

23.3 Summarization

23.4 Single Document Summarization 

23.4.1 Unsupervised Content Selection

23.4.2 Unsupervised Summarization Based on Rhetorical Parsing

23.4.3 Supervised Content Selection

23.4.4 Sentence Simplification

23.5 Multi-Document Summarization

23.5.1 Content Selection in Multi-Document Summarization

23.5.2 Information Ordering in Multi-Document Summarization

23.6 Focused Summarization and Question Answering

23.7 Summarization Evaluation

23.8 Summary

Bibliographical and Historical Notes

Exercises

 

24 Dialogue and Conversational Agents

24.1 Properties of Human Conversations

24.1.1 Turns and Turn-Taking

24.1.2 Language as Action: Speech Acts

24.1.3 Language as Joint Action: Grounding

24.1.4 Conversational Structure

24.1.5 Conversational Implicature

24.2 Basic Dialogue Systems

24.2.1 ASR component

24.2.2 NLU component

24.2.3 Generation and TTS components

24.2.4 Dialogue Manager

24.2.5 Dealing with Errors: Confirmation and Rejection

24.3 VoiceXML

24.4 Dialogue System Design and Evaluation

24.4.1 Designing Dialogue Systems

24.4.2 Evaluating Dialogue Systems

24.5 Information-State and Dialogue Acts

24.5.1 Using Dialogue Acts

24.5.2 Interpreting Dialogue Acts

24.5.3 Detecting Correction Acts

24.5.4 Generating Dialogue Acts: Confirmation and Rejection

24.6 Markov Decision Process Architecture

24.7 Advanced: Plan-Based Dialogue Agents

24.7.1 Plan-Inferential Interpretation and Production

24.7.2 The Intentional Structure of Dialogue

24.8 Summary

Bibliographical and Historical Notes

Exercises

 

25 Machine Translation

25.1 Why Machine Translation Is Hard

25.1.1 Typology

25.1.2 Other Structural Divergences

25.1.3 Lexical Divergences

25.2 Classical MT and the Vauquois Triangle

25.2.1 Direct Translation

25.2.2 Transfer

25.2.3 Combined Direct and Tranfer Approaches in Classic MT

25.2.4 The Interlingua Idea: Using Meaning

25.3 Statistical MT

25.4 P(F|E): the Phrase-Based Translation Model

25.5 Alignment in MT

25.5.1 IBM Model 1

25.5.2 HMM Alignment

25.6 Training Alignment Models

25.6.1 EM for Training Alignment Models

25.7 Symmetrizing Alignments for Phrase-Based MT

25.8 Decoding for Phrase-Based Statistical MT

25.9 MT Evaluation

25.9.1 Using Human Raters

25.9.2 Automatic Evaluation: BLEU

25.10 Advanced: Syntactic Models for MT

25.11 Advanced: IBM Model 3 and Fertitlity

25.11.1 Training for Model 3

25.12 Advanced: Log-linearModels for MT . . . . . . . . . . . . . . . . 903

25.13 Summary

Bibliographical and Historical Notes

Exercises

 

Bibliography

Author Index

Subject Index

 

 

 

Dan Jurafsky is an associate professor in the Department of Linguistics, and by courtesy in Department of Computer Science, at Stanford University. Previously, he was on the faculty of the University of Colorado, Boulder, in the Linguistics and Computer Science departments and the Institute of Cognitive Science. He was born in Yonkers, New York, and received a B.A. in Linguistics in 1983 and a Ph.D. in Computer Science in 1992, both from the University of California at Berkeley. He received the National Science Foundation CAREER award in 1998 and the MacArthur Fellowship in 2002. He has published over 90 papers on a wide range of topics in speech and language processing.

James H. Martin is a professor in the Department of Computer Science and in the Department of Linguistics, and a fellow in the Institute of Cognitive Science at the University of Colorado at Boulder. He was born in New York City, received a B.S. in Comoputer Science from Columbia University in 1981 and a Ph.D. in Computer Science from the University of California at Berkeley in  1988. He has authored over 70 publications in computer science including the book A Computational Model of Metaphor Interpretation.

Need help? Get in touch

Video
Play
Privacy and cookies
By watching, you agree Pearson can share your viewership data for marketing and analytics for one year, revocable upon changing cookie preferences. Disabling cookies may affect video functionality. More info...

Pearson eTextbook: What’s on the inside just might surprise you

They say you can’t judge a book by its cover. It’s the same with your students. Meet each one right where they are with an engaging, interactive, personalized learning experience that goes beyond the textbook to fit any schedule, any budget, and any lifestyle.Â