FiniteState and the Noisy Channel - PowerPoint PPT Presentation

About This Presentation
Title:

FiniteState and the Noisy Channel

Description:

Given L = a 'lexicon' FSA that matches all English words. How to apply to this problem? ... uygar las, tir ama dik lar imiz dan mis, siniz casina ... – PowerPoint PPT presentation

Number of Views:77
Avg rating:3.0/5.0
Slides: 19
Provided by: jasone2
Learn more at: https://www.cs.jhu.edu
Category:

less

Transcript and Presenter's Notes

Title: FiniteState and the Noisy Channel


1
Finite-State and the Noisy Channel
2
Word Segmentation
theprophetsaidtothecity
  • What does this say?
  • And what other words are substrings?
  • Could segment with parsing (how?), but slow.
  • Given L a lexicon FSA that matches all
    English words.
  • How to apply to this problem?
  • What if Lexicon is weighted?
  • From unigrams to bigrams?
  • Smooth L to include unseen words?

3
Spelling correction
  • Spelling correction also needs a lexicon L
  • But there is distortion
  • Let T be a transducer that models common typos
    and other spelling errors
  • ance ? ence (deliverance, ...)
  • e ? e (deliverance, ...)
  • e ? e // Cons _ Cons (athlete, ...)
  • rr ? r (embarrass, occurrence, )
  • ge ? dge (privilege, )
  • etc.
  • Now what can you do with L .o. T ?
  • Should T and L have probabilities?
  • Want T to include all possible errors

4
Noisy Channel Model
real language X
noisy channel X ? Y
yucky language Y
want to recover X from Y
5
Noisy Channel Model
real language X
correct spelling
typos
noisy channel X ? Y
yucky language Y
misspelling
want to recover X from Y
6
Noisy Channel Model
real language X
(lexicon space)
delete spaces
noisy channel X ? Y
yucky language Y
text w/o spaces
want to recover X from Y
7
Noisy Channel Model
real language X
(lexicon space)
pronunciation
noisy channel X ? Y
yucky language Y
speech
want to recover X from Y
8
Noisy Channel Model
real language X
tree
probabilistic CFG
delete everythingbut terminals
noisy channel X ? Y
yucky language Y
text
want to recover X from Y
9
Noisy Channel Model
real language X
p(X)

p(Y X)
noisy channel X ? Y

yucky language Y
p(X,Y)
want to recover x?X from y?Y choose x that
maximizes p(x y) or equivalently p(x,y)
10
Noisy Channel Model
p(X)

p(Y X)

p(X,Y)
Note p(x,y) sums to 1. Suppose yC what is
best x?
11
Noisy Channel Model
aa/0.7
bb/0.3
p(X)
.o.

aC/0.1
bC/0.8
p(Y X)
aD/0.9
bD/0.2


p(X,Y)
aC/0.07
bC/0.24
aD/0.63
bD/0.06
Suppose yC what is best x?
12
Noisy Channel Model
aa/0.7
bb/0.3
p(X)
.o.

aC/0.1
bC/0.8
p(Y X)
aD/0.9
bD/0.2
restrict just to paths compatible with output C


p(X, y)
aC/0.07
bC/0.24
best path
13
Morpheme Segmentation
  • Let Lexicon be a machine that matches all Turkish
    words
  • Same problem as word segmentation
  • Just at a lower level morpheme segmentation
  • Turkish word uygarlas,tiramadiklarimizdanmis,sini
    zcasina uygarlas,tiramadiklarimizdanmis,
    sinizcasina(behaving) as if you are among
    those whom we could not cause to become civilized
  • Some constraints on morpheme sequence bigram
    probs
  • Generative model concatenate then fix up joints
  • stop -ing stopping, fly -s flies
  • Use a cascade of transducers to handle all the
    fixups
  • But this is just morphology!
  • Can use probabilities here too (but people often
    dont)

14
Edit Distance Transducer
O(k) deletion arcs
be
ae
O(k2) substitution arcs
ab
ba
ea
O(k) insertionarcs
aa
eb
bb
O(k) no-change arcs
15
Edit Distance Transducer
Stochastic
O(k) deletion arcs
be
ae
O(k2) substitution arcs
ab
ba
ea
O(k) insertionarcs
aa
eb
bb
O(k) identity arcs
Likely edits high-probability arcs
16
Edit Distance Transducer
Stochastic
Best path (by Dijkstras algorithm)
clara
le
ae
re
ae
ce
.o.
ac
lc
cc
ac
rc
ec
ec
ec
ec
ec
ec
ce
le
ae
re
ae
la
ca
aa
ra
aa

ea
ea
ea
ea
ea
ea
ce
le
ae
re
ae
lc
cc
ac
rc
ac
ec
ec
ec
ec
ec
ec
ce
le
ae
re
ae
la
.o.
ca
aa
aa
ra
ea
ea
ea
ea
ea
ea
caca
ce
le
ae
re
ae
17
Speech Recognition by FST Composition (Pereira
Riley 1996)
p(word seq)
trigram language model
p(phone seq word seq)
pronunciation model
p(acoustics phone seq)
acoustic model
.o.
observed acoustics
18
Speech Recognition by FST Composition (Pereira
Riley 1996)
p(word seq)
trigram language model
p(phone seq word seq)
CATk æ t
p(acoustics phone seq)
?
phone context
phone context
.o.
Write a Comment
User Comments (0)
About PowerShow.com