Algorithmic Aspects of Finite Metric Spaces - PowerPoint PPT Presentation

1 / 41
About This Presentation
Title:

Algorithmic Aspects of Finite Metric Spaces

Description:

d(x,y) = 0 iff x=y. d(x,y) = d(y,x) Symmetric. d(x,z) d(x,y) d(y,z) Triangle ... for optimization problem (metric labelling) yield log n approximate estimator ... – PowerPoint PPT presentation

Number of Views:104
Avg rating:3.0/5.0
Slides: 42
Provided by: csPrin7
Category:

less

Transcript and Presenter's Notes

Title: Algorithmic Aspects of Finite Metric Spaces


1
Algorithmic Aspects of Finite Metric Spaces
  • Moses Charikar
  • Princeton University

2
Metric Space
  • A set of points X
  • Distance function d(x,y)d X ?0??)
  • d(x,y) 0 iff xy
  • d(x,y) d(y,x) Symmetric
  • d(x,z) d(x,y) d(y,z) Triangle inequality
  • Metric space M(X,d)

3
Example Metrics Normed spaces
  • x (x1, x2, , xd) y (y1, y2, , yd)
  • lp norm l1 l2 (Euclidean) l?
  • lpd lp norm in Rd
  • Hamming cube 0,1d

4
Example Metrics domain specific
  • Shortest path distances on graph
  • Symmetric difference on sets
  • Edit distance on strings
  • Hausdorff distance, Earth Mover Distance on sets
    of n points

5
Metric Embeddings
  • General idea Map complex metrics to simple
    metrics
  • Why ? richer algorithmic toolkit for simple
    metrics
  • Simple metrics
  • normed spaces lp
  • low dimensional normed spaces lpd
  • tree metrics
  • Mapping should not change distances much (low
    distortion)

6
Low Distortion Embeddings
  • Metric spaces (X1,d1) (X2,d2),embedding f X1
    ? X2 has distortion D if ratio of distances
    changes by D ? x,y ? X1

f
http//humanities.ucsd.edu/courses/kuchtahum4/pix/
earth.jpg
http//www.physast.uga.edu/jss/1010/ch10/earth.jp
g
7
Applications
  • High dimensional ? Low dimensional(Dimension
    reduction)
  • Algorithmic efficiency (running time)
  • Compact representation (storage space)
  • Streaming algorithms
  • Specific metrics ? normed spaces
  • Nearest neighbor search
  • Optimization problems
  • General metrics ? tree metrics
  • Optimization problems, online algorithms

Solve problems on very large data sets in one
pass using a very small amount of storage
8
A (very) Brief Historyfundamental results
  • Metric spaces studied in functional analysis
  • n point metric embeds into l?n with no distortion
    Frechet
  • n point metric embeds into lp with distortion log
    n Bourgain 85
  • Dimension reduction for n point Euclidean metric
    with distortion 1e Johnson, Lindenstrauss 84

9
A (very) Brief Historyapplications in Computer
Science
  • Optimization problems
  • Application to graph partitioning Linial,
    London, Rabinovich 95Arora, Rao, Vazirani
    04
  • n point metrics into tree metrics Bartal 96
    98 FRT 03
  • Efficient algorithms
  • Dimension reduction
  • Nearest neighbor search, Streaming algorithms

10
Outline
metric as data
  • Dimension reduction
  • Streaming data model
  • Compact representation
  • Finite metrics in optimization
  • graph partitioning and clustering

Embedding theorems for finite metrics
metric as model
11
Disclaimer
  • This is not an attempt at a survey
  • Biased by my own interests
  • Much more relevant and related work than I can do
    justice do in limited time.
  • Goal Give glimpse of different applications of
    finite metric spaces
  • Core ideas, no messy details

12
Disclaimer Community Bias
  • Theoretical viewpoint
  • Focus on algorithmic techniques with performance
    guarantees
  • Worst case guarantees

13
Outline
metric as data
  • Dimension reduction
  • Streaming data model
  • Compact representation
  • Finite metrics in optimization
  • graph partitioning and clustering

Embedding theorems for finite metrics
metric as model
14
Metric as data
  • What is the data ?
  • Mathematical representation of objects (e.g.
    documents, images, customer profiles, queries).
  • Sets, vectors, points in Euclidean space, points
    in a metric space, vertices of a graph.
  • Metric is part of data

15
Johnson Lindenstrauss JL84
  • n points in Euclidean space (l2 norm) can be
    mapped down to O((log n)/?2) dimensions with
    distortion at most 1?.
  • Quite simple JL84, FM88, IM98, AV99, DG99,
    Ach01
  • Project onto random unit vectors
  • projection of (u-v) onto one random vector
    behaves like Gaussian scaled by u-v2
  • Need log n dimensions for tight concentration
    bounds
  • Even a random -1,1 vector works

16
Dimension reduction for l2
  • Two interesting properties
  • Linear mapping
  • Oblivious choice of linear mapping does not
    depend on point set
  • Many applications
  • Making high dimensional problems tractable
  • Streaming algorithms
  • Learning mixtures of gaussians Dasgupta 99
  • Learning robust concepts Arriaga,Vempala 99
    Klivans,Servedio 04

17
Dimension reduction for l1
  • C,Sahai 02Linear embeddings are not good for
    dimension reduction in l1
  • There exist n points in l1 in n dimensions, such
    that any linear mapping with distortion ? needs
    n/?2 dimensions

18
Dimension reduction for l1
  • C, Brinkman 03Strong lower bounds for
    dimension reduction in l1
  • There exist n points in l1 , such that any
    embedding with constant distortion ? needs n1/?2
    dimensions
  • Alternate, simpler proof Lee, Naor 03

19
Outline
metric as data
  • Dimension reduction
  • Streaming data model
  • Compact representation
  • Finite metrics in optimization
  • graph partitioning and clustering

Embedding theorems for finite metrics
Solve problems on very large data sets in one
pass using a very small amount of storage
metric as model
20
Frequency Moments
Alon,Matias,Szegedy 99
  • Data stream is sequence of elements in n
  • ni frequency of element i
  • Fk ? nik kth frequency moment
  • F0 number of distinct elements
  • F2 skewness measure of data stream
  • Goal
  • Given a data stream, estimate Fk in one pass and
    sub-linear space

21
Estimating F2
  • Consider a single counter c and randomly chosen
    xi ? 1, -1 for each i in n
  • On seeing each element i, update c xi
  • c ? ni xi
  • Claim Ec2 ? ni2 F2 Varc2 ?
    2(F2)2 (4-wise independence)
  • Average 1/?2 copies of this estimator to get
    (1?) approximation

22
Differences between data streams
  • ni frequency of element i in stream 1
  • mi frequency of element i in stream 2
  • Goal measure ? (ni mi)2
  • F2 sketches are additive? ni xi - ? mi xi ?
    (ni mi)xi
  • Basically, dimension reduction in l2 norm
  • Very useful primitivee.g. frequent items
    C, Chen, Farach-Colton 02

23
Estimate l1 norms ?
  • Indyk 00
  • p-stable distributionDistribution over R such
    that? ni xi distributed as (? nip )1/p X
  • Cauchy distribution c(x)1/?(1x2) 1-stable
  • Gaussian distribution 2-stable
  • As before, c ? ni xi
  • Cauchy does not have finite expectation !
  • Estimate scale factor by taking median

24
Outline
metric as data
  • Dimension reduction
  • Streaming data model
  • Compact representation
  • Finite metrics in optimization
  • graph partitioning and clustering

Embedding theorems for finite metrics
metric as model
25
Similarity Preserving Hash Functions
  • Similarity function sim(x,y)
  • Family of hash functions F with probability
    distribution such that

26
Applications
  • Compact representation scheme for estimating
    similarity
  • Approximate nearest neighbor search
    Indyk,Motwani 98 Kushilevitz,Ostrovsky,Rabani
    98

27
Estimating Set Similarity
  • Broder,Manasse,Glassman,Zweig,97
  • Broder,C,Frieze,Mitzenmacher,98
  • Collection of subsets

28
Minwise Independent Permutations
29
Existence of SPH schemes C 02
  • sim(x,y) admits an SPH scheme if? family of
    hash functions F such that
  • Theorem If sim(x,y) admits an SPH scheme then
    1-sim(x,y) satisfies triangle inequality embeds
    into l1
  • Rounding procedures for LPs and SDPs yield
    similarity and distance preserving hashing
    schemes.

30
Random Hyperplane Rounding based SPH
  • Collection of vectors
  • Pick random hyperplane through origin (normal
    )
  • Goemans,Williamson

31
Earth Mover Distance (EMD)
LP Rounding algorithms for optimization problem
(metric labelling) yield log n approximate
estimator for EMD on n points. Implies that EMD
embeds into l1 with distortion log n
P
Q
EMD(P,Q)
32
Outline
metric as data
  • Dimension reduction
  • Streaming data model
  • Compact representation
  • Finite metrics in optimization
  • graph partitioning and clustering

Embedding theorems for finite metrics
metric as model
33
Graph partitioning problems
  • Given graph, partition into U,V
  • Maximum cut maximize E(U,V)
  • Sparsest cut
  • minimize

34
Correlation clustering
Cohen,Richman,02Bansal,Blum,Chawla,02
Similar ()
Dissimilar (-)
example courtesy Shuchi Chawla
Mr. Rumsfeld
his
The secretary
he
Saddam Hussein
35
Graph partitioning as metric problem
  • Partitioning is equivalent to finding appropriate
    0,1 metric
  • possibly additional constraints
  • Objective function linear in metric
  • Find best 0,1 metric

cut metric
relaxation
36
Metric relaxation approaches
  • Max Cut Goemans,Williamson 94
  • map vertices to points on unit sphere (SDP)
  • exploit geometry to get good solution(random
    hyperplane cut)
  • Sparsest Cut Linial,London,Rabinovich 95
  • LP gives best metric need l1 metric
  • Bourgain 84 embeds any metric into l1 with
    distortion log n
  • Existential theorem can be made algorithmic
  • log n approximation
  • recent SDP based ?log n approximationArora,Rao,V
    azirani 04

37
Metric relaxation approaches
  • Correlation clustering C,Guruswami,Wirth,03
    Emanuel,Fiat,03 Immorlica,Karger,03
  • Find best 0,1 metric from similarity/dissimilari
    ty data via LP
  • Use metric to guide clustering
  • close points in same cluster
  • distant points in different clusters
  • Learning best metric ?
  • Note In many cases, LP/SDP can be eliminated to
    yield efficient algorithms

38
Outline
metric as data
  • Dimension reduction
  • Streaming data model
  • Compact representation
  • Finite metrics in optimization
  • graph partitioning and clustering

Embedding theorems for finite metrics
metric as model
39
Some connections to learning
  • Dimension reduction in l2
  • Learning mixtures of Gaussians Dasgupta
    99Random projections make skewed gaussians
    more spherical, making learning easier
  • Learning with large marginArriaga,Vempala 99
    Klivans,Servedio 04Random projections
    preserve margin,large margin ? few dimensions
  • Kernel methods for SVMs
  • mappings to l2

40
Ongoing developments
  • Notion of intrinsic dimensionality of metric
    spaceGupta,Krauthgamer,Lee,03Krauthgamer,Lee
    ,Mendel,Naor,04
  • Doubling dimension How many balls of radius R
    needed to cover ball of radius 2R ?
  • Complexity measure of metric space
  • natural parameter for embeddings
  • Open Can every metric of constant doubling
    dimension in l2 be embedded into l2 with O(1)
    dimensions and O(1) distortion ?
  • Not true for l1
  • related to learning low dimension manifolds,
    PCA, MDS, LLE, Isomap

41
Some things I didnt mention
  • Approximating general metrics via tree metrics
  • modified notion of distortion
  • useful for approximation, online algorithms
  • Many mathematically appealing questions
  • Embeddings between normed spaces
  • Spectral methods for approximating matrices
    (SVD, LSI)
  • PCA, MDS, LLE, Isomap

42
Conclusions
  • Whirlwind tour of finite metrics
  • Rich algorithmic toolkit for finite metric spaces
  • Synergy between Computer Science and Mathematics
  • Exciting area of active research
  • range from practical applications to deep
    theoretical questions
  • Many more applications to be discovered
Write a Comment
User Comments (0)
About PowerShow.com