Model-driven Data Acquisition in Sensor Networks - PowerPoint PPT Presentation

1 / 25
About This Presentation
Title:

Model-driven Data Acquisition in Sensor Networks

Description:

Model-driven Data Acquisition in Sensor Networks Amol Deshpande1,4 Carlos Guestrin4,2 Sam Madden4,3 Joe Hellerstein1,4 Wei Hong4 1UC Berkeley 2Carnegie Mellon ... – PowerPoint PPT presentation

Number of Views:199
Avg rating:3.0/5.0
Slides: 26
Provided by: S66
Category:

less

Transcript and Presenter's Notes

Title: Model-driven Data Acquisition in Sensor Networks


1
Model-driven Data Acquisition in Sensor Networks
  • Amol Deshpande1,4 Carlos Guestrin4,2 Sam
    Madden4,3
  • Joe Hellerstein1,4 Wei Hong4
  • 1UC Berkeley 2Carnegie Mellon University
  • 3MIT 4Intel Research -
    Berkeley

2
Sensor networks and distributed systems
  • A collection of devices that can sense, actuate,
    and communicate over a wireless network
  • Sensors for temperature, humidity, pressure,
    sound, magnetic fields, acceleration, visible and
    ultraviolet light, etc.
  • Available resources
  • 4 MHz, 8 bit CPU
  • 40 Kbps wireless
  • 3V battery (lasts days or months)
  • Analogous issues in other distributed systems,
    including streams and the Internet

3
Real deployments
  • Great Duck Island

4
Example Intel Berkeley Lab deployment
5
AnalogySensor net as a database
  • Data aggregation
  • Can reduce communication

TinyDB
Query
SQL-style query
  • Declarative interface
  • Sensor nets are not just for PhDs
  • Decrease deployment time

Every time step
6
Limitations of existing approach
  • Data collection
  • Every node must wake up at every time step
  • Data loss ignored
  • No quality guarantees
  • Data inefficient ignoring correlations

TinyDB
  • Query distribution
  • Every node must receive query

Query
New Query
SQL-style query
Every time step
7
Sensor net data is correlated
  • Data is not i.i.d. ? shouldnt ignore
    missing data
  • Observing one sensor ? information about other
    sensors (and future values)
  • Observing one attribute ? information about other
    attributes

8
Model-driven data acquisition overview
posterior belief
Probabilistic Model
  • Strengths of model-based data acquisition
  • Observe fewer attributes
  • Exploit correlations
  • Reuse information between queries
  • Directly deal with missing data
  • Answer more complex (probabilistic) queries

Query
New Query
SQL-style query with desired confidence
9
Probabilistic models and queries
Users perspective
Query SELECT nodeId, temp 0.5C, conf(.95) FROM
sensors WHERE nodeId in 1..8
System selects and observes subset of
nodes Observed nodes 3,6,8
Query result
Node 1 2 3 4 5 6 7 8
Temp. 17.3 18.1 17.4 16.1 19.2 21.3 17.5 16.3
Conf. 98 95 100 99 95 100 98 100
10
Probabilistic models and queries
Joint distribution P(X1,,Xn)
Prob. below 1-??
  • Learn from historical data

Higher prob., could answer query
11
Dynamic models filtering
Joint distribution at time t
Fewer obs. in future queries
  • Example Kalman filter
  • Learn from historical data

Observe attributes Example Observe X118
12
Supported queries
  • Value query
  • Xi ? with prob. at least 1-?
  • SELECT and Range query
  • Xi?a,b with prob. at least 1-?
  • which sensors have temperature greater than 25C
    ?
  • Aggregation
  • average ? of subset of attribs. with prob. gt
    1-?
  • combine aggregation and selection
  • probability gt 10 sensors have temperature greater
    than 25C ?
  • Queries require solution to integrals
  • Many queries computed in closed-form
  • Some require numerical integration/sampling

13
Model-driven data acquisition overview
posterior belief
Probabilistic Model
What sensors do we observe ? How do we collect
observations?
Query
SQL-style query with desired confidence
Data gathering plan
14
Acquisition costs
  • Attributes have different acquisition costs
  • Exploit correlation through probabilistic model
  • Must consider networking cost

15
Network model and plan format
2
8
7
Goal Find subset S that is sufficient to answer
query at minimum cost C(S )
1
9
6
3
12
4
5
10
11
Cost of collecting subset S of sensor values
C(S )
Ca(S )
Ct(S )
  • Assume known (quasi-static) network topology
  • Define traversal using (1.5-approximate) TSP
  • Ct(S ) is expected cost of TSP (lossy
    communication)

16
Choosing observation plan
Is a subset S sufficient?
Xi2a,b with prob. gt 1-?
17
BBQ system
  • Multivariate Gaussians
  • Learn from historical data

posterior belief
Probabilistic Model
  • Exhaustive or greedy search
  • Factor 1.5 TSP approximation
  • Simple matrix operations

Query
SQL-style query with desired confidence
Data gathering plan
  • Value
  • Range
  • Average
  • Equivalent to Kalman filter
  • Simple matrix operations

18
Experimental results
  • Redwood trees and Intel Lab datasets
  • Learned models from data
  • Static model
  • Dynamic model Kalman filter, time-indexed
    transition probabilities
  • Evaluated on a wide range of queries

19
Cost versus Confidence level
20
Obtaining approximate values
Query True temperature value epsilon with
confidence 95
21
Approximate range queries
Query Temperature in T1,T2 with confidence 95
22
Comparison to other methods
23
Intel Lab traversals
24
BBQ system
  • Multivariate Gaussians
  • Learn from historical data

posterior belief
Probabilistic Model
  • Exhaustive or greedy search
  • Factor 1.5 TSP approximation
  • Simple matrix operations

Query
  • Extensions
  • More complex queries
  • Other probabilistic models
  • More advanced planning
  • Outlier detection
  • Dynamic networks
  • Continuous queries

SQL-style query with desired confidence
Data gathering plan
  • Value
  • Range
  • Average
  • Equivalent to Kalman filter
  • Simple matrix operations

25
Conclusions
  • Model-driven data acquisition
  • Observe fewer attributes
  • Exploit correlations
  • Reuse information between queries
  • Directly deal with missing data
  • Answer more complex (probabilistic) queries
  • Basis for future sensor network systems
Write a Comment
User Comments (0)
About PowerShow.com