Title: Minimally Supervised Learning of Semantic Knowledge from Query Logs
1Minimally Supervised Learning of Semantic
Knowledge from Query Logs
Mamoru Komachi() and Hisami Suzuki() () Nara
Institute of Science and Technology, Japan ()
Microsoft Research, USA
- IJCNLP-08, Hyderabad, India
2Task
similar
similar
Darjeeling
Kombucha (Japanese tea)
Chai (Indian tea)
- Learn semantic categories from web search query
logs by bootstrapping with minimal supervision - Semantic category a set of words which are
interrelated - Named entities, technical terms, paraphrases,
- Can be useful for search ads, etc
2012/2/24
2
3Approach
- Semantic categories
- The objects search users frequently ask (cf.
Pasca and Durme 2007) - Query logs
- Reflect the interest of search users
- Short but relevant for word categorization
- Include word segmentation specified by users
- Bootstrapping
- Adapted by many binary relation extraction tasks
(Brin 1998 Collins and Singer 1999 Etzioni et
al. 2005) - Can start from small set of instances (cf. Sekine
and Suzuki 2007)
4Our Contribution
- First to use the Japanese query logs for the task
of learning of named entities - Propose an efficient method suited for query
logs, based on the general-purpose Espresso
(Pantel and Pennacchiotti 2006) algorithm
5Table of Contents
- Related work
- Bootstrapping techniques for relation extraction
- Scoring metrics
- The Tchai algorithm
- Problems of Espresso
- Extension to Espresso
- Experiment
- System performance and comparison to other
algorithms - Samples of extracted instances and patterns
6Bootstrapping
- Iteratively conduct pattern induction and
instance extraction starting from seed instances - Can fertilize small set of seed instances
Query log (Corpus)
Instances
Contextual patterns
vaio
Compare vaio laptop
Compare laptop
Compare toshiba satellite laptop
Toshiba satellite
slot
Compare HP xb3000 laptop
HP xb3000
7Instance lookup and pattern induction
ANA ??
ANA
??
query log
extracted pattern
instance
Restaurant reservation?
Flight reservation?
Broad coverage, Noisy patterns
Use all strings but instances Require no
segmentation
- Semantic drift
- Computational efficency
Generic patterns
8Instance/Pattern Scoring Metrics
- Sekine Suzuki (2007)
- Starts from a large named entity dictionary
- Assign low scores to generic patterns and ignore
- Basilisk (Thelen and Riloff, 2002)
- Balance the recall and precision of generic
patterns - Espresso (Pantel and Pennacchiotti, 2006)
PMI is normalized by the maximum of all P and I
P patterns in corpus I instances in corpus PMI
pointwise mutual information r reliability score
Reliability of an instance and a pattern is
mutually defined
9Problems of Espresso
- Generic patterns/instances
- Generic patterns require a lot of computation
- Computational efficiency
- Espresso computes the reliability for all
patterns to rank in each iteration
10The Tchai Algorithm
- Filter generic patterns/instances
- Not to select generic patterns and instances
- Replace scaling factor in reliability scores
- Take the maximum PMI for a given instance/pattern
rather than the maximum for all instances and
patterns - This modification shows a large impact on the
effectiveness of our algorithm - Only induce patterns at the beginning
- Tchai runs 400X faster than Espresso
11Comparison of methods
12Experiments
- Japanese query logs from 2007/01-02
- Unique one million (166 millions in token)
- Target categories
- Manually classified 10,000 most frequent search
words (in the log of 2006/12) -- hereafter
referred to as 10K list - Travel the largest category (712 words)
- Finance the smallest category (240 words)
13Results
High precision (92.1)
Learned 251 novel words
Due to the ambiguity of hand labeling (e.g. Tokyo
Disney Land)
Include common nouns related to Travel (e.g.
Rental car)
14Sample of Instances (Travel category)
Able to learn several sub-categories in which no
seed words given
15Impact of Pattern Induction
No degradation without pattern induction
Can run 400X faster without any cost!
16Effect of each modification
Filtering outperforms no-filtering constantly
Scaling factor has the most impact
17System Performance
High precision and recall
High precision but low relative recall due to
strict filtering
Relative Recall (Pantel et al., 2004)
18Cumulative precision Travel
Tchai achieved the best precision
19Cumulative precision Finance
Both Basilisk and Espresso suffered from
acquiring generic pattern in early stages of
iteration
20Sample Extracted Patterns
Basilisk and Espresso extracted location names as
context patterns, which may be too generic for
Travel domain
Tchai found context patterns that are
characteristic to the domain
21Conclusion and future work
- Conclusion
- Use of query logs for semantic category learning
- Improved Espresso algorithm in both precision and
performance - Future work
- Generalize bootstrapping method by graph-based
matrix calculation
22Thank you for listening!
Tchai