Hypertext Categoriza.. - School of Computer Science

Download Report

Transcript Hypertext Categoriza.. - School of Computer Science

Hypertext Categorization
Rayid Ghani
IR Seminar - 10/3/00
“Standard” Approach
 Apply traditional text learning algorithms
 In many cases, goal is not to classify
hypertext but to test the algorithms
 Is it actually the right approach?
Results?
 Mixed results


Positive results in most cases BUT the goal was
to test the algorithms
Negative in few e.g. Chakrabarti BUT the goal
was to motivate their own algorithm
How is hypertext different?
 Link Information
 Diverse Authorship
 Short text - topic not obvious from the text
 Structure / position within the web graph
 Author-supplied features(meta-tags)
 Bold , italics, heading etc.
How to use those extra
features?
Specific approaches to classify
hypertext
 Chakrabarti et al SIGMOD 98
 Oh et al. SIGIR 00
 Slattery & Mitchell ICML 00
 Goal is not classification but retrieval


Bharat & Henzinger SIGIR 98
Croft & Turtle 93
Chakrabarti et al. SIGMOD 98
 Use the page and linkage information
 Add words from the “neighbors” and treat
them as belonging to the page itself


Decrease in performance (not surprising)
Link information is very noisy
 Use topic information from neighbors instead
Data Sets
 IBM Patent Database

12 classes (630 train, 300 test for each class)
 Yahoo

13 classes , 20000 docs (for expts involving
hypertext, only 900 documents were used) (?)
Experiments
 Using text from neighbors


Local+Neighbor_Text: 
Local+Neighbor_Text_Tagged: 
 Assume Neighbors are Pre-classified




Text – 36%
Link – 34%
Prefix – 22.1% (words in class heirarchy used)
Text+Prefix – 21%
Oh et al. SIGIR 2000
 Relationship b/w class and neighbors of a
web page in the training set is not
consistent/useful (?)
 Instead, Use the class and neighbor info of
the page being classified (use regularities in
the test set)
Classification
 Classify test instance d by:
arg max [ P (C | G , T )]
c
 arg max [ P (C | T ) P (C | G )]
c
|T |
 arg max [( P (c) P(ti | c) N ( ti |d ) ) Neighbord (c)]
c
i 1
ld (c )
Neighbord (c) 
wL
Ld
Algorithm
 For each test document d, generate a set A of
“trustable” neighbors
 For all terms ti in d, adjust the term weight using the
term weights from A
 For each doc a in A, assign a max confidence value
if its class is known otherwise assign a class
probabilistically and give it partial confidence
weight
 Classify d using the equation given earlier.
Experiments
 Reuters used to assess the algorithm on
datasets without hyperlinks – only varying
the size of the training set & # of features (?)

Results not directly comparable but numbers
similar to reported results
 Articles from an encyclopedia – 76 classes,
20836 documents
Results
 Terms+Classes > Only Classes > Only Terms
> No use of inlinks
Other issues
 Link discrimination
 Knowledge of neighbor classes
 Use of links in training set
 Inclusion of new terms from neighbors
Comparison
Chakrabarti Oh et al. Improvement
Links in training Y
set
Link
N
discrimination
Knowledge of
Y
neighbor class
N
5%
Y
6.7%
Y
Iteration
Y
N
6.6%
1.9%
1.5%
Using new terms Y
from neighbors
N
31.4%
Slattery & Mitchell ICML 00
 Given a problem setting in which the test set
contains structural regularities, How can we
find and use them?
Hubs and Authorities
Kleinberg (1998)
“.. a good hub is a page
that points to many
good authorities;
a good authority is a
page pointed to by
many good hubs.”
Hubs
Authorities
Hubs and Authorities
Kleinberg (1998)
“Hubs and authorities
exhibit what could be
called a mutually
reinforcing relationship”
Iterative relaxation:
Hub ( p )   Authority (q )
q: p  q
Authority ( p ) 
Hubs
Authorities
 Hub (q)
q:q  p
The Plan
 Take an existing learning algorithm
 Extend it to exploit structural regularities in
the test set

Using Hubs and Authorities as inspiration
FOIL
Quinlan & Cameron-Jones (1993)
Learns relational rules like:
target_page(A) :- has_research(A), link(A,B),
has_publications(B).
For each test example
 Pick matching rule with best training set
performance p.
 Predict positive with confidence p
FOIL-Hubs Representation
Add two rules to a learned rule set
 target_page(A):-link(B,A),target_hub(B).
 target_hub(A):link(A,B),target_page(B).
Talk about confidence rather than truth
 target_page(page15) = 0.75
Evaluate by summing instantiations
target_pag e(page15) 
 target_hub (B)
B : link(B, page15)
FOIL-Hubs Algorithm
1. Apply learned FOIL rules: learned(A)
2. Iterate
1.
2.
3.
Evaluate target_hub(A)
Evaluate target_page(A)
Set target_page(A) =
target_pag e(A)  s  learned(A)
3. Report target_page(A)
FOIL-Hubs Algorithm
Learned FOIL rules
foil(A)
target_page(A)
target_hub(A)
1.
Apply learned FOIL rules to test set
2.
Initialise target_page(A) confidence from foil(A)
3.
Evaluate target_hub(A)
4.
Evaluate target_page(A)
5. target_page(A)=target_page(A)s+foil(A)
Data Set
 4127 pages from Computer Science
departments of four universities:
Cornell University
University of Texas at Austin
University of Washington University of Wisconsin
• Hand labeled into:
Student
558 Web pages
Course
243 Web pages
Faculty
153 Web pages
Experiment
Three binary classification tasks
1. Student Home Page
2. Course Home Page
3. Faculty Home Page
Leave two university out cross-validation
Student Home Page
100
Precision
80
60
40
FOIL-Hubs
FOIL
20
0
0
20
40
60
Recall
80
100
Course Home Page
100
FOIL-Hubs
FOIL
Precision
80
60
40
20
0
0
20
40
60
Recall
80
100
More Detailed Results
Partition the test data into
 Examples covered by some learned FOIL
rule
 Examples covered by no learned FOIL rule
Student – FOIL covered
100
Precision
80
60
40
FOIL-Hubs
FOIL
20
0
0
20
40
60
Recall
80
100
Student – FOIL uncovered
100
Precision
80
60
FOIL-Hubs
FOIL
40
20
0
0
20
40
60
Recall
80
100
Course – FOIL covered
100
Precision
80
60
40
FOIL-Hubs
FOIL
20
0
0
20
40
60
Recall
80
100
Course – FOIL uncovered
100
FOIL-Hubs
FOIL
Precision
80
60
40
20
0
0
20
40
60
Recall
80
100
Recap
 We’ve searched for regularities of the form
student_page(A):link(Web->KB members page,A)
in the test set.
 We consider this an instance of a regularity schema
student_page(A):link(<page constant>,A)
Conclusions
 Test set regularities can be used to improve
classification performance
 FOIL-Hubs used such regularities to
outperform FOIL on three Web page
classification problems
 We can potentially search for other regularity
schemas using FOIL
Other work
 Using the structure of HTML to improve
retrieval. Michal Cutler, Yungming Shih,
Weiyi Meng. USENIX 1997

Use tfidf - different different weights to text in
different html tags