Definition of Evaluation

Download Report

Transcript Definition of Evaluation

College of Science & Technology
Dep. Of Computer Science & IT
BCs of Information Technology
Data Mining
Chapter 5: Evaluation
2013
Prepared by: Mahmoud Rafeek Al-Farra
www.cst.ps/staff/mfarra
Course’s Out Lines
2











Introduction
Data Preparation and Preprocessing
Data Representation
Classification Methods
Evaluation
Clustering Methods
Mid Exam
Association Rules
Knowledge Representation
Special Case study : Document clustering
Discussion of Case studies by students
Out Lines
3




Definition of Evaluation
Measure of interestingness
Training versus Testing
Cluster evaluation
Definition of Evaluation
4

After examining the data and applying automated
methods for data mining, we must carefully
consider the quality of the end-product of our
effort. This step is evaluation.

Evaluation evaluates the performance of the a
proposed solution to the data mining task.
Definition of Evaluation
5

A large number of patterns
Pattern Evaluation
and
rules
exist
in
database. Many of them
has no interest to the user. Data Mining
Task-relevant Data
Data Warehouse
Data Cleaning
Data Integration
Databases
Selection
Measure of interestingness
6

Measure of interestingness has two approaches:
 Objective:
where the interestingness is measured in
term of its structure and underlying data used in the
discovery process.
Measure of interestingness
7

Measure of interestingness has two approaches:
 Subjective:
Subjective measure do not depended only
in the structure of a rule and the data used , but also on
the user who examines the pattern. These measures
recognize that a pattern of interest to one user , may be
no interest to another user.
Training versus Testing
8

“Just trust me!” does not work in evaluation.

Error on the training data is not a good indicator of
performance on future data.

Simple solution probably not be exactly the same
as the training that can be used if lots of (labeled)
data is available.

Split data into training and test set.
Training versus Testing
9

A strong and effective way to evaluate results is to
hide some data and then do a fair comparison of
training results to unseen data.

In this way it prevents poor results and gives the
developers time to extract the best performance
from the application system.

Many kinds of splitting data into training and
testing most common holdout and cross validation
Cluster evaluation
10

One type of measure allows us to evaluate different
sets of clusters without external knowledge and is
called an internal quality measure; it is used when
we don't have external knowledge about the
clustering data.

Overall similarity is an example for internal
quality measure and will be discussed below.
Cluster evaluation
11

The second type of measures lets us evaluate the
quality of clustering by comparing the clusters
produced by the clustering techniques to known
classes (external knowledge).

This type of measure is called an external quality
measure and we will discuss two external qualities
which are entropy and F-measure.
Cluster evaluation
12

There are many different quality measures and the
performance and relative ranking of different
clustering algorithms can vary substantially
depending on which measure is used.
Thanks
13