CS计算机代考程序代写 database Bayesian data mining decision tree Excel Bayesian network algorithm Classification & Prediction: Bayes Classifiers

Classification & Prediction: Bayes Classifiers

Skip to main content

Print book

Classification & Prediction: Bayes Classifiers

Site:

Wattle

Course:

COMP3425/COMP8410 – Data Mining – Sem 1 2021

Book:

Classification & Prediction: Bayes Classifiers

Printed by:

Zizuo Xiao

Date:

Saturday, 8 May 2021, 11:05 PM

Description

Naive Bayes and Bayesian Belief Networks

Table of contents
1. Introduction
2. Probabilistic Classifier (Text:8.3.1) 2.1. Basic Probabilities (not in text)
2.2. Bayes’ Theorem (Text 8.3.1)
2.3. Limitation (Text 8.3.2)

3. Naive Bayes (Text:8.3.2) 3.1. Laplacian Correction
3.2. Numerical attributes

4. Bayesian Belief Networks (Text: 9.1) 4.1. Training a Belief Network (Text 9.1.2)

5. Reading and Exercises

1. Introduction
Most
of this material is derived from the text, Han, Kamber and Pei, Chapter
8 and 9, or the corresponding powerpoint slides made available by the
publisher.  Where a source other than the text or its slides was
used for the material, attribution is given. Unless otherwise stated,
images are copyright of the publisher, Elsevier.

Here,
we will discuss the probabilistic classifiers derived from Bayes’
theorem, including Bayes classifier, naive Bayes classifier and Bayesian
belief networks. 

2. Probabilistic Classifier (Text:8.3.1)
 What is Bayesian classifier?

A statistical (probabilistic) classifier: Predicts the probability of a given tuple belonging to a particular class
Foundation: Based on Bayes’ theorem. Bayes was a mid-18th century monk (apparently).

Performance: Comparable accuracy performance to decision tree and neural network classifiers
Computational performance is much enhanced by assuming class-conditional independence, in which case the method is called Naive Bayes.

Incremental:
Each training example can incrementally contribute to the
classification probabilities, so this allows adapting over time to
gradual or incremental changes in (labelled)  training data.
It is not really possible to humanly interpret  the results (i.e. it is known as a  “black box” method), although it’s relationship to its training data is straightforward to understand. 

2.1. Basic Probabilities (not in text)

Basic Probability Theory

Before discussing probabilistic classifiers, we recap basic probability theory first.

Event : A subset of outcomes of an experiment (a subset of event space).
Let’s assume that we roll a dice with six faces. If we observe number 3 from a single roll, then 3 is the event,

A set of observations can also be an event, signifying
any of the observations in the set.  For example, an event from a
dice roll can signify the outcome that either 1, 3, or 5 is rolled.

Event space (sample space): the set of all possible outcomes
e.g. {1,2,3,4,5,6} with a six-faced dice

Probability of event : probability of observing an event
e.g. probability of observing 5 from a single dice roll,

Joint probability : probability of observing multiple distinguishable events.
e.g. roll a dice and flip a coin, simultaneously. What would
be the probability of observing 3 from the dice and HEAD from the coin ?

Example

For an experiment, we roll a dice and flip a coin simultaneously, and record the first six trials as follows:

Trial #
Dice
Coin

1
1
H

2
2
T

3
1
T

4
3
H

5
4
H

6
1
T

Q: Given the above experiments, what is the probability of observing 3 from the dice? 

A:

Q: Given above experiments, what is the probability of observing Dice={1,2} from the dice? 

A:

Q: Given above experiments, what is the probability of observing 1 and TAIL from a single execution?

A:

Conditional probability

A conditional probability measures the probability of event given that another event has occurred. If and are events with , the conditional probability of given is . 

Example: Drug test

Let’s assume that we have 4000 patients who have taken a
drug test. The following table summarises the result of the drug
test. We categorise the result based on gender and test result.

Women
Men

Success
200
1800

Failure
1800
200

Let

represent gender

represent a result of a drug test

Then what is the probability of a patient being a woman when the patient fails on a drug test, i.e., ?

From these probabilities, we can compute the conditional probability

 

2.2. Bayes’ Theorem (Text 8.3.1)
Terminology

A running example: Let’s
assume that you are a owner of a computer shop. You may want to
identify which customers buy a computer for  targeting your
advertising. So you decide to record a customer’s age and credit rating whether the customer buys a computer or not for future predictions.

Evidence : A Bayesian term for observed data tuple, described by measurements made on a set of attributes.
E.g., record of customer’s information such as age and credit rating.

Sometimes the probability is also called evidence.

Hypothesis : A target of the classification. Hypothesis such that belongs to a specified class .
E.g.,  = buy computer, = not buy computer

Prior probability, :  the a priori probability of  
E.g., = the probability that any given customer will buy a computer regardless of age, or credit rating.

Likelihood, : the probability of observing the sample  given that the hypothesis holds.
E.g., Given that a customer, , will buy a computer, the probability that the customer is 35 years old and has fair credit rating.

Posterior probability, : the a posteriori probability, that is the probability that the hypothesis holds given the observed data .
E.g., Given that a customer, is 35 years old and has fair credit rating, the probability that

will buy a computer.

The prediction of a class for some new tuple   for which the class is unknown, is determined by the class which has the highest posterior probability.

Bayes’ Theorem

In
many cases, it is easy to estimate the posterior probabilty through
estimating the prior and likelihood of given problem from historical
data (i.e a training set).
E.g., to estimate the prior , we can count the number of customers who bought a computer and divide it by the total number of customers.

E.g., to estimate the likelihood , we can measure the proportion of customers whose age is 35 and have fair credit rating among the customers who bought a computer.

E.g., to estimate the evidence we can measure the proportion of customers whose age is 35 and have fair credit rating amongst all the
customers, irrespective of computer-buying. 

The posterior probability can then be computed from the prior and likelihood through Bayes’ theorem.

Bayes’ theorem provides a way to relate likelihood, prior, and posterior probabilities in the following way, when

Informally, this equation can be interpreted as

Posterior = likelihood x prior / evidence

Bayes’ theorem is used to predict  belongs to  iff the posterior   is the highest among all other for all the k classes. We can also state the probability that  belongs to is . Because we can give  this probability, we call Bayes classification a probablistic classifier. 

For determining the classification of some , we are looking to find the that maximises yet is the same for every ,
so can be ignored in all the calculations as long as we don’t need to know the probability.

ACTION: Bayes’ Theorem can be derived straightforwardly from  conditional probability. The derivation is given here if
you want to know.

Example: with training data

Let’s
assume that you are a owner of a computer shop. You may want to
identify which customers buy a computer for a targeted advertisement. So
the owner decided to record a customers’s age and credit rating no
matter the customer buys a computer or not. The following table shows a set of customer records in the computer shop. What is the probability of a customer who is youth and has fair credit rating buying a computer?

age
credit
buys_computer

youth
fair
no

youth
fair
yes

middle_aged
excellent
yes

middle_aged
fair
no

youth
fair
no

middle_aged
excellent
no

middle_aged
fair
yes

Prior: probability of a customer buying a computer regardless of their information.

Likelihood

Evidence

Posterior

Therefore, the customer would not buy a computer
When
computing a posterior, the evidence term is the same for all hypothesis
classes. Since our goal is to find the highest class,  the
evidence  term is often ignored in practice.

Example: with estimated probabilities

You might be interested in finding out a probability of patients having liver cancer if they are an alcoholic. In this scenario, we discover by using Bayes’
Theorem that  “being an alcoholic” is a useful diagnostic examination for liver cancer.

Prior:  means
the event “Patient has liver cancer.” Past data tells you that 1% of patients entering your clinic have liver disease. means the event “Patient does not have liver disease”.
,

Evidence: could mean the examination that “Patient is an alcoholic.” Five percent of the clinic’s patients are alcoholics.

Likelihood: You may also know from the medical literature that among those patients diagnosed with liver cancer, 70% are alcoholics.
; the probability that a patient is alcoholic, given that they have liver cancer, is
70%.

Bayes’
theorem tells you: If the patient is an alcoholic, their chances of
having liver cancer is 0.14 (14%). This is  much more than the
1%  prior probability suggested by past data.

ACTION: This 6.5 minute video explains the application of Bayes’ Theorem by example if you want more. https://www.khanacademy.org/partner-content/wi-phi/wiphi-critical-thinking/wiphi-fundamentals/v/bayes-theorem

2.3. Limitation (Text 8.3.2)
In
the following example, we would like to classify whether a certain
customer would buy a computer or not. We have a customer purchase
history as follows:

age credit buys_computer
youth fair no
youth fair yes
middle_aged excellent yes
middle_aged fair no
youth excellent no
middle_aged excellent no
middle_aged fair yes

What is the probability of (youth, excellent) customer buying a computer?

If we compute the likelihood :As we can see, we observe 0 likelihood for buying a computer with attribute (age=youth, credit=excellent).

Therefore, posterior probability of tuples with (age=youth, credit=excellent) will be 0:

This does not mean that every buyer with (age=youth, credit=excellent) would not buy a computer.
The data contains some information about customers who are youth or have excellent credit.
But the classifier ignores it because there  are no who are youth and have excellent credit.

It
is usual  to interpret this to mean that the number of
observations is too small to obtain a reliable posterior probability.
This tendency toward having zero probability will increase as we incorporate more and more attributes.
Because we need at least one observation for every possible combination of attributes and target classes.

In the next section, we will see that this problem is mitigated somewhat with naive Bayes  that assumes class conditional independence, but we will still need the Laplacian correction when there is some attribute value which has not been seen in some class in the training data.

3. Naive Bayes (Text:8.3.2)

Naive Bayes Classification method

Let
be a training set of tuples and their associated class labels, and each
tuple is represented by an n-Dimensional attribute vector

Suppose there are classes

Classification aims  to derive the maximum posteriori, i.e., the maximal using Bayes’ theorem 

Since P(X) is constant for all classes, we only need to maximise

For
Naive Bayes, we simplify Bayes’ theorem to reduce the computation cost
of each likelihood in the training phase. Instead of a computing and
recording a likelihood for each tuple for each class in our training
set, we summarise by computing a likelihood
for each attribute value for each class, that is, the class distribution
for each attribute value. Statistically, we are making an assumption
that, within each class, each attribute is independent of all the
others.

Class conditional independence:  We assume the object’s attribute values are conditionally independent of each other given a class label, so we can write

In other words, we factorise each attribute in the likelihood function, by assuming that there are no dependence relationships amongst the attributes.

This greatly reduces the computation cost as it only counts the class distribution

If is categorical, is the number of tuples in having value for divided by (number of tuples of in )

Blithely assuming class conditional independence of attributes is naive, hence the name of the method. It is not checked, and is commonly even known to be untrue, however, it seems to work, mostly.

Example

Let’s compute the likelihood of the previous example using the assumption of class conditional independence

age
credit
buys_computer

youth
fair
no

youth
fair
yes

middle_aged
excellent
yes

middle_aged
fair
no

youth
excellent
no

middle_aged
excellent
no

middle_aged
fair
yes

 

With the conditional independence assumption, the likelihood of tuple (youth, excellent) is

We can also see here that we have  mitigated the limitation observed earlier caused by the lack of observations for (youth, excellent) actually buying a computer.

Example 2

Here we have some more complex customer history with four different attributes.

age
income
student
credit
buys_computer

youth
high
no
fair
no

youth
high
no
excellent
no

middle_aged
high
no
fair
yes

senior
medium
no
fair
yes

senior
low
yes
fair
yes

senior
low
yes
excellent
no

middle_aged
low
yes
excellent
yes

youth
medium
no
fair
no

youth
low
yes
fair
yes

senior
medium
yes
fair
yes

youth
medium
yes
excellent
yes

middle_aged
medium
no
excellent
yes

middle_aged
high
yes
fair
yes

senior
medium
no
excellent
no

Compute prior probability on hypothesis:

Compute conditional probability  for each class
Attribute ‘age’

Attribute ‘ income’

Attribute ‘student’

Attribute ‘credit’

Predict probability of  buying computer 

Compute likelihood 

Compute

Therefore,  belongs to class 

3.1. Laplacian Correction
Zero-probability problem

Naïve
Bayesian prediction requires each class conditional probability to be
non-zero, as otherwise the predicted probability will be zero.

Example

Let’s assume that we extract following two tables for student and credit attributes from a customer history, where each entry represents a number of customers:

Buy computer \ Student
Yes
No

Yes
0
5

No
3
7

Buy computer \ credit
Fair
Excellent

Yes
4
1

No
6
4

Using naive Bayes, let’s classify the probability of a student with fair credit buying a computer. First, we need to compute the likelihood:

Therefore, the classifier will classify that the student will not buy a computer irrespective of the prior.
This is because no student has bought a computer ever before. In other words, the likelihood of student buying a computer:
,
indicates irrespective of the other attributes, the classifier will always classify a student tuple as not buy a computer. During the classification of an unlabelled tuple, all the other attributes have no effect if the student attribute is Yes.
This is not wrong, but inconvenient, as in some cases, the other
attributes may have a different opinion to contribute to the
classification of the tuple.

Laplace correction

To avoid the zero probability in the likelihood, we can simply add a small constant to the summary table as follows:

Buy computer \ Student
Yes
No

Yes
0+
5+

No
3
7

If we let , which is the usual value, then the likelihoods of naive Bayes are:

Using
the Laplacian correction with of 1, we pretend that we have
1 more tuple for each possible value for Student (i.e., Yes and No,
here)  but we only pretend this while computing the likelihood factors for the attribute and class combination which has a zero count in the data for at least one of its values.

Likelihood
for alternative(non-zero count) values of the affected attribute are
also affected, but this will come into play when we are predicting a different customer at a different time: e.g.

The likelihood for the other class (for the same student with fair credit)  is unchanged as before:

The “corrected” probability estimates are close to their “uncorrected” counterparts, yet the zero probability value is avoided.

3.2. Numerical attributes
So far, we’ve only considered the case when every attribute is a categorical or binary variable. However, numerical variables are common.

In
this section, we will show how to use a naive-Bayes classifier with a
continuous (numerical)  attribute. This approach can also be used
for ordinal variables, although depending on the application, and where
the range of possible values is small, it may be more useful to treat
ordinals as categorical  even though the information of the order
will not be used for prediction.

It is common to assume that a continuous attribute follows a Gaussian distribution (also called normal, or bell curve). 

Two parameters define a Gaussian distribution mean: and standard deviation

Probability density function of Gaussian:

Class conditional likelihood of th-continuous attribute given class is

To solve the equation for class conditional likelihood, we only need and , which are calculated as given earlier.

Example

Let’s assume that the attribute age is not discretized in the following example:

age
credit_rating
buys_computer

22
fair
no

23
fair
yes

35
excellent
yes

31
fair
no

20
excellent
no

38
excellent
no

40
fair
yes

Let buys_computer be a class label, then and .

The class conditional mean and variance of attribute age are:

Let be attributes of a future customer, the class conditional probability of this customer is:

This likelihood for each continuous variable can be used directly in the calculation of class conditional likelihood for Naive Bayes, combined with the likelihoods for discrete attributes. Via   Bayes theorem, we can then predict the probability of the customer buying a computer.

ACTION:
If you want more, here is a 40-minute youtube video working through a
small example of Naive Bayes with Laplacian correction and continuous
variables.

Naive Bayes classification with Laplace Correction

4. Bayesian Belief Networks (Text: 9.1)
Concept and Mechanism

Bayesian belief networks—probabilistic graphical models, which unlike naive Bayesian classifiers allow the representation of dependencies among subsets of attributes.
The
naive Bayesian classifier makes the assumption of class conditional
independence, that is, given the class label of a tuple, the values of
the attributes are assumed to be conditionally independent of one
another.
In practice, however, dependencies can exist between variables (attributes).
Bayesian belief networks provide a graphical model of causal relationships between attributes.
A belief network is defined by two components

a directed acyclic graph
Node: represents a random variable (attribute), can be discrete- or continuous-valued
Edge: represents
a probabilistic dependence, If an arc is drawn from a node Y to a
node Z, then Y is a parent or immediate predecessor of Z.

a set of conditional probability tables

Example

Simple
Bayesian belief network with six boolean variables. (a) A proposed
causal(graphical) model, represented by a directed acyclic graph. (b)
The conditional probability table for the values of the variable LungCancer (LC) showing each possible combination of the values of its parent nodes, FamilyHistory (FH) and Smoker (S). 

Causal relations:

having lung cancer is influenced by a person’s family history of lung cancer, as well as whether or not the person is a smoker.
Variable PositiveXRay
is independent of whether the patient has a family history of lung
cancer or is a smoker, given that we know the patient has lung cancer.
Once we know the outcome of the variable LungCancer, then the variables FamilyHistory and Smoker do not provide any additional information regarding PositiveXRay.

Variable LungCancer is conditionally independent of Emphysema, given its parents, FamilyHistory and Smoker.

Conditional probability table (CPT):

The CPT for a variable specifies the conditional distribution , where are the parents of . Figure (b) shows a CPT for the variable LungCancer. The conditional probability for each known value of LungCancer
is given for each possible combination of the values of its
parents. For instance, we can interpret the upper leftmost and
bottom rightmost entries as

More formally, let  be
a data tuple described by the variables. Recall that each variable
is conditionally independent of its nondescendants in the network
graph, given its parents. This allows the network to provide a complete
representation of the existing joint probability distribution with the
following equation:

,

where is the probability of a particular combination of values of , and the values for correspond to the entries in the CPT for .

4.1. Training a Belief Network (Text 9.1.2)
How to construct a directed network?

The network topology (or “layout” of nodes and arcs) may be constructed by human experts or alternatively inferred from the data.

The network variables may be observable or hidden in all or some of the training tuples. The hidden data case is also referred to as missing values or incomplete data.
Several algorithms exist for learning the network topology from the training data given observable variables. 
Human experts usually have a good grasp of the direct conditional dependencies that hold in the domain
under analysis, and can design the network topology. Typically, these
conditional dependencies are thought of causal relationships, e.g. that
Smoking causes LungCancer. Experts must specify conditional probabilities for some of the nodes that participate in these direct dependencies (some of the CPTs). These probabilities can then be used to compute the remaining probability values.

How to learn the network? 

If the network topology is known and all the variables are observable in the training data
Computing the CPT entries is straightforward (very like naive Bayes)

When the network topology is given and some of the variables are hidden
Several heuristic methods exist: many software packages provide solutions
The gradient descent method is well known: it works by treating each conditional probability as a weight.
It initialises the weights randomly up front and then iteratively
adjusts each one by a small amount to raise the product of the computed
probabilites of each datapoint in the training set. It stops when it is
not increasing the product any more.
This
is computationally demanding, but it has the benefit that human domain
knowledge is employed in the solution to design the network structure
and thereby to assign initial probability values.

5. Reading and Exercises
ACTION: Recommended, for a fuller explanation of Bayesian approaches, read

Hidden from students:FileDavid Heckerman, Bayesian Networks for Data Mining, 1997

ACTION: Try out these exercises

Hidden from students:PageExercise: Classification with employ databasePage

When you have had a go, you can check your answer against these worked answers:

Hidden from students:PageSolution to Exercise: Classification with employ database