程序代写代做代考 data mining database information retrieval algorithm IT enabled Business Intelligence, CRM, Database Applications

IT enabled Business Intelligence, CRM, Database Applications

Sep-18
Other Techniques
Prof. Vibs Abhishek
The Paul Merage School of Business
University of California, Irvine
BANA 273 Session 9

1

Agenda
Term Project Presentations next week
Upload presentation file to Canvas at least 1 hour before class
Overview of other techniques
Wiki for contributing final exam questions
https://docs.google.com/document/d/1LFkkveDdileus5zJOg-LfU5siOZT8ObUR0GrsbF3iVE/edit?usp=sharing

2

Attribute Selection
Weka – Correlation Based Feature (CFS) Selection
CfsSubsetEval
A good feature subset is one that contains features highly correlated with (predictive of) the class, yet uncorrelated with (not predictive of) each other.
CFS is a fully automatic algorithm — it does not require the user to specify any thresholds or the number of features to be selected, although both are simple to incorporate if desired

Other Methods
Text Mining
KNN
Collaborative filtering
Logistic Regression
Support Vector Machines (SVM)
Neural Nets
Bagging
Boosting

Three Forms of Web Mining
Content Data
Link Structure
Web usage data
Text Mining
Link Analysis
Web usage
mining
Data Available on the web
Three forms of web mining

5

Why Text Mining?
What can be discovered from text?
Significant proportion of information of great potential value is stored in documents:
News stories pertaining to competition, customers & the business environment at large
Technical reports on new technology
Email communications with customer, partners, and within the organization
Corporate documents embodying corporate knowledge and expertise
Legal documents — automatic reasoning

6

Opportunities
Finding patterns in text:
Identify and track trends in industry – associations
What are my competitors doing?
What relevant products are being developed?
What are the potential usage of my products?
Identify emerging themes in collections of documents -cluster
Customer communications: cluster messages, each segment identifies a common theme such as complaints about a certain problem, or queries about product features.
Automated categorization of e-mails (Spam Filter!), web pages, and news stories – classification

7

Structuring Textual Information
Many methods designed to analyze structured data
If documents can be represented by a set of attributes – can use existing data mining methods
How to represent a document ?
Structured
representation

Apply DM methods
to find patterns
among documents

8

Text Mining Concepts
Document
Token or term
Corpus
Bag of Words
Stop word elimination; Stemming; all lower case
Term Frequency (TF)
Inverse Document Frequency (IDF)
TFIDF
N-gram sequences
Named entity extraction
Topic models

Document Representation
A document representation aims to capture what the document is about
One possible approach:
Each row in the table represents a document
Attribute describes whether or not a term appears in the document
Example
Terms
Camera Digital Memory Pixel …
Document 1 1 1 0 1
Document 2 1 1 0 0
… … … … …

10

Document Representation using TF
Term Frequency:
Attributes represent the frequency in which a term appears in the document
TF(t, d)
May impose upper and lower limits on TF because the dimensionality is too high

Terms
Camera Digital Memory Print …
Document 1 3 2 0 1
Document 2 0 4 0 3
… … … … …

11

Inverse Document Frequency (IDF)
But a term is mentioned more times in longer documents
Therefore, use relative frequency (% of document): IDF(t) = 1 + log(Total # of docs/# docs containing t)
Terms
Camera Digital Memory Print …
Document 1 3 2 1 2
Document 2 1 1.4 1 3
… … … … …

12

Combining TF and IDF
TFIDF(t, d) = TF(t, d) * IDF (t)
Each row represents a document
Each column is an attribute (term)
You can use classifier, clustering etc. on this data

N-gram sequences
“The quick brown fox jumps”
2-grams or bi-grams:
{quick, brown, fox, jumps, quick_brown, brown-Fox, fox_jumps}
You can see that the number of columns can quickly get out of hand

Named entity extraction

Example “Silicon Valley”, “LA Lakers”, “Merage School of Business”

Topic Models

Topic Models
Examining the Impact of Keyword Ambiguity on Search Advertising Performance: A Topic Model Approach, Gong, Abhishek and Li (MISQ 2018)

Text Mining Application 1: Association Rules
Doc No. Microsoft antitrust France …
100 0 0 1
101 0 0 0
102 0 0 0
103 1 1 0
104 1 1 0

Document
No. Item 1 Item 2 Item 3 …
100 France Iraq US
101 NASDAQ NYSE job
102 Iraq US UK
103 Microsoft antitrust OS
104 Microsoft Antitrust windows

Microsoft

Antitrust
After proper representation, data mining techniques can be applied to text, e.g. association rules, clustering, classification.
Keyword-based Association Rules: treat keywords as items.
OR
Sentiment Analysis

18

Personalized Web Ad Delivery
Objective:
Improve effectiveness of Web ads
Customize ad delivery so that ad corresponds to the context user is exploring
Web content is dynamic  need automated ad placement
Example: Gmail
Solution:
Represent each ad as a document with a set of keywords.
For example: ad for hybrid car is represented by the following set of keyword: car, electric, environment, etc.
Then deliver ads to viewers of pages (i.e., documents) that resemble this description.

19

Link Structure Analysis to rank Web pages
Traditional Information Retrieval methods only examine the appearance of relevant terms, and often fail to account for
The quality of the information in the retrieved documents.
The reliability of the source

From the retrieved documents, want to rank authoritative documents higher

Approach: Mining the Web’s link structure to identify authoritative web pages

20

Identify Authoritative Web Pages
The Web includes pages and hyperlinks
A lot of information is in the structure of web page linkages. Hyperlinks contain rich latent human information
An author creates hyperlink pointing to another page — can be viewed as endorsement
The collective endorsement of a given page by different authors can help discover authoritative pages
Google uses link structure of the Web to rank documents (PageRank)

21

Using Hubs to identify Authoritative Web Pages
A hub is a page pointing to many good authorities.
E.g., a web page pointing to many good sources of information on business intelligence
A hub may not be an authority, and have very few links pointing to it.
Yet a link from a hub to a page is valued more than a link from a regular page
An authority is a page pointed to by many good hubs
Page
Hub
Page
Page
Page
Page
Page
Hub

Authority
Authority

Authorities are pages containing useful information
course home pages
home pages of auto manufacturers
Hubs are pages that link to authorities
course bulletin
list of US auto manufacturers

22

Personalization
Personalization/customization tailors certain offerings by providers to consumers based on knowledge about them with certain goals in mind.
Customer

Personalized
offerings
How?

23

Classifier: Logistic Regression
This is not a regression
Uses logistic function and hinge loss function

© Prof. V Choudhary, September 18

K Nearest Neighbor (KNN)
K-Nearest Neighbor can be used for classification/prediction tasks.

Step 1: Using a chosen distance metric, compute the distance between the new example and all past examples.

Step 2: Choose the k past examples that are closest to the new example.

Step 3: Work out the predominant class of those k nearest neighbors – the predominant class is your prediction for the new example. i.e. classification is done by majority vote of the k nearest neighbors. For prediction problem with numeric target variable, the (weighted) average of the k nearest neighbors is used as the predicted target value.

25

Each example is represented with a set of numerical attributes

“Closeness” is defined in terms of the Euclidean distance between two examples.
The Euclidean distance between X=(x1, x2, x3,…xn) and Y =(y1,y2, y3,…yn) is defined as:

John:
Age=35
Income=95K
No. of credit cards=3

Rachel:
Age=41
Income=215K
No. of credit cards=2
How do we determine our neighbors?

26

Example : 3-Nearest Neighbors
K-Nearest Neighbor Classifier
Customer Age Income No. credit cards Response
John 35 35K 3 No
Rachel 22 50K 2 Yes
Hannah 63 200K 1 No
Tom 59 170K 1 No
Nellie 25 40K 4 Yes
David 37 50K 2 ?

27

One seeks recommendations about movies, restaurants, books etc. from people with similar tastes
Automate the process of “word-of-mouth” by which people recommend products or services to one another.

Collaborative Filtering:
Finding like-minded people

28

Collaborative Filtering
Starts with a history of people’s personal preferences
Uses a distance function – people who like the same things are “close”
Determine a neighborhood size (say k closest data points). We will examine recommendations from this neighborhood only.
Typically k is between 20 and 50
Uses “votes” which are weighted by distances, so close neighbor votes count more

29

Example:

Implicit rating

30

Artificial Neural Networks

An artificial neural network (ANN), usually called neural network (NN), is a mathematical model or computational model that is inspired by the structure and/or functional aspects of biological neural networks. — Wikipedia

A neural network consists of an interconnected group of artificial neurons, and it processes information using a connectionist approach to computation. — Wikipedia

Neural Nets learn complex functions Y=f(X) from data.

ANN can approximate any function (e.g. logistic regression, linear regression).

31

32
Components of Neural Nets
Neural Nets are composed of
Nodes, and
Arcs
Each arc specifies a weight.
Each node (other than the input nodes) contains a Transfer Function which converts its inputs to outputs. The input to a node is the weighted sum of the inputs from its arcs.

32

Recommender Systems
Collaborative Filtering

Content Based Recommendation
Use document content to create a description (tags)
Create user profile with weights for different tags
Example Books: Genre, Author, Length, Pictures etc.
Knowledge Based Recommendation
When we do not have history of purchases (Camera)
Examine customer needs and match to product features

© Prof. V Choudhary, September 18

33

Bagging
Combining predictions by voting/averaging
Each model receives equal weight
“Idealized” version:
Sample several training sets of size n
(instead of just having one training set of size n)
Build a classifier for each training set
Combine the classifiers’ predictions

© Prof. V Choudhary, September 18

34

Bagging classifiers
Let n be the number of instances in the training data
For each of t iterations:
Sample n instances from training set
(with replacement)
Apply learning algorithm to the sample
Store resulting model
For each of the t models:
Predict class of instance using model
Return class that is predicted most often
Model generation
Classification

Boosting
Also uses voting/averaging
Weights models according to performance
Several variants
Read text for AdaBoost

© Prof. V Choudhary, September 18

36

Link Analysis is used for …
A: Identifying similar consumers for product recommendations
B: Highly non-linear classification
C: Replicating logistic regression
D: Determining which web sites or documents are more authoritative and credible.
E: None of the above

© Prof. V Choudhary, September 18

37

Next Session
Project Presentations
All Students must attend
Please upload slides/files to drop-box on Canvas at least 1 hour before class.

© Prof. V Choudhary, September 18

38

å
=

=
n
i
i
i
y
x
Y
X
D
1
2
)
(
)
,
(

/docProps/thumbnail.jpeg