Text Preprocessing For Unsupervised Learning: Why It Matters, When ...

6 downloads 194 Views 1MB Size Report
10-fold cross validation. ▷ Split data into train/test sets (80/20). ▷ Find minimum perplexity over num. topics. ▷
Text Preprocessing For Unsupervised Learning: Why It Matters, When It Misleads, And What To Do About It Matthew J. Denny1 Penn State University

Arthur Spirling New York University

October 15, 20016

1

Work supported by NSF Grant: DGE-1144860

Text-As-Data Research 1. Awesome Research Design! 2. Collect Awesome Text Data! 3. ... 4. Perform Awesome Analysis! 5. Publish Awesome Paper!

... Raw Text

Preprocessing Document-Term Matrix amend

federal

section

spending



56

34

20

75



24

13

41

0













Common Preprocessing Decisions P N L S W I ‘3’

– – – – – – –

Punctuation Removal Number Removal Lowercasing Stemming Stopword Removal Infrequent Term Removal n-gram Inclusion

7 binary choices −→ 27 = 128 specifications.

Supervised Learning

Unsupervised Learning

What Could Possibly Go Wrong?

Motivating Example I

UK Manifestos Corpus (1918–2001)

I

Labour, Liberal, Conservative Parties

I

Wordfish I

I

Place documents in ideological space

Process: 1. Select preprocessing specification 2. Run Wordfish

1983 Labour Manifesto

A-Priori Rankings I

Focus on 8 Manifestos. 1. Four general elections (1983–1997) 2. Labour and Conservative parties

I

Lab 1983: “longest suicide note in history”, extremely left–wing.

Lab

1983

Con

< Lab

1992

1987

< Con

< Lab

1997

1992

< Con

< Lab

1987

1997

< Con

Suggest Documents