Tuesday, November 8, 2011

10/27/2011

Representations of text are very high dimensional, methods that sum evidence from many or all features (e.g. naïve Bayes, KNN, neural-net) tend to work better than ones that try to isolate just a few relevant features.
Naive Bayes classifier based on bayes networks and it works better if smoothing probability sometimes and we can use M-estimates to improve probablity estimates.