Changes between Version 4 and Version 5 of tutorial/ProbabilisticLearningModels


Ignore:
Timestamp:
08/14/09 09:40:11 (15 years ago)
Author:
horak
Comment:

--

Legend:

Unmodified
Added
Removed
Modified
  • tutorial/ProbabilisticLearningModels

    v4 v5  
    99== Introduction == 
    1010Classification in general is the problem of deciding for a given input to which class it belongs. Usually classification can be subdivided into a learning phase (aka training phase) and a classification phase (aka test phase). (TODO: offline, online, reinforcement, ... learning) 
     11 
     12=== Basics === 
     13 
     14There are several basics to concern and understand before diving into probabilistic learning models. 
     15 
     16==== Example / Instance ==== 
     17Examples or also called instances are the basic entities in this field. 
     18They occur as training examples, as validation or test examples, and finally as real data. 
     19 
     20{{{ 
     21E.g., In a document classification scenario, examples are documents.  
     22Already classified documents are used for training or evaluation purpose. 
     23}}} 
     24 
     25==== Feature ==== 
     26A feature is a descriptive property of an example. Features are processible by machines.  
     27{{{ 
     28E.g., In a document classification scenario, features might be the words of a document. 
     29In consequence, single features might describe multiple examples (here documents) 
     30}}} 
     31==== Feature Extraction ==== 
     32Feature extraction is the task of extracting features from examples. 
     33{{{ 
     34E.g., In our document classification scenario, a tokenizer that extracts words from text might be used for feature extraction. 
     35}}} 
     36In more sophisticated scenarios, feature extraction can be hierarchically nested by extracting new features from existing feature lists. 
     37{{{ 
     38E.g., In our document classification scenario, a word n-gram algorithm extracts n-gram features from extracted word sequences.  
     39}}} 
     40 
     41==== Feature Selection ==== 
     42Each feature for each example has be processed by model trainers or executors. 
     43There are several reasons for selecting just subsets of existing features. 
     44First, not all features are useful for separating different classes. In details, there is no statistically significant dependency between class and feature occurance. 
     45{{{ 
     46E.g., In our document classification scenario, stop words or high frequent words are not useful for separating e.g., spam mails from ham mails. 
     47}}} 
     48Second, just a small set of features might be enough for classifiying examples successfully. Adding more just decreases  performance.  
    1149 
    1250=== Relational Classification === 
     
    4280 
    4381 
     82 
     83 
    4484== Appendix == 
    4585=== Mathematical Foundations ===