📙 Language Model
= a 📊 Probability distribution over words or word sequences
How?
Unigrams
- Only use word features and use all words → 🤦♂️ Naive Bayes
N-Grams (2+)
-
Compute the conditional 🎲 Probability for all words w of W:
-
Choose the word combination with the highest conditional probability (= most likely follow up word)
🔗 Links