Train bigram and trigram models on NLTK corpora.
Calculate conditional probabilities of words.
Interactive console-based predictions.
Works with multiple corpora (Gutenberg, Brown, Webtext, etc.).
Shows how probability-based text prediction works in NLP.
Clone the repository:
git clone https://github.com/WasifAsad/N-Gram-Language-Model-to-predict-next-word.git
cd N-Gram-Language-Model-to-predict-next-wordAdd smoothing techniques (Laplace, Kneser-Ney).
Support larger n-grams (4-grams, 5-grams).
Create a simple GUI or web app version.
Allow combining multiple corpora for training.