Python implementation of an N-gram language model with Laplace smoothing and sentence generation.
-
Updated
Feb 9, 2018 - Python
Python implementation of an N-gram language model with Laplace smoothing and sentence generation.
Ngrams with Basic Smoothings
Programming for NLP Project - Implement a basic n-gram language model and generate sentence using beam search
Ngram language model implemented in Pharo
N-Gram language model that learns n-gram probabilities from a given corpus and generates new sentences from it based on the conditional probabilities from the generated words and phrases.
Generating text using N-gram models that were trained on shakespeare.txt and warpeace.txt. n-gram models are widely used in statistical natural language processing. In speech recognition, phonemes and sequences of phonemes are modeled using a n-gram distribution.
Ngrams with Basic Smoothings
Markov model for generating fake headlines ✏️
Next-token prediction in JavaScript — build fast language and diffusion models.
Built a system from scratch in Python which can detect spelling and grammatical errors in a word and sentence respectively using N-gram based Smoothed-Language Model, Levenshtein Distance, Hidden Markov Model and Naive Bayes Classifier.
A general emoji-text translator which translates emoji-text to chinese
This project is an auto-filling text program implemented in Python using N-gram models. The program suggests the next word based on the input given by the user. It utilizes N-gram models, specifically Trigrams and Bigrams, to generate predictions.
Ngrams with Basic Smoothings
It's a python based n-gram langauage model which calculates bigrams, probability and smooth probability (laplace) of a sentence using bi-gram and perplexity of the model.
A C++ library implementing fast language models estimation using the 1-Sort algorithm.
Basic Language Models , Bag of Words, Ngram Models Etc NLP modelling and associated tasks
Language identification toolkit for identifying what language a document is writen in
Prediction of the following words using N-gram technique.
Next Word Prediction model in UZBEK language
Part of a semester project this grammarchecker uses a n-gram language model to detect grammatical errors and a bert model is used to generate suggestions.
Add a description, image, and links to the ngram-language-model topic page so that developers can more easily learn about it.
To associate your repository with the ngram-language-model topic, visit your repo's landing page and select "manage topics."