These code are used to train the model used for the competition "Real or Not? NLP with Disaster Tweets"
https://www.kaggle.com/c/nlp-getting-started/overview/description
- Normal Machine Learning
- Topic Modeling output as features to train Supervised Machine Learning
- Neural Network (GloVe, LSTM)
- Bernoulli Naive Bayes with Lemmatization, OneHotEncoder, CountVectorizer -> 78.92 %
- 222 Topic Unigram with Random Forest Classifier -> 61.87 %
- GloVe With LSTM -> 79.20 %
By far the approach that generate the best accuracy is Neural Network using GloVe with LSTM