#!/usr/bin/env python # coding: utf-8 # In[3]: import os from nltk.tokenize import word_tokenize from nltk.corpus import stopwords from nltk import ngrams import string from nltk.stem import PorterStemmer import matplotlib.pyplot as plt from wordcloud import WordCloud from collections import Counter from sklearn.feature_extraction.text import TfidfVectorizer from sklearn.svm import LinearSVC from sklearn.feature_extraction.text import TfidfVectorizer from sklearn.svm import LinearSVC from pprint import pprint from sklearn.preprocessing import MultiLabelBinarizer from sklearn.multiclass import OneVsRestClassifier from sklearn.metrics import f1_score, precision_score, recall_score from nltk.corpus import reuters from operator import itemgetter # Test that verifies that the imports don't fail and that we have the correct files from nltk downloaded reuters.fileids() stopwords.words('english') word_tokenize('This is just a test') print('You are all set') # In[2]: get_ipython().system(' pip install wordcloud') # In[ ]: