def tokenize_and_normalize(string, lower=True): if lower: return nltk.wordpunct_tokenize(normalize(string).lower().strip()) else: return nltk.wordpunct_tokenize(normalize(string).strip())