def tokenize_to_sentence(text):
parser = nltk.data.load('tokenizers/punkt/english.pickle')
# split into sentences
sentences = parser.tokenize(text.strip())
return [lemmatize_sentence(sentence) for sentence in sentences]
smell_datamine_multiprocessing.py 文件源码
python
阅读 43
收藏 0
点赞 0
评论 0
评论列表
文章目录