def word_tokenization(tick_blog_list):
'''
word tokenization by jieba to list
return list : [[,], [,], ...]
'''
count = 0
seg_list = []
try:
for blog in tick_blog_list:
if blog != '':
count += 1
segments = jieba.cut(blog)
tmp = []
for seg in segments:
tmp.append(seg)
seg_list.append(tmp)
except IOError as e:
logging.error('IOError %s' % e)
finally:
return seg_list
# Python????????
sentiment.py 文件源码
python
阅读 29
收藏 0
点赞 0
评论 0
评论列表
文章目录