def tokenize(sentence):
cn_sent = get_cnstr(sentence)
term_list = jieba.lcut(cn_sent, cut_all=False)
final_term_list = [term for term in term_list if len(term)>1 and is_cn_char(term)]
return final_term_list
评论列表
文章目录