def extractSentences(document):
# sent_detector = nltk.data.load('tokenizers/punkt/english.pickle')
# sentenceTokens = sent_detector.tokenize(text.strip())
sentenceTokens = document.sentences()
graph = buildGraph(sentenceTokens)
calculated_page_rank = nx.pagerank(graph, weight='weight')
#most important sentences in ascending order of importance
sentences = sorted(calculated_page_rank, key=calculated_page_rank.get, reverse=True)
#return a 100 word summary
summary = ' '.join(sentences)
summaryWords = summary.split()
summaryWords = summaryWords[0:101]
summary = ' '.join(summaryWords)
return summary
评论列表
文章目录