jieba 分词. 西游记相关的分词,出现次数最高的20个。

 1 import  jieba
 2  
 3 txt = open("D:\\西游记.txt", "r", encoding='gb18030').read()
 4 words = jieba.lcut(txt)     # 使用精确模式对文本进行分词
 5 counts = {}     # 通过键值对的形式存储词语及其出现的次数
 6  
 7 for word in words:
 8     if len(word) == 1:
 9         continue   
10     elif word == "大圣" or word=="老孙" or word=="行者" or word=="孙大圣" or word=="孙行者" or word=="猴王" or word=="悟空" or word=="齐天大圣" or word=="猴子":
11         rword = "孙悟空"
12     elif word == "师父" or word == "三藏" or word=="圣僧":
13         rword = "唐僧"
14     elif word == "呆子" or word=="八戒" or word=="老猪":
15         rword = "猪八戒"
16     elif word=="沙和尚":
17         rword="沙僧"
18     elif word == "妖精" or word=="妖魔" or word=="妖道":
19         rword = "妖怪"
20     elif word=="佛祖":
21         rword="如来"
22     elif word=="三太子":
23         rword="白马"
24     else:
25         rword = word
26     counts[rword] = counts.get(rword,0) + 1
27          
28 items = list(counts.items())#将键值对转换成列表
29 items.sort(key=lambda x: x[1], reverse=True)    # 根据词语出现的次数进行从大到小排序
30  
31 for i in range(20):
32     word, count = items[i]
33     print("{0:<10}{1:>5}".format(word, count))

jieba 分词.      西游记相关的分词,出现次数最高的20个。

上一篇:jieba分词


下一篇:【E-21】AttributeError: 'float' object has no attribute 'decode'