import jiebatxt = open("wuxi.txt", encoding="utf-8").read() #'wuxi.txt' 更换你的文件(txt格式)
def jiebafenci(txt,wordslist):jieba.load_userdict('tingcibiao.txt')words = jieba.lcut(txt) counts = {} for word in words: counts[word] = counts.get(word,0) + 1 lst=[]for i in range(len(wordslist)):try :print(wordslist[i],counts[wordslist[i]])except:lst.append(wordslist[i])print('不存在的词:',lst)
if __name__=='__main__':txt = open("wuxi.txt", encoding="utf-8").read() #'wuxi.txt' 更换你的文件(txt格式)need_words = open("tingcibiao.txt", encoding="utf-8").read() #这个是要查找的词的txt文件 每个词一行find=need_words.split()jiebafenci(txt,find)
首先安装jieba库,打开Anaconda Prompt (或其他编辑器)输入pip install jieba
tingcibiao.txt 文件内容如下图(停词表另有含义,此处为不规范命名)
:
统计结果为
