如何获得每一篇文章的相关性评分进行分类,Python NLP



这里有一个代码,这个代码将文本分为10类,它在最后显示了算法的整体准确性:

import numpy as np
from sklearn.naive_bayes import MultinomialNB
from sklearn.feature_extraction.text import CountVectorizer
import pandas as pd
from collections import Counter
from sklearn.model_selection import train_test_split
from sklearn.feature_extraction.text import TfidfTransformer
df = pd.read_csv('data/wine_data.csv')
counter = Counter(df['variety'].tolist())
top_10_varieties = {i[0]: idx for idx, i in enumerate(counter.most_common(10))}
df = df[df['variety'].map(lambda x: x in top_10_varieties)]
description_list = df['description'].tolist()
varietal_list = [top_10_varieties[i] for i in df['variety'].tolist()]
varietal_list = np.array(varietal_list)
count_vect = CountVectorizer()
x_train_counts = count_vect.fit_transform(description_list)

tfidf_transformer = TfidfTransformer()
x_train_tfidf = tfidf_transformer.fit_transform(x_train_counts)
train_x, test_x, train_y, test_y = train_test_split(x_train_tfidf,   varietal_list,test_size=0.3)
clf = MultinomialNB().fit(train_x, train_y)
y_score = clf.predict(test_x)
n_right = 0
for i in range(len(y_score)):
if y_score[i] == test_y[i]:
n_right += 1
print("Accuracy: %.2f%%" % ((n_right/float(len(test_y)) * 100))) code here

我的问题是,如何获得数据集中每篇文章的相关性分数,如下所示:

相关性得分

您可以使用predict_proba方法查看测试集返回的概率估计。用_classes压缩它应该会给您相应的相关性。

probs = clf.predict_proba(test_x)
for i in range(len(test_x)):
probs_classes = list(zip(clf._classes, probs[i]))
print(f"X = {test_x[i]}, Predicted = {probs_classes}")

相关内容

  • 没有找到相关文章

最新更新