scikit-learn CountVectorizer UnicodeDecodeError



我有以下代码片段,我试图列出术语频率,其中first_textsecond_text.tex文档:

from sklearn.feature_extraction.text import CountVectorizer
training_documents = (first_text, second_text)  
vectorizer = CountVectorizer()
vectorizer.fit_transform(training_documents)
print "Vocabulary:", vectorizer.vocabulary 

当我运行脚本时,我得到以下内容:

File "test.py", line 19, in <module>
    vectorizer.fit_transform(training_documents)
  File "/usr/local/lib/python2.7/site-packages/sklearn/feature_extraction/text.py", line 817, in fit_transform
    self.fixed_vocabulary_)
  File "/usr/local/lib/python2.7/site-packages/sklearn/feature_extraction/text.py", line 752, in _count_vocab
    for feature in analyze(doc):
  File "/usr/local/lib/python2.7/site-packages/sklearn/feature_extraction/text.py", line 238, in <lambda>
    tokenize(preprocess(self.decode(doc))), stop_words)
  File "/usr/local/lib/python2.7/site-packages/sklearn/feature_extraction/text.py", line 115, in decode
    doc = doc.decode(self.encoding, self.decode_error)
  File "/usr/local/Cellar/python/2.7.11/Frameworks/Python.framework/Versions/2.7/lib/python2.7/encodings/utf_8.py", line 16, in decode
    return codecs.utf_8_decode(input, errors, True)
UnicodeDecodeError: 'utf8' codec can't decode byte 0xa2 in position 200086: invalid start byte

如何解决这个问题?

谢谢。

如果您可以计算出您的文档的编码是什么(也许它们是latin-1),您可以将其传递给CountVectorizer

vectorizer = CountVectorizer(encoding='latin-1')

否则,您可以直接跳过包含有问题字节的令牌,使用

vectorizer = CountVectorizer(decode_error='ignore')

相关内容

  • 没有找到相关文章

最新更新