Hadoop二进制文件输入错误



我得到一个错误像kkslct得到这里:http://eldadlevy.wordpress.com/2011/02/05/hadoop-binary-files-processing-entroduced-by-image-duplicates-finder/kkslct解决了这个问题,但我不能。什么好主意吗?

代码是:

public class BinaryFilesToHadoopSequenceFileMapper extends Mapper<Object, Text, Text, BytesWritable> {
public void map(Object key, Text value, Context context) 
throws IOException, InterruptedException {

    String uri = value.toString();
    Configuration conf = new Configuration();
    FSDataInputStream in = null;
    try {
        FileSystem fs = FileSystem.get(URI.create(uri), conf); ERROR HERE

得到如下错误:错误BinaryFilesToHadoopSequenceFileMapper错误,uri: L���2���,fR���T�Ӑ% 2���板式换热器Z��әh��T�����z�/�#��:� null class java.lang.IllegalArgumentException with Cause: java.net.URISyntaxException: Illegal character in path at index 13: L���2���;fR���ӐT�%2���PHe�Z�әh��T������Z/�#��:�

它曾经工作!!

真是浪费了半天时间。我忘记了:输入是一个文件,包含所有的图像文件作为HDFS的文件名。例如:

hdfs://localhost: 8022/user/elevy/smallArchiveImages WonderWoman.jpg

http://eldadlevy.wordpress.com/2011/02/05/hadoop-binary-files-processing-entroduced-by-image-duplicates-finder/

最新更新