part-r-00000没有这样的文件或目录。

问题描述 投票:0回答:1

我试图运行AggregateWordCount的例子,但我得到了这个错误,当我使用这个代码运行Wordcount时,一切都很好。

!/bin/bash
# test the hadoop cluster by running wordcount

# create input files
mkdir input
echo "Hello World" >input/file2.txt
echo "Hello Hadoop" >input/file1.txt

# create input directory on HDFS
hadoop fs -mkdir -p input
put input files to HDFS
hdfs dfs -put ./input/* input

# run wordcount
#hadoop jar $HADOOP_HOME/share/hadoop/mapreduce/sources/hadoop-mapreduce-examples-2.7.7-sources.jar org.apache.hadoop.examples.WordCount input output

hadoop jar $HADOOP_HOME/share/hadoop/mapreduce/sources/hadoop-mapreduce-examples-2.7.7-sources.jar org.apache.hadoop.examples.AggregateWordCount input output

# print the input files
echo -e "\ninput file1.txt:"
hdfs dfs -cat input/file1.txt
echo -e "\ninput file2.txt:"
hdfs dfs -cat input/file2.txt

# print the output of wordcount
echo -e "\nwordcount output:"
hdfs dfs -cat output/part-r-00000
hadoop mapreduce
1个回答
1
投票

你在hdfs上用下面的命令做一个递归查找,你写的文件与 output/part-r-00000

hadoop fs -ls -R /user/your_directory grep -i "output/part-r-00000"

应该递归地列出目录。

调整你的代码或脚本指向这个。

© www.soinside.com 2019 - 2024. All rights reserved.