大数据 爬虫Python AI Sql玩转大数据大数据

Spark local mode 报Input path doe

2018-02-28  本文已影响0人  sparkle123
cd C:\Users\Administrator\IdeaProjects\SparkSQLProject>
mvn clean package -DskipTests
spark-submit \
  --class  com.xxx.cn.SQLContextApp \
  --master local[2]  \
    home/hadoop/Downloads/sparksql-1.0.jar \
    home/hadoop/app/spark-2.2.0-bin-2.6.0-cdh5.7.0/examples/src/main/resources/people.json

Exception in thread "main" org.apache.spark.sql.AnalysisException: Path does not exist: hdfs://hadoop000:8020/home/hadoop/app/...

Try explicitly specify sc.textFile("file:///path to the file/"). The error occurs when Hadoop environment is set.

SparkContext.textFile internally calls org.apache.hadoop.mapred.FileInputFormat.getSplits, which in turn uses org.apache.hadoop.fs.getDefaultUri if schema is absent. This method reads "fs.defaultFS" parameter of Hadoop conf. If you set HADOOP_CONF_DIR environment variable, the parameter is usually set as "hdfs://..."; otherwise "file://".

export  JAVA_HOME=/home/hadoop/app/jdk1.8.0_144
export  SPARK_MASTER_HOST=hadoop000
export  SPARK_WORKER_CORES=2
export  SPARK_WORKER_MEMORY=1g
export  SPARK_WORKER_INSTANCES=2
export HADOOP_HOME=/home/hadoop/app/hadoop-2.6.0-cdh5.7.0
#export HADOOP_CONF_DIR=/home/hadoop/app/hadoop-2.6.0-cdh5.7.0/etc/hadoop
上一篇 下一篇

猜你喜欢

热点阅读