栏目分类:
子分类:
返回
名师互学网用户登录
快速导航关闭
当前搜索
当前分类
子分类
实用工具
热门搜索
名师互学网 > IT > 前沿技术 > 大数据 > 大数据系统

python 使用 SPARK Hadoop

python 使用 SPARK Hadoop

下载相应包。然后放到linux 相关目录,然后配置环境变量,配置文件如下

vim ~/.bash_profile

# .bash_profile

# Get the aliases and functions
if [ -f ~/.bashrc ]; then
        . ~/.bashrc
fi

# User specific environment and startup programs

PATH=$PATH:$HOME/.local/bin:$HOME/bin

export PATH

#java setting
export JAVA_HOME=/home/handoop/app/jdk1.8.0_91
export PATH=$JAVA_HOME/bin:$PATH

#scala setting
export SCALA_HOME=/home/handoop/app/scala-2.11.8
export PATH=$SCALA_HOME/bin:$PATH

#hadoop setting
export HADOOP_HOME=/home/handoop/app/hadoop-2.6.0-cdh5.7.0
export PATH=$HADOOP_HOME/bin:$PATH

#maven setting
export MAVEN_HOME=/home/handoop/app/apache-maven-3.3.9
export PATH=$MAVEN_HOME/bin:$PATH

#spak setting
export SPARK_HOME=/home/handoop/app/spark-2.3.0-bin-2.6.0-cdh5.7.0
export PATH=$SPARK_HOME/bin:$PATH

# pthon qidong
export PYSPARK_PYTHON=python3
#build run .py code
export PYTHonPATH=$SPARK_HOME/python:$SPARK_HOME/python/lib/py4j-0.10.6-src.zip

windows 下使用pyCharm 中使用spark

配置环境变量

1,Javahome:

JAVA_HOME:C:Program FilesJavajdk1.8.0_321  (java 必须是1.8及以上。我用1.5踩坑)

系统变量:

PATH:

C:Program FilesJavajdk1.8.0_321bin

E:WooPythonPython_Sparkhadoop-2.6.0bin

HADOOP_HOME:E:WooPythonPython_Sparkhadoop-2.6.0

pycharm 的Edit Configuration 设置Environment variables:

PYTHonUNBUFFERED=1;SPARK_HOME=E:WooPythonPython_Sparkspark-2.3.0-bin-2.6.0-cdh5.7.0;PYTHonPATH=E:WooPythonPython_Sparkspark-2.3.0-bin-2.6.0-cdh5.7.0python

工具/File/Settings/Project Structure /Add Content Root

E:WooPythonPython_Sparkspark-2.3.0-bin-2.6.0-cdh5.7.0pythonlibpy4j-0.10.6-src

E:WooPythonPython_Sparkspark-2.3.0-bin-2.6.0-cdh5.7.0pythonlibpyspark.zip

转载请注明:文章转载自 www.mshxw.com
本文地址:https://www.mshxw.com/it/780887.html
我们一直用心在做
关于我们 文章归档 网站地图 联系我们

版权所有 (c)2021-2022 MSHXW.COM

ICP备案号:晋ICP备2021003244-6号