安装高可用Hadoop生态 (四) 安装Spark
2024-10-04 10:47:48
4. 安装Spark
4.1. 准备目录
tar -xf /home/hadoop/install/spark-2.1.-bin-without-hadoop.tgz -C /opt/cloud/packages/
ln -s /opt/cloud/packages/spark-2.1.-bin-without-hadoop /opt/cloud/bin/spark
ln -s /opt/cloud/packages/spark-2.1.-bin-without-hadoop/conf /opt/cloud/etc/spark
mkdir -p /opt/cloud/logs/spark
mkdir -p /opt/cloud/data/spark
4.2. 设置环境变量
vi ~/.bashrc
export SPARK_HOME=/opt/cloud/bin/spark
export PATH=$SPARK_HOME/bin:$PATH
即刻生效
source ~/.bashrc
4.3. 修改配置文件
4.3.1. spark-env.sh
cd /opt/cloud/etc/spark
mv spark-env.sh.template spark-env.sh
vi spark-env.sh
export JAVA_HOME=/usr/lib/jvm/java
export HADOOP_HOME=/opt/cloud/bin/hadoop
export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop
export SPARK_DIST_CLASSPATH=$(hadoop classpath)[]
export SPARK_LOCAL_DIRS=/opt/cloud/data/spark
export SPARK_LOG_DIR=/opt/cloud/logs/spark
export SPARK_PID_DIR=/opt/cloud/hdfs/tmp
4.3.2. spark-defaults.conf
mv spark-defaults.conf.template spark-defaults.conf
vi spark-defaults.conf
spark.driver.memory 512m
4.4. 测试
通过修改MASTER临时变量,测试yarn模式
export MASTER=yarn
./bin/run-example SparkPi
4.5. 卸载
复原环境变量,删除spark相关行
vi ~/.bashrc
删除临时数据和目录
rm /opt/cloud/bin/spark
rm /opt/cloud/etc/spark -rf
rm /opt/cloud/packages/spark-2.1.-bin-without-hadoop/ -rf
rm /opt/cloud/data/spark/ -rf
rm /opt/cloud/logs/spark/ -rf
[1] without hadoop 时需要
最新文章
- ES6的Iterator,jquery Fn
- Servers
- Jenkins 2.26 发布,可扩展的持续集成引擎
- 2013 ACM/ICPC 长沙网络赛J题
- cookie文件在电脑的保存位置
- [BZOJ 1072] [SCOI2007] 排列perm 【状压DP】
- Spark GraphX的函数源码分析及应用实例
- 同一台电脑启动两个或多个tomcat
- iMac 无线键盘 无法配对
- 转换器3:手写PHP转Python编译器,词法部分
- yum错误,Cannot find a valid baseurl for repo: base 和 No more mirrors to try
- Aplication的意义和生命周期,与Context的关系,以及关于Aplication和Context相关问题的记录和解决办法
- 决胜 Poker
- 【大数据系列】windows搭建hadoop开发环境
- 【LeetCode】48. Rotate Image (2 solutions)
- JSF多列预选择
- [转]MVP+WCF+三层结构搭建项目框架
- (工具类)MD5算法|时间格式转换|字符串转数字
- bugfree登录后报错PHP Fatal error: Call-time pass-by-reference has been removed in
- C# - Generics泛型,一图话c#泛型