日韩性视频-久久久蜜桃-www中文字幕-在线中文字幕av-亚洲欧美一区二区三区四区-撸久久-香蕉视频一区-久久无码精品丰满人妻-国产高潮av-激情福利社-日韩av网址大全-国产精品久久999-日本五十路在线-性欧美在线-久久99精品波多结衣一区-男女午夜免费视频-黑人极品ⅴideos精品欧美棵-人人妻人人澡人人爽精品欧美一区-日韩一区在线看-欧美a级在线免费观看

歡迎訪問 生活随笔!

生活随笔

當前位置: 首頁 > 编程资源 > 编程问答 >内容正文

编程问答

spark集群详细搭建过程及遇到的问题解决(四)

發布時間:2023/12/9 编程问答 35 豆豆
生活随笔 收集整理的這篇文章主要介紹了 spark集群详细搭建过程及遇到的问题解决(四) 小編覺得挺不錯的,現在分享給大家,幫大家做個參考.

spark集群詳細搭建過程及遇到的問題解決(三)中,我們將講述了hadoop的安裝過程,在本文中將主要講述spark的安裝配置過程。

spark@master:~/spark$?cd?hadoop spark@master:~/spark/hadoop$?cd?$SPARK_HOME/conf spark@master:~/spark/spark/conf$?cp?slaves.template?slaves? spark@master:~/spark/spark/conf$?vim?slaves

添加以下內容

spark@master:~/spark/spark/conf$?cp?spark-env.sh.template?spark-env.sh

spark-env.sh?Spark進程啟動時需要加載的配置

改模板配置中有選項的具體說明,此處參考稍微加入了一些配置:

spark@master:~/spark/spark/conf$?vim?spark-env.sh

添加以下內容

export?SPARK_PID_DIR=/home/spark/spark/spark/tmp/pid export?SCALA_HOME=/home/spark/spark/scalaexport?JAVA_HOME=/home/spark/spark/jdk export?HADOOP_HOME=/home/spark/spark/hadoop export?SPARK_MASTER_IP=master export?SPARK_MASTER_PORT=7077 export?SPARK_WORKER_MERMORY=2G export?HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop

注意更改相應的目錄,并保存。

spark@master:~/spark/spark/conf$??hadoop?fs?-mkdir?hdfs://master:9000/sparkHistoryLogs mkdir:?Cannot?create?directory?/sparkHistoryLogs.?Name?node?is?in?safe?mode.

發現不能創建,并提示Name node處于安全模式,因此我們先關掉安全模式

spark@master:~/spark/spark/conf$?hdfs?dfsadmin?-safemode?leave Safe?mode?is?OFF

重新建立

spark@master:~/spark/spark/conf$??hadoop?fs?-mkdir?hdfs://master:9000/sparkHistoryLogsspark@master:~/spark/spark/conf$?cp?spark-defaults.conf.template?spark-defaults.conf

配置spark-defaults.conf,該文件為spark提交任務時默認讀取的配置文件

spark@master:~/spark/spark/conf$?vim?spark-defaults.conf

添加以下內容

spark.master?????????????????????spark://master:7077 spark.eventLog.enabled???????????true spark.eventLog.dir???????????????hdfs://master:9000/sparkHistoryLogs spark.eventLog.compress??????????true spark.history.updateInterval?????5 spark.history.ui.port????????????7777 spark.history.fs.logDirectory????hdfs://master:9000/sparkHistoryLogs


將配置好的spark文件復制到worker1、worker2節點中

切換到worker1節點中,執行

spark@worker1:~/spark$?scp?-r?spark@master:/home/spark/spark/spark?./spark

注意復制的目錄是放在spark目錄下


切換到worker2節點中,執行

spark@worker2:~/spark$?scp?-r?spark@master:/home/spark/spark/spark?./spark

注意復制的目錄是放在spark目錄下


切換到master中


接著啟動spark

spark@master:~/spark/spark/conf$?$SPARK_HOME/sbin/start-all.shstarting?org.apache.spark.deploy.master.Master,?logging?to?/home/spark/spark/spark/logs/spark-spark-org.apache.spark.deploy.master.Master-1-master.out master:?starting?org.apache.spark.deploy.worker.Worker,?logging?to?/home/spark/spark/spark/logs/spark-spark-org.apache.spark.deploy.worker.Worker-1-master.out worker2:?starting?org.apache.spark.deploy.worker.Worker,?logging?to?/home/spark/spark/spark/logs/spark-spark-org.apache.spark.deploy.worker.Worker-1-worker2.out worker1:?starting?org.apache.spark.deploy.worker.Worker,?logging?to?/home/spark/spark/spark/logs/spark-spark-org.apache.spark.deploy.worker.Worker-1-worker1.out

可以看到啟動成功,

停止spark 使用

$SPARK_HOME/sbin/stop-all.sh


啟動Spark歷史任務記錄:

spark@master:~/spark/spark/conf$?$SPARK_HOME/sbin/start-history-server.shstarting?org.apache.spark.deploy.history.HistoryServer,?logging?to?/home/spark/spark/spark/logs/spark-spark-org.apache.spark.deploy.history.HistoryServer-1-master.out

查看SparkHadoop相關的所有進程

spark@master:~/spark/spark/conf$?jps?-l6711?org.apache.hadoop.hdfs.server.namenode.NameNode 18863?org.apache.spark.deploy.master.Master 7053?org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode 18966?org.apache.spark.deploy.worker.Worker 19122?sun.tools.jps.Jps 19070?org.apache.spark.deploy.history.HistoryServer 15529?org.apache.hadoop.hdfs.server.datanode.DataNode 7352?org.apache.hadoop.yarn.server.nodemanager.NodeManager 7222?org.apache.hadoop.yarn.server.resourcemanager.ResourceManager

至此Spark集群也已經運行成功。


Spark-shell測試Spark集群:

需要先執行

hdfs?dfsadmin?-safemode?leave

將安全模式關閉

spark@master:~/spark/spark/conf$?$SPARK_HOME/bin/spark-shell?--master?spark://master:7077

可以看到啟動成功

測試一下:


一些web瀏覽界面:

集群節點信息:http://master:8080,可以將master換成ip地址


歷史任務:http://master:7777,因為沒有執行任務,所以看不到

Hadoop?集群信息:?http://master:50070/

圖中顯示安全模式已經關閉,接下來重新打開hadoop集群,則需要執行下列命令

spark@master:~/spark/spark/conf$?hdfs?dfsadmin?-safemode?enter

至此,已經全部安裝完畢。

若想停止spark集群則執行

spark@master:~/spark/spark/conf$?$SPARK_HOME/sbin/stop-all.shmaster:?stopping?org.apache.spark.deploy.worker.Worker worker2:?stopping?org.apache.spark.deploy.worker.Worker worker1:?stopping?org.apache.spark.deploy.worker.Worker stopping?org.apache.spark.deploy.master.Master

若想停止hadoop集群則執行

spark@master:~/spark/spark/conf$?$HADOOP_HOME/sbin/stop-all.shThis?script?is?Deprecated.?Instead?use?stop-dfs.sh?and?stop-yarn.sh Stopping?namenodes?on?[master] master:?no?namenode?to?stop master:?stopping?datanode worker1:?stopping?datanode worker2:?stopping?datanode Stopping?secondary?namenodes?[0.0.0.0] 0.0.0.0:?no?secondarynamenode?to?stop stopping?yarn?daemons no?resourcemanager?to?stop master:?no?nodemanager?to?stop worker1:?no?nodemanager?to?stop worker2:?no?nodemanager?to?stop no?proxyserver?to?stop



最后附一些常用檢測命令:來自(http://ciscolinux.blog.51cto.com/746827/1313110)

1.查看端口是否開啟

netstat -tupln | grep 9000

netstat -tupln | grep 9001

2.訪問masterNameNode)和slaveJobTracker)啟動是否正常http://192.168.0.202:5007050030

3.jps查看守護進程是否運行

master顯示:Job TrackerJpsSecondaryNameNod NameNode

slave顯示:DataNode JpsTaskTracker

4.查看集群狀態統計信息(hadoopdfsadmin -report

masterslave輸入信息:

九、常用命令

hadoop dfs -ls #列出HDFS下文件

hadoop dfs -ls in #列出HDFS下某個文檔中的文件

hadoop dfs -put test.txt test #上傳文件到指定目錄并且重新命名,只有所有的DataNode都接收完數據才算成功

hadoop dfs -get in getin #HDFS獲取文件并且重新命名為getin,同put一樣可操作文件也可操作目錄

hadoop dfs -rmr out #刪除HDFS上的out目錄

hadoop dfs -cat in/* #查看HDFSin目錄的內容

hadoop dfsadmin -safemode leave #退出安全模式

hadoop dfsadmin -safemode enter #進入安全模式


添加一個新的節點

請按照worker1和worker2的配置步驟,即可


轉載于:https://blog.51cto.com/lefteva/1874268

總結

以上是生活随笔為你收集整理的spark集群详细搭建过程及遇到的问题解决(四)的全部內容,希望文章能夠幫你解決所遇到的問題。

如果覺得生活随笔網站內容還不錯,歡迎將生活随笔推薦給好友。