清理ambari安装的集群

ambari

2016-11-04

1128

0

本文针对redhat或者centos

对于测试集群,如果通过ambari安装的集群后,想重新再来一次的话,需要清理集群。

对于安装了很多hadoop组件的话,这个工作很繁琐。接下来是我整理的清理过程。

1、通过ambari将集群中的所用组件都关闭,如果关闭不了,直接kill -9 XXX

2、关闭ambari-server,ambari-agent

ambari-server stop  
ambari-agent stop 

3、卸载安装的软件

yum remove hadoop_2* hdp-select* ranger_2* zookeeper_* bigtop* atlas-metadata* ambari* postgresql spark*  slider* storm* snappy*

以上命令可能不全,执行完一下命令后,再执行

yum list | grep @HDP  

查看是否还有没有卸载的,如果有,继续通过#yum remove XXX卸载

4、删除postgresql的数据

postgresql软件卸载后,其数据还保留在硬盘中,需要把这部分数据删除掉,如果不删除掉,重新安装ambari-server后,有可能还应用以前的安装数据,而这些数据时错误数据,所以需要删除掉。

rm -rf /var/lib/pgsql  

5、删除用户

ambari安装hadoop集群会创建一些用户,清除集群时有必要清除这些用户,并删除对应的文件夹。这样做可以避免集群运行时出现的文件访问权限错误的问题。  

  

userdel oozie  
userdel hive  
userdel ambari-qa  
userdel flume    
userdel hdfs    
userdel knox    
userdel storm    
userdel mapred  
userdel hbase    
userdel tez    
userdel zookeeper  
userdel kafka    
userdel falcon  
userdel sqoop    
userdel yarn    
userdel hcat  
userdel atlas  
userdel spark  
userdel ams
rm -rf /home/atlas  
rm -rf /home/accumulo  
rm -rf /home/hbase  
rm -rf /home/hive  
rm -rf /home/oozie  
rm -rf /home/storm  
rm -rf /home/yarn  
rm -rf /home/ambari-qa  
rm -rf /home/falcon  
rm -rf /home/hcat  
rm -rf /home/kafka  
rm -rf /home/mahout  
rm -rf /home/spark  
rm -rf /home/tez  
rm -rf /home/zookeeper  
rm -rf /home/flume  
rm -rf /home/hdfs  
rm -rf /home/knox  
rm -rf /home/mapred  
rm -rf /home/sqoop  

6、删除ambari遗留数据

rm -rf /var/lib/ambari*  
rm -rf /usr/lib/python2.6/site-packages/ambari_*  
rm -rf /usr/lib/python2.6/site-packages/resource_management  
rm -rf /usr/lib/ambri-*  

7,删除其他hadoop组件遗留数据

rm -rf /etc/falcon 
rm -rf /etc/knox 
rm -rf /etc/hive-webhcat 
rm -rf /etc/kafka 
rm -rf /etc/slider 
rm -rf /etc/storm-slider-client 
rm -rf /etc/spark 
rm -rf /etc/hadoop
rm -rf /etc/hbase
rm -rf /etc/zookeeper
rm -rf /etc/storm
rm -rf /etc/flume
rm -rf /var/run/spark 
rm -rf /var/run/hadoop 
rm -rf /var/run/hbase 
rm -rf /var/run/zookeeper 
rm -rf /var/run/flume 
rm -rf /var/run/storm 
rm -rf /var/run/webhcat 
rm -rf /var/run/hadoop-yarn 
rm -rf /var/run/hadoop-mapreduce 
rm -rf /var/run/kafka 
rm -rf /var/log/hadoop 
rm -rf /var/log/hbase 
rm -rf /var/log/flume 
rm -rf /var/log/storm 
rm -rf /var/log/hadoop-yarn 
rm -rf /var/log/hadoop-mapreduce 
rm -rf /var/log/knox 
rm -rf /usr/lib/flume 
rm -rf /usr/lib/storm 
rm -rf /var/lib/hive 
rm -rf /var/lib/oozie 
rm -rf /var/lib/flume 
rm -rf /var/lib/hadoop-hdfs 
rm -rf /var/lib/knox 
rm -rf /var/log/hive 
rm -rf /var/log/oozie 
rm -rf /var/log/zookeeper 
rm -rf /var/log/falcon 
rm -rf /var/log/webhcat 
rm -rf /var/log/spark
rm -rf /var/log/kafka
rm -rf /var/tmp/oozie 
rm -rf /tmp/ambari-qa 
rm -rf /var/hadoop 
rm -rf /hadoop/falcon 
rm -rf /tmp/hadoop 
rm -rf /tmp/hadoop-hdfs 
rm -rf /usr/hdp 
rm -rf /usr/hadoop 
rm -rf /opt/hadoop 
rm -rf /opt/hadoop2
rm -rf /tmp/hadoop 
rm -rf /var/hadoop 
rm -rf /hadoop 
rm -rf /etc/ambari-metrics-collector
rm -rf /etc/ambari-metrics-monitor
rm -rf /var/run/ambari-metrics-collector
rm -rf /var/run/ambari-metrics-monitor
rm -rf /var/log/ambari-metrics-collector
rm -rf /var/log/ambari-metrics-monitor
rm -rf /var/lib/hadoop-yarn
rm -rf /var/lib/hadoop-mapreduce

8、清理yum数据源

#yum clean all  

转载请注明出处: http://www.julyme.com/20161104/16.html

发表评论

全部评论:0条

Julyme

感觉还行吧。

Julyme的IT技术分享



/sitemap