Hadoop错误:java.io.IOException: Incompatible clusterIDs

问题:

配置Hadoop集群时,一个节点的DataNode无法启动

排查:

查看hadoop-root-datanode-bigdata114.log文件,错误信息如下:

java.io.IOException: Incompatible clusterIDs in /root/training/hadoop-2.7.3/tmp/dfs/data:

namenode clusterID = CID-947a48a2-56aa-4566-85d6-b5987d0bfeca; datanode clusterID = CID-8c75bd62-98f6-4482-85f3-865cb9a0f70f

解决:

打开 /root/training/hadoop-2.7.3/tmp/dfs/data/current 下的VERSION文件(/tmp 是在 core-site.xml 中设置的hadoop临时目录):

#Thu Dec 13 00:00:49 CST 2018
storageID=DS-e242ac88-5b0f-402c-b138-7e29f583ec82
clusterID=CID-8c75bd62-98f6-4482-85f3-865cb9a0f70f
cTime=0
datanodeUuid=2ddcdd2a-9f76-4e3e-8cc3-9b9d44448936
storageType=DATA_NODE
layoutVersion=-56

将clusterID修改为CID-947a48a2-56aa-4566-85d6-b5987d0bfeca 即可

原文地址:https://www.cnblogs.com/cxc1357/p/10134817.html

时间: 2024-10-05 05:50:28

Hadoop错误:java.io.IOException: Incompatible clusterIDs的相关文章

重新格式化namenode后,出现java.io.IOException Incompatible clusterIDs

错误: java.io.IOException: Incompatible clusterIDs in /data/dfs/data: namenode clusterID = CID-d1448b9e-da0f-499e-b1d4-78cb18ecdebb; datanode clusterID = CID-ff0faa40-2940-4838-b321-98272eb0dee3 原因: 每次namenode format会重新创建一个namenodeId,而data目录包含了上次format

hadoop格式化:java.io.IOException: Incompatible clusterIDs in /home/lxh/hadoop/hdfs/data: namenode clusterID

1 概述  解决hadoop启动hdfs时,datanode无法启动的问题.错误为: java.io.IOException: Incompatible clusterIDs in /home/lxh/hadoop/hdfs/data: namenode clusterID = CID-a3938a0b-57b5-458d-841c-d096e2b7a71c; datanode clusterID = CID-200e6206-98b5-44b2-9e48-262871884eeb 2 问题描述

hadoop错误java.io.IOException Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try

错误: java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try 原因: 无法写入:我的环境中有3个datanode,备份数量设置的是3.在写操作时,它会在pipeline中写3个机器.默认replace-datanode-on-failure.policy是DEFAULT,如果系统中的dat

java.io.IOException: Incompatible namespaceIDs

问题描述:   在实验的时候,需要往以前的集群中添加一台datanode,在添加之前,由于在调式namenode的时候,格式化了dfs,这就导致了namenode上的namespaceID和以前集群上datanode的namespaceID不一致,结果就会导致在启动hadoop集群的时候,旧的的datanode就没有办法启动hdfs服务,但是旧的dadanode可以启动tasktracker. 2015-01-28 11:01:27,561 ERROR org.apache.hadoop.hdf

keytool 错误 java.io.IOException: incorrect AVA format

给一个APK做签名,选择新建一个key并填写相关信息,但在Finish时,keytool报出了一个错误:keytool error: java.io.IOException: Incorrect AVA format这是因为在创建key的时候,organization和organizational unit等信息都不能包含 ”,”删除","可以正常签名.

hadoop异常 java.io.IOException: Job status not available

[[email protected] conf]$ hadoop jar /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar wordcount /user/lizeyi/people.txt  /user/lizeyi/wordcount7 15/06/08 18:36:16 INFO client.RMProxy: Connecting to ResourceManager at master.hadoop/10.3.4.35:80

hadoop错误FATAL org.apache.hadoop.hdfs.server.namenode.NameNode Exception in namenode join java.io.IOException There appears to be a gap in the edit log

错误: FATAL org.apache.hadoop.hdfs.server.namenode.NameNode Exception in namenode join java.io.IOException There appears to be a gap in the edit log 原因: namenode元数据被破坏,需要修复 解决:     恢复一下namenode hadoop namenode –recover 一路选择c,一般就OK了 如果,您认为阅读这篇博客让您有些收获,不

hive执行query语句时提示错误:org.apache.hadoop.ipc.RemoteException: java.io.IOException: java.io.IOException:

hive> select product_id, track_time from trackinfo limit 5; Total MapReduce jobs = 1 Launching Job 1 out of 1 Number of reduce tasks is set to 0 since there's no reduce operator org.apache.hadoop.ipc.RemoteException: java.io.IOException: java.io.IOEx

hive运行query语句时提示错误:org.apache.hadoop.ipc.RemoteException: java.io.IOException: java.io.IOException:

hive> select product_id, track_time from trackinfo limit 5; Total MapReduce jobs = 1 Launching Job 1 out of 1 Number of reduce tasks is set to 0 since there's no reduce operator org.apache.hadoop.ipc.RemoteException: java.io.IOException: java.io.IOEx