hadoop伪分布配置执行名称节点格式化错误,求大神看看怎么解决。

qq_41138218 2018-03-20 10:30:03
18/03/20 07:14:56 INFO namenode.NameNode: registered UNIX signal handlers for [TERM, HUP, INT]
18/03/20 07:14:56 INFO namenode.NameNode: createNameNode [-format]
Formatting using clusterid: CID-d9ce35a4-978b-4f1f-b484-a3c2871bd058
18/03/20 07:14:58 INFO namenode.FSNamesystem: No KeyProvider found.
18/03/20 07:14:58 INFO namenode.FSNamesystem: fsLock is fair:true
18/03/20 07:14:59 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit=1000
18/03/20 07:14:59 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true
18/03/20 07:14:59 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
18/03/20 07:14:59 INFO blockmanagement.BlockManager: The block deletion will start around 2018 Mar 20 07:14:59
18/03/20 07:14:59 INFO util.GSet: Computing capacity for map BlocksMap
18/03/20 07:14:59 INFO util.GSet: VM type = 64-bit
18/03/20 07:14:59 INFO util.GSet: 2.0% max memory 889 MB = 17.8 MB
18/03/20 07:14:59 INFO util.GSet: capacity = 2^21 = 2097152 entries
18/03/20 07:14:59 INFO blockmanagement.BlockManager: dfs.block.access.token.enable=false
18/03/20 07:14:59 INFO blockmanagement.BlockManager: defaultReplication = 1
18/03/20 07:14:59 INFO blockmanagement.BlockManager: maxReplication = 512
18/03/20 07:14:59 INFO blockmanagement.BlockManager: minReplication = 1
18/03/20 07:14:59 INFO blockmanagement.BlockManager: maxReplicationStreams = 2
18/03/20 07:14:59 INFO blockmanagement.BlockManager: shouldCheckForEnoughRacks = false
18/03/20 07:14:59 INFO blockmanagement.BlockManager: replicationRecheckInterval = 3000
18/03/20 07:14:59 INFO blockmanagement.BlockManager: encryptDataTransfer = false
18/03/20 07:14:59 INFO blockmanagement.BlockManager: maxNumBlocksToLog = 1000
18/03/20 07:14:59 INFO namenode.FSNamesystem: fsOwner = amour (auth:SIMPLE)
18/03/20 07:14:59 INFO namenode.FSNamesystem: supergroup = supergroup
18/03/20 07:14:59 INFO namenode.FSNamesystem: isPermissionEnabled = true
18/03/20 07:14:59 INFO namenode.FSNamesystem: HA Enabled: false
18/03/20 07:14:59 INFO namenode.FSNamesystem: Append Enabled: true
18/03/20 07:14:59 INFO util.GSet: Computing capacity for map INodeMap
18/03/20 07:14:59 INFO util.GSet: VM type = 64-bit
18/03/20 07:14:59 INFO util.GSet: 1.0% max memory 889 MB = 8.9 MB
18/03/20 07:14:59 INFO util.GSet: capacity = 2^20 = 1048576 entries
18/03/20 07:14:59 INFO namenode.FSDirectory: ACLs enabled? false
18/03/20 07:14:59 INFO namenode.FSDirectory: XAttrs enabled? true
18/03/20 07:14:59 INFO namenode.FSDirectory: Maximum size of an xattr: 16384
18/03/20 07:14:59 INFO namenode.NameNode: Caching file names occuring more than 10 times
18/03/20 07:14:59 INFO util.GSet: Computing capacity for map cachedBlocks
18/03/20 07:14:59 INFO util.GSet: VM type = 64-bit
18/03/20 07:14:59 INFO util.GSet: 0.25% max memory 889 MB = 2.2 MB
18/03/20 07:14:59 INFO util.GSet: capacity = 2^18 = 262144 entries
18/03/20 07:14:59 INFO namenode.FSNamesystem: dfs.namenode.safemode.threshold-pct = 0.9990000128746033
18/03/20 07:14:59 INFO namenode.FSNamesystem: dfs.namenode.safemode.min.datanodes = 0
18/03/20 07:14:59 INFO namenode.FSNamesystem: dfs.namenode.safemode.extension = 30000
18/03/20 07:15:00 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10
18/03/20 07:15:00 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10
18/03/20 07:15:00 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25
18/03/20 07:15:00 INFO namenode.FSNamesystem: Retry cache on namenode is enabled
18/03/20 07:15:00 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
18/03/20 07:15:00 INFO util.GSet: Computing capacity for map NameNodeRetryCache
18/03/20 07:15:00 INFO util.GSet: VM type = 64-bit
18/03/20 07:15:00 INFO util.GSet: 0.029999999329447746% max memory 889 MB = 273.1 KB
18/03/20 07:15:00 INFO util.GSet: capacity = 2^15 = 32768 entries
18/03/20 07:15:00 INFO namenode.FSImage: Allocated new BlockPoolId: BP-1509097366-127.0.1.1-1521555300129
18/03/20 07:15:00 WARN namenode.NameNode: Encountered exception during format:
java.io.IOException: Cannot create directory /usr/local/hadoop/tmp/dfs/name/current
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:337)
at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:548)
at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:569)
at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:161)
at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:991)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1429)
at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1554)
18/03/20 07:15:00 ERROR namenode.NameNode: Failed to start namenode.
java.io.IOException: Cannot create directory /usr/local/hadoop/tmp/dfs/name/current
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:337)
at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:548)
at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:569)
at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:161)
at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:991)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1429)
at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1554)
18/03/20 07:15:00 INFO util.ExitUtil: Exiting with status 1
18/03/20 07:15:00 INFO namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1
************************************************************/
...全文
736 3 打赏 收藏 转发到动态 举报
写回复
用AI写文章
3 条回复
切换为时间正序
请发表友善的回复…
发表回复
qq_41138218 2018-03-29
  • 打赏
  • 举报
回复
引用 1 楼 zycxnanwang的回复:
不清楚你的hadoop配置 /usr/local/hadoop/tmp/dfs/name/current 上面那个文件路径, 你是配置的是那个文件的目录 你首先要建立这个目录
我解决了,我一怒之下直接把hadoop chmod 777,就解决了
zycxnanwang 2018-03-22
  • 打赏
  • 举报
回复
不清楚你的hadoop配置 /usr/local/hadoop/tmp/dfs/name/current 上面那个文件路径, 你是配置的是那个文件的目录 你首先要建立这个目录

20,808

社区成员

发帖
与我相关
我的任务
社区描述
Hadoop生态大数据交流社区,致力于有Hadoop,hive,Spark,Hbase,Flink,ClickHouse,Kafka,数据仓库,大数据集群运维技术分享和交流等。致力于收集优质的博客
社区管理员
  • 分布式计算/Hadoop社区
  • 涤生大数据
加入社区
  • 近7日
  • 近30日
  • 至今
社区公告
暂无公告

试试用AI创作助手写篇文章吧