大神帮忙看下哪里错了,找了一个小时实在找不出来了

cherrx 2017-10-31 01:31:43
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

17/10/31 05:56:11 INFO namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG: host = hadoop01/192.168.188.101
STARTUP_MSG: args = [-format]
STARTUP_MSG: version = 2.7.1
STARTUP_MSG:
STARTUP_MSG: build = https://git-wip-us.apache.org/repos/asf/hadoop.git -r 15ecc87ccf4a0228f35af08fc56de536e6ce657a; compiled by 'jenkins' on 2015-06-29T06:04Z
STARTUP_MSG: java = 1.8.0_65
************************************************************/
17/10/31 05:56:11 INFO namenode.NameNode: registered UNIX signal handlers for [TERM, HUP, INT]
17/10/31 05:56:11 INFO namenode.NameNode: createNameNode [-format]
17/10/31 05:56:11 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Formatting using clusterid: CID-a403c531-f69d-4bd2-9ee2-b82b8e6e863a
17/10/31 05:56:11 INFO namenode.FSNamesystem: No KeyProvider found.
17/10/31 05:56:11 INFO namenode.FSNamesystem: fsLock is fair:true
17/10/31 05:56:11 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit=1000
17/10/31 05:56:11 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true
17/10/31 05:56:11 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
17/10/31 05:56:11 INFO blockmanagement.BlockManager: The block deletion will start around 2017 Oct 31 05:56:11
17/10/31 05:56:11 INFO util.GSet: Computing capacity for map BlocksMap
17/10/31 05:56:11 INFO util.GSet: VM type = 64-bit
17/10/31 05:56:11 INFO util.GSet: 2.0% max memory 966.7 MB = 19.3 MB
17/10/31 05:56:11 INFO util.GSet: capacity = 2^21 = 2097152 entries
17/10/31 05:56:11 INFO blockmanagement.BlockManager: dfs.block.access.token.enable=false
17/10/31 05:56:11 INFO blockmanagement.BlockManager: defaultReplication = 1
17/10/31 05:56:11 INFO blockmanagement.BlockManager: maxReplication = 512
17/10/31 05:56:11 INFO blockmanagement.BlockManager: minReplication = 1
17/10/31 05:56:11 INFO blockmanagement.BlockManager: maxReplicationStreams = 2
17/10/31 05:56:11 INFO blockmanagement.BlockManager: shouldCheckForEnoughRacks = false
17/10/31 05:56:11 INFO blockmanagement.BlockManager: replicationRecheckInterval = 3000
17/10/31 05:56:11 INFO blockmanagement.BlockManager: encryptDataTransfer = false
17/10/31 05:56:11 INFO blockmanagement.BlockManager: maxNumBlocksToLog = 1000
17/10/31 05:56:11 INFO namenode.FSNamesystem: fsOwner = root (auth:SIMPLE)
17/10/31 05:56:11 INFO namenode.FSNamesystem: supergroup = supergroup
17/10/31 05:56:11 INFO namenode.FSNamesystem: isPermissionEnabled = true
17/10/31 05:56:11 INFO namenode.FSNamesystem: HA Enabled: false
17/10/31 05:56:11 INFO namenode.FSNamesystem: Append Enabled: true
17/10/31 05:56:11 INFO util.GSet: Computing capacity for map INodeMap
17/10/31 05:56:11 INFO util.GSet: VM type = 64-bit
17/10/31 05:56:11 INFO util.GSet: 1.0% max memory 966.7 MB = 9.7 MB
17/10/31 05:56:11 INFO util.GSet: capacity = 2^20 = 1048576 entries
17/10/31 05:56:11 INFO namenode.FSDirectory: ACLs enabled? false
17/10/31 05:56:11 INFO namenode.FSDirectory: XAttrs enabled? true
17/10/31 05:56:11 INFO namenode.FSDirectory: Maximum size of an xattr: 16384
17/10/31 05:56:11 INFO namenode.NameNode: Caching file names occuring more than 10 times
17/10/31 05:56:11 INFO util.GSet: Computing capacity for map cachedBlocks
17/10/31 05:56:11 INFO util.GSet: VM type = 64-bit
17/10/31 05:56:11 INFO util.GSet: 0.25% max memory 966.7 MB = 2.4 MB
17/10/31 05:56:11 INFO util.GSet: capacity = 2^18 = 262144 entries
17/10/31 05:56:11 INFO namenode.FSNamesystem: dfs.namenode.safemode.threshold-pct = 0.9990000128746033
17/10/31 05:56:11 INFO namenode.FSNamesystem: dfs.namenode.safemode.min.datanodes = 0
17/10/31 05:56:11 INFO namenode.FSNamesystem: dfs.namenode.safemode.extension = 30000
17/10/31 05:56:11 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10
17/10/31 05:56:11 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10
17/10/31 05:56:11 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25
17/10/31 05:56:11 INFO namenode.FSNamesystem: Retry cache on namenode is enabled
17/10/31 05:56:11 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
17/10/31 05:56:11 INFO util.GSet: Computing capacity for map NameNodeRetryCache
17/10/31 05:56:11 INFO util.GSet: VM type = 64-bit
17/10/31 05:56:11 INFO util.GSet: 0.029999999329447746% max memory 966.7 MB = 297.0 KB
17/10/31 05:56:11 INFO util.GSet: capacity = 2^15 = 32768 entries
17/10/31 05:56:11 INFO namenode.FSImage: Allocated new BlockPoolId: BP-798510957-192.168.188.101-1509454571883
17/10/31 05:56:11 INFO common.Storage: Storage directory /root/work/hadoop-2.7.1/tmp/dfs/name has been successfully formatted.
17/10/31 05:56:12 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0
17/10/31 05:56:12 INFO util.ExitUtil: Exiting with status 0
17/10/31 05:56:12 INFO namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at hadoop01/192.168.188.101
************************************************************/
...全文
386 4 打赏 收藏 转发到动态 举报
写回复
用AI写文章
4 条回复
切换为时间正序
请发表友善的回复…
发表回复
shiter 2017-11-01
  • 打赏
  • 举报
回复
以我的经验,很有可能是磁盘空间满了?或者是什么造成nodemanager停的原因?
luoyoumou 2017-10-31
  • 打赏
  • 举报
回复
记住:不要以为收到 SHUTDOWN_MSG 就以为是遇到错误了。
luoyoumou 2017-10-31
  • 打赏
  • 举报
回复
不都是INFO么? 格式化后,就可以运行“启动”脚本了撒!

20,807

社区成员

发帖
与我相关
我的任务
社区描述
Hadoop生态大数据交流社区,致力于有Hadoop,hive,Spark,Hbase,Flink,ClickHouse,Kafka,数据仓库,大数据集群运维技术分享和交流等。致力于收集优质的博客
社区管理员
  • 分布式计算/Hadoop社区
  • 涤生大数据
加入社区
  • 近7日
  • 近30日
  • 至今
社区公告
暂无公告

试试用AI创作助手写篇文章吧