分享

执行MR,slaves2节点日志报错RECEIVED SIGNAL 15: SIGTERM

跪求大神指教,度娘了好久没有解决!!
刚搭建的三个节点的cdh5.4.1集群,启动集群无异常,节点nodemanage都可以正常启动,
但是执行wordcount的时候(有的时候wordcount可以执行,未报错),map执行完,slaves2节点直接挂掉了,查看slaves2日志,RECEIVED SIGNAL 15: SIGTERM
详细日志如下:
2017-09-22 15:46:03,656 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: registered UNIX signal handlers for [TERM, HUP, INT]
2017-09-22 15:46:04,075 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties
2017-09-22 15:46:04,154 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s).
2017-09-22 15:46:04,154 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system started
2017-09-22 15:46:04,160 INFO org.apache.hadoop.hdfs.server.datanode.BlockScanner: Disabled block scanner.
2017-09-22 15:46:04,163 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Configured hostname is slaves2
2017-09-22 15:46:04,172 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting DataNode with maxLockedMemory = 0
2017-09-22 15:46:04,197 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened streaming server at /0.0.0.0:50010
2017-09-22 15:46:04,198 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is 1048576 bytes/s
2017-09-22 15:46:04,198 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Number threads for balancing is 5
2017-09-22 15:46:04,309 INFO org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
2017-09-22 15:46:04,316 INFO org.apache.hadoop.http.HttpRequestLog: Http request log for http.requests.datanode is not defined
2017-09-22 15:46:04,328 INFO org.apache.hadoop.http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
2017-09-22 15:46:04,330 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context datanode
2017-09-22 15:46:04,330 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs
2017-09-22 15:46:04,330 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static
2017-09-22 15:46:04,345 INFO org.apache.hadoop.http.HttpServer2: addJerseyResourcePackage: packageName=org.apache.hadoop.hdfs.server.datanode.web.resources;org.apache.hadoop.hdfs.web.resources, pathSpec=/webhdfs/v1/*
2017-09-22 15:46:04,348 INFO org.apache.hadoop.http.HttpServer2: Jetty bound to port 50075
2017-09-22 15:46:04,348 INFO org.mortbay.log: jetty-6.1.26.cloudera.4
2017-09-22 15:46:04,525 INFO org.mortbay.log: Started HttpServer2$SelectChannelConnectorWithSafeStartup@0.0.0.0:50075
2017-09-22 15:46:04,694 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: dnUserName = administrator
2017-09-22 15:46:04,694 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: supergroup = supergroup
2017-09-22 15:46:04,732 INFO org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue
2017-09-22 15:46:04,753 INFO org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 50020
2017-09-22 15:46:04,784 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened IPC server at /0.0.0.0:50020
2017-09-22 15:46:04,795 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Refresh request received for nameservices: null
2017-09-22 15:46:04,819 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting BPOfferServices for nameservices: <default>
2017-09-22 15:46:04,830 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool <registering> (Datanode Uuid unassigned) service to master/172.26.11.103:9000 starting to offer service
2017-09-22 15:46:04,862 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting
2017-09-22 15:46:04,863 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 50020: starting
2017-09-22 15:46:05,957 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: master/172.26.11.103:9000. Already tried 0 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2017-09-22 15:46:06,339 INFO org.apache.hadoop.hdfs.server.common.Storage: Lock on /home/administrator/hadoop-2.6.0-cdh5.4.1/dfs/data/in_use.lock acquired by nodename 16887@slaves2
2017-09-22 15:46:06,368 INFO org.apache.hadoop.hdfs.server.common.Storage: Analyzing storage directories for bpid BP-157856151-172.26.11.103-1504264452458
2017-09-22 15:46:06,368 INFO org.apache.hadoop.hdfs.server.common.Storage: Locking is disabled for /home/administrator/hadoop-2.6.0-cdh5.4.1/dfs/data/current/BP-157856151-172.26.11.103-1504264452458
2017-09-22 15:46:06,369 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Setting up storage: nsid=1475826562;bpid=BP-157856151-172.26.11.103-1504264452458;lv=-56;nsInfo=lv=-60;cid=CID-a08ab648-1a3a-4c04-8c0d-8bfdff704881;nsid=1475826562;c=0;bpid=BP-157856151-172.26.11.103-1504264452458;dnuuid=9fc577ff-7966-46ed-9030-4e8369b0fd4f
2017-09-22 15:46:06,398 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Added new volume: DS-a4f5f0ab-8f0c-472e-917d-9cbba6739c65
2017-09-22 15:46:06,398 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Added volume - /home/administrator/hadoop-2.6.0-cdh5.4.1/dfs/data/current, StorageType: DISK
2017-09-22 15:46:06,419 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Registered FSDatasetState MBean
2017-09-22 15:46:06,420 INFO org.apache.hadoop.hdfs.server.datanode.DirectoryScanner: Periodic Directory Tree Verification scan starting at 1506066527420 with interval 600000
2017-09-22 15:46:06,421 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding block pool BP-157856151-172.26.11.103-1504264452458
2017-09-22 15:46:06,421 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Scanning block pool BP-157856151-172.26.11.103-1504264452458 on volume /home/administrator/hadoop-2.6.0-cdh5.4.1/dfs/data/current...
2017-09-22 15:46:06,436 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Cached dfsUsed found for /home/administrator/hadoop-2.6.0-cdh5.4.1/dfs/data/current/BP-157856151-172.26.11.103-1504264452458/current: 4114286
2017-09-22 15:46:06,438 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Time taken to scan block pool BP-157856151-172.26.11.103-1504264452458 on /home/administrator/hadoop-2.6.0-cdh5.4.1/dfs/data/current: 16ms
2017-09-22 15:46:06,438 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Total time to scan all replicas for block pool BP-157856151-172.26.11.103-1504264452458: 17ms
2017-09-22 15:46:06,439 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding replicas to map for block pool BP-157856151-172.26.11.103-1504264452458 on volume /home/administrator/hadoop-2.6.0-cdh5.4.1/dfs/data/current...
2017-09-22 15:46:06,448 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Time to add replicas to map for block pool BP-157856151-172.26.11.103-1504264452458 on volume /home/administrator/hadoop-2.6.0-cdh5.4.1/dfs/data/current: 8ms
2017-09-22 15:46:06,448 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Total time to add all replicas to map: 9ms
2017-09-22 15:46:06,450 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool BP-157856151-172.26.11.103-1504264452458 (Datanode Uuid null) service to master/172.26.11.103:9000 beginning handshake with NN
2017-09-22 15:46:06,520 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool Block pool BP-157856151-172.26.11.103-1504264452458 (Datanode Uuid null) service to master/172.26.11.103:9000 successfully registered with NN
2017-09-22 15:46:06,520 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: For namenode master/172.26.11.103:9000 using DELETEREPORT_INTERVAL of 300000 msec  BLOCKREPORT_INTERVAL of 600000msec CACHEREPORT_INTERVAL of 10000msec Initial delay: 0msec; heartBeatInterval=3000
2017-09-22 15:46:06,634 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Namenode Block pool BP-157856151-172.26.11.103-1504264452458 (Datanode Uuid 9fc577ff-7966-46ed-9030-4e8369b0fd4f) service to master/172.26.11.103:9000 trying to claim ACTIVE state with txid=47384
2017-09-22 15:46:06,634 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Acknowledging ACTIVE Namenode Block pool BP-157856151-172.26.11.103-1504264452458 (Datanode Uuid 9fc577ff-7966-46ed-9030-4e8369b0fd4f) service to master/172.26.11.103:9000
2017-09-22 15:46:06,730 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Successfully sent block report 0xbeb6a6509fe0d,  containing 1 storage report(s), of which we sent 1. The reports had 85 total blocks and used 1 RPC(s). This took 3 msec to generate and 93 msecs for RPC and NN processing. Got back one command: FinalizeCommand/5.
2017-09-22 15:46:06,730 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Got finalize command for block pool BP-157856151-172.26.11.103-1504264452458
2017-09-22 15:47:48,533 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Successfully sent block report 0xbeb821e33d701,  containing 1 storage report(s), of which we sent 1. The reports had 85 total blocks and used 1 RPC(s). This took 0 msec to generate and 5 msecs for RPC and NN processing. Got back one command: FinalizeCommand/5.
2017-09-22 15:47:48,533 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Got finalize command for block pool BP-157856151-172.26.11.103-1504264452458
2017-09-22 15:48:47,450 INFO org.apache.hadoop.hdfs.server.datanode.DirectoryScanner: BlockPool BP-157856151-172.26.11.103-1504264452458 Total blocks: 85, missing metadata files:0, missing block files:0, missing blocks in memory:0, mismatched blocks:0
2017-09-22 15:49:02,750 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-157856151-172.26.11.103-1504264452458:blk_1073742324_1500 src: /172.26.11.103:47304 dest: /172.26.11.104:50010
2017-09-22 15:49:02,945 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /172.26.11.103:47304, dest: /172.26.11.104:50010, bytes: 276076, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_1791086058_1, offset: 0, srvID: 9fc577ff-7966-46ed-9030-4e8369b0fd4f, blockid: BP-157856151-172.26.11.103-1504264452458:blk_1073742324_1500, duration: 65343859
2017-09-22 15:49:02,946 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-157856151-172.26.11.103-1504264452458:blk_1073742324_1500, type=HAS_DOWNSTREAM_IN_PIPELINE terminating
2017-09-22 15:49:08,023 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-157856151-172.26.11.103-1504264452458:blk_1073742325_1501 src: /172.26.11.105:54290 dest: /172.26.11.104:50010
2017-09-22 15:49:08,071 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /172.26.11.105:54290, dest: /172.26.11.104:50010, bytes: 276076, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_-655849944_1, offset: 0, srvID: 9fc577ff-7966-46ed-9030-4e8369b0fd4f, blockid: BP-157856151-172.26.11.103-1504264452458:blk_1073742325_1501, duration: 45526598
2017-09-22 15:49:08,071 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-157856151-172.26.11.103-1504264452458:blk_1073742325_1501, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2017-09-22 15:49:08,158 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-157856151-172.26.11.103-1504264452458:blk_1073742326_1502 src: /172.26.11.103:47312 dest: /172.26.11.104:50010
2017-09-22 15:49:08,172 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /172.26.11.103:47312, dest: /172.26.11.104:50010, bytes: 107, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_-655849944_1, offset: 0, srvID: 9fc577ff-7966-46ed-9030-4e8369b0fd4f, blockid: BP-157856151-172.26.11.103-1504264452458:blk_1073742326_1502, duration: 8951625
2017-09-22 15:49:08,172 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-157856151-172.26.11.103-1504264452458:blk_1073742326_1502, type=HAS_DOWNSTREAM_IN_PIPELINE terminating
2017-09-22 15:49:08,191 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-157856151-172.26.11.103-1504264452458:blk_1073742327_1503 src: /172.26.11.105:54292 dest: /172.26.11.104:50010
2017-09-22 15:49:08,198 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /172.26.11.105:54292, dest: /172.26.11.104:50010, bytes: 30, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_-655849944_1, offset: 0, srvID: 9fc577ff-7966-46ed-9030-4e8369b0fd4f, blockid: BP-157856151-172.26.11.103-1504264452458:blk_1073742327_1503, duration: 5002928
2017-09-22 15:49:08,198 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-157856151-172.26.11.103-1504264452458:blk_1073742327_1503, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2017-09-22 15:49:08,272 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-157856151-172.26.11.103-1504264452458:blk_1073742328_1504 src: /172.26.11.105:54294 dest: /172.26.11.104:50010
2017-09-22 15:49:08,282 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /172.26.11.105:54294, dest: /172.26.11.104:50010, bytes: 89862, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_-655849944_1, offset: 0, srvID: 9fc577ff-7966-46ed-9030-4e8369b0fd4f, blockid: BP-157856151-172.26.11.103-1504264452458:blk_1073742328_1504, duration: 7674763
2017-09-22 15:49:08,282 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-157856151-172.26.11.103-1504264452458:blk_1073742328_1504, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2017-09-22 15:49:14,194 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-157856151-172.26.11.103-1504264452458:blk_1073742329_1505 src: /172.26.11.105:54310 dest: /172.26.11.104:50010
2017-09-22 15:49:14,216 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /172.26.11.105:54310, dest: /172.26.11.104:50010, bytes: 106912, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_1694406125_1, offset: 0, srvID: 9fc577ff-7966-46ed-9030-4e8369b0fd4f, blockid: BP-157856151-172.26.11.103-1504264452458:blk_1073742329_1505, duration: 19939134
2017-09-22 15:49:14,216 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-157856151-172.26.11.103-1504264452458:blk_1073742329_1505, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2017-09-22 15:49:14,258 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-157856151-172.26.11.103-1504264452458:blk_1073742330_1506 src: /172.26.11.105:54314 dest: /172.26.11.104:50010
2017-09-22 15:49:14,263 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /172.26.11.105:54314, dest: /172.26.11.104:50010, bytes: 297, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_1694406125_1, offset: 0, srvID: 9fc577ff-7966-46ed-9030-4e8369b0fd4f, blockid: BP-157856151-172.26.11.103-1504264452458:blk_1073742330_1506, duration: 3324221
2017-09-22 15:49:14,263 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP-157856151-172.26.11.103-1504264452458:blk_1073742330_1506, type=LAST_IN_PIPELINE, downstreams=0:[] terminating
2017-09-22 15:49:19,318 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: RECEIVED SIGNAL 15: SIGTERM
2017-09-22 15:49:19,321 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at slaves2/172.26.11.104
************************************************************/

已有(2)人评论

跳转到指定楼层
qcbb001 发表于 2017-9-22 17:50:27
可能是多次format namenode节点导致  namdenode 的 namespaceID 与  datanode的namespaceID 不一致造成的。

有两种方法可以解决,第一种方法是删除DataNode的所有资料(及将集群中每个datanode的/hdfs/data/current中的VERSION删掉,然后执行hadoop namenode -format重启集群,错误消失。<推荐>);第二种方法是修改每个DataNode的namespaceID(位于/hdfs/data/current/VERSION文件中)<优先>或修改NameNode的namespaceID(位于/hdfs/name/current/VERSION文件中),使其一致。

来自:
hadoop datanode启动不起来:解决办法二
http://www.aboutyun.com/forum.php?mod=viewthread&tid=7931



回复

使用道具 举报

云云众生奈我何 发表于 2017-9-22 21:04:21
@qcbb001 按照第一个步骤操作了,结果还是一样!
这个解决方法并不能解决,,,
因为我启动集群,namenode是可以启动的,只是在MR运行时,datanode节点报错
ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: RECEIVED SIGNAL 15: SIGTERM。
不过还是谢谢你
回复

使用道具 举报

您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

关闭

推荐上一条 /2 下一条