Created
May 30, 2014 08:03
-
-
Save fivesmallq/930edfe512ed9b532a0a to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF-8 | |
14/05/30 16:02:47 WARN conf.Configuration: fs.default.name is deprecated. Instead, use fs.defaultFS | |
2014-05-30 16:02:47.807 java[27723:1003] Unable to load realm info from SCDynamicStore | |
2014-05-30 16:02:47.809 java[27723:1003] Unable to load realm info from SCDynamicStore | |
14/05/30 16:02:47 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable | |
14/05/30 16:02:47 WARN conf.Configuration: fs.default.name is deprecated. Instead, use fs.defaultFS | |
14/05/30 16:02:48 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. | |
14/05/30 16:02:48 WARN mapred.JobClient: No job jar file set. User classes may not be found. See JobConf(Class) or JobConf#setJar(String). | |
14/05/30 16:02:57 INFO mapred.FileInputFormat: Total input paths to process : 1 | |
14/05/30 16:02:57 INFO hdfs.DFSClient: Exception in createBlockOutputStream | |
java.net.ConnectException: Connection refused | |
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) | |
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:599) | |
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:207) | |
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:528) | |
at org.apache.hadoop.hdfs.DFSOutputStream.createSocketForPipeline(DFSOutputStream.java:1254) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:1080) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1040) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:488) | |
14/05/30 16:02:57 INFO hdfs.DFSClient: Abandoning BP-1628426775-127.0.0.1-1394440676844:blk_-7191553566950434627_1302 | |
14/05/30 16:02:57 INFO hdfs.DFSClient: Excluding datanode 127.0.0.1:50010 | |
14/05/30 16:02:57 WARN hdfs.DFSClient: DataStreamer Exception | |
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/cloudera/.staging/job_201405300019_0008/job.split could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. | |
at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1340) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2296) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:501) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:299) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44954) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1752) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1748) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1438) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1746) | |
at org.apache.hadoop.ipc.Client.call(Client.java:1238) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) | |
at java.lang.reflect.Method.invoke(Method.java:597) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:291) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1177) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1030) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:488) | |
14/05/30 16:02:57 INFO mapred.JobClient: Cleaning up the staging area hdfs://localhost.localdomain:8020/user/cloudera/.staging/job_201405300019_0008 | |
14/05/30 16:02:57 ERROR security.UserGroupInformation: PriviledgedActionException as:cloudera (auth:SIMPLE) cause:org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/cloudera/.staging/job_201405300019_0008/job.split could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. | |
at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1340) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2296) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:501) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:299) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44954) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1752) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1748) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1438) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1746) | |
14/05/30 16:02:57 ERROR security.UserGroupInformation: PriviledgedActionException as:cloudera (auth:SIMPLE) cause:org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/cloudera/.staging/job_201405300019_0008/job.split could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. | |
at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1340) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2296) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:501) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:299) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44954) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1752) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1748) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1438) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1746) | |
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/cloudera/.staging/job_201405300019_0008/job.split could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. | |
at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1340) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2296) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:501) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:299) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44954) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1752) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1748) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1438) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1746) | |
at org.apache.hadoop.ipc.Client.call(Client.java:1238) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) | |
at java.lang.reflect.Method.invoke(Method.java:597) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:291) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1177) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1030) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:488) | |
14/05/30 16:02:57 ERROR hdfs.DFSClient: Failed to close file /user/cloudera/.staging/job_201405300019_0008/job.split | |
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/cloudera/.staging/job_201405300019_0008/job.split could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. | |
at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1340) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2296) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:501) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:299) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44954) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1752) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1748) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1438) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1746) | |
at org.apache.hadoop.ipc.Client.call(Client.java:1238) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) | |
at java.lang.reflect.Method.invoke(Method.java:597) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:291) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1177) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1030) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:488) |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment