Created
May 29, 2014 08:21
-
-
Save fivesmallq/e31b003d9e6c98d15933 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF-8 | |
2014-05-29 16:21:20.007 java[3866:1003] Unable to load realm info from SCDynamicStore | |
14/05/29 16:21:20 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable | |
2014-05-29 16:21:20.091 java[3866:1003] Unable to load realm info from SCDynamicStore | |
14/05/29 16:21:21 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. | |
14/05/29 16:21:21 WARN mapred.JobClient: No job jar file set. User classes may not be found. See JobConf(Class) or JobConf#setJar(String). | |
14/05/29 16:21:21 INFO input.FileInputFormat: Total input paths to process : 1 | |
14/05/29 16:21:21 INFO hdfs.DFSClient: Exception in createBlockOutputStream | |
java.net.ConnectException: Connection refused | |
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) | |
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:599) | |
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:207) | |
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:528) | |
at org.apache.hadoop.hdfs.DFSOutputStream.createSocketForPipeline(DFSOutputStream.java:1254) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:1080) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1040) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:488) | |
14/05/29 16:21:21 INFO hdfs.DFSClient: Abandoning BP-780931682-127.0.0.1-1381159027878:blk_5246535241558601811_2269 | |
14/05/29 16:21:21 INFO hdfs.DFSClient: Excluding datanode 127.0.0.1:50010 | |
14/05/29 16:21:21 WARN hdfs.DFSClient: DataStreamer Exception | |
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/fivesmallq/.staging/job_201405282359_0002/job.split could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. | |
at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1339) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2198) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:501) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:299) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44954) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1751) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1747) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1745) | |
at org.apache.hadoop.ipc.Client.call(Client.java:1238) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) | |
at java.lang.reflect.Method.invoke(Method.java:597) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:291) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1177) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1030) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:488) | |
14/05/29 16:21:21 INFO mapred.JobClient: Cleaning up the staging area hdfs://hadoop-server:8020/user/fivesmallq/.staging/job_201405282359_0002 | |
14/05/29 16:21:21 ERROR security.UserGroupInformation: PriviledgedActionException as:fivesmallq (auth:SIMPLE) cause:org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/fivesmallq/.staging/job_201405282359_0002/job.split could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. | |
at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1339) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2198) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:501) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:299) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44954) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1751) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1747) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1745) | |
Exception in thread "main" org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/fivesmallq/.staging/job_201405282359_0002/job.split could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. | |
at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1339) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2198) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:501) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:299) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44954) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1751) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1747) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1745) | |
at org.apache.hadoop.ipc.Client.call(Client.java:1238) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) | |
at java.lang.reflect.Method.invoke(Method.java:597) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:291) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1177) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1030) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:488) | |
14/05/29 16:21:21 ERROR hdfs.DFSClient: Failed to close file /user/fivesmallq/.staging/job_201405282359_0002/job.split | |
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/fivesmallq/.staging/job_201405282359_0002/job.split could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. | |
at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1339) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2198) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:501) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:299) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44954) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1751) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1747) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1745) | |
at org.apache.hadoop.ipc.Client.call(Client.java:1238) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) | |
at java.lang.reflect.Method.invoke(Method.java:597) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83) | |
at com.sun.proxy.$Proxy9.addBlock(Unknown Source) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:291) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1177) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1030) | |
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:488) |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment