Created
January 16, 2014 02:35
-
-
Save alphastorm/8448873 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
2014-01-15 17:05:09,943 INFO org.apache.hadoop.hive.metastore.HiveMetaStore: 131: source:/10.1.128.51 drop_table : db=transaction_cost tbl=batmobile_payments_deduped | |
2014-01-15 17:05:09,944 INFO org.apache.hadoop.hive.metastore.HiveMetaStore.audit: ugi=anonymous ip=/10.1.128.51 cmd=source:/10.1.128.51 drop_table : db=transaction_cost tbl=batmobile_payments_deduped | |
2014-01-15 17:05:09,944 INFO org.apache.hadoop.hive.metastore.HiveMetaStore: 131: source:/10.1.128.51 get_table : db=transaction_cost tbl=batmobile_payments_deduped | |
2014-01-15 17:05:09,944 INFO org.apache.hadoop.hive.metastore.HiveMetaStore.audit: ugi=anonymous ip=/10.1.128.51 cmd=source:/10.1.128.51 get_table : db=transaction_cost tbl=batmobile_payments_deduped | |
2014-01-15 17:05:09,982 INFO hive.metastore.hivemetastoressimpl: deleting hdfs://nameservice1/user/hive/warehouse/transaction_cost.db/batmobile_payments_deduped | |
2014-01-15 17:05:09,990 WARN org.apache.hadoop.fs.TrashPolicyDefault: Can't create trash directory: hdfs://nameservice1/user/anonymous/.Trash/Current/user/hive/warehouse/transaction_cost.db | |
2014-01-15 17:05:10,091 ERROR hive.log: Got exception: java.io.IOException Failed to move to trash: hdfs://nameservice1/user/hive/warehouse/transaction_cost.db/batmobile_payments_deduped | |
2014-01-15 17:05:10,092 ERROR hive.log: java.io.IOException: Failed to move to trash: hdfs://nameservice1/user/hive/warehouse/transaction_cost.db/batmobile_payments_deduped | |
at org.apache.hadoop.fs.TrashPolicyDefault.moveToTrash(TrashPolicyDefault.java:154) | |
at org.apache.hadoop.fs.Trash.moveToTrash(Trash.java:109) | |
at org.apache.hadoop.hive.metastore.HiveMetaStoreFsImpl.deleteDir(HiveMetaStoreFsImpl.java:49) | |
at org.apache.hadoop.hive.metastore.Warehouse.deleteDir(Warehouse.java:209) | |
at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.deleteTableData(HiveMetaStore.java:1212) | |
at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.drop_table_core(HiveMetaStore.java:1195) | |
at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.drop_table(HiveMetaStore.java:1310) | |
at sun.reflect.GeneratedMethodAccessor30.invoke(Unknown Source) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) | |
at java.lang.reflect.Method.invoke(Method.java:597) | |
at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:105) | |
at $Proxy5.drop_table(Unknown Source) | |
at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$drop_table.getResult(ThriftHiveMetastore.java:5891) | |
at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$drop_table.getResult(ThriftHiveMetastore.java:5875) | |
at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) | |
at org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:115) | |
at org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:112) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) | |
at org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:582) | |
at org.apache.hadoop.hive.metastore.TUGIBasedProcessor.process(TUGIBasedProcessor.java:123) | |
at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:244) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) | |
at java.lang.Thread.run(Thread.java:662) | |
Caused by: org.apache.hadoop.security.AccessControlException: Permission denied: user=anonymous, access=WRITE, inode="/user":hdfs:supergroup:drwxr-xr-x | |
at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:224) | |
at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:204) | |
at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:149) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:4846) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:4828) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:4802) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3130) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3094) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3075) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:669) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:419) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44970) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1752) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1748) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1746) | |
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) | |
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) | |
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) | |
at java.lang.reflect.Constructor.newInstance(Constructor.java:513) | |
at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:90) | |
at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:57) | |
at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2153) | |
at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2122) | |
at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:545) | |
at org.apache.hadoop.fs.TrashPolicyDefault.moveToTrash(TrashPolicyDefault.java:130) | |
... 25 more | |
Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Permission denied: user=anonymous, access=WRITE, inode="/user":hdfs:supergroup:drwxr-xr-x | |
at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:224) | |
at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:204) | |
at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:149) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:4846) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:4828) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:4802) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3130) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3094) | |
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3075) | |
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:669) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:419) | |
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44970) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) | |
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1752) | |
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1748) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:396) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) | |
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1746) | |
at org.apache.hadoop.ipc.Client.call(Client.java:1238) | |
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202) | |
at $Proxy12.mkdirs(Unknown Source) | |
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:426) | |
at sun.reflect.GeneratedMethodAccessor38.invoke(Unknown Source) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) | |
at java.lang.reflect.Method.invoke(Method.java:597) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164) | |
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83) | |
at $Proxy13.mkdirs(Unknown Source) | |
at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2151) | |
... 28 more | |
2014-01-15 17:05:10,097 ERROR org.apache.hadoop.hive.metastore.HiveMetaStore: Failed to delete table directory: hdfs://nameservice1/user/hive/warehouse/transaction_cost.db/batmobile_payments_deduped Got exception: java.io.IOException Failed to move to trash: hdfs://nameservice1/user/hive/warehouse/transaction_cost.db/batmobile_payments_deduped |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment