Created
January 21, 2018 14:03
-
-
Save mattazend/b611d0232d94ade4bc4c16bcb79f73a8 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
2018-01-21T13:02:26.333Z --conf spark.hadoop.yarn.resourcemanager.connect.max-wait.ms=60000 --conf spark.hadoop.fs.defaultFS=hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020 --conf spark.hadoop.yarn.resourcemanager.address=ip-10-0-1-48.us-west-2.compute.internal:8032 --conf spark.dynamicAllocation.enabled=true --conf spark.shuffle.service.enabled=true --conf spark.dynamicAllocation.minExecutors=1 --conf spark.dynamicAllocation.maxExecutors=18 --conf spark.executor.memory=5g --conf spark.executor.cores=4 --JOB_ID j_e5de928b745ac948b9e0299b9c1983bea68d9383a53d12d9c9738efd77128f04 --extra-py-files s3://###CLIPPED####/jobs/test_trunc/python-libs/pyRserve.zip --JOB_RUN_ID jr_13d318641534349c645a5622c62d56d0804929911f5c326a19cc872be957a9a4 --scriptLocation s3://###CLIPPED####/gluescripts/test_trunc --job-bookmark-option job-bookmark-disable --job-language python --TempDir s3://###CLIPPED####/jobs/test_trunc/scripts --JOB_NAME test_trunc | |
2018-01-21T13:02:26.340Z YARN_RM_DNS=ip-10-0-1-48.us-west-2.compute.internal | |
2018-01-21T13:02:26.357Z Detected region us-west-2 | |
2018-01-21T13:02:26.357Z JOB_NAME = test_trunc | |
2018-01-21T13:02:26.359Z Specifying us-west-2 while copying script. | |
2018-01-21T13:02:27.181Z Completed 229 Bytes/229 Bytes (3.7 KiB/s) with 1 file(s) remaining | |
2018-01-21T13:02:27.181Z download: s3://###CLIPPED####/gluescripts/test_trunc to ./script_2018-01-21-13-02-26.py | |
2018-01-21T13:02:27.254Z SCRIPT_URL = /tmp/g-06dbb7fa17bfe5fff90406afac80ef44367183a9-3027215820432691858/script_2018-01-21-13-02-26.py | |
2018-01-21T13:02:27.254Z ------------------EXECUTING SCRIPT------------------ | |
2018-01-21T13:02:27.254Z import sys | |
from awsglue.transforms import * | |
from awsglue.utils import getResolvedOptions | |
from pyspark.context import SparkContext | |
from awsglue.context import GlueContext | |
from awsglue.job import Job | |
import pprint | |
import pyRserve | |
2018-01-21T13:02:27.254Z ---------------------------------------------------- | |
2018-01-21T13:02:27.270Z /usr/lib/spark/bin/spark-submit --conf spark.hadoop.yarn.resourcemanager.connect.max-wait.ms=60000 --conf spark.hadoop.fs.defaultFS=hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020 --conf spark.hadoop.yarn.resourcemanager.address=ip-10-0-1-48.us-west-2.compute.internal:8032 --conf spark.dynamicAllocation.enabled=true --conf spark.shuffle.service.enabled=true --conf spark.dynamicAllocation.minExecutors=1 --conf spark.dynamicAllocation.maxExecutors=18 --conf spark.executor.memory=5g --conf spark.executor.cores=4 --name tape --master yarn --deploy-mode cluster --jars /opt/amazon/superjar/glue-assembly.jar --files /tmp/glue-default.conf,/tmp/glue-override.conf,/opt/amazon/certs/InternalAndExternalAndAWSTrustStore.jks,/opt/amazon/certs/rds-combined-ca-bundle.pem,/tmp/g-06dbb7fa17bfe5fff90406afac80ef44367183a9-3027215820432691858/script_2018-01-21-13-02-26.py --py-files /tmp/PyGlue.zip,s3://###CLIPPED####/jobs/test_trunc/python-libs/pyRserve.zip --driver-memory 5g --executor-memory 5g /tmp/runscript.py script_2018-01-21-13-02-26.py --JOB_NAME test_trunc --JOB_ID j_e5de928b745ac948b9e0299b9c1983bea68d9383a53d12d9c9738efd77128f04 --JOB_RUN_ID jr_13d318641534349c645a5622c62d56d0804929911f5c326a19cc872be957a9a4 --job-bookmark-option job-bookmark-disable --TempDir s3://###CLIPPED####/jobs/test_trunc/scripts | |
2018-01-21T13:02:41.253Z 18/01/21 13:02:28 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable | |
18/01/21 13:02:29 INFO RMProxy: Connecting to ResourceManager at ip-10-0-1-48.us-west-2.compute.internal/10.0.1.48:8032 | |
18/01/21 13:02:29 INFO Client: Requesting a new application from cluster with 10 NodeManagers | |
18/01/21 13:02:29 INFO Client: Verifying our application has not requested more than the maximum memory capability of the cluster (12288 MB per container) | |
18/01/21 13:02:29 INFO Client: Will allocate AM container, with 5632 MB memory including 512 MB overhead | |
18/01/21 13:02:29 INFO Client: Setting up container launch context for our AM | |
18/01/21 13:02:29 INFO Client: Setting up the launch environment for our AM container | |
18/01/21 13:02:29 DEBUG Client: Using the default YARN application classpath: $HADOOP_CONF_DIR,$HADOOP_COMMON_HOME/share/hadoop/common/*,$HADOOP_COMMON_HOME/share/hadoop/common/lib/*,$HADOOP_HDFS_HOME/share/hadoop/hdfs/*,$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*,$HADOOP_YARN_HOME/share/hadoop/yarn/*,$HADOOP_YARN_HOME/share/hadoop/yarn/lib/* | |
18/01/21 13:02:29 DEBUG Client: Using the default MR application classpath: $HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*,$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/* | |
18/01/21 13:02:29 INFO Client: Preparing resources for our AM container | |
18/01/21 13:02:30 DEBUG Client: | |
18/01/21 13:02:30 WARN Client: Neither spark.yarn.jars nor spark.yarn.archive is set, falling back to uploading libraries under SPARK_HOME. | |
18/01/21 13:02:33 INFO Client: Uploading resource file:/tmp/spark-e753e2e5-a1f0-4c6f-a6df-52614abda117/__spark_libs__4940228112749105289.zip -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/__spark_libs__4940228112749105289.zip | |
18/01/21 13:02:35 INFO Client: Uploading resource file:/opt/amazon/superjar/glue-assembly.jar -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/glue-assembly.jar | |
18/01/21 13:02:39 INFO Client: Uploading resource file:/tmp/glue-default.conf -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/glue-default.conf | |
18/01/21 13:02:39 INFO Client: Uploading resource file:/tmp/glue-override.conf -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/glue-override.conf | |
18/01/21 13:02:39 INFO Client: Uploading resource file:/opt/amazon/certs/InternalAndExternalAndAWSTrustStore.jks -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/InternalAndExternalAndAWSTrustStore.jks | |
18/01/21 13:02:39 INFO Client: Uploading resource file:/opt/amazon/certs/rds-combined-ca-bundle.pem -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/rds-combined-ca-bundle.pem | |
18/01/21 13:02:39 INFO Client: Uploading resource file:/tmp/g-06dbb7fa17bfe5fff90406afac80ef44367183a9-3027215820432691858/script_2018-01-21-13-02-26.py -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/script_2018-01-21-13-02-26.py | |
18/01/21 13:02:39 INFO Client: Uploading resource file:/tmp/runscript.py -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/runscript.py | |
18/01/21 13:02:39 INFO Client: Uploading resource file:/usr/lib/spark/python/lib/pyspark.zip -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/pyspark.zip | |
18/01/21 13:02:39 INFO Client: Uploading resource file:/usr/lib/spark/python/lib/py4j-0.10.4-src.zip -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/py4j-0.10.4-src.zip | |
18/01/21 13:02:39 INFO Client: Uploading resource file:/tmp/PyGlue.zip -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/PyGlue.zip | |
18/01 | |
2018-01-21T13:02:41.264Z /21 13:02:40 INFO PlatformInfo: Unable to read clusterId from http://localhost:8321/configuration, trying extra instance data file: /var/lib/instance-controller/extraInstanceData.json | |
18/01/21 13:02:40 INFO PlatformInfo: Unable to read clusterId from /var/lib/instance-controller/extraInstanceData.json, trying EMR job-flow data file: /var/lib/info/job-flow.json | |
18/01/21 13:02:40 INFO PlatformInfo: Unable to read clusterId from /var/lib/info/job-flow.json, out of places to look | |
18/01/21 13:02:41 INFO Client: Uploading resource s3://###CLIPPED####/jobs/test_trunc/python-libs/pyRserve.zip -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/pyRserve.zip | |
18/01/21 13:02:41 INFO S3NativeFileSystem: Opening 's3://###CLIPPED####/jobs/test_trunc/python-libs/pyRserve.zip' for reading | |
18/01/21 13:02:41 INFO Client: Uploading resource file:/tmp/spark-e753e2e5-a1f0-4c6f-a6df-52614abda117/__spark_conf__1785103944074202664.zip -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010/__spark_conf__.zip | |
18/01/21 13:02:41 DEBUG Client: =============================================================================== | |
18/01/21 13:02:41 DEBUG Client: YARN AM launch context: | |
18/01/21 13:02:41 DEBUG Client: user class: org.apache.spark.deploy.PythonRunner | |
18/01/21 13:02:41 DEBUG Client: env: | |
18/01/21 13:02:41 DEBUG Client: CLASSPATH -> ./*:/usr/lib/hadoop-lzo/lib/*:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*<CPS>{{PWD}}<CPS>{{PWD}}/__spark_conf__<CPS>{{PWD}}/__spark_libs__/*<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/* | |
18/01/21 13:02:41 DEBUG Client: SPARK_YARN_STAGING_DIR -> hdfs://ip-10-0-1-48.us-west-2.compute.internal:8020/user/root/.sparkStaging/application_1516530473599_0010 | |
18/01/21 13:02:41 DEBUG Client: SPARK_USER -> root | |
18/01/21 13:02:41 DEBUG Client: SPARK_YARN_MODE -> true | |
18/01/21 13:02:41 DEBUG Client: PYTHONPATH -> {{PWD}}/pyspark.zip<CPS>{{PWD}}/py4j-0.10.4-src.zip<CPS>{{PWD}}/PyGlue.zip<CPS>{{PWD}}/pyRserve.zip | |
18/01/21 13:02:41 DEBUG Client: resources: | |
18/01/21 13:02:41 DEBUG Client: py4j-0.10.4-src.zip -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/py4j-0.10.4-src.zip" } size: 74096 timestamp: 1516539759729 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: glue-assembly.jar -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/glue-assembly.jar" } size: 377465495 timestamp: 1516539759567 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: pyspark.zip -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/pyspark.zip" } size: 452353 timestamp: 1516539759710 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: __spark_libs__ -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/__spark_libs__4940228112749105289.zip" } size: 196716025 timestamp: 1516539755676 type: ARCHIVE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: rds-combined-ca-bundle.pem -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/rds-combined-ca-bundle.pem" } size: 21672 timestamp: 1516539759653 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG | |
2018-01-21T13:02:50.337Z Client: glue-default.conf -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/glue-default.conf" } size: 224 timestamp: 1516539759588 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: runscript.py -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/runscript.py" } size: 2370 timestamp: 1516539759689 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: glue-override.conf -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/glue-override.conf" } size: 271 timestamp: 1516539759609 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: InternalAndExternalAndAWSTrustStore.jks -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/InternalAndExternalAndAWSTrustStore.jks" } size: 118806 timestamp: 1516539759630 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: PyGlue.zip -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/PyGlue.zip" } size: 99983 timestamp: 1516539759757 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: script_2018-01-21-13-02-26.py -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/script_2018-01-21-13-02-26.py" } size: 229 timestamp: 1516539759671 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: __spark_conf__ -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/__spark_conf__.zip" } size: 7501 timestamp: 1516539761171 type: ARCHIVE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: pyRserve.zip -> resource { scheme: "hdfs" host: "ip-10-0-1-48.us-west-2.compute.internal" port: 8020 file: "/user/root/.sparkStaging/application_1516530473599_0010/pyRserve.zip" } size: 59979 timestamp: 1516539761107 type: FILE visibility: PRIVATE | |
18/01/21 13:02:41 DEBUG Client: command: | |
18/01/21 13:02:41 DEBUG Client: LD_LIBRARY_PATH="/usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native:$LD_LIBRARY_PATH" {{JAVA_HOME}}/bin/java -server -Xmx5120m -Djava.io.tmpdir={{PWD}}/tmp '-XX:+UseConcMarkSweepGC' '-XX:CMSInitiatingOccupancyFraction=70' '-XX:MaxHeapFreeRatio=70' '-XX:+CMSClassUnloadingEnabled' '-XX:OnOutOfMemoryError=kill -9 %p' '-Djavax.net.ssl.trustStore=InternalAndExternalAndAWSTrustStore.jks' '-Djavax.net.ssl.trustStoreType=JKS' '-Djavax.net.ssl.trustStorePassword=amazon' '-DRDS_ROOT_CERT_PATH=rds-combined-ca-bundle.pem' -Dspark.yarn.app.container.log.dir=<LOG_DIR> org.apache.spark.deploy.yarn.ApplicationMaster --class 'org.apache.spark.deploy.PythonRunner' --primary-py-file runscript.py --arg 'script_2018-01-21-13-02-26.py' --arg '--JOB_NAME' --arg 'test_trunc' --arg '--JOB_ID' --arg 'j_e5de928b745ac948b9e0299b9c1983bea68d9383a53d12d9c9738efd77128f04' --arg '--JOB_RUN_ID' --arg 'jr_13d318641534349c645a5622c62d56d0804929911f5c326a19cc872be957a9a4' --arg '--job-bookmark-option' --arg 'job-bookmark-disable' --arg '--TempDir' --arg 's3://###CLIPPED####/jobs/test_trunc/scripts' --properties-file {{PWD}}/__spark_conf__/__spark_conf__.properties 1> <LOG_DIR>/stdout 2> <LOG_DIR>/stderr | |
18/01/21 13:02:41 DEBUG Client: =============================================================================== | |
18/01/21 13:02:41 INFO SecurityManager: Changing view acls to: root | |
18/01/21 13:02:41 INFO SecurityManager: Changing modify acls to: root | |
18/01/21 13:02:41 INFO SecurityManager: Changing view acls groups to: | |
18/01/21 13:02:41 INFO SecurityManager: Changing modify acls groups to: | |
18/01/21 13:02:41 INFO SecurityMa | |
2018-01-21T13:02:50.338Z nager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); groups with view permissions: Set(); users with modify permissions: Set(root); groups with modify permissions: Set() | |
18/01/21 13:02:41 DEBUG Client: spark.yarn.maxAppAttempts is not set. Cluster's default value will be used. | |
18/01/21 13:02:41 INFO Client: Submitting application application_1516530473599_0010 to ResourceManager | |
18/01/21 13:02:41 INFO YarnClientImpl: Submitted application application_1516530473599_0010 | |
18/01/21 13:02:42 INFO Client: Application report for application_1516530473599_0010 (state: ACCEPTED) | |
2018-01-21T13:02:50.339Z applicationid is application_1516530473599_0010, yarnRMDNS is ip-10-0-1-48.us-west-2.compute.internal | |
Application info reporting is enabled. | |
2018-01-21T13:02:50.339Z ----------Recording application Id and Yarn RM DNS for cancellation----------------- | |
2018-01-21T13:02:50.340Z ) | |
18/01/21 13:02:50 DEBUG Client: | |
client token: N/A | |
diagnostics: N/A | |
ApplicationMaster host: N/A | |
ApplicationMaster RPC port: -1 | |
queue: default | |
start time: 1516539761299 | |
final status: UNDEFINED | |
tracking URL: http://ip-10-0-1-48.us-west-2.compute.internal:20888/proxy/application_1516530473599_0010/ | |
user: root | |
2018-01-21T13:02:51.339Z 18/01/21 13:02:51 INFO Client: Application report for application_1516530473599_0010 (state: FAILED) | |
2018-01-21T13:02:51.339Z 18/01/21 13:02:51 DEBUG Client: | |
client token: N/A | |
diagnostics: Application application_1516530473599_0010 failed 2 times due to AM Container for appattempt_1516530473599_0010_000002 exited with exitCode: 1 | |
For more detailed output, check application tracking page:http://169.254.76.1:8088/cluster/app/application_1516530473599_0010Then, click on links to logs of each attempt. | |
Diagnostics: Exception from container-launch. | |
Container id: container_1516530473599_0010_02_000001 | |
Exit code: 1 | |
Stack trace: ExitCodeException exitCode=1: | |
at org.apache.hadoop.util.Shell.runCommand(Shell.java:582) | |
at org.apache.hadoop.util.Shell.run(Shell.java:479) | |
at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:773) | |
at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:212) | |
at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:302) | |
at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82) | |
at java.util.concurrent.FutureTask.run(FutureTask.java:266) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) | |
at java.lang.Thread.run(Thread.java:748) | |
Container exited with a non-zero exit code 1 | |
Failing this attempt. Failing the application. | |
ApplicationMaster host: N/A | |
ApplicationMaster RPC port: -1 | |
queue: default | |
start time: 1516539761299 | |
final status: FAILED | |
tracking URL: http://169.254.76.1:8088/cluster/app/application_1516530473599_0010 | |
user: root | |
2018-01-21T13:02:51.348Z Exception in thread "main" org.apache.spark.SparkException: Application application_1516530473599_0010 finished with failed status | |
at org.apache.spark.deploy.yarn.Client.run(Client.scala:1167) | |
at org.apache.spark.deploy.yarn.Client$.main(Client.scala:1213) | |
at org.apache.spark.deploy.yarn.Client.main(Client.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:738) | |
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187) | |
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212) | |
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126) | |
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) | |
2018-01-21T13:02:51.349Z 18/01/21 13:02:51 INFO ShutdownHookManager: Shutdown hook called | |
2018-01-21T13:02:51.350Z 18/01/21 13:02:51 INFO ShutdownHookManager: Deleting directory /tmp/spark-e753e2e5-a1f0-4c6f-a6df-52614abda117 | |
2018-01-21T13:02:53.111Z | |
Container: container_1516530473599_0010_02_000001 on ip-10-0-1-230.us-west-2.compute.internal_8041 | |
==================================================================================================== | |
LogType:stderr | |
Log Upload Time:Sun Jan 21 13:02:52 +0000 2018 | |
LogLength:3934 | |
Log Contents: | |
SLF4J: Class path contains multiple SLF4J bindings. | |
SLF4J: Found binding in [jar:file:/mnt/yarn/usercache/root/filecache/22/glue-assembly.jar!/org/slf4j/impl/StaticLoggerBinder.class] | |
SLF4J: Found binding in [jar:file:/mnt/yarn/usercache/root/filecache/30/__spark_libs__4940228112749105289.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] | |
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. | |
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] | |
18/01/21 13:02:48 INFO SignalUtils: Registered signal handler for TERM | |
18/01/21 13:02:48 INFO SignalUtils: Registered signal handler for HUP | |
18/01/21 13:02:48 INFO SignalUtils: Registered signal handler for INT | |
18/01/21 13:02:49 INFO ApplicationMaster: Preparing Local resources | |
18/01/21 13:02:49 INFO ApplicationMaster: ApplicationAttemptId: appattempt_1516530473599_0010_000002 | |
18/01/21 13:02:49 INFO SecurityManager: Changing view acls to: yarn,root | |
18/01/21 13:02:49 INFO SecurityManager: Changing modify acls to: yarn,root | |
18/01/21 13:02:49 INFO SecurityManager: Changing view acls groups to: | |
18/01/21 13:02:49 INFO SecurityManager: Changing modify acls groups to: | |
18/01/21 13:02:49 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(yarn, root); groups with view permissions: Set(); users with modify permissions: Set(yarn, root); groups with modify permissions: Set() | |
18/01/21 13:02:49 INFO ApplicationMaster: Starting the user application in a separate Thread | |
18/01/21 13:02:49 INFO ApplicationMaster: Waiting for spark context initialization... | |
18/01/21 13:02:50 ERROR ApplicationMaster: User application exited with status 1 | |
18/01/21 13:02:50 INFO ApplicationMaster: Final app status: FAILED, exitCode: 1, (reason: User application exited with status 1) | |
18/01/21 13:02:50 ERROR ApplicationMaster: Uncaught exception: | |
org.apache.spark.SparkException: Exception thrown in awaitResult: | |
at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:194) | |
at org.apache.spark.deploy.yarn.ApplicationMaster.runDriver(ApplicationMaster.scala:401) | |
at org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:254) | |
at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$main$1.apply$mcV$sp(ApplicationMaster.scala:766) | |
at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:67) | |
at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:66) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:422) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) | |
at org.apache.spark.deploy.SparkHadoopUtil.runAsSparkUser(SparkHadoopUtil.scala:66) | |
at org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:764) | |
at org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala) | |
Caused by: org.apache.spark.SparkUserAppException: User application exited with 1 | |
at org.apache.spark.deploy.PythonRunner$.main(PythonRunner.scala:96) | |
at org.apache.spark.deploy.PythonRunner.main(PythonRunner.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:637) | |
18/01/21 13:02:50 INFO ApplicationMaster: Unregistering ApplicationMaster with FAILED (diag message: User application exited with status 1) | |
18/01/21 13:02:50 INFO ApplicationMaster: Deleting staging directory hdfs://ip-10-0-1-48.us-west-2.compute.int | |
2018-01-21T13:02:53.434Z ernal:8020/user/root/.sparkStaging/application_1516530473599_0010 | |
18/01/21 13:02:50 INFO ShutdownHookManager: Shutdown hook called | |
End of LogType:stderr | |
LogType:stdout | |
Log Upload Time:Sun Jan 21 13:02:52 +0000 2018 | |
LogLength:206 | |
Log Contents: | |
Traceback (most recent call last): | |
File "script_2018-01-21-13-02-26.py", line 8, in <module> | |
import pyRserve | |
ImportError: No module named pyRserve | |
End of LogType:stdout | |
Container: container_1516530473599_0010_01_000001 on ip-10-0-1-56.us-west-2.compute.internal_8041 | |
=================================================================================================== | |
LogType:stderr | |
Log Upload Time:Sun Jan 21 13:02:52 +0000 2018 | |
LogLength:3618 | |
Log Contents: | |
SLF4J: Class path contains multiple SLF4J bindings. | |
SLF4J: Found binding in [jar:file:/mnt/yarn/usercache/root/filecache/62/glue-assembly.jar!/org/slf4j/impl/StaticLoggerBinder.class] | |
SLF4J: Found binding in [jar:file:/mnt/yarn/usercache/root/filecache/70/__spark_libs__4940228112749105289.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] | |
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. | |
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] | |
18/01/21 13:02:43 INFO SignalUtils: Registered signal handler for TERM | |
18/01/21 13:02:43 INFO SignalUtils: Registered signal handler for HUP | |
18/01/21 13:02:43 INFO SignalUtils: Registered signal handler for INT | |
18/01/21 13:02:44 INFO ApplicationMaster: Preparing Local resources | |
18/01/21 13:02:44 INFO ApplicationMaster: ApplicationAttemptId: appattempt_1516530473599_0010_000001 | |
18/01/21 13:02:44 INFO SecurityManager: Changing view acls to: yarn,root | |
18/01/21 13:02:44 INFO SecurityManager: Changing modify acls to: yarn,root | |
18/01/21 13:02:44 INFO SecurityManager: Changing view acls groups to: | |
18/01/21 13:02:44 INFO SecurityManager: Changing modify acls groups to: | |
18/01/21 13:02:44 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(yarn, root); groups with view permissions: Set(); users with modify permissions: Set(yarn, root); groups with modify permissions: Set() | |
18/01/21 13:02:44 INFO ApplicationMaster: Starting the user application in a separate Thread | |
18/01/21 13:02:44 INFO ApplicationMaster: Waiting for spark context initialization... | |
18/01/21 13:02:45 ERROR ApplicationMaster: User application exited with status 1 | |
18/01/21 13:02:45 INFO ApplicationMaster: Final app status: FAILED, exitCode: 1, (reason: User application exited with status 1) | |
18/01/21 13:02:45 ERROR ApplicationMaster: Uncaught exception: | |
org.apache.spark.SparkException: Exception thrown in awaitResult: | |
at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:194) | |
at org.apache.spark.deploy.yarn.ApplicationMaster.runDriver(ApplicationMaster.scala:401) | |
at org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:254) | |
at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$main$1.apply$mcV$sp(ApplicationMaster.scala:766) | |
at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:67) | |
at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:66) | |
at java.security.AccessController.doPrivileged(Native Method) | |
at javax.security.auth.Subject.doAs(Subject.java:422) | |
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) | |
at org.apache.spark.deploy.SparkHadoopUtil.runAsSparkUser(SparkHadoopUtil.scala:66) | |
at org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:764) | |
at org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala) | |
Caused by: org.apache.spark.SparkUserAppException: User application exited with 1 | |
at org.apache.spark.deploy.PythonRunner$.main(PythonRunner.scala:96) | |
at org.apache.spark.deploy.PythonRunner.main(PythonRunner.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:637) | |
18/01/21 13:02:45 INFO ShutdownHookManager: Shutdown hook called | |
End of LogType:stderr | |
LogType:stdout | |
Log Upload Time:Sun Jan 21 13:02:52 +0000 2018 | |
LogLength:206 | |
Log Contents: | |
Traceback (most recent call last): | |
File "script_2018-01-21-13-02-26.py", line 8, in <module> | |
import pyRserve | |
ImportError: No module named pyRserve | |
End of LogType:stdout | |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment