Last active
March 30, 2016 18:06
-
-
Save dkirrane/8485d8d6f4c422310ec69d0e89271b35 to your computer and use it in GitHub Desktop.
Spark Streaming DataFrame to Elasticsearch (https://discuss.elastic.co/t/sparkstreaming-to-elasticesrahc-error-networkclient-connection-timed-out-connect/45834)
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
16/03/30 18:46:51 ERROR NetworkClient: Node [172.18.0.2:9200] failed (Connection timed out: connect); selected next node [10.123.45.67:9200] | |
16/03/30 18:48:15 ERROR NetworkClient: Node [172.18.0.2:9200] failed (Connection timed out: connect); no other nodes left - aborting... | |
16/03/30 18:48:15 ERROR Executor: Exception in task 0.0 in stage 12.0 (TID 609) | |
org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.18.0.2:9200]] | |
at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:142) | |
at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:423) | |
at org.elasticsearch.hadoop.rest.RestClient.executeNotFoundAllowed(RestClient.java:431) | |
at org.elasticsearch.hadoop.rest.RestClient.exists(RestClient.java:507) | |
at org.elasticsearch.hadoop.rest.RestClient.touch(RestClient.java:513) | |
at org.elasticsearch.hadoop.rest.RestRepository.touch(RestRepository.java:491) | |
at org.elasticsearch.hadoop.rest.RestService.initSingleIndex(RestService.java:412) | |
at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:400) | |
at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:40) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
16/03/30 18:48:15 WARN TaskSetManager: Lost task 0.0 in stage 12.0 (TID 609, localhost): org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.18.0.2:9200]] | |
at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:142) | |
at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:423) | |
at org.elasticsearch.hadoop.rest.RestClient.executeNotFoundAllowed(RestClient.java:431) | |
at org.elasticsearch.hadoop.rest.RestClient.exists(RestClient.java:507) | |
at org.elasticsearch.hadoop.rest.RestClient.touch(RestClient.java:513) | |
at org.elasticsearch.hadoop.rest.RestRepository.touch(RestRepository.java:491) | |
at org.elasticsearch.hadoop.rest.RestService.initSingleIndex(RestService.java:412) | |
at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:400) | |
at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:40) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
16/03/30 18:48:15 ERROR TaskSetManager: Task 0 in stage 12.0 failed 1 times; aborting job | |
16/03/30 18:48:15 ERROR JobScheduler: Error running job streaming job 1459359917000 ms.0 | |
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 12.0 failed 1 times, most recent failure: Lost task 0.0 in stage 12.0 (TID 609, localhost): org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.18.0.2:9200]] | |
at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:142) | |
at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:423) | |
at org.elasticsearch.hadoop.rest.RestClient.executeNotFoundAllowed(RestClient.java:431) | |
at org.elasticsearch.hadoop.rest.RestClient.exists(RestClient.java:507) | |
at org.elasticsearch.hadoop.rest.RestClient.touch(RestClient.java:513) | |
at org.elasticsearch.hadoop.rest.RestRepository.touch(RestRepository.java:491) | |
at org.elasticsearch.hadoop.rest.RestService.initSingleIndex(RestService.java:412) | |
at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:400) | |
at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:40) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Driver stacktrace: | |
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418) | |
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) | |
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) | |
at scala.Option.foreach(Option.scala:236) | |
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588) | |
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) | |
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620) | |
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832) | |
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845) | |
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1922) | |
at org.elasticsearch.spark.sql.EsSparkSQL$.saveToEs(EsSparkSQL.scala:55) | |
at org.elasticsearch.spark.sql.EsSparkSQL$.saveToEs(EsSparkSQL.scala:43) | |
at org.elasticsearch.spark.sql.api.java.JavaEsSparkSQL$.saveToEs(JavaEsSparkSQL.scala:27) | |
at org.elasticsearch.spark.sql.api.java.JavaEsSparkSQL.saveToEs(JavaEsSparkSQL.scala) | |
at com.my.spark.app.JsonFileStream.lambda$main$1282d8df$1(JsonFileStream.java:102) | |
at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$3.apply(JavaDStreamLike.scala:335) | |
at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$3.apply(JavaDStreamLike.scala:335) | |
at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661) | |
at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:50) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50) | |
at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:426) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:49) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49) | |
at scala.util.Try$.apply(Try.scala:161) | |
at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:224) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224) | |
at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:223) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Caused by: org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.18.0.2:9200]] | |
at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:142) | |
at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:423) | |
at org.elasticsearch.hadoop.rest.RestClient.executeNotFoundAllowed(RestClient.java:431) | |
at org.elasticsearch.hadoop.rest.RestClient.exists(RestClient.java:507) | |
at org.elasticsearch.hadoop.rest.RestClient.touch(RestClient.java:513) | |
at org.elasticsearch.hadoop.rest.RestRepository.touch(RestRepository.java:491) | |
at org.elasticsearch.hadoop.rest.RestService.initSingleIndex(RestService.java:412) | |
at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:400) | |
at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:40) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) | |
... 3 more | |
Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 12.0 failed 1 times, most recent failure: Lost task 0.0 in stage 12.0 (TID 609, localhost): org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.18.0.2:9200]] | |
at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:142) | |
at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:423) | |
at org.elasticsearch.hadoop.rest.RestClient.executeNotFoundAllowed(RestClient.java:431) | |
at org.elasticsearch.hadoop.rest.RestClient.exists(RestClient.java:507) | |
at org.elasticsearch.hadoop.rest.RestClient.touch(RestClient.java:513) | |
at org.elasticsearch.hadoop.rest.RestRepository.touch(RestRepository.java:491) | |
at org.elasticsearch.hadoop.rest.RestService.initSingleIndex(RestService.java:412) | |
at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:400) | |
at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:40) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Driver stacktrace: | |
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418) | |
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) | |
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) | |
at scala.Option.foreach(Option.scala:236) | |
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588) | |
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) | |
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620) | |
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832) | |
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845) | |
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1922) | |
at org.elasticsearch.spark.sql.EsSparkSQL$.saveToEs(EsSparkSQL.scala:55) | |
at org.elasticsearch.spark.sql.EsSparkSQL$.saveToEs(EsSparkSQL.scala:43) | |
at org.elasticsearch.spark.sql.api.java.JavaEsSparkSQL$.saveToEs(JavaEsSparkSQL.scala:27) | |
at org.elasticsearch.spark.sql.api.java.JavaEsSparkSQL.saveToEs(JavaEsSparkSQL.scala) | |
at com.my.spark.app.JsonFileStream.lambda$main$1282d8df$1(JsonFileStream.java:102) | |
at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$3.apply(JavaDStreamLike.scala:335) | |
at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$3.apply(JavaDStreamLike.scala:335) | |
at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661) | |
at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:50) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50) | |
at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:426) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:49) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49) | |
at scala.util.Try$.apply(Try.scala:161) | |
at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:224) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224) | |
at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:223) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Caused by: org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.18.0.2:9200]] | |
at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:142) | |
at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:423) | |
at org.elasticsearch.hadoop.rest.RestClient.executeNotFoundAllowed(RestClient.java:431) | |
at org.elasticsearch.hadoop.rest.RestClient.exists(RestClient.java:507) | |
at org.elasticsearch.hadoop.rest.RestClient.touch(RestClient.java:513) | |
at org.elasticsearch.hadoop.rest.RestRepository.touch(RestRepository.java:491) | |
at org.elasticsearch.hadoop.rest.RestService.initSingleIndex(RestService.java:412) | |
at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:400) | |
at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:40) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.elasticsearch.spark.sql.EsSparkSQL$$anonfun$saveToEs$1.apply(EsSparkSQL.scala:55) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) | |
... 3 more | |
16/03/30 18:48:15 ERROR ReceiverTracker: Deregistered receiver for stream 0: Stopped by driver | |
RDD count: '276' | |
16/03/30 18:48:15 WARN SocketReceiver: Error receiving data | |
org.apache.spark.SparkException: Cannot add data as BlockGenerator has not been started or has been stopped | |
at org.apache.spark.streaming.receiver.BlockGenerator.addData(BlockGenerator.scala:172) | |
at org.apache.spark.streaming.receiver.ReceiverSupervisorImpl.pushSingle(ReceiverSupervisorImpl.scala:119) | |
at org.apache.spark.streaming.receiver.Receiver.store(Receiver.scala:119) | |
at org.apache.spark.streaming.dstream.SocketReceiver.receive(SocketInputDStream.scala:77) | |
at org.apache.spark.streaming.dstream.SocketReceiver$$anon$2.run(SocketInputDStream.scala:59) | |
16/03/30 18:48:15 WARN ReceiverSupervisorImpl: Restarting receiver with delay 2000 ms: Error receiving data | |
org.apache.spark.SparkException: Cannot add data as BlockGenerator has not been started or has been stopped | |
at org.apache.spark.streaming.receiver.BlockGenerator.addData(BlockGenerator.scala:172) | |
at org.apache.spark.streaming.receiver.ReceiverSupervisorImpl.pushSingle(ReceiverSupervisorImpl.scala:119) | |
at org.apache.spark.streaming.receiver.Receiver.store(Receiver.scala:119) | |
at org.apache.spark.streaming.dstream.SocketReceiver.receive(SocketInputDStream.scala:77) | |
at org.apache.spark.streaming.dstream.SocketReceiver$$anon$2.run(SocketInputDStream.scala:59) | |
16/03/30 18:48:15 WARN ReceiverSupervisorImpl: Receiver has been stopped | |
Exception in thread "receiver-supervisor-future-0" java.lang.Error: java.lang.InterruptedException: sleep interrupted | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1148) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Caused by: java.lang.InterruptedException: sleep interrupted | |
at java.lang.Thread.sleep(Native Method) | |
at org.apache.spark.streaming.receiver.ReceiverSupervisor$$anonfun$restartReceiver$1.apply$mcV$sp(ReceiverSupervisor.scala:195) | |
at org.apache.spark.streaming.receiver.ReceiverSupervisor$$anonfun$restartReceiver$1.apply(ReceiverSupervisor.scala:188) | |
at org.apache.spark.streaming.receiver.ReceiverSupervisor$$anonfun$restartReceiver$1.apply(ReceiverSupervisor.scala:188) | |
at scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24) | |
at scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
... 2 more | |
+----------+----------+------------------+-------------+-----+ | |
|sourceId |agentId |accountId |channel |state| | |
+----------+----------+------------------+-------------+-----+ | |
|Source_123|Agent_Log3|Account_Agent_Log3|Channel.Voice|READY| | |
|Source_123|Agent_Log1|Account_Agent_Log1|Channel.Voice|READY| | |
|Source_123|Agent_Log2|Account_Agent_Log2|Channel.Voice|READY| | |
|Source_123|Agent_Log3|Account_Agent_Log3|Channel.Voice|READY| | |
|Source_123|Agent_Log1|Account_Agent_Log1|Channel.Voice|READY| | |
|Source_123|Agent_Log2|Account_Agent_Log2|Channel.Voice|READY| | |
|Source_123|Agent_Log3|Account_Agent_Log3|Channel.Voice|READY| | |
|Source_123|Agent_Log1|Account_Agent_Log1|Channel.Voice|READY| | |
|Source_123|Agent_Log2|Account_Agent_Log2|Channel.Voice|READY| | |
|Source_123|Agent_Log3|Account_Agent_Log3|Channel.Voice|READY| | |
|Source_123|Agent_Log1|Account_Agent_Log1|Channel.Voice|READY| | |
|Source_123|Agent_Log2|Account_Agent_Log2|Channel.Voice|READY| | |
|Source_123|Agent_Log3|Account_Agent_Log3|Channel.Voice|READY| | |
|Source_123|Agent_Log1|Account_Agent_Log1|Channel.Voice|READY| | |
|Source_123|Agent_Log2|Account_Agent_Log2|Channel.Voice|READY| | |
|Source_123|Agent_Log3|Account_Agent_Log3|Channel.Voice|READY| | |
|Source_123|Agent_Log1|Account_Agent_Log1|Channel.Voice|READY| | |
|Source_123|Agent_Log2|Account_Agent_Log2|Channel.Voice|READY| | |
|Source_123|Agent_Log3|Account_Agent_Log3|Channel.Voice|READY| | |
|Source_123|Agent_Log1|Account_Agent_Log1|Channel.Voice|READY| | |
+----------+----------+------------------+-------------+-----+ | |
only showing top 20 rows | |
Exception in thread "streaming-job-executor-0" java.lang.Error: java.lang.InterruptedException | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1148) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Caused by: java.lang.InterruptedException | |
at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedNanos(AbstractQueuedSynchronizer.java:1039) | |
at java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1328) | |
at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208) | |
at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218) | |
at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) | |
at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) | |
at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) | |
at scala.concurrent.Await$.result(package.scala:107) | |
at org.apache.spark.sql.execution.joins.BroadcastHashJoin.doExecute(BroadcastHashJoin.scala:107) | |
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132) | |
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) | |
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130) | |
at org.apache.spark.sql.execution.Project.doExecute(basicOperators.scala:46) | |
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132) | |
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) | |
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130) | |
at org.apache.spark.sql.execution.ConvertToSafe.doExecute(rowFormatConverters.scala:56) | |
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132) | |
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) | |
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130) | |
at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:187) | |
at org.apache.spark.sql.execution.Limit.executeCollect(basicOperators.scala:165) | |
at org.apache.spark.sql.execution.SparkPlan.executeCollectPublic(SparkPlan.scala:174) | |
at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$execute$1$1.apply(DataFrame.scala:1499) | |
at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$execute$1$1.apply(DataFrame.scala:1499) | |
at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:56) | |
at org.apache.spark.sql.DataFrame.withNewExecutionId(DataFrame.scala:2086) | |
at org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$execute$1(DataFrame.scala:1498) | |
at org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$collect(DataFrame.scala:1505) | |
at org.apache.spark.sql.DataFrame$$anonfun$head$1.apply(DataFrame.scala:1375) | |
at org.apache.spark.sql.DataFrame$$anonfun$head$1.apply(DataFrame.scala:1374) | |
at org.apache.spark.sql.DataFrame.withCallback(DataFrame.scala:2099) | |
at org.apache.spark.sql.DataFrame.head(DataFrame.scala:1374) | |
at org.apache.spark.sql.DataFrame.take(DataFrame.scala:1456) | |
at org.elasticsearch.spark.sql.EsSparkSQL$.saveToEs(EsSparkSQL.scala:46) | |
at org.elasticsearch.spark.sql.EsSparkSQL$.saveToEs(EsSparkSQL.scala:43) | |
at org.elasticsearch.spark.sql.api.java.JavaEsSparkSQL$.saveToEs(JavaEsSparkSQL.scala:27) | |
at org.elasticsearch.spark.sql.api.java.JavaEsSparkSQL.saveToEs(JavaEsSparkSQL.scala) | |
at com.my.spark.app.JsonFileStream.lambda$main$1282d8df$1(JsonFileStream.java:102) | |
at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$3.apply(JavaDStreamLike.scala:335) | |
at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$3.apply(JavaDStreamLike.scala:335) | |
at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661) | |
at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:50) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50) | |
at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:426) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:49) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49) | |
at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49) | |
at scala.util.Try$.apply(Try.scala:161) | |
at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:224) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224) | |
at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57) | |
at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:223) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
... 2 more | |
16/03/30 18:48:17 ERROR LiveListenerBus: SparkListenerBus has already stopped! Dropping event SparkListenerTaskEnd(20,0,ResultTask,Success,org.apache.spark.scheduler.TaskInfo@5e10ae8a,org.apache.spark.executor.TaskMetrics@12f2ba2c) | |
16/03/30 18:48:17 ERROR LiveListenerBus: SparkListenerBus has already stopped! Dropping event SparkListenerTaskEnd(20,0,ResultTask,Success,org.apache.spark.scheduler.TaskInfo@6b15b244,org.apache.spark.executor.TaskMetrics@1a74ae7) | |
16/03/30 18:48:17 ERROR LiveListenerBus: SparkListenerBus has already stopped! Dropping event SparkListenerTaskStart(20,0,org.apache.spark.scheduler.TaskInfo@5a98d0a7) | |
16/03/30 18:48:17 ERROR LiveListenerBus: SparkListenerBus has already stopped! Dropping event SparkListenerStageCompleted(org.apache.spark.scheduler.StageInfo@7926472b) | |
16/03/30 18:48:17 ERROR LiveListenerBus: SparkListenerBus has already stopped! Dropping event SparkListenerJobEnd(15,1459360097733,JobFailed(org.apache.spark.SparkException: Job 15 cancelled because SparkContext was shut down)) | |
16/03/30 18:48:17 ERROR TaskSchedulerImpl: Exception in statusUpdate | |
java.util.concurrent.RejectedExecutionException: Task org.apache.spark.scheduler.TaskResultGetter$$anon$2@64dd778f rejected from java.util.concurrent.ThreadPoolExecutor@3ba27c46[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 1010] | |
at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2047) | |
at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:823) | |
at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1369) | |
at org.apache.spark.scheduler.TaskResultGetter.enqueueSuccessfulTask(TaskResultGetter.scala:49) | |
at org.apache.spark.scheduler.TaskSchedulerImpl.liftedTree2$1(TaskSchedulerImpl.scala:354) | |
at org.apache.spark.scheduler.TaskSchedulerImpl.statusUpdate(TaskSchedulerImpl.scala:331) | |
at org.apache.spark.scheduler.local.LocalEndpoint$$anonfun$receive$1.applyOrElse(LocalBackend.scala:66) | |
at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:116) | |
at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:204) | |
at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:100) | |
at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:215) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
16/03/30 18:48:17 ERROR TaskSchedulerImpl: Exception in statusUpdate | |
java.util.concurrent.RejectedExecutionException: Task org.apache.spark.scheduler.TaskResultGetter$$anon$2@3d8de772 rejected from java.util.concurrent.ThreadPoolExecutor@3ba27c46[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 1010] | |
at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2047) | |
at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:823) | |
at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1369) | |
at org.apache.spark.scheduler.TaskResultGetter.enqueueSuccessfulTask(TaskResultGetter.scala:49) | |
at org.apache.spark.scheduler.TaskSchedulerImpl.liftedTree2$1(TaskSchedulerImpl.scala:354) | |
at org.apache.spark.scheduler.TaskSchedulerImpl.statusUpdate(TaskSchedulerImpl.scala:331) | |
at org.apache.spark.scheduler.local.LocalEndpoint$$anonfun$receive$1.applyOrElse(LocalBackend.scala:66) | |
at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:116) | |
at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:204) | |
at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:100) | |
at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:215) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
16/03/30 18:48:17 ERROR TaskSchedulerImpl: Exception in statusUpdate | |
java.util.concurrent.RejectedExecutionException: Task org.apache.spark.scheduler.TaskResultGetter$$anon$2@b5fdb19 rejected from java.util.concurrent.ThreadPoolExecutor@3ba27c46[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 1010] | |
at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2047) | |
at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:823) | |
at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1369) | |
at org.apache.spark.scheduler.TaskResultGetter.enqueueSuccessfulTask(TaskResultGetter.scala:49) | |
at org.apache.spark.scheduler.TaskSchedulerImpl.liftedTree2$1(TaskSchedulerImpl.scala:354) | |
at org.apache.spark.scheduler.TaskSchedulerImpl.statusUpdate(TaskSchedulerImpl.scala:331) | |
at org.apache.spark.scheduler.local.LocalEndpoint$$anonfun$receive$1.applyOrElse(LocalBackend.scala:66) | |
at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:116) | |
at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:204) | |
at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:100) | |
at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:215) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
16/03/30 18:48:17 ERROR TaskSchedulerImpl: Exception in statusUpdate | |
java.util.concurrent.RejectedExecutionException: Task org.apache.spark.scheduler.TaskResultGetter$$anon$2@7f7fe36b rejected from java.util.concurrent.ThreadPoolExecutor@3ba27c46[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 1010] | |
at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2047) | |
at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:823) | |
at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1369) | |
at org.apache.spark.scheduler.TaskResultGetter.enqueueSuccessfulTask(TaskResultGetter.scala:49) | |
at org.apache.spark.scheduler.TaskSchedulerImpl.liftedTree2$1(TaskSchedulerImpl.scala:354) | |
at org.apache.spark.scheduler.TaskSchedulerImpl.statusUpdate(TaskSchedulerImpl.scala:331) | |
at org.apache.spark.scheduler.local.LocalEndpoint$$anonfun$receive$1.applyOrElse(LocalBackend.scala:66) | |
at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:116) | |
at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:204) | |
at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:100) | |
at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:215) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
16/03/30 18:48:17 ERROR TaskSchedulerImpl: Exception in statusUpdate | |
java.util.concurrent.RejectedExecutionException: Task org.apache.spark.scheduler.TaskResultGetter$$anon$2@787ce6a4 rejected from java.util.concurrent.ThreadPoolExecutor@3ba27c46[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 1010] | |
at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2047) | |
at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:823) | |
at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1369) | |
at org.apache.spark.scheduler.TaskResultGetter.enqueueSuccessfulTask(TaskResultGetter.scala:49) | |
at org.apache.spark.scheduler.TaskSchedulerImpl.liftedTree2$1(TaskSchedulerImpl.scala:354) | |
at org.apache.spark.scheduler.TaskSchedulerImpl.statusUpdate(TaskSchedulerImpl.scala:331) | |
at org.apache.spark.scheduler.local.LocalEndpoint$$anonfun$receive$1.applyOrElse(LocalBackend.scala:66) | |
at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:116) | |
at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:204) | |
at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:100) | |
at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:215) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
16/03/30 18:48:17 ERROR MapOutputTrackerMaster: Error communicating with MapOutputTracker | |
java.lang.NullPointerException | |
at org.apache.spark.MapOutputTracker.askTracker(MapOutputTracker.scala:110) | |
at org.apache.spark.MapOutputTracker.getStatuses(MapOutputTracker.scala:212) | |
at org.apache.spark.MapOutputTracker.getMapSizesByExecutorId(MapOutputTracker.scala:152) | |
at org.apache.spark.shuffle.BlockStoreShuffleReader.read(BlockStoreShuffleReader.scala:47) | |
at org.apache.spark.sql.execution.ShuffledRowRDD.compute(ShuffledRowRDD.scala:166) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
16/03/30 18:48:17 ERROR Executor: Exception in task 199.0 in stage 20.0 (TID 1015) | |
org.apache.spark.SparkException: Error communicating with MapOutputTracker | |
at org.apache.spark.MapOutputTracker.askTracker(MapOutputTracker.scala:114) | |
at org.apache.spark.MapOutputTracker.getStatuses(MapOutputTracker.scala:212) | |
at org.apache.spark.MapOutputTracker.getMapSizesByExecutorId(MapOutputTracker.scala:152) | |
at org.apache.spark.shuffle.BlockStoreShuffleReader.read(BlockStoreShuffleReader.scala:47) | |
at org.apache.spark.sql.execution.ShuffledRowRDD.compute(ShuffledRowRDD.scala:166) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Caused by: java.lang.NullPointerException | |
at org.apache.spark.MapOutputTracker.askTracker(MapOutputTracker.scala:110) | |
... 27 more |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment