Created
September 13, 2015 18:27
-
-
Save natewave/830515f2322d9523892b to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
15/09/13 20:26:54 INFO KafkaRDD: Computing topic defectRequests, partition 0 offsets -1 -> 290 | |
15/09/13 20:26:54 ERROR TaskContextImpl: Error in TaskCompletionListener | |
java.lang.NullPointerException | |
at org.apache.spark.streaming.kafka.KafkaRDD$KafkaRDDIterator.close(KafkaRDD.scala:157) | |
at org.apache.spark.util.NextIterator.closeIfNeeded(NextIterator.scala:63) | |
at org.apache.spark.streaming.kafka.KafkaRDD$KafkaRDDIterator$$anonfun$1.apply(KafkaRDD.scala:100) | |
at org.apache.spark.streaming.kafka.KafkaRDD$KafkaRDDIterator$$anonfun$1.apply(KafkaRDD.scala:100) | |
at org.apache.spark.TaskContextImpl$$anon$1.onTaskCompletion(TaskContextImpl.scala:56) | |
at org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:75) | |
at org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:73) | |
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) | |
at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:73) | |
at org.apache.spark.scheduler.Task.run(Task.scala:72) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
15/09/13 20:26:54 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID 0) | |
org.apache.spark.util.TaskCompletionListenerException | |
at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:83) | |
at org.apache.spark.scheduler.Task.run(Task.scala:72) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
15/09/13 20:26:54 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.spark.util.TaskCompletionListenerException | |
at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:83) | |
at org.apache.spark.scheduler.Task.run(Task.scala:72) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
15/09/13 20:26:54 ERROR TaskSetManager: Task 0 in stage 0.0 failed 1 times; aborting job | |
15/09/13 20:26:54 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks have all completed, from pool | |
15/09/13 20:26:54 INFO TaskSchedulerImpl: Cancelling stage 0 | |
15/09/13 20:26:54 INFO DAGScheduler: ResultStage 0 (print at KafkaCommons.scala:127) failed in 0.582 s | |
15/09/13 20:26:54 INFO DAGScheduler: Job 0 failed: print at KafkaCommons.scala:127, took 0.729425 s | |
15/09/13 20:26:54 ERROR JobScheduler: Error running job streaming job 1442168814000 ms.0 | |
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.spark.util.TaskCompletionListenerException | |
at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:83) | |
at org.apache.spark.scheduler.Task.run(Task.scala:72) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Driver stacktrace: | |
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1273) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1264) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1263) | |
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) | |
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1263) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:730) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:730) | |
at scala.Option.foreach(Option.scala:236) | |
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:730) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1457) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1418) | |
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) | |
Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.spark.util.TaskCompletionListenerException | |
at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:83) | |
at org.apache.spark.scheduler.Task.run(Task.scala:72) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Driver stacktrace: | |
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1273) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1264) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1263) | |
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) | |
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1263) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:730) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:730) | |
at scala.Option.foreach(Option.scala:236) | |
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:730) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1457) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1418) | |
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) |
The topic and partition exist in my case, -1 is normal since there is no consumer group that has offset metadata associated with this topic-partition...but once I start consuming it will change to reflect my progress.
I fear it has to do with deserialization issues, as I am using confluent.io schema registry and a custom kafka decoder. Probably a mismatch between the producer and consumer
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
you have this here