Created
March 12, 2016 00:03
-
-
Save ramv-dailymotion/19fa292bf3f2bc1d62cc to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
16/03/10 00:44:17 INFO com.google.cloud.hadoop.io.bigquery.ShardedExportToCloudStorage: Table 'dailymotion-data:video_catalog.snapshot2016030700' to be exported has 127841354 rows and 40712201343 bytes | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:> (0 + 30) / 30] | |
[Stage 0:=> (1 + 29) / 30] | |
[Stage 0:===> (2 + 28) / 30] | |
[Stage 0:=====> (3 + 27) / 30] | |
[Stage 0:=======> (4 + 26) / 30] | |
[Stage 0:=========> (5 + 25) / 30] | |
[Stage 0:===========> (6 + 24) / 30] | |
[Stage 0:=============> (7 + 23) / 30] | |
[Stage 0:===============> (8 + 22) / 30] | |
[Stage 0:=================> (9 + 21) / 30] | |
[Stage 0:==================> (10 + 20) / 30] | |
[Stage 0:====================> (11 + 19) / 30] | |
[Stage 0:======================> (12 + 18) / 30] | |
[Stage 0:========================> (13 + 17) / 30] | |
[Stage 0:==========================> (14 + 16) / 30] | |
[Stage 0:============================> (15 + 15) / 30] | |
[Stage 0:=============================> (16 + 14) / 30] | |
[Stage 0:===============================> (17 + 13) / 30] | |
[Stage 0:=================================> (18 + 12) / 30] | |
[Stage 0:===================================> (19 + 11) / 30] | |
[Stage 0:=====================================> (20 + 10) / 30] | |
[Stage 0:=======================================> (21 + 9) / 30] | |
[Stage 0:=========================================> (22 + 8) / 30] | |
[Stage 0:===========================================> (23 + 7) / 30] | |
[Stage 0:=============================================> (24 + 6) / 30] | |
[Stage 0:===============================================> (25 + 5) / 30] | |
[Stage 0:=================================================> (26 + 4) / 30] | |
[Stage 0:=================================================> (26 + 4) / 30] | |
[Stage 0:===================================================> (27 + 3) / 30] | |
[Stage 0:===================================================> (27 + 3) / 30] | |
[Stage 0:=====================================================> (28 + 2) / 30] | |
[Stage 0:=====================================================> (28 + 2) / 30] | |
[Stage 0:=======================================================> (29 + 1) / 30] | |
[Stage 0:=======================================================> (29 + 1) / 30] | |
[Stage 0:=======================================================> (29 + 1) / 30] | |
[Stage 1:> (0 + 0) / 1000] | |
[Stage 1:> (0 + 8) / 1000]16/03/10 01:08:14 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 6.0 in stage 1.0 (TID 36, rc-spark-stage-w-1.c.dailymotion-data.internal): com.esotericsoftware.kryo.KryoException: java.io.IOException: failed to uncompress the chunk: PARSING_ERROR(2) | |
Serialization trace: | |
value (com.google.gson.JsonPrimitive) | |
members (com.google.gson.JsonObject) | |
at com.esotericsoftware.kryo.io.Input.fill(Input.java:142) | |
at com.esotericsoftware.kryo.io.Input.require(Input.java:169) | |
at com.esotericsoftware.kryo.io.Input.readUtf8_slow(Input.java:524) | |
at com.esotericsoftware.kryo.io.Input.readUtf8(Input.java:517) | |
at com.esotericsoftware.kryo.io.Input.readString(Input.java:447) | |
at com.esotericsoftware.kryo.serializers.DefaultSerializers$StringSerializer.read(DefaultSerializers.java:157) | |
at com.esotericsoftware.kryo.serializers.DefaultSerializers$StringSerializer.read(DefaultSerializers.java:146) | |
at com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:648) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer$ObjectField.read(FieldSerializer.java:605) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:221) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at com.esotericsoftware.kryo.serializers.MapSerializer.read(MapSerializer.java:134) | |
at com.esotericsoftware.kryo.serializers.MapSerializer.read(MapSerializer.java:17) | |
at com.esotericsoftware.kryo.Kryo.readObjectOrNull(Kryo.java:699) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer$ObjectField.read(FieldSerializer.java:611) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:221) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:42) | |
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:33) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at org.apache.spark.serializer.KryoDeserializationStream.readObject(KryoSerializer.scala:228) | |
at org.apache.spark.serializer.DeserializationStream.readValue(Serializer.scala:171) | |
at org.apache.spark.serializer.DeserializationStream$$anon$2.getNext(Serializer.scala:201) | |
at org.apache.spark.serializer.DeserializationStream$$anon$2.getNext(Serializer.scala:198) | |
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) | |
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:32) | |
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) | |
at org.apache.spark.storage.MemoryStore.unrollSafely(MemoryStore.scala:282) | |
at org.apache.spark.CacheManager.putInBlockManager(CacheManager.scala:171) | |
at org.apache.spark.CacheManager.getOrCompute(CacheManager.scala:78) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:268) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Caused by: java.io.IOException: failed to uncompress the chunk: PARSING_ERROR(2) | |
at org.xerial.snappy.SnappyInputStream.hasNextChunk(SnappyInputStream.java:361) | |
at org.xerial.snappy.SnappyInputStream.rawRead(SnappyInputStream.java:158) | |
at org.xerial.snappy.SnappyInputStream.read(SnappyInputStream.java:142) | |
at com.esotericsoftware.kryo.io.Input.fill(Input.java:140) | |
... 43 more | |
16/03/10 01:08:14 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 7.0 in stage 1.0 (TID 37, rc-spark-stage-w-1.c.dailymotion-data.internal): com.esotericsoftware.kryo.KryoException: java.io.IOException: failed to uncompress the chunk: PARSING_ERROR(2) | |
Serialization trace: | |
members (com.google.gson.JsonObject) | |
at com.esotericsoftware.kryo.io.Input.fill(Input.java:142) | |
at com.esotericsoftware.kryo.io.Input.require(Input.java:169) | |
at com.esotericsoftware.kryo.io.Input.readInt(Input.java:337) | |
at com.esotericsoftware.kryo.Kryo.readReferenceOrNull(Kryo.java:747) | |
at com.esotericsoftware.kryo.Kryo.readObjectOrNull(Kryo.java:697) | |
at com.esotericsoftware.kryo.serializers.MapSerializer.read(MapSerializer.java:122) | |
at com.esotericsoftware.kryo.serializers.MapSerializer.read(MapSerializer.java:17) | |
at com.esotericsoftware.kryo.Kryo.readObjectOrNull(Kryo.java:699) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer$ObjectField.read(FieldSerializer.java:611) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:221) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:42) | |
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:33) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at org.apache.spark.serializer.KryoDeserializationStream.readObject(KryoSerializer.scala:228) | |
at org.apache.spark.serializer.DeserializationStream.readValue(Serializer.scala:171) | |
at org.apache.spark.serializer.DeserializationStream$$anon$2.getNext(Serializer.scala:201) | |
at org.apache.spark.serializer.DeserializationStream$$anon$2.getNext(Serializer.scala:198) | |
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) | |
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:32) | |
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) | |
at org.apache.spark.storage.MemoryStore.unrollSafely(MemoryStore.scala:282) | |
at org.apache.spark.CacheManager.putInBlockManager(CacheManager.scala:171) | |
at org.apache.spark.CacheManager.getOrCompute(CacheManager.scala:78) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:268) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Caused by: java.io.IOException: failed to uncompress the chunk: PARSING_ERROR(2) | |
at org.xerial.snappy.SnappyInputStream.hasNextChunk(SnappyInputStream.java:361) | |
at org.xerial.snappy.SnappyInputStream.rawRead(SnappyInputStream.java:158) | |
at org.xerial.snappy.SnappyInputStream.read(SnappyInputStream.java:142) | |
at com.esotericsoftware.kryo.io.Input.fill(Input.java:140) | |
... 37 more | |
16/03/10 01:08:14 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 2.0 in stage 1.0 (TID 32, rc-spark-stage-w-1.c.dailymotion-data.internal): com.esotericsoftware.kryo.KryoException: java.io.IOException: failed to uncompress the chunk: PARSING_ERROR(2) | |
at com.esotericsoftware.kryo.io.Input.fill(Input.java:142) | |
at com.esotericsoftware.kryo.io.Input.require(Input.java:169) | |
at com.esotericsoftware.kryo.io.Input.readAscii_slow(Input.java:580) | |
at com.esotericsoftware.kryo.io.Input.readAscii(Input.java:558) | |
at com.esotericsoftware.kryo.io.Input.readString(Input.java:436) | |
at com.esotericsoftware.kryo.util.DefaultClassResolver.readName(DefaultClassResolver.java:132) | |
at com.esotericsoftware.kryo.util.DefaultClassResolver.readClass(DefaultClassResolver.java:115) | |
at com.esotericsoftware.kryo.Kryo.readClass(Kryo.java:610) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:721) | |
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:41) | |
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:33) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at org.apache.spark.serializer.KryoDeserializationStream.readObject(KryoSerializer.scala:228) | |
at org.apache.spark.serializer.DeserializationStream.readValue(Serializer.scala:171) | |
at org.apache.spark.serializer.DeserializationStream$$anon$2.getNext(Serializer.scala:201) | |
at org.apache.spark.serializer.DeserializationStream$$anon$2.getNext(Serializer.scala:198) | |
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) | |
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:32) | |
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) | |
at org.apache.spark.storage.MemoryStore.unrollSafely(MemoryStore.scala:282) | |
at org.apache.spark.CacheManager.putInBlockManager(CacheManager.scala:171) | |
at org.apache.spark.CacheManager.getOrCompute(CacheManager.scala:78) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:268) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Caused by: java.io.IOException: failed to uncompress the chunk: PARSING_ERROR(2) | |
at org.xerial.snappy.SnappyInputStream.hasNextChunk(SnappyInputStream.java:361) | |
at org.xerial.snappy.SnappyInputStream.rawRead(SnappyInputStream.java:158) | |
at org.xerial.snappy.SnappyInputStream.read(SnappyInputStream.java:142) | |
at com.esotericsoftware.kryo.io.Input.fill(Input.java:140) | |
... 35 more | |
[Stage 1:> (0 + 16) / 1000] | |
[Stage 1:> (0 + 24) / 1000] | |
[Stage 1:> (0 + 32) / 1000]16/03/10 01:08:18 ERROR org.apache.spark.scheduler.TaskSetManager: Task 6 in stage 1.0 failed 4 times; aborting job | |
Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 6 in stage 1.0 failed 4 times, most recent failure: Lost task 6.3 in stage 1.0 (TID 63, rc-spark-stage-w-1.c.dailymotion-data.internal): com.esotericsoftware.kryo.KryoException: java.io.IOException: failed to uncompress the chunk: PARSING_ERROR(2) | |
Serialization trace: | |
value (com.google.gson.JsonPrimitive) | |
members (com.google.gson.JsonObject) | |
at com.esotericsoftware.kryo.io.Input.fill(Input.java:142) | |
at com.esotericsoftware.kryo.io.Input.require(Input.java:169) | |
at com.esotericsoftware.kryo.io.Input.readUtf8_slow(Input.java:524) | |
at com.esotericsoftware.kryo.io.Input.readUtf8(Input.java:517) | |
at com.esotericsoftware.kryo.io.Input.readString(Input.java:447) | |
at com.esotericsoftware.kryo.serializers.DefaultSerializers$StringSerializer.read(DefaultSerializers.java:157) | |
at com.esotericsoftware.kryo.serializers.DefaultSerializers$StringSerializer.read(DefaultSerializers.java:146) | |
at com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:648) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer$ObjectField.read(FieldSerializer.java:605) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:221) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at com.esotericsoftware.kryo.serializers.MapSerializer.read(MapSerializer.java:134) | |
at com.esotericsoftware.kryo.serializers.MapSerializer.read(MapSerializer.java:17) | |
at com.esotericsoftware.kryo.Kryo.readObjectOrNull(Kryo.java:699) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer$ObjectField.read(FieldSerializer.java:611) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:221) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:42) | |
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:33) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at org.apache.spark.serializer.KryoDeserializationStream.readObject(KryoSerializer.scala:228) | |
at org.apache.spark.serializer.DeserializationStream.readValue(Serializer.scala:171) | |
at org.apache.spark.serializer.DeserializationStream$$anon$2.getNext(Serializer.scala:201) | |
at org.apache.spark.serializer.DeserializationStream$$anon$2.getNext(Serializer.scala:198) | |
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) | |
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:32) | |
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) | |
at org.apache.spark.storage.MemoryStore.unrollSafely(MemoryStore.scala:282) | |
at org.apache.spark.CacheManager.putInBlockManager(CacheManager.scala:171) | |
at org.apache.spark.CacheManager.getOrCompute(CacheManager.scala:78) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:268) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Caused by: java.io.IOException: failed to uncompress the chunk: PARSING_ERROR(2) | |
at org.xerial.snappy.SnappyInputStream.hasNextChunk(SnappyInputStream.java:361) | |
at org.xerial.snappy.SnappyInputStream.rawRead(SnappyInputStream.java:158) | |
at org.xerial.snappy.SnappyInputStream.read(SnappyInputStream.java:142) | |
at com.esotericsoftware.kryo.io.Input.fill(Input.java:140) | |
... 43 more | |
Driver stacktrace: | |
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418) | |
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) | |
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) | |
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) | |
at scala.Option.foreach(Option.scala:236) | |
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599) | |
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588) | |
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) | |
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620) | |
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832) | |
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845) | |
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1922) | |
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1213) | |
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1.apply(PairRDDFunctions.scala:1156) | |
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1.apply(PairRDDFunctions.scala:1156) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) | |
at org.apache.spark.rdd.RDD.withScope(RDD.scala:316) | |
at org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopDataset(PairRDDFunctions.scala:1156) | |
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$4.apply$mcV$sp(PairRDDFunctions.scala:1060) | |
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$4.apply(PairRDDFunctions.scala:1026) | |
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$4.apply(PairRDDFunctions.scala:1026) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) | |
at org.apache.spark.rdd.RDD.withScope(RDD.scala:316) | |
at org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopFile(PairRDDFunctions.scala:1026) | |
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$1.apply$mcV$sp(PairRDDFunctions.scala:952) | |
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$1.apply(PairRDDFunctions.scala:952) | |
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$1.apply(PairRDDFunctions.scala:952) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) | |
at org.apache.spark.rdd.RDD.withScope(RDD.scala:316) | |
at org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopFile(PairRDDFunctions.scala:951) | |
at org.apache.spark.rdd.RDD$$anonfun$saveAsTextFile$1.apply$mcV$sp(RDD.scala:1443) | |
at org.apache.spark.rdd.RDD$$anonfun$saveAsTextFile$1.apply(RDD.scala:1422) | |
at org.apache.spark.rdd.RDD$$anonfun$saveAsTextFile$1.apply(RDD.scala:1422) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) | |
at org.apache.spark.rdd.RDD.withScope(RDD.scala:316) | |
at org.apache.spark.rdd.RDD.saveAsTextFile(RDD.scala:1422) | |
at org.apache.spark.api.java.JavaRDDLike$class.saveAsTextFile(JavaRDDLike.scala:507) | |
at org.apache.spark.api.java.AbstractJavaRDDLike.saveAsTextFile(JavaRDDLike.scala:46) | |
at com.dailymotion.rc.recommender.VideoCatalogStoreGenerator.main(VideoCatalogStoreGenerator.java:208) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) | |
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) | |
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) | |
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) | |
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) | |
Caused by: com.esotericsoftware.kryo.KryoException: java.io.IOException: failed to uncompress the chunk: PARSING_ERROR(2) | |
Serialization trace: | |
value (com.google.gson.JsonPrimitive) | |
members (com.google.gson.JsonObject) | |
at com.esotericsoftware.kryo.io.Input.fill(Input.java:142) | |
at com.esotericsoftware.kryo.io.Input.require(Input.java:169) | |
at com.esotericsoftware.kryo.io.Input.readUtf8_slow(Input.java:524) | |
at com.esotericsoftware.kryo.io.Input.readUtf8(Input.java:517) | |
at com.esotericsoftware.kryo.io.Input.readString(Input.java:447) | |
at com.esotericsoftware.kryo.serializers.DefaultSerializers$StringSerializer.read(DefaultSerializers.java:157) | |
at com.esotericsoftware.kryo.serializers.DefaultSerializers$StringSerializer.read(DefaultSerializers.java:146) | |
at com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:648) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer$ObjectField.read(FieldSerializer.java:605) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:221) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at com.esotericsoftware.kryo.serializers.MapSerializer.read(MapSerializer.java:134) | |
at com.esotericsoftware.kryo.serializers.MapSerializer.read(MapSerializer.java:17) | |
at com.esotericsoftware.kryo.Kryo.readObjectOrNull(Kryo.java:699) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer$ObjectField.read(FieldSerializer.java:611) | |
at com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:221) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:42) | |
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:33) | |
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729) | |
at org.apache.spark.serializer.KryoDeserializationStream.readObject(KryoSerializer.scala:228) | |
at org.apache.spark.serializer.DeserializationStream.readValue(Serializer.scala:171) | |
at org.apache.spark.serializer.DeserializationStream$$anon$2.getNext(Serializer.scala:201) | |
at org.apache.spark.serializer.DeserializationStream$$anon$2.getNext(Serializer.scala:198) | |
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) | |
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:32) | |
at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) | |
at org.apache.spark.storage.MemoryStore.unrollSafely(MemoryStore.scala:282) | |
at org.apache.spark.CacheManager.putInBlockManager(CacheManager.scala:171) | |
at org.apache.spark.CacheManager.getOrCompute(CacheManager.scala:78) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:268) | |
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) | |
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) | |
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) | |
at org.apache.spark.scheduler.Task.run(Task.scala:89) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Caused by: java.io.IOException: failed to uncompress the chunk: PARSING_ERROR(2) | |
at org.xerial.snappy.SnappyInputStream.hasNextChunk(SnappyInputStream.java:361) | |
at org.xerial.snappy.SnappyInputStream.rawRead(SnappyInputStream.java:158) | |
at org.xerial.snappy.SnappyInputStream.read(SnappyInputStream.java:142) | |
at com.esotericsoftware.kryo.io.Input.fill(Input.java:140) | |
... 43 more | |
16/03/10 01:08:18 WARN org.apache.spark.ExecutorAllocationManager: No stages are running, but numRunningTasks != 0 | |
16/03/10 01:08:18 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 1.3 in stage 1.0 (TID 66, rc-spark-stage-w-1.c.dailymotion-data.internal): TaskKilled (killed intentionally) | |
16/03/10 01:08:18 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 7.3 in stage 1.0 (TID 69, rc-spark-stage-w-1.c.dailymotion-data.internal): TaskKilled (killed intentionally) | |
16/03/10 01:08:18 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 0.2 in stage 1.0 (TID 65, rc-spark-stage-w-1.c.dailymotion-data.internal): TaskKilled (killed intentionally) | |
16/03/10 01:08:18 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 16.1 in stage 1.0 (TID 87, rc-spark-stage-w-1.c.dailymotion-data.internal): TaskKilled (killed intentionally) | |
16/03/10 01:08:18 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 32.0 in stage 1.0 (TID 86, rc-spark-stage-w-1.c.dailymotion-data.internal): TaskKilled (killed intentionally) | |
16/03/10 01:08:18 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 4.3 in stage 1.0 (TID 68, rc-spark-stage-w-1.c.dailymotion-data.internal): TaskKilled (killed intentionally) | |
16/03/10 01:08:18 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 5.3 in stage 1.0 (TID 64, rc-spark-stage-w-1.c.dailymotion-data.internal): TaskKilled (killed intentionally) | |
16/03/10 01:08:18 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 8.2 in stage 1.0 (TID 67, rc-spark-stage-w-1.c.dailymotion-data.internal): TaskKilled (killed intentionally) | |
16/03/10 01:08:18 INFO akka.remote.RemoteActorRefProvider$RemotingTerminator: Shutting down remote daemon. | |
16/03/10 01:08:18 INFO akka.remote.RemoteActorRefProvider$RemotingTerminator: Remote daemon shut down; proceeding with flushing remote transports. | |
16/03/10 01:08:19 INFO akka.remote.RemoteActorRefProvider$RemotingTerminator: Remoting shut down. | |
Job output is complete |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment