Created
March 21, 2018 14:18
-
-
Save greebie/a9cb0b3186ef7d574916a00db5c89898 to your computer and use it in GitHub Desktop.
Java Heap Error.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
18/03/21 10:06:03 INFO TaskSetManager: Finished task 9.0 in stage 0.0 (TID 9) in 92758 ms on localhost (executor driver) (26/33) | |
[Stage 0:============================================> (26 + 7) / 33]18/03/21 10:06:10 ERROR Executor: Exception in task 17.0 in stage 0.0 (TID 17) | |
java.lang.OutOfMemoryError: Java heap space | |
at java.util.Arrays.copyOf(Arrays.java:3332) | |
at java.lang.StringCoding.safeTrim(StringCoding.java:89) | |
at java.lang.StringCoding.access$100(StringCoding.java:50) | |
at java.lang.StringCoding$StringDecoder.decode(StringCoding.java:154) | |
at java.lang.StringCoding.decode(StringCoding.java:193) | |
at java.lang.StringCoding.decode(StringCoding.java:254) | |
at java.lang.String.<init>(String.java:534) | |
at java.lang.String.<init>(String.java:554) | |
at io.archivesunleashed.spark.archive.io.ArchiveRecord.<init>(ArchiveRecord.scala:68) | |
at io.archivesunleashed.spark.matchbox.RecordLoader$$anonfun$2.apply(RecordLoader.scala:37) | |
at io.archivesunleashed.spark.matchbox.RecordLoader$$anonfun$2.apply(RecordLoader.scala:37) | |
at scala.collection.Iterator$$anon$11.next(Iterator.scala:409) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:462) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:461) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408) | |
at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:191) | |
at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:63) | |
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) | |
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) | |
at org.apache.spark.scheduler.Task.run(Task.scala:108) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
18/03/21 10:06:10 ERROR SparkUncaughtExceptionHandler: Uncaught exception in thread Thread[Executor task launch worker for task 17,5,main] | |
java.lang.OutOfMemoryError: Java heap space | |
at java.util.Arrays.copyOf(Arrays.java:3332) | |
at java.lang.StringCoding.safeTrim(StringCoding.java:89) | |
at java.lang.StringCoding.access$100(StringCoding.java:50) | |
at java.lang.StringCoding$StringDecoder.decode(StringCoding.java:154) | |
at java.lang.StringCoding.decode(StringCoding.java:193) | |
at java.lang.StringCoding.decode(StringCoding.java:254) | |
at java.lang.String.<init>(String.java:534) | |
at java.lang.String.<init>(String.java:554) | |
at io.archivesunleashed.spark.archive.io.ArchiveRecord.<init>(ArchiveRecord.scala:68) | |
at io.archivesunleashed.spark.matchbox.RecordLoader$$anonfun$2.apply(RecordLoader.scala:37) | |
at io.archivesunleashed.spark.matchbox.RecordLoader$$anonfun$2.apply(RecordLoader.scala:37) | |
at scala.collection.Iterator$$anon$11.next(Iterator.scala:409) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:462) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:461) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408) | |
at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:191) | |
at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:63) | |
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) | |
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) | |
at org.apache.spark.scheduler.Task.run(Task.scala:108) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
18/03/21 10:06:10 WARN TaskSetManager: Lost task 17.0 in stage 0.0 (TID 17, localhost, executor driver): java.lang.OutOfMemoryError: Java heap space | |
at java.util.Arrays.copyOf(Arrays.java:3332) | |
at java.lang.StringCoding.safeTrim(StringCoding.java:89) | |
at java.lang.StringCoding.access$100(StringCoding.java:50) | |
at java.lang.StringCoding$StringDecoder.decode(StringCoding.java:154) | |
at java.lang.StringCoding.decode(StringCoding.java:193) | |
at java.lang.StringCoding.decode(StringCoding.java:254) | |
at java.lang.String.<init>(String.java:534) | |
at java.lang.String.<init>(String.java:554) | |
at io.archivesunleashed.spark.archive.io.ArchiveRecord.<init>(ArchiveRecord.scala:68) | |
at io.archivesunleashed.spark.matchbox.RecordLoader$$anonfun$2.apply(RecordLoader.scala:37) | |
at io.archivesunleashed.spark.matchbox.RecordLoader$$anonfun$2.apply(RecordLoader.scala:37) | |
at scala.collection.Iterator$$anon$11.next(Iterator.scala:409) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:462) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:461) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408) | |
at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:191) | |
at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:63) | |
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) | |
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) | |
at org.apache.spark.scheduler.Task.run(Task.scala:108) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
18/03/21 10:06:10 ERROR TaskSetManager: Task 17 in stage 0.0 failed 1 times; aborting job | |
[Stage 0:============================================> (26 + 6) / 33]18/03/21 10:06:10 INFO SparkContext: Invoking stop() from shutdown hook | |
18/03/21 10:06:10 INFO TaskSchedulerImpl: Cancelling stage 0 | |
18/03/21 10:06:10 INFO TaskSchedulerImpl: Stage 0 was cancelled | |
18/03/21 10:06:10 INFO Executor: Executor is trying to kill task 16.0 in stage 0.0 (TID 16), reason: stage cancelled | |
18/03/21 10:06:10 INFO Executor: Executor is trying to kill task 31.0 in stage 0.0 (TID 31), reason: stage cancelled | |
18/03/21 10:06:10 INFO Executor: Executor is trying to kill task 10.0 in stage 0.0 (TID 10), reason: stage cancelled | |
18/03/21 10:06:10 INFO Executor: Executor is trying to kill task 28.0 in stage 0.0 (TID 28), reason: stage cancelled | |
18/03/21 10:06:10 INFO Executor: Executor is trying to kill task 14.0 in stage 0.0 (TID 14), reason: stage cancelled | |
18/03/21 10:06:10 INFO Executor: Executor is trying to kill task 8.0 in stage 0.0 (TID 8), reason: stage cancelled | |
18/03/21 10:06:10 INFO DAGScheduler: ShuffleMapStage 0 (map at RecordRDD.scala:38) failed in 99.809 s due to Job aborted due to stage failure: Task 17 in stage 0.0 failed 1 times, most recent failure: Lost task 17.0 in stage 0.0 (TID 17, localhost, executor driver): java.lang.OutOfMemoryError: Java heap space | |
at java.util.Arrays.copyOf(Arrays.java:3332) | |
at java.lang.StringCoding.safeTrim(StringCoding.java:89) | |
at java.lang.StringCoding.access$100(StringCoding.java:50) | |
at java.lang.StringCoding$StringDecoder.decode(StringCoding.java:154) | |
at java.lang.StringCoding.decode(StringCoding.java:193) | |
at java.lang.StringCoding.decode(StringCoding.java:254) | |
at java.lang.String.<init>(String.java:534) | |
at java.lang.String.<init>(String.java:554) | |
at io.archivesunleashed.spark.archive.io.ArchiveRecord.<init>(ArchiveRecord.scala:68) | |
at io.archivesunleashed.spark.matchbox.RecordLoader$$anonfun$2.apply(RecordLoader.scala:37) | |
at io.archivesunleashed.spark.matchbox.RecordLoader$$anonfun$2.apply(RecordLoader.scala:37) | |
at scala.collection.Iterator$$anon$11.next(Iterator.scala:409) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:462) | |
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:461) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408) | |
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408) | |
at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:191) | |
at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:63) | |
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) | |
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) | |
at org.apache.spark.scheduler.Task.run(Task.scala:108) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
Driver stacktrace: | |
18/03/21 10:06:10 INFO DAGScheduler: Job 0 failed: sortBy at RecordRDD.scala:40, took 100.157088 s | |
18/03/21 10:06:10 INFO Executor: Executor killed task 16.0 in stage 0.0 (TID 16), reason: stage cancelled | |
18/03/21 10:06:10 INFO SparkUI: Stopped Spark web UI at http://192.168.2.12:4040 | |
18/03/21 10:06:10 WARN TaskSetManager: Lost task 16.0 in stage 0.0 (TID 16, localhost, executor driver): TaskKilled (stage cancelled) | |
18/03/21 10:06:10 INFO Executor: Executor killed task 31.0 in stage 0.0 (TID 31), reason: stage cancelled | |
18/03/21 10:06:10 WARN TaskSetManager: Lost task 31.0 in stage 0.0 (TID 31, localhost, executor driver): TaskKilled (stage cancelled) | |
18/03/21 10:06:10 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped! | |
18/03/21 10:06:11 INFO MemoryStore: MemoryStore cleared | |
18/03/21 10:06:11 INFO BlockManager: BlockManager stopped | |
18/03/21 10:06:11 INFO BlockManagerMaster: BlockManagerMaster stopped | |
18/03/21 10:06:11 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped! | |
18/03/21 10:06:11 INFO SparkContext: Successfully stopped SparkContext | |
18/03/21 10:06:11 INFO ShutdownHookManager: Shutdown hook called | |
18/03/21 10:06:11 INFO ShutdownHookManager: Deleting directory /private/var/folders/6g/r87hcdl560j9dxxm0_jb0rpr0000gn/T/spark-499fb32c-b7e7-4e7c-a422-549babc0ddd9/repl-edefbd5a-a47d-45f7-9580-5beb646964f9 | |
18/03/21 10:06:11 ERROR TaskContextImpl: Error in TaskCompletionListener | |
java.lang.IllegalStateException: Block broadcast_0 not found | |
at org.apache.spark.storage.BlockInfoManager$$anonfun$2.apply(BlockInfoManager.scala:293) | |
at org.apache.spark.storage.BlockInfoManager$$anonfun$2.apply(BlockInfoManager.scala:293) | |
at scala.Option.getOrElse(Option.scala:121) | |
at org.apache.spark.storage.BlockInfoManager.unlock(BlockInfoManager.scala:292) | |
at org.apache.spark.storage.BlockManager.releaseLock(BlockManager.scala:720) | |
at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$org$apache$spark$broadcast$TorrentBroadcast$$releaseLock$1.apply(TorrentBroadcast.scala:250) | |
at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$org$apache$spark$broadcast$TorrentBroadcast$$releaseLock$1.apply(TorrentBroadcast.scala:250) | |
at org.apache.spark.TaskContext$$anon$1.onTaskCompletion(TaskContext.scala:128) | |
at org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:117) | |
at org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:117) | |
at org.apache.spark.TaskContextImpl$$anonfun$invokeListeners$1.apply(TaskContextImpl.scala:130) | |
at org.apache.spark.TaskContextImpl$$anonfun$invokeListeners$1.apply(TaskContextImpl.scala:128) | |
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) | |
at org.apache.spark.TaskContextImpl.invokeListeners(TaskContextImpl.scala:128) | |
at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:116) | |
at org.apache.spark.scheduler.Task.run(Task.scala:118) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
18/03/21 10:06:11 ERROR TaskContextImpl: Error in TaskCompletionListener | |
java.lang.IllegalStateException: Block broadcast_1 not found | |
at org.apache.spark.storage.BlockInfoManager$$anonfun$2.apply(BlockInfoManager.scala:293) | |
at org.apache.spark.storage.BlockInfoManager$$anonfun$2.apply(BlockInfoManager.scala:293) | |
at scala.Option.getOrElse(Option.scala:121) | |
at org.apache.spark.storage.BlockInfoManager.unlock(BlockInfoManager.scala:292) | |
at org.apache.spark.storage.BlockManager.releaseLock(BlockManager.scala:720) | |
at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$org$apache$spark$broadcast$TorrentBroadcast$$releaseLock$1.apply(TorrentBroadcast.scala:250) | |
at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$org$apache$spark$broadcast$TorrentBroadcast$$releaseLock$1.apply(TorrentBroadcast.scala:250) | |
at org.apache.spark.TaskContext$$anon$1.onTaskCompletion(TaskContext.scala:128) | |
at org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:117) | |
at org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:117) | |
at org.apache.spark.TaskContextImpl$$anonfun$invokeListeners$1.apply(TaskContextImpl.scala:130) | |
at org.apache.spark.TaskContextImpl$$anonfun$invokeListeners$1.apply(TaskContextImpl.scala:128) | |
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) | |
at org.apache.spark.TaskContextImpl.invokeListeners(TaskContextImpl.scala:128) | |
at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:116) | |
at org.apache.spark.scheduler.Task.run(Task.scala:118) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
18/03/21 10:06:11 ERROR TaskContextImpl: Error in TaskCompletionListener | |
java.lang.IllegalStateException: Block broadcast_0 not found | |
at org.apache.spark.storage.BlockInfoManager$$anonfun$2.apply(BlockInfoManager.scala:293) | |
at org.apache.spark.storage.BlockInfoManager$$anonfun$2.apply(BlockInfoManager.scala:293) | |
at scala.Option.getOrElse(Option.scala:121) | |
at org.apache.spark.storage.BlockInfoManager.unlock(BlockInfoManager.scala:292) | |
at org.apache.spark.storage.BlockManager.releaseLock(BlockManager.scala:720) | |
at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$org$apache$spark$broadcast$TorrentBroadcast$$releaseLock$1.apply(TorrentBroadcast.scala:250) | |
at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$org$apache$spark$broadcast$TorrentBroadcast$$releaseLock$1.apply(TorrentBroadcast.scala:250) | |
at org.apache.spark.TaskContext$$anon$1.onTaskCompletion(TaskContext.scala:128) | |
at org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:117) | |
at org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:117) | |
at org.apache.spark.TaskContextImpl$$anonfun$invokeListeners$1.apply(TaskContextImpl.scala:130) | |
at org.apache.spark.TaskContextImpl$$anonfun$invokeListeners$1.apply(TaskContextImpl.scala:128) | |
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) | |
at org.apache.spark.TaskContextImpl.invokeListeners(TaskContextImpl.scala:128) | |
at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:116) | |
at org.apache.spark.scheduler.Task.run(Task.scala:118) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
18/03/21 10:06:11 ERROR TaskContextImpl: Error in TaskCompletionListener | |
java.lang.IllegalStateException: Block broadcast_1 not found | |
at org.apache.spark.storage.BlockInfoManager$$anonfun$2.apply(BlockInfoManager.scala:293) | |
at org.apache.spark.storage.BlockInfoManager$$anonfun$2.apply(BlockInfoManager.scala:293) | |
at scala.Option.getOrElse(Option.scala:121) | |
at org.apache.spark.storage.BlockInfoManager.unlock(BlockInfoManager.scala:292) | |
at org.apache.spark.storage.BlockManager.releaseLock(BlockManager.scala:720) | |
at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$org$apache$spark$broadcast$TorrentBroadcast$$releaseLock$1.apply(TorrentBroadcast.scala:250) | |
at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$org$apache$spark$broadcast$TorrentBroadcast$$releaseLock$1.apply(TorrentBroadcast.scala:250) | |
at org.apache.spark.TaskContext$$anon$1.onTaskCompletion(TaskContext.scala:128) | |
at org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:117) | |
at org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:117) | |
at org.apache.spark.TaskContextImpl$$anonfun$invokeListeners$1.apply(TaskContextImpl.scala:130) | |
at org.apache.spark.TaskContextImpl$$anonfun$invokeListeners$1.apply(TaskContextImpl.scala:128) | |
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) | |
at org.apache.spark.TaskContextImpl.invokeListeners(TaskContextImpl.scala:128) | |
at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:116) | |
at org.apache.spark.scheduler.Task.run(Task.scala:118) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
18/03/21 10:06:11 ERROR Utils: Uncaught exception in thread Executor task launch worker for task 14 | |
java.lang.NullPointerException | |
at org.apache.spark.scheduler.Task$$anonfun$run$1.apply$mcV$sp(Task.scala:129) | |
at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1317) | |
at org.apache.spark.scheduler.Task.run(Task.scala:127) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
18/03/21 10:06:11 ERROR Utils: Uncaught exception in thread Executor task launch worker for task 10 | |
java.lang.NullPointerException | |
at org.apache.spark.scheduler.Task$$anonfun$run$1.apply$mcV$sp(Task.scala:129) | |
at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1317) | |
at org.apache.spark.scheduler.Task.run(Task.scala:127) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
at java.lang.Thread.run(Thread.java:745) | |
18/03/21 10:06:11 INFO Executor: Executor interrupted and killed task 14.0 in stage 0.0 (TID 14), reason: stage cancelled | |
18/03/21 10:06:11 INFO Executor: Executor interrupted and killed task 10.0 in stage 0.0 (TID 10), reason: stage cancelled | |
18/03/21 10:06:11 ERROR ShutdownHookManager: Exception while deleting Spark temp dir: /private/var/folders/6g/r87hcdl560j9dxxm0_jb0rpr0000gn/T/spark-499fb32c-b7e7-4e7c-a422-549babc0ddd9/repl-edefbd5a-a47d-45f7-9580-5beb646964f9 | |
java.io.IOException: Failed to delete: /private/var/folders/6g/r87hcdl560j9dxxm0_jb0rpr0000gn/T/spark-499fb32c-b7e7-4e7c-a422-549babc0ddd9/repl-edefbd5a-a47d-45f7-9580-5beb646964f9 | |
at org.apache.spark.util.Utils$.deleteRecursively(Utils.scala:1031) | |
at org.apache.spark.util.ShutdownHookManager$$anonfun$1$$anonfun$apply$mcV$sp$3.apply(ShutdownHookManager.scala:65) | |
at org.apache.spark.util.ShutdownHookManager$$anonfun$1$$anonfun$apply$mcV$sp$3.apply(ShutdownHookManager.scala:62) | |
at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) | |
at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186) | |
at org.apache.spark.util.ShutdownHookManager$$anonfun$1.apply$mcV$sp(ShutdownHookManager.scala:62) | |
at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:216) | |
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ShutdownHookManager.scala:188) | |
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:188) | |
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:188) | |
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1948) | |
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply$mcV$sp(ShutdownHookManager.scala:188) | |
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:188) | |
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:188) | |
at scala.util.Try$.apply(Try.scala:192) | |
at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188) | |
at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178) | |
at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54) | |
18/03/21 10:06:11 INFO ShutdownHookManager: Deleting directory /private/var/folders/6g/r87hcdl560j9dxxm0_jb0rpr0000gn/T/spark-499fb32c-b7e7-4e7c-a422-549babc0ddd9 | |
sh: /users/ryandeschamps/auk-download/739/4867/1/derivatives/all-domains/4867-fullurls.txt: No such file or directory | |
sh: /users/ryandeschamps/auk-download/739/4867/1/derivatives/all-text/4867-fulltext.txt: No such file or directory |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment