Last active
August 29, 2015 14:25
-
-
Save retronym/979da8069e04fd14b448 to your computer and use it in GitHub Desktop.
Grepping for forkjoin over a few GH orgs
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
/code/orgs % (for d in $(find . -name .git); do (cd $d/..; R=$(git --no-pager grep forkjoin); [[ $? -eq 0 ]] && printf "\n$PWD\n=========================\n$R\n") done) | |
/code/orgs/akka/akka | |
========================= | |
akka-actor-tests/src/test/scala/akka/actor/SchedulerSpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-actor-tests/src/test/scala/akka/pattern/CircuitBreakerStressSpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-actor/src/main/scala/akka/actor/ActorCell.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-actor/src/main/scala/akka/dispatch/AbstractDispatcher.scala:import scala.concurrent.forkjoin.{ ForkJoinPool, ForkJoinTask } | |
akka-actor/src/main/scala/akka/dispatch/Dispatcher.scala:import scala.concurrent.forkjoin.ForkJoinPool | |
akka-actor/src/main/scala/akka/dispatch/Mailbox.scala:import scala.concurrent.forkjoin.ForkJoinTask | |
akka-actor/src/main/scala/akka/dispatch/ThreadPoolBuilder.scala:import scala.concurrent.forkjoin._ | |
akka-actor/src/main/scala/akka/pattern/BackoffSupervisor.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-actor/src/main/scala/akka/routing/Random.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-actor/src/main/scala/akka/routing/SmallestMailbox.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsCollector.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsRouting.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster-metrics/src/test/scala/akka/cluster/metrics/EWMASpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster-tools/src/main/scala/akka/cluster/pubsub/DistributedPubSubMediator.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster/src/main/scala/akka/cluster/ClusterDaemon.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster/src/main/scala/akka/cluster/ClusterMetricsCollector.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster/src/main/scala/akka/cluster/routing/AdaptiveLoadBalancing.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster/src/multi-jvm/scala/akka/cluster/StressSpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster/src/multi-jvm/scala/akka/cluster/SurviveNetworkInstabilitySpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster/src/test/scala/akka/cluster/ClusterHeartbeatSenderStateSpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-cluster/src/test/scala/akka/cluster/EWMASpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-contrib/docs/peek-mailbox.rst: at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:262) | |
akka-contrib/docs/peek-mailbox.rst: at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:975) | |
akka-contrib/docs/peek-mailbox.rst: at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1478) | |
akka-contrib/docs/peek-mailbox.rst: at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:104) | |
akka-distributed-data/src/main/scala/akka/cluster/ddata/Replicator.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-distributed-data/src/multi-jvm/scala/akka/cluster/ddata/JepsenInspiredInsertSpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-distributed-data/src/test/scala/akka/cluster/ddata/LotsOfDataBot.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-docs/rst/java/code/docs/ddata/DataBot.java:import scala.concurrent.forkjoin.ThreadLocalRandom; | |
akka-docs/rst/java/code/docs/ddata/DistributedDataDocTest.java:import scala.concurrent.forkjoin.ThreadLocalRandom; | |
akka-docs/rst/scala/code/docs/ddata/DistributedDataDocSpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-docs/rst/scala/code/docs/ddata/DistributedDataDocSpec.scala: import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-docs/rst/scala/code/docs/routing/CustomRouterDocSpec.scala: import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-persistence/src/test/scala/akka/persistence/AtLeastOnceDeliveryFailureSpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-persistence/src/test/scala/akka/persistence/journal/chaos/ChaosJournal.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-remote/src/main/scala/akka/remote/AddressUidExtension.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-remote/src/main/scala/akka/remote/transport/FailureInjectorTransportAdapter.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-remote/src/test/scala/akka/remote/AckedDeliverySpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-remote/src/test/scala/akka/remote/RemotingSpec.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-samples/akka-sample-cluster-java/src/main/java/sample/cluster/stats/StatsSampleClient.java:import scala.concurrent.forkjoin.ThreadLocalRandom; | |
akka-samples/akka-sample-cluster-scala/src/main/scala/sample/cluster/stats/StatsSample.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
akka-samples/akka-sample-fsm-scala/src/main/scala/sample/redelivery/FsmSimpleRedelivery.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/akka/assembla-migration | |
========================= | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:262) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:975) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1478) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:104) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) | |
report-1-Rollins.xml: at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) | |
/code/orgs/sbt/sbt-atmos | |
========================= | |
sample/abc/src/main/scala/com/typesafe/atmos/sample/Sample.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
sample/abc/src/test/scala/com/typesafe/atmos/sample/TestSample.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/sbt/sbt-proguard | |
========================= | |
src/sbt-test/proguard/akka/project/SampleBuild.scala:-keepclassmembernames class scala.concurrent.forkjoin.ForkJoinPool { | |
src/sbt-test/proguard/akka/project/SampleBuild.scala:-keepclassmembernames class scala.concurrent.forkjoin.ForkJoinPool$WorkQueue { | |
src/sbt-test/proguard/akka/project/SampleBuild.scala:-keepclassmembernames class scala.concurrent.forkjoin.LinkedTransferQueue { | |
src/sbt-test/proguard/akka/project/SampleBuild.scala: scala.concurrent.forkjoin.LinkedTransferQueue$Node head; | |
src/sbt-test/proguard/akka/project/SampleBuild.scala: scala.concurrent.forkjoin.LinkedTransferQueue$Node tail; | |
src/sbt-test/proguard/akka/project/SampleBuild.scala:-keepclassmembernames class scala.concurrent.forkjoin.LinkedTransferQueue$Node { | |
src/sbt-test/proguard/akka/project/SampleBuild.scala: scala.concurrent.forkjoin.LinkedTransferQueue$Node next; | |
src/sbt-test/proguard/akka/project/SampleBuild.scala:-dontnote scala.concurrent.forkjoin.ForkJoinPool | |
/code/orgs/scala/async | |
========================= | |
src/test/scala/scala/async/run/futures/FutureSpec.scala: implicit val ec = scala.concurrent.ExecutionContext.fromExecutor(new scala.concurrent.forkjoin.ForkJoinPool(), { | |
/code/orgs/scala/legacy-svn-scala | |
========================= | |
build.detach.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.detach.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.detach.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <target name="newforkjoin" | |
build.xml: description="Requires forkjoin library to be rebuilt. Add this target before any other if class file format is incompatible."> | |
build.xml: <property name="forkjoin.outdated" value="yes"/> | |
build.xml: <include name="forkjoin.jar"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <target name="quick.newforkjoin" depends="quick.libs" if="forkjoin.outdated"> | |
build.xml: <antcall target="forkjoin.done" inheritRefs="true"/> | |
build.xml: <property name="forkjoin.jar" value="${build-libs.dir}/forkjoin.jar"/> | |
build.xml: <target name="quick.forkjoin" depends="quick.newforkjoin" unless="forkjoin.outdated"> | |
build.xml: <property name="forkjoin.jar" value="${lib.dir}/forkjoin.jar"/> | |
build.xml: <target name="quick.pre-comp" depends="quick.forkjoin"> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <zipfileset dirmode="755" filemode="644" src="${forkjoin.jar}"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <target name="libs.pre-forkjoin" depends="libs.start"> | |
build.xml: <fail message="Compiling forkjoin.jar requires java 1.6. Please set the property `java6.home` in build.properties or using `-Djava6.home=/path/to/java6`"> | |
build.xml: <uptodate property="libs.forkjoin.available" targetfile="${build-libs.dir}/forkjoin.complete"> | |
build.xml: <srcfiles dir="${src.dir}/forkjoin"> | |
build.xml: <target name="libs.forkjoin" depends="libs.pre-forkjoin" unless="libs.forkjoin.available"> | |
build.xml: <mkdir dir="${build-libs.dir}/classes/forkjoin"/> | |
build.xml: srcdir="${src.dir}/forkjoin" | |
build.xml: destdir="${build-libs.dir}/classes/forkjoin" | |
build.xml: classpath="${build-libs.dir}/classes/forkjoin" | |
build.xml: <touch file="${build-libs.dir}/forkjoin.complete" verbose="no"/> | |
build.xml: <target name="libs.pre-forkjoinpack" depends="libs.forkjoin"> | |
build.xml: <target name="libs.forkjoinpack" depends="libs.pre-forkjoinpack" unless="libs.forkjoinpack.available"> | |
build.xml: <jar destfile="${build-libs.dir}/forkjoin.jar"> | |
build.xml: <fileset dir="${build-libs.dir}/classes/forkjoin"/> | |
build.xml: <target name="forkjoin.done" depends="libs.forkjoinpack"/> | |
build.xml: <include name="forkjoin.jar"/> | |
classpath.SAMPLE: <classpathentry kind="lib" path="lib/forkjoin.jar"/> | |
docs/development/scala.tools.nsc/zipfile-bug.txt: at scala.concurrent.forkjoin.ForkJoinPool.awaitBlocker(ForkJoinPool.java:1791) | |
docs/development/scala.tools.nsc/zipfile-bug.txt: at scala.concurrent.forkjoin.ForkJoinPool.managedBlock(ForkJoinPool.java:1781) | |
Binary file lib/forkjoin.jar matches | |
project/build/ScalaBuildProject.scala: lazy val forkJoinJar = lib / forkjoinJarName | |
project/build/ScalaBuildProject.scala: val forkjoinJarName = "forkjoin.jar" | |
project/build/ScalaSBTBuilder.scala: * An additional subproject used to build new version of forkjoin, fjbg and msil | |
project/build/ScalaSBTBuilder.scala: lazy val newForkjoin = copyJar(forkJoinConfig, forkjoinJarName).dependsOn(buildForkjoin) | |
project/build/ScalaSBTBuilder.scala: lazy val forkJoinConfig = new CompilationStep("forkjoin", pathLayout, log) with Packaging { | |
project/build/ScalaSBTBuilder.scala: def label = "new forkjoin library" | |
project/build/ScalaSBTBuilder.scala: lazy val packagingConfig = new PackagingConfiguration(libsDestination/forkjoinJarName, List(outputDirectory ##)) | |
project/build/ScalaSBTBuilder.scala: val buildForkjoinTaskDescription = "create all the jar needed to make a new forkjoin. It does not replace the current library and compiler jars in the libs folder, but the products of the task are instead located in target/libs." | |
project/build/ScalaSBTBuilder.scala: val newForkjoinTaskDescription = "create a new forkjoin and replace the corresponding jar in the libs folder." | |
src/actors/scala/actors/ReactorTask.scala:import scala.concurrent.forkjoin.RecursiveAction | |
src/actors/scala/actors/scheduler/DrainableForkJoinPool.scala:import scala.concurrent.forkjoin.{ForkJoinPool, ForkJoinTask} | |
src/actors/scala/actors/scheduler/ForkJoinScheduler.scala:import scala.concurrent.forkjoin._ | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinPool.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinTask.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinWorkerThread.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/LinkedTransferQueue.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/RecursiveAction.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/RecursiveTask.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ThreadLocalRandom.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/TransferQueue.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/package-info.java:package scala.concurrent.forkjoin; | |
src/intellij/scala-lang.ipr.SAMPLE: <root url="jar://$PROJECT_DIR$/../../lib/forkjoin.jar!/" /> | |
src/intellij/scala-lang.ipr.SAMPLE: <root url="jar://$PROJECT_DIR$/../../lib/forkjoin.jar!/" /> | |
src/library/scala/collection/parallel/Tasks.scala:import scala.concurrent.forkjoin._ | |
src/library/scala/collection/parallel/Tasks.scala: val defaultForkJoinPool: ForkJoinPool = new ForkJoinPool() // scala.parallel.forkjoinpool | |
src/library/scala/parallel/package.scala.disabled:import scala.concurrent.forkjoin._ | |
src/library/scala/parallel/package.scala.disabled: private[scala] val forkjoinpool = new ForkJoinPool() | |
src/library/scala/parallel/package.scala.disabled: else forkjoinpool.execute(task) | |
src/partest-alternative/scala/tools/partest/Config.scala: val forkjoin = pathForLibrary("forkjoin") | |
src/partest-alternative/scala/tools/partest/Config.scala: val fjMarker = "scala.concurrent.forkjoin.ForkJoinTask" | |
src/partest-alternative/scala/tools/partest/Config.scala: lazy val forkJoinPath: List[Path] = if (needsForkJoin) List(forkjoin) else Nil | |
src/partest-alternative/scala/tools/partest/category/Runner.scala: import build.{ scalacheck, forkjoin } | |
src/partest-alternative/scala/tools/partest/category/Runner.scala: override def classpathPaths = super.classpathPaths ::: List(scalacheck, forkjoin) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/generic/ParallelBenches.scala: def forkJoinPool: scala.concurrent.forkjoin.ForkJoinPool = self.forkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/generic/ParallelBenches.scala: def forkJoinPool: scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtables/ParallelHashTableSets.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtables/ParallelHashTables.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtries/ParallelHashTries.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/misc/Coder.scala: collection.parallel.tasksupport.environment.asInstanceOf[concurrent.forkjoin.ForkJoinPool].setParallelism(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/misc/Loader.scala: collection.parallel.tasksupport.environment.asInstanceOf[concurrent.forkjoin.ForkJoinPool].setParallelism(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/MatrixMultiplication.scala: collection.parallel.tasksupport.environment = forkjoinpool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: val forkjoinpool = new scala.concurrent.forkjoin.ForkJoinPool(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: forkjoinpool.setMaximumPoolSize(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: collection.parallel.tasksupport.environment = forkjoinpool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Parallelism: " + forkjoinpool.getParallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Active threads: " + forkjoinpool.getActiveThreadCount) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Work stealings: " + forkjoinpool.getStealCount) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_range/RangeBenches.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_view/SeqViewBenches.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/disabled/coder/Coder.scala: // import scala.concurrent.forkjoin.ForkJoinPool | |
tools/lockercp:echo ${cp}:$LIBDIR/fjbg.jar:$LIBDIR/msil.jar:$LIBDIR/forkjoin.jar:$LIBDIR/jline.jar:$LIBDIR/extra/'*' | |
tools/quickcp:echo ${cp}:$LIBDIR/fjbg.jar:$LIBDIR/msil.jar:$LIBDIR/forkjoin.jar:$LIBDIR/jline.jar:$LIBDIR/extra/'*' | |
/code/orgs/scala/scala | |
========================= | |
build-ant-macros.xml: <!-- project: forkjoin --> | |
build.sbt: .dependsOn (forkjoin) | |
build.sbt:lazy val forkjoin = configureAsForkOfJavaProject(project) | |
build.sbt: aggregate(library, forkjoin, reflect, compiler, interactive, repl, | |
build.sbt: * we depend on. At the moment there's just forkjoin. | |
build.xml: <property name="build-forkjoin.dir" value="${build-libs.dir}"/> | |
build.xml: <path id="forkjoin.classpath" path="${build-forkjoin.dir}/classes/forkjoin"/> | |
build.xml: <property name="forkjoin-classes" refid="forkjoin.classpath"/> | |
build.xml: <!-- projects without project-specific options: forkjoin, manual, bin, repl --> | |
build.xml: <path refid="forkjoin.classpath"/> | |
build.xml: <path refid="forkjoin.classpath"/> | |
build.xml: <path refid="forkjoin.classpath"/> | |
build.xml: <fileset dir="${forkjoin-classes}"/> | |
build.xml: <path refid="forkjoin.classpath"/> | |
build.xml: <path refid="forkjoin.classpath"/> | |
build.xml: <target name="forkjoin.done" depends="init"> <simple-javac project="forkjoin" args="-XDignore.symbol.file" jar="no"/></target> | |
build.xml: <target name="locker.start" depends="forkjoin.done"> | |
build.xml: <target name="pack.lib" depends="quick.lib, forkjoin.done"> <staged-pack project="library"/></target> | |
doc/LICENSE.md: * forkjoin | |
doc/License.rtf: '95 forkjoin} | |
lib/forkjoin.jar.desired.sha1:ddd7d5398733c4fbbb8355c049e258d47af636cf ?forkjoin.jar | |
src/actors/scala/actors/ReactorTask.scala:import scala.concurrent.forkjoin.RecursiveAction | |
src/actors/scala/actors/scheduler/DrainableForkJoinPool.scala:import scala.concurrent.forkjoin.{ForkJoinPool, ForkJoinTask} | |
src/actors/scala/actors/scheduler/ForkJoinScheduler.scala:import scala.concurrent.forkjoin._ | |
src/build/dbuild-meta-json-gen.scala: Seq.empty), // TODO: forkjoin | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinPool.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinTask.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinWorkerThread.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/LinkedTransferQueue.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/RecursiveAction.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/RecursiveTask.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ThreadLocalRandom.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/TransferQueue.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/package-info.java:package scala.concurrent.forkjoin; | |
src/intellij/actors.iml.SAMPLE: <orderEntry type="module" module-name="forkjoin" /> | |
src/intellij/forkjoin.iml.SAMPLE: <content url="file://$MODULE_DIR$/../forkjoin"> | |
src/intellij/forkjoin.iml.SAMPLE: <sourceFolder url="file://$MODULE_DIR$/../forkjoin" isTestSource="false" /> | |
src/intellij/library.iml.SAMPLE: <orderEntry type="module" module-name="forkjoin" /> | |
src/intellij/scala.ipr.SAMPLE: <module fileurl="file://$PROJECT_DIR$/forkjoin.iml" filepath="$PROJECT_DIR$/forkjoin.iml" /> | |
src/intellij/test-junit.iml.SAMPLE: <orderEntry type="module" module-name="forkjoin" /> | |
src/intellij/test.iml.SAMPLE: <orderEntry type="module" module-name="forkjoin" /> | |
src/library/scala/collection/concurrent/TrieMap.scala: scala.concurrent.forkjoin.ThreadLocalRandom.current.nextInt(0, array.length) | |
src/library/scala/collection/parallel/ParIterableLike.scala: * new scala.concurrent.forkjoin.ForkJoinPool(2)) | |
src/library/scala/collection/parallel/TaskSupport.scala:import scala.concurrent.forkjoin.ForkJoinPool | |
src/library/scala/collection/parallel/TaskSupport.scala: * new scala.concurrent.forkjoin.ForkJoinPool(2)) | |
src/library/scala/collection/parallel/TaskSupport.scala: * forkjoin based task support or a thread pool executor one, depending on | |
src/library/scala/collection/parallel/Tasks.scala:import scala.concurrent.forkjoin._ | |
src/library/scala/concurrent/impl/ExecutionContextImpl.scala:import scala.concurrent.forkjoin._ | |
test/benchmarking/ParCtrie-map.scala: parctrie.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
test/benchmarking/ParCtrie-size.scala: val ts = new collection.parallel.ForkJoinTaskSupport(new concurrent.forkjoin.ForkJoinPool(par)) | |
test/benchmarking/ParHashMap.scala: phm.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
test/benchmarking/ParVector-reduce.scala: parvector.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/generic/ParallelBenches.scala: def forkJoinPool: scala.concurrent.forkjoin.ForkJoinPool = self.forkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/generic/ParallelBenches.scala: def forkJoinPool: scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtables/ParallelHashTableSets.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtables/ParallelHashTables.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtries/ParallelHashTries.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/misc/Coder.scala: collection.parallel.tasksupport.environment.asInstanceOf[concurrent.forkjoin.ForkJoinPool].setParallelism(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/misc/Loader.scala: collection.parallel.tasksupport.environment.asInstanceOf[concurrent.forkjoin.ForkJoinPool].setParallelism(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/MatrixMultiplication.scala: collection.parallel.tasksupport.environment = forkjoinpool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: val forkjoinpool = new scala.concurrent.forkjoin.ForkJoinPool(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: forkjoinpool.setMaximumPoolSize(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: collection.parallel.tasksupport.environment = forkjoinpool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Parallelism: " + forkjoinpool.getParallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Active threads: " + forkjoinpool.getActiveThreadCount) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Work stealings: " + forkjoinpool.getStealCount) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_range/RangeBenches.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_view/SeqViewBenches.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/disabled/coder/Coder.scala: // import scala.concurrent.forkjoin.ForkJoinPool | |
test/files/jvm/future-spec/FutureTests.scala: implicit val ec = scala.concurrent.ExecutionContext.fromExecutor(new scala.concurrent.forkjoin.ForkJoinPool(), { | |
test/files/jvm/scala-concurrent-tck.scala: assert(bc.isInstanceOf[scala.concurrent.forkjoin.ForkJoinWorkerThread]) | |
test/files/jvm/t7146.check:should be scala.concurrent.forkjoin.ForkJoinPool == true | |
test/files/jvm/t7146.scala: println("should be scala.concurrent.forkjoin.ForkJoinPool == " + | |
test/files/jvm/t7146.scala: i.executor.toString.startsWith("scala.concurrent.forkjoin.ForkJoinPool")) | |
/code/orgs/scala/scala-lang | |
========================= | |
news/_posts/2013-03-20-gsoc13.md:Currently, the Scala Parallel Collections incur some abstraction penalties, that lead to boxing, virtual calls, and reliance on iterators. With the arrival of Scala Macros, there exists a plethora of optimization opportunities at the library level that allows us to eliminate these inefficiencies. Orthogonally, the scheduling based purely on the fork-join framework ([http://docs.oracle.com/javase/tutorial/essential/concurrency/forkjoin.html](http://docs.oracle.com/javase/tutorial/essential/concurrency/forkjoin.html "http://docs.oracle.com/javase/tutorial/essential/concurrency/forkjoin.html")) suffers from several abstraction penalties, and needs to be specialized to achieve better performance and load-balancing. | |
news/_posts/2013-05-23-release-notes-v2.10.2-RC1.md:[https://github.com/scala/scala/commit/ac1a0f0](ac1a0f0) | <notextile>Remove duplication in java builds of fjbg/asm/forkjoin</notextile> | |
news/_posts/2013-05-23-release-notes-v2.10.2-RC1.md:[https://github.com/scala/scala/commit/d85c3f1](d85c3f1) | <notextile>Formatting. Introduce {asm,forkjoin,fjbg}-classes props.</notextile> | |
news/_posts/2013-05-29-release-notes-v2.11.0-M3.md:[https://github.com/scala/scala/commit/ac1a0f0](ac1a0f0) | <notextile>Remove duplication in java builds of fjbg/asm/forkjoin</notextile> | |
news/_posts/2013-05-29-release-notes-v2.11.0-M3.md:[https://github.com/scala/scala/commit/d85c3f1](d85c3f1) | <notextile>Formatting. Introduce {asm,forkjoin,fjbg}-classes props.</notextile> | |
news/_posts/2013-05-31-release-notes-v2.10.2-RC2.md:[https://github.com/scala/scala/commit/ac1a0f0](ac1a0f0) | <notextile>Remove duplication in java builds of fjbg/asm/forkjoin</notextile> | |
news/_posts/2013-05-31-release-notes-v2.10.2-RC2.md:[https://github.com/scala/scala/commit/d85c3f1](d85c3f1) | <notextile>Formatting. Introduce {asm,forkjoin,fjbg}-classes props.</notextile> | |
news/_posts/2013-06-06-release-notes-v2.10.2.md:[https://github.com/scala/scala/commit/ac1a0f0](ac1a0f0) | <notextile>Remove duplication in java builds of fjbg/asm/forkjoin</notextile> | |
news/_posts/2013-06-06-release-notes-v2.10.2.md:[https://github.com/scala/scala/commit/d85c3f1](d85c3f1) | <notextile>Formatting. Introduce {asm,forkjoin,fjbg}-classes props.</notextile> | |
news/_posts/2014-06-30-2.12-roadmap.md:* Use the JDK’s forkjoin library instead of embedding our own. (Switch the global default ExecutionContext to be backed by the ForkJoinPool.commonPool().) | |
/code/orgs/scala/scala.github.com | |
========================= | |
es/overviews/parallel-collections/configuration.md: scala> pc.tasksupport = new ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(2)) | |
es/overviews/parallel-collections/performance.md: partrie.tasksupport = new ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
es/overviews/parallel-collections/performance.md: parvector.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
es/overviews/parallel-collections/performance.md: phm.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
ja/overviews/parallel-collections/configuration.md: scala> pc.tasksupport = new ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(2)) | |
ja/overviews/parallel-collections/performance.md: partrie.tasksupport = new ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
ja/overviews/parallel-collections/performance.md: parvector.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
ja/overviews/parallel-collections/performance.md: phm.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
overviews/parallel-collections/configuration.md: scala> pc.tasksupport = new ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(2)) | |
overviews/parallel-collections/performance.md: partrie.tasksupport = new ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
overviews/parallel-collections/performance.md: parvector.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
overviews/parallel-collections/performance.md: phm.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
/code/orgs/scala-ide/build-tools | |
========================= | |
src/test/resources/site_scala-ide-stub_01/content.xml: <provided namespace='java.package' name='scala.concurrent.forkjoin' version='2.10.3.v20130923-060037-VFINAL-e2fec6b28d'/> | |
/code/orgs/scala-ide/scala | |
========================= | |
build.detach.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.detach.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.detach.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <target name="newforkjoin" | |
build.xml: description="Requires forkjoin library to be rebuilt. Add this target before any other if class file format is incompatible."> | |
build.xml: <property name="forkjoin.outdated" value="yes"/> | |
build.xml: <include name="forkjoin.jar"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <target name="quick.newforkjoin" depends="quick.libs" if="forkjoin.outdated"> | |
build.xml: <antcall target="forkjoin.done" inheritRefs="true"/> | |
build.xml: <property name="forkjoin.jar" value="${build-libs.dir}/forkjoin.jar"/> | |
build.xml: <target name="quick.forkjoin" depends="quick.newforkjoin" unless="forkjoin.outdated"> | |
build.xml: <property name="forkjoin.jar" value="${lib.dir}/forkjoin.jar"/> | |
build.xml: <target name="quick.pre-comp" depends="quick.forkjoin"> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <zipfileset dirmode="755" filemode="644" src="${forkjoin.jar}"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <pathelement location="${lib.dir}/forkjoin.jar"/> | |
build.xml: <target name="libs.pre-forkjoin" depends="libs.start"> | |
build.xml: <fail message="Compiling forkjoin.jar requires java 1.6. Please set the property `java6.home` in build.properties or using `-Djava6.home=/path/to/java6`"> | |
build.xml: <uptodate property="libs.forkjoin.available" targetfile="${build-libs.dir}/forkjoin.complete"> | |
build.xml: <srcfiles dir="${src.dir}/forkjoin"> | |
build.xml: <target name="libs.forkjoin" depends="libs.pre-forkjoin" unless="libs.forkjoin.available"> | |
build.xml: <mkdir dir="${build-libs.dir}/classes/forkjoin"/> | |
build.xml: srcdir="${src.dir}/forkjoin" | |
build.xml: destdir="${build-libs.dir}/classes/forkjoin" | |
build.xml: classpath="${build-libs.dir}/classes/forkjoin" | |
build.xml: <touch file="${build-libs.dir}/forkjoin.complete" verbose="no"/> | |
build.xml: <target name="libs.pre-forkjoinpack" depends="libs.forkjoin"> | |
build.xml: <target name="libs.forkjoinpack" depends="libs.pre-forkjoinpack" unless="libs.forkjoinpack.available"> | |
build.xml: <jar destfile="${build-libs.dir}/forkjoin.jar"> | |
build.xml: <fileset dir="${build-libs.dir}/classes/forkjoin"/> | |
build.xml: <target name="forkjoin.done" depends="libs.forkjoinpack"/> | |
build.xml: <include name="forkjoin.jar"/> | |
classpath.SAMPLE: <classpathentry kind="lib" path="lib/forkjoin.jar"/> | |
docs/development/scala.tools.nsc/zipfile-bug.txt: at scala.concurrent.forkjoin.ForkJoinPool.awaitBlocker(ForkJoinPool.java:1791) | |
docs/development/scala.tools.nsc/zipfile-bug.txt: at scala.concurrent.forkjoin.ForkJoinPool.managedBlock(ForkJoinPool.java:1781) | |
lib/forkjoin.jar.desired.sha1:12c479a33ee283599fdb7aa91d6a1df0197a52cf ?forkjoin.jar | |
src/actors/scala/actors/ReactorTask.scala:import scala.concurrent.forkjoin.RecursiveAction | |
src/actors/scala/actors/scheduler/DrainableForkJoinPool.scala:import scala.concurrent.forkjoin.{ForkJoinPool, ForkJoinTask} | |
src/actors/scala/actors/scheduler/ForkJoinScheduler.scala:import scala.concurrent.forkjoin._ | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinPool.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinTask.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinWorkerThread.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/LinkedTransferQueue.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/RecursiveAction.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/RecursiveTask.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ThreadLocalRandom.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/TransferQueue.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/package-info.java:package scala.concurrent.forkjoin; | |
src/intellij/scala-lang.ipr.SAMPLE: <root url="jar://$PROJECT_DIR$/../../lib/forkjoin.jar!/" /> | |
src/intellij/scala-lang.ipr.SAMPLE: <root url="jar://$PROJECT_DIR$/../../lib/forkjoin.jar!/" /> | |
src/library/scala/collection/parallel/Tasks.scala:import scala.concurrent.forkjoin._ | |
src/library/scala/collection/parallel/Tasks.scala: val defaultForkJoinPool: ForkJoinPool = new ForkJoinPool() // scala.parallel.forkjoinpool | |
src/library/scala/parallel/package.scala.disabled:import scala.concurrent.forkjoin._ | |
src/library/scala/parallel/package.scala.disabled: private[scala] val forkjoinpool = new ForkJoinPool() | |
src/library/scala/parallel/package.scala.disabled: else forkjoinpool.execute(task) | |
src/partest-alternative/scala/tools/partest/Config.scala: val forkjoin = pathForLibrary("forkjoin") | |
src/partest-alternative/scala/tools/partest/Config.scala: val fjMarker = "scala.concurrent.forkjoin.ForkJoinTask" | |
src/partest-alternative/scala/tools/partest/Config.scala: lazy val forkJoinPath: List[Path] = if (needsForkJoin) List(forkjoin) else Nil | |
src/partest-alternative/scala/tools/partest/category/Runner.scala: import build.{ scalacheck, forkjoin } | |
src/partest-alternative/scala/tools/partest/category/Runner.scala: override def classpathPaths = super.classpathPaths ::: List(scalacheck, forkjoin) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/generic/ParallelBenches.scala: def forkJoinPool: scala.concurrent.forkjoin.ForkJoinPool = self.forkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/generic/ParallelBenches.scala: def forkJoinPool: scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtables/ParallelHashTableSets.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtables/ParallelHashTables.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtries/ParallelHashTries.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/misc/Coder.scala: collection.parallel.tasksupport.environment.asInstanceOf[concurrent.forkjoin.ForkJoinPool].setParallelism(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/misc/Loader.scala: collection.parallel.tasksupport.environment.asInstanceOf[concurrent.forkjoin.ForkJoinPool].setParallelism(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/MatrixMultiplication.scala: collection.parallel.tasksupport.environment = forkjoinpool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: val forkjoinpool = new scala.concurrent.forkjoin.ForkJoinPool(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: forkjoinpool.setMaximumPoolSize(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: collection.parallel.tasksupport.environment = forkjoinpool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Parallelism: " + forkjoinpool.getParallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Active threads: " + forkjoinpool.getActiveThreadCount) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Work stealings: " + forkjoinpool.getStealCount) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_range/RangeBenches.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_view/SeqViewBenches.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/disabled/coder/Coder.scala: // import scala.concurrent.forkjoin.ForkJoinPool | |
tools/lockercp:echo ${cp}:$LIBDIR/fjbg.jar:$LIBDIR/msil.jar:$LIBDIR/forkjoin.jar:$LIBDIR/jline.jar:$LIBDIR/extra/'*' | |
tools/quickcp:echo ${cp}:$LIBDIR/fjbg.jar:$LIBDIR/msil.jar:$LIBDIR/forkjoin.jar:$LIBDIR/jline.jar:$LIBDIR/extra/'*' | |
/code/orgs/scala-ide/scala-ide | |
========================= | |
org.scala-ide.sdt.core.tests/test-workspace/scala-compiler/.classpath: <classpathentry kind="lib" path="lib/forkjoin.jar"/> | |
/code/orgs/typesafehub/activator-akka-clustering | |
========================= | |
src/main/scala/chat/RandomUser.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/typesafehub/activator-akka-distributed-workers | |
========================= | |
src/main/scala/worker/WorkProducer.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/typesafehub/activator-akka-distributed-workers-java | |
========================= | |
src/main/java/worker/WorkProducer.java:import scala.concurrent.forkjoin.ThreadLocalRandom; | |
/code/orgs/typesafehub/activator-akka-stream-java8 | |
========================= | |
src/main/java/sample/stream/WritePrimes.java:import scala.concurrent.forkjoin.ThreadLocalRandom; | |
/code/orgs/typesafehub/activator-akka-stream-scala | |
========================= | |
src/main/scala/sample/stream/WritePrimes.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/typesafehub/activator-akka-supervision | |
========================= | |
src/main/scala/supervision/FlakyExpressionCalculator.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/typesafehub/akka-contrib-extra | |
========================= | |
src/test/scala/akka/contrib/process/BlockingProcessSpec.scala: val nameSeed = scala.concurrent.forkjoin.ThreadLocalRandom.current().nextLong() | |
/code/orgs/typesafehub/akka-workshop | |
========================= | |
cluster-patterns/src/main/scala/ScaleOutWithAkkaCluster.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/typesafehub/backpressure | |
========================= | |
modules/frontend/src/main/scala/frontend/WorkProducer.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/typesafehub/echo | |
========================= | |
trace/akka/2.1.x/src/main/scala/com/typesafe/trace/DispatcherMonitor.scala:import scala.concurrent.forkjoin.ForkJoinPool | |
trace/akka/2.2.x/src/main/scala/com/typesafe/trace/DispatcherMonitor.scala:import scala.concurrent.forkjoin.ForkJoinPool | |
trace/akka/2.3.x/src/main/scala/com/typesafe/trace/DispatcherMonitor.scala:import scala.concurrent.forkjoin.ForkJoinPool | |
/code/orgs/typesafehub/elucidator | |
========================= | |
analytics/src/main/scala/activator/analytics/analyzer/StatsAnalyzer.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/typesafehub/intel-unca | |
========================= | |
src/main/scala/gainlo/unca/Unca.scala:import scala.concurrent.forkjoin.ForkJoinPool | |
/code/orgs/typesafehub/playframework | |
========================= | |
framework/src/play/src/main/scala/play/core/system/Execution.scala:import scala.concurrent.forkjoin.{ ForkJoinWorkerThread, ForkJoinPool } | |
framework/src/play/src/main/scala/play/core/system/Execution.scala:import scala.concurrent.forkjoin.ForkJoinPool.ForkJoinWorkerThreadFactory | |
/code/orgs/typesafehub/profsrv-architecture | |
========================= | |
jvm/threadpool-tuning.md:* [Fork-Join Pool](http://blog.jessitron.com/2014/02/forkjoinpool-other-executorservice.html) | |
jvm/threads.md:* [Fork-Join Pool](http://blog.jessitron.com/2014/02/forkjoinpool-other-executorservice.html) | |
/code/orgs/typesafehub/ReactiveMaps | |
========================= | |
app/actors/GeoJsonBot.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
app/backend/BotManager.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/typesafehub/sbt-echo | |
========================= | |
sample/abc/src/main/scala/com/typesafe/echo/sample/Sample.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
sample/abc/src/test/scala/com/typesafe/echo/sample/TestSample.scala:import scala.concurrent.forkjoin.ThreadLocalRandom | |
/code/orgs/typesafehub/scala | |
========================= | |
build-ant-macros.xml: <!-- project: asm/forkjoin --> | |
build.sbt: .dependsOn (forkjoin) | |
build.sbt:lazy val forkjoin = configureAsForkOfJavaProject(project) | |
build.sbt: aggregate(library, forkjoin, reflect, compiler, asm, interactive, repl, | |
build.sbt: * we depend on. At the moment there are just two: asm and forkjoin. | |
build.xml: <property name="build-forkjoin.dir" value="${build-libs.dir}"/> | |
build.xml: <path id="forkjoin.classpath" path="${build-forkjoin.dir}/classes/forkjoin"/> | |
build.xml: <property name="forkjoin-classes" refid="forkjoin.classpath"/> | |
build.xml: <!-- projects without project-specific options: asm, forkjoin, manual, bin, repl --> | |
build.xml: <path refid="forkjoin.classpath"/> | |
build.xml: <path refid="forkjoin.classpath"/> | |
build.xml: <path refid="forkjoin.classpath"/> | |
build.xml: <fileset dir="${forkjoin-classes}"/> | |
build.xml: <path refid="forkjoin.classpath"/> | |
build.xml: <path refid="forkjoin.classpath"/> | |
build.xml: <target name="forkjoin.done" depends="init"> <simple-javac project="forkjoin" args="-XDignore.symbol.file" jar="no"/></target> | |
build.xml: <target name="locker.start" depends="asm.done, forkjoin.done"> | |
build.xml: <target name="pack.lib" depends="quick.lib, forkjoin.done"> <staged-pack project="library"/></target> | |
doc/LICENSE.md: * forkjoin | |
doc/License.rtf: '95 forkjoin} | |
lib/forkjoin.jar.desired.sha1:ddd7d5398733c4fbbb8355c049e258d47af636cf ?forkjoin.jar | |
src/actors/scala/actors/ReactorTask.scala:import scala.concurrent.forkjoin.RecursiveAction | |
src/actors/scala/actors/scheduler/DrainableForkJoinPool.scala:import scala.concurrent.forkjoin.{ForkJoinPool, ForkJoinTask} | |
src/actors/scala/actors/scheduler/ForkJoinScheduler.scala:import scala.concurrent.forkjoin._ | |
src/build/dbuild-meta-json-gen.scala: Seq.empty), // TODO: forkjoin | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinPool.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinTask.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ForkJoinWorkerThread.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/LinkedTransferQueue.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/RecursiveAction.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/RecursiveTask.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/ThreadLocalRandom.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/TransferQueue.java:package scala.concurrent.forkjoin; | |
src/forkjoin/scala/concurrent/forkjoin/package-info.java:package scala.concurrent.forkjoin; | |
src/intellij/actors.iml.SAMPLE: <orderEntry type="module" module-name="forkjoin" /> | |
src/intellij/forkjoin.iml.SAMPLE: <content url="file://$MODULE_DIR$/../forkjoin"> | |
src/intellij/forkjoin.iml.SAMPLE: <sourceFolder url="file://$MODULE_DIR$/../forkjoin" isTestSource="false" /> | |
src/intellij/library.iml.SAMPLE: <orderEntry type="module" module-name="forkjoin" /> | |
src/intellij/scala.ipr.SAMPLE: <module fileurl="file://$PROJECT_DIR$/forkjoin.iml" filepath="$PROJECT_DIR$/forkjoin.iml" /> | |
src/intellij/test-junit.iml.SAMPLE: <orderEntry type="module" module-name="forkjoin" /> | |
src/intellij/test.iml.SAMPLE: <orderEntry type="module" module-name="forkjoin" /> | |
src/library/scala/collection/concurrent/TrieMap.scala: scala.concurrent.forkjoin.ThreadLocalRandom.current.nextInt(0, array.length) | |
src/library/scala/collection/parallel/ParIterableLike.scala: * new scala.concurrent.forkjoin.ForkJoinPool(2)) | |
src/library/scala/collection/parallel/TaskSupport.scala:import scala.concurrent.forkjoin.ForkJoinPool | |
src/library/scala/collection/parallel/TaskSupport.scala: * new scala.concurrent.forkjoin.ForkJoinPool(2)) | |
src/library/scala/collection/parallel/TaskSupport.scala: * forkjoin based task support or a thread pool executor one, depending on | |
src/library/scala/collection/parallel/Tasks.scala:import scala.concurrent.forkjoin._ | |
src/library/scala/concurrent/impl/ExecutionContextImpl.scala:import scala.concurrent.forkjoin._ | |
test/benchmarking/ParCtrie-map.scala: parctrie.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
test/benchmarking/ParCtrie-size.scala: val ts = new collection.parallel.ForkJoinTaskSupport(new concurrent.forkjoin.ForkJoinPool(par)) | |
test/benchmarking/ParHashMap.scala: phm.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
test/benchmarking/ParVector-reduce.scala: parvector.tasksupport = new collection.parallel.ForkJoinTaskSupport(new scala.concurrent.forkjoin.ForkJoinPool(par)) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/generic/ParallelBenches.scala: def forkJoinPool: scala.concurrent.forkjoin.ForkJoinPool = self.forkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/generic/ParallelBenches.scala: def forkJoinPool: scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtables/ParallelHashTableSets.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtables/ParallelHashTables.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/hashtries/ParallelHashTries.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/misc/Coder.scala: collection.parallel.tasksupport.environment.asInstanceOf[concurrent.forkjoin.ForkJoinPool].setParallelism(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/misc/Loader.scala: collection.parallel.tasksupport.environment.asInstanceOf[concurrent.forkjoin.ForkJoinPool].setParallelism(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/MatrixMultiplication.scala: collection.parallel.tasksupport.environment = forkjoinpool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: val forkjoinpool = new scala.concurrent.forkjoin.ForkJoinPool(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: forkjoinpool.setMaximumPoolSize(parallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: collection.parallel.tasksupport.environment = forkjoinpool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Parallelism: " + forkjoinpool.getParallelism) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Active threads: " + forkjoinpool.getActiveThreadCount) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_array/Resettable.scala: println("Work stealings: " + forkjoinpool.getStealCount) | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_range/RangeBenches.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/benchmarks/src/scala/collection/parallel/benchmarks/parallel_view/SeqViewBenches.scala: val forkJoinPool = new scala.concurrent.forkjoin.ForkJoinPool | |
test/disabled/coder/Coder.scala: // import scala.concurrent.forkjoin.ForkJoinPool | |
test/files/jvm/future-spec/FutureTests.scala: implicit val ec = scala.concurrent.ExecutionContext.fromExecutor(new scala.concurrent.forkjoin.ForkJoinPool(), { | |
test/files/jvm/scala-concurrent-tck.scala: assert(bc.isInstanceOf[scala.concurrent.forkjoin.ForkJoinWorkerThread]) | |
test/files/jvm/t7146.check:should be scala.concurrent.forkjoin.ForkJoinPool == true | |
test/files/jvm/t7146.scala: println("should be scala.concurrent.forkjoin.ForkJoinPool == " + | |
test/files/jvm/t7146.scala: i.executor.toString.startsWith("scala.concurrent.forkjoin.ForkJoinPool")) | |
/code/orgs/typesafehub/spark | |
========================= | |
core/src/test/resources/spark-events/local-1422981780767/EVENT_LOG_1:{"Event":"SparkListenerJobEnd","Job ID":1,"Job Result":{"Result":"JobFailed","Exception":{"Message":"Job aborted due to stage failure: Task 3 in stage 2.0 failed 1 times, most recent failure: Lost task 3.0 in stage 2.0 (TID 19, localhost): java.lang.RuntimeException: got a 3, failing | |
at $line11.$read$$iwC$$iwC$$iwC$$iwC$$anonfun$1.apply(<console>:18) | |
at $line11.$read$$iwC$$iwC$$iwC$$iwC$$anonfun$1.apply(<console>:17) | |
at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) | |
at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1311) | |
at org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:910) | |
at org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:910) | |
at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1314) | |
at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1314) | |
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61) | |
at org.apache.spark.scheduler.Task.run(Task.scala:56) | |
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196) | |
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) | |
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) | |
at java.lang.Thread.run(Thread.java:745) | |
Driver stacktrace:","Stack Trace":[{"Declaring Class":"org.apache.spark.scheduler.DAGScheduler","Method Name":"org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages","File Name":"DAGScheduler.scala","Line Number":1214},{"Declaring Class":"org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1","Method Name":"apply","File Name":"DAGScheduler.scala","Line Number":1203},{"Declaring Class":"org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1","Method Name":"apply","File Name":"DAGScheduler.scala","Line Number":1202},{"Declaring Class":"scala.collection.mutable.ResizableArray$class","Method Name":"foreach","File Name":"ResizableArray.scala","Line Number":59},{"Declaring Class":"scala.collection.mutable.ArrayBuffer","Method Name":"foreach","File Name":"ArrayBuffer.scala","Line Number":47},{"Declaring Class":"org.apache.spark.scheduler.DAGScheduler","Method Name":"abortStage","File Name":"DAGScheduler.scala","Line Number":1202},{"Declaring Class":"org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1","Method Name":"apply","File Name":"DAGScheduler.scala","Line Number":696},{"Declaring Class":"org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1","Method Name":"apply","File Name":"DAGScheduler.scala","Line Number":696},{"Declaring Class":"scala.Option","Method Name":"foreach","File Name":"Option.scala","Line Number":236},{"Declaring Class":"org.apache.spark.scheduler.DAGScheduler","Method Name":"handleTaskSetFailed","File Name":"DAGScheduler.scala","Line Number":696},{"Declaring Class":"org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2","Method Name":"applyOrElse","File Name":"DAGScheduler.scala","Line Number":1420},{"Declaring Class":"akka.actor.Actor$class","Method Name":"aroundReceive","File Name":"Actor.scala","Line Number":465},{"Declaring Class":"org.apache.spark.scheduler.DAGSchedulerEventProcessActor","Method Name":"aroundReceive","File Name":"DAGScheduler.scala","Line Number":1375},{"Declaring Class":"akka.actor.ActorCell","Method Name":"receiveMessage","File Name":"ActorCell.scala","Line Number":516},{"Declaring Class":"akka.actor.ActorCell","Method Name":"invoke","File Name":"ActorCell.scala","Line Number":487},{"Declaring Class":"akka.dispatch.Mailbox","Method Name":"processMailbox","File Name":"Mailbox.scala","Line Number":238},{"Declaring Class":"akka.dispatch.Mailbox","Method Name":"run","File Name":"Mailbox.scala","Line Number":220},{"Declaring Class":"akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask","Method Name":"exec","File Name":"AbstractDispatcher.scala","Line Number":393},{"Declaring Class":"scala.concurrent.forkjoin.ForkJoinTask","Method Name":"doExec","File Name":"ForkJoinTask.java","Line Number":260},{"Declaring Class":"scala.concurrent.forkjoin.ForkJoinPool$WorkQueue","Method Name":"runTask","File Name":"ForkJoinPool.java","Line Number":1339},{"Declaring Class":"scala.concurrent.forkjoin.ForkJoinPool","Method Name":"runWorker","File Name":"ForkJoinPool.java","Line Number":1979},{"Declaring Class":"scala.concurrent.forkjoin.ForkJoinWorkerThread","Method Name":"run","File Name":"ForkJoinWorkerThread.java","Line Number":107}]}}} | |
/code/orgs/typesafehub/www.playframework.org | |
========================= | |
Binary file modules/scala-0.1.zip matches | |
Binary file modules/scala-0.2.zip matches | |
Binary file modules/scala-0.3.zip matches | |
Binary file modules/scala-0.6.zip matches | |
Binary file modules/scala-0.7.1.zip matches | |
Binary file modules/scala-0.7.2.zip matches | |
Binary file modules/scala-0.7.zip matches | |
Binary file modules/scala-head.zip matches | |
/code/orgs |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment