Created
August 6, 2015 18:14
-
-
Save urlgrey/76da3e01741c3c4ee5ef to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# Template for a Spark Job Server configuration file | |
# When deployed these settings are loaded when job server starts | |
# | |
# Spark Cluster / Job Server configuration | |
spark { | |
# spark.master will be passed to each job's JobContext | |
# master = "local[4]" | |
master = "spark://spark-master-host:7077" | |
# master = "mesos://vm28-hulk-pub:5050" | |
# master = "yarn-client" | |
# Default # of CPUs for jobs to use for Spark standalone cluster | |
job-number-cpus = 4 | |
jobserver { | |
port = 8090 | |
jar-store-rootdir = /tmp/jobserver/jars | |
jobdao = spark.jobserver.io.JobFileDAO | |
filedao { | |
rootdir = /tmp/spark-job-server/filedao/data | |
} | |
} | |
# predefined Spark contexts | |
# contexts { | |
# my-low-latency-context { | |
# num-cpu-cores = 1 # Number of cores to allocate. Required. | |
# memory-per-node = 512m # Executor memory per node, -Xmx style eg 512m, 1G, etc. | |
# } | |
# # define additional contexts here | |
# } | |
contexts { | |
my-app-context { | |
num-cpu-cores = 2 # Number of cores to allocate. Required. | |
memory-per-node = 8G # Executor memory per node, -Xmx style eg 512m, #1G, etc. | |
} | |
} | |
# universal context configuration. These settings can be overridden, see README.md | |
context-settings { | |
num-cpu-cores = 2 # Number of cores to allocate. Required. | |
memory-per-node = 512m # Executor memory per node, -Xmx style eg 512m, #1G, etc. | |
# in case spark distribution should be accessed from HDFS (as opposed to being installed on every mesos slave) | |
# spark.executor.uri = "hdfs://namenode:8020/apps/spark/spark.tgz" | |
# uris of jars to be loaded into the classpath for this context. Uris is a string list, or a string separated by commas ',' | |
# dependent-jar-uris = ["file:///some/path/present/in/each/mesos/slave/somepackage.jar"] | |
# If you wish to pass any settings directly to the sparkConf as-is, add them here in passthrough, | |
# such as hadoop connection settings that don't use the "spark." prefix | |
passthrough { | |
#es.nodes = "192.1.1.1" | |
} | |
} | |
# This needs to match SPARK_HOME for cluster SparkContexts to be created successfully | |
# home = "/home/spark/spark" | |
home = "/root/spark" | |
} | |
spray.can.server { | |
idle-timeout = infinite | |
request-timeout = infinite | |
pipelining-limit = 2 # for maximum performance (prevents StopReading / ResumeReading messages to the IOBridge) | |
default-host-header = "spray.io:8765" | |
parsing.max-content-length = 400m | |
client { | |
request-timeout = infinite | |
} | |
} | |
bu |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment