Skip to content

Instantly share code, notes, and snippets.

@pindlebot
Last active May 29, 2019 05:08
Show Gist options
  • Save pindlebot/e8492a97b54bdb3fa2d8dff71b38f9c3 to your computer and use it in GitHub Desktop.
Save pindlebot/e8492a97b54bdb3fa2d8dff71b38f9c3 to your computer and use it in GitHub Desktop.
# Template for a Spark Job Server configuration file
# When deployed these settings are loaded when job server starts
#
# Spark Cluster / Job Server configuration
spark {
# Spark Master will be automatically learned via the DSE
# spark.master will be passed to each job's JobContext
# master = "local[4]"
# master = "mesos://vm28-hulk-pub:5050"
# master = "yarn-client"
# Default # of CPUs for jobs to use for Spark standalone cluster
job-number-cpus = 4
jobserver {
port = 8090
}
# predefined Spark contexts
# contexts {
# my-low-latency-context {
# num-cpu-cores = 1 # Number of cores to allocate. Required.
# memory-per-node = 512m # Executor memory per node, -Xmx style eg 512m, 1G, etc.
# }
# # define additional contexts here
# }
# universal context configuration. These settings can be overridden, see README.md
context-settings {
num-cpu-cores = 2 # Number of cores to allocate. Required.
memory-per-node = 512m # Executor memory per node, -Xmx style eg 512m, #1G, etc.
# in case spark distribution should be accessed from HDFS (as opposed to being installed on every mesos slave)
# spark.executor.uri = "hdfs://namenode:8020/apps/spark/spark.tgz"
# uris of jars to be loaded into the classpath for this context. Uris is a string list, or a string separated by commas ','
# dependent-jar-uris = ["file:///some/path/present/in/each/mesos/slave/somepackage.jar"]
# If you wish to pass any settings directly to the sparkConf as-is, add them here in passthrough,
# such as hadoop connection settings that don't use the "spark." prefix
passthrough {
#es.nodes = "192.1.1.1"
}
}
# This needs to match SPARK_HOME for cluster SparkContexts to be created successfully
# home = "/home/spark/spark"
}
# Note that you can use this file to define settings not only for job server,
# but for your Spark jobs as well. Spark job configuration merges with this configuration file as defaults.
deploy {
manager-start-cmd = "dse spark-jobserver context-per-jvm-managed-start"
}
spray.can.server {
verbose-error-messages = on
verbose-error-logging = on
pipelining-limit = 10
stats-support = on
request-timeout = 10 s
idle-timeout = 30 s
max-connections = 10
pipelining = on
}
akka {
# Log the complete configuration at INFO level when the actor system is started.
# This is useful when you are uncertain of what configuration is used.
log-config-on-start = on
debug {
router-misconfiguration = on
}
}
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment