Last active
May 29, 2019 05:08
-
-
Save pindlebot/e8492a97b54bdb3fa2d8dff71b38f9c3 to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# Template for a Spark Job Server configuration file | |
# When deployed these settings are loaded when job server starts | |
# | |
# Spark Cluster / Job Server configuration | |
spark { | |
# Spark Master will be automatically learned via the DSE | |
# spark.master will be passed to each job's JobContext | |
# master = "local[4]" | |
# master = "mesos://vm28-hulk-pub:5050" | |
# master = "yarn-client" | |
# Default # of CPUs for jobs to use for Spark standalone cluster | |
job-number-cpus = 4 | |
jobserver { | |
port = 8090 | |
} | |
# predefined Spark contexts | |
# contexts { | |
# my-low-latency-context { | |
# num-cpu-cores = 1 # Number of cores to allocate. Required. | |
# memory-per-node = 512m # Executor memory per node, -Xmx style eg 512m, 1G, etc. | |
# } | |
# # define additional contexts here | |
# } | |
# universal context configuration. These settings can be overridden, see README.md | |
context-settings { | |
num-cpu-cores = 2 # Number of cores to allocate. Required. | |
memory-per-node = 512m # Executor memory per node, -Xmx style eg 512m, #1G, etc. | |
# in case spark distribution should be accessed from HDFS (as opposed to being installed on every mesos slave) | |
# spark.executor.uri = "hdfs://namenode:8020/apps/spark/spark.tgz" | |
# uris of jars to be loaded into the classpath for this context. Uris is a string list, or a string separated by commas ',' | |
# dependent-jar-uris = ["file:///some/path/present/in/each/mesos/slave/somepackage.jar"] | |
# If you wish to pass any settings directly to the sparkConf as-is, add them here in passthrough, | |
# such as hadoop connection settings that don't use the "spark." prefix | |
passthrough { | |
#es.nodes = "192.1.1.1" | |
} | |
} | |
# This needs to match SPARK_HOME for cluster SparkContexts to be created successfully | |
# home = "/home/spark/spark" | |
} | |
# Note that you can use this file to define settings not only for job server, | |
# but for your Spark jobs as well. Spark job configuration merges with this configuration file as defaults. | |
deploy { | |
manager-start-cmd = "dse spark-jobserver context-per-jvm-managed-start" | |
} | |
spray.can.server { | |
verbose-error-messages = on | |
verbose-error-logging = on | |
pipelining-limit = 10 | |
stats-support = on | |
request-timeout = 10 s | |
idle-timeout = 30 s | |
max-connections = 10 | |
pipelining = on | |
} | |
akka { | |
# Log the complete configuration at INFO level when the actor system is started. | |
# This is useful when you are uncertain of what configuration is used. | |
log-config-on-start = on | |
debug { | |
router-misconfiguration = on | |
} | |
} |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment