Skip to content

Instantly share code, notes, and snippets.

@jimbobhickville
Last active December 11, 2015 15:12
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save jimbobhickville/1f10b3508ef946eccb92 to your computer and use it in GitHub Desktop.
Save jimbobhickville/1f10b3508ef946eccb92 to your computer and use it in GitHub Desktop.
spark-submit --properties-file "../conf/app.conf" --packages "com.databricks:spark-csv_2.11:1.2.0" --jars "hdfs://master-1.local:8020/apps/twitter_sentiment/lucene-core-5.3.0.jar,hdfs://master-1.local:8020/apps/twitter_sentiment/lucene-analyzers-common-5.3.0.jar" --class "com.rackspace.spark.Sentiment" --supervise --deploy-mode "cluster" --master spark://master-1.local:7077 hdfs://master-1.local:8020/apps/twitter_sentiment/sentiment-project_2.10-1.0.jar
Launch Command: "/usr/lib/jvm/java-1.8.0-openjdk-1.8.0.60-2.b27.el7_1.x86_64/jre/bin/java" "-cp" "/usr/lib/spark/lib/spark-examples-1.5.0-hadoop2.7.1.jar:/etc/spark/conf/:/usr/lib/spark/lib/spark-assembly-1.5.0-hadoop2.7.1.jar" "-Xms1024M" "-Xmx1024M" "-Dspark.sentimentApp.numKafkaPartitions=3" "-Dakka.loglevel=WARNING" "-Dspark.sentimentApp.accessToken=[redacted]" "-Dspark.sentimentApp.twitterRefresh=10" "-Dspark.submit.deployMode=cluster" "-Dspark.sentimentApp.clientRefreshInterval=30" "-Dspark.sentimentApp.zookeeperHosts=zookeeper-1.local:2181,zookeeper-2.local:2181,zookeeper-3.local:2181" "-Dspark.sentimentApp.apiSecret=[redacted]" "-Dspark.rpc.askTimeout=10" "-Dspark.sentimentApp.apiKey=[redacted]" "-Dspark.jars=hdfs://master-1.local:8020/apps/twitter_sentiment/lucene-core-5.3.0.jar,hdfs://master-1.local:8020/apps/twitter_sentiment/lucene-analyzers-common-5.3.0.jar,file:/root/.ivy2/jars/com.databricks_spark-csv_2.11-1.2.0.jar,file:/root/.ivy2/jars/org.apache.commons_commons-csv-1.1.jar,file:/root/.ivy2/jars/com.univocity_univocity-parsers-1.5.1.jar,file:/root/.ivy2/jars/com.databricks_spark-csv_2.11-1.2.0.jar,file:/root/.ivy2/jars/org.apache.commons_commons-csv-1.1.jar,file:/root/.ivy2/jars/com.univocity_univocity-parsers-1.5.1.jar,hdfs://master-1.local:8020/apps/twitter_sentiment/sentiment-project_2.10-1.0.jar" "-Dspark.sentimentApp.kafkaQueue=scored-tweets" "-Dspark.driver.supervise=true" "-Dspark.sentimentApp.keywords=[redacted]" "-Dspark.sentimentApp.clientQueueCache=/tmp/tweet_cache.json" "-Dspark.sentimentApp.accessTokenSecret=[redacted]" "-Dspark.sentimentApp.clientQueueSize=150" "-Dspark.sentimentApp.kafkaBrokers=slave-1.local:6667,slave-2.local:6667,slave-3.local:6667" "-Dspark.master=spark://master-1.local:7077" "-Dspark.app.name=com.rackspace.spark.Sentiment" "org.apache.spark.deploy.worker.DriverWrapper" "akka.tcp://sparkWorker@10.223.112.235:59117/user/Worker" "/var/lib/spark/driver-20151210153926-0009/sentiment-project_2.10-1.0.jar" "com.rackspace.spark.Sentiment"
========================================
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/lib/spark/lib/spark-examples-1.5.0-hadoop2.7.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/lib/spark/lib/spark-assembly-1.5.0-hadoop2.7.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
15/12/10 15:50:50 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
15/12/10 15:50:50 INFO SecurityManager: Changing view acls to: spark
15/12/10 15:50:50 INFO SecurityManager: Changing modify acls to: spark
15/12/10 15:50:50 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(spark); users with modify permissions: Set(spark)
15/12/10 15:50:51 INFO Slf4jLogger: Slf4jLogger started
15/12/10 15:50:51 INFO Utils: Successfully started service 'Driver' on port 33551.
15/12/10 15:50:51 INFO WorkerWatcher: Connecting to worker akka.tcp://sparkWorker@10.223.112.235:59117/user/Worker
15/12/10 15:50:51 INFO SparkContext: Running Spark version 1.5.0
15/12/10 15:50:51 WARN SparkConf:
SPARK_CLASSPATH was detected (set to '/usr/lib/spark/lib/spark-examples-1.5.0-hadoop2.7.1.jar').
This is deprecated in Spark 1.0+.
Please instead use:
- ./spark-submit with --driver-class-path to augment the driver classpath
- spark.executor.extraClassPath to augment the executor classpath
15/12/10 15:50:51 WARN SparkConf: Setting 'spark.executor.extraClassPath' to '/usr/lib/spark/lib/spark-examples-1.5.0-hadoop2.7.1.jar' as a work-around.
15/12/10 15:50:51 WARN SparkConf: Setting 'spark.driver.extraClassPath' to '/usr/lib/spark/lib/spark-examples-1.5.0-hadoop2.7.1.jar' as a work-around.
15/12/10 15:50:51 INFO SecurityManager: Changing view acls to: spark
15/12/10 15:50:51 INFO SecurityManager: Changing modify acls to: spark
15/12/10 15:50:51 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(spark); users with modify permissions: Set(spark)
15/12/10 15:50:51 INFO WorkerWatcher: Successfully connected to akka.tcp://sparkWorker@10.223.112.235:59117/user/Worker
15/12/10 15:50:51 INFO Slf4jLogger: Slf4jLogger started
15/12/10 15:50:51 INFO Utils: Successfully started service 'sparkDriver' on port 57077.
15/12/10 15:50:51 INFO SparkEnv: Registering MapOutputTracker
15/12/10 15:50:51 INFO SparkEnv: Registering BlockManagerMaster
15/12/10 15:50:51 INFO DiskBlockManager: Created local directory at /data1/spark/blockmgr-e7cc1f76-ad3d-4f50-9337-6d34b29cf952
15/12/10 15:50:51 INFO MemoryStore: MemoryStore started with capacity 530.0 MB
15/12/10 15:50:51 INFO HttpFileServer: HTTP File server directory is /data1/spark/spark-ee667b90-7be7-49f0-b542-9349c6f06224/httpd-192c340c-2242-4c87-ba7b-80af868ebbc3
15/12/10 15:50:51 INFO HttpServer: Starting HTTP Server
15/12/10 15:50:52 INFO Utils: Successfully started service 'HTTP file server' on port 32802.
15/12/10 15:50:52 INFO SparkEnv: Registering OutputCommitCoordinator
15/12/10 15:50:52 INFO Utils: Successfully started service 'SparkUI' on port 4040.
15/12/10 15:50:52 INFO SparkUI: Started SparkUI at http://gateway-1.local:4040
15/12/10 15:50:52 INFO SparkContext: Added JAR hdfs://master-1.local:8020/apps/twitter_sentiment/lucene-core-5.3.0.jar at hdfs://master-1.local:8020/apps/twitter_sentiment/lucene-core-5.3.0.jar with timestamp 1449762652181
15/12/10 15:50:52 INFO SparkContext: Added JAR hdfs://master-1.local:8020/apps/twitter_sentiment/lucene-analyzers-common-5.3.0.jar at hdfs://master-1.local:8020/apps/twitter_sentiment/lucene-analyzers-common-5.3.0.jar with timestamp 1449762652185
15/12/10 15:50:52 ERROR SparkContext: Jar not found at file:/root/.ivy2/jars/com.databricks_spark-csv_2.11-1.2.0.jar
15/12/10 15:50:52 ERROR SparkContext: Jar not found at file:/root/.ivy2/jars/org.apache.commons_commons-csv-1.1.jar
15/12/10 15:50:52 ERROR SparkContext: Jar not found at file:/root/.ivy2/jars/com.univocity_univocity-parsers-1.5.1.jar
15/12/10 15:50:52 ERROR SparkContext: Jar not found at file:/root/.ivy2/jars/com.databricks_spark-csv_2.11-1.2.0.jar
15/12/10 15:50:52 ERROR SparkContext: Jar not found at file:/root/.ivy2/jars/org.apache.commons_commons-csv-1.1.jar
15/12/10 15:50:52 ERROR SparkContext: Jar not found at file:/root/.ivy2/jars/com.univocity_univocity-parsers-1.5.1.jar
15/12/10 15:50:52 INFO SparkContext: Added JAR hdfs://master-1.local:8020/apps/twitter_sentiment/sentiment-project_2.10-1.0.jar at hdfs://master-1.local:8020/apps/twitter_sentiment/sentiment-project_2.10-1.0.jar with timestamp 1449762652198
15/12/10 15:50:52 WARN MetricsSystem: Using default name DAGScheduler for source because spark.app.id is not set.
15/12/10 15:50:52 INFO AppClient$ClientEndpoint: Connecting to master spark://master-1.local:7077...
15/12/10 15:50:52 INFO SparkDeploySchedulerBackend: Connected to Spark cluster with app ID app-20151210155052-0096
15/12/10 15:50:52 INFO AppClient$ClientEndpoint: Executor added: app-20151210155052-0096/0 on worker-20151029032924-10.223.112.199-48093 (10.223.112.199:48093) with 2 cores
15/12/10 15:50:52 INFO SparkDeploySchedulerBackend: Granted executor ID app-20151210155052-0096/0 on hostPort 10.223.112.199:48093 with 2 cores, 1024.0 MB RAM
15/12/10 15:50:52 INFO AppClient$ClientEndpoint: Executor added: app-20151210155052-0096/1 on worker-20151029032656-10.223.112.209-39515 (10.223.112.209:39515) with 2 cores
15/12/10 15:50:52 INFO SparkDeploySchedulerBackend: Granted executor ID app-20151210155052-0096/1 on hostPort 10.223.112.209:39515 with 2 cores, 1024.0 MB RAM
15/12/10 15:50:52 INFO AppClient$ClientEndpoint: Executor added: app-20151210155052-0096/2 on worker-20151029033207-10.223.112.235-59117 (10.223.112.235:59117) with 1 cores
15/12/10 15:50:52 INFO SparkDeploySchedulerBackend: Granted executor ID app-20151210155052-0096/2 on hostPort 10.223.112.235:59117 with 1 cores, 1024.0 MB RAM
15/12/10 15:50:52 INFO AppClient$ClientEndpoint: Executor updated: app-20151210155052-0096/1 is now LOADING
15/12/10 15:50:52 INFO AppClient$ClientEndpoint: Executor updated: app-20151210155052-0096/0 is now LOADING
15/12/10 15:50:52 INFO AppClient$ClientEndpoint: Executor updated: app-20151210155052-0096/2 is now LOADING
15/12/10 15:50:52 INFO AppClient$ClientEndpoint: Executor updated: app-20151210155052-0096/0 is now RUNNING
15/12/10 15:50:52 INFO AppClient$ClientEndpoint: Executor updated: app-20151210155052-0096/1 is now RUNNING
15/12/10 15:50:52 INFO AppClient$ClientEndpoint: Executor updated: app-20151210155052-0096/2 is now RUNNING
15/12/10 15:50:52 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 43519.
15/12/10 15:50:52 INFO NettyBlockTransferService: Server created on 43519
15/12/10 15:50:52 INFO BlockManagerMaster: Trying to register BlockManager
15/12/10 15:50:52 INFO BlockManagerMasterEndpoint: Registering block manager 10.223.112.235:43519 with 530.0 MB RAM, BlockManagerId(driver, 10.223.112.235, 43519)
15/12/10 15:50:52 INFO BlockManagerMaster: Registered BlockManager
15/12/10 15:50:53 INFO SparkDeploySchedulerBackend: SchedulerBackend is ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.0
15/12/10 15:50:55 INFO SparkDeploySchedulerBackend: Registered executor: AkkaRpcEndpointRef(Actor[akka.tcp://sparkExecutor@10.223.112.199:43065/user/Executor#734925163]) with ID 0
Exception in thread "main" java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.spark.deploy.worker.DriverWrapper$.main(DriverWrapper.scala:58)
at org.apache.spark.deploy.worker.DriverWrapper.main(DriverWrapper.scala)
Caused by: java.lang.ClassNotFoundException: Failed to load class for data source: com.databricks.spark.csv.
at org.apache.spark.sql.execution.datasources.ResolvedDataSource$.lookupDataSource(ResolvedDataSource.scala:67)
at org.apache.spark.sql.execution.datasources.ResolvedDataSource$.apply(ResolvedDataSource.scala:87)
at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:114)
at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:104)
at com.rackspace.spark.Sentiment$.prepareNBModel(Sentiment.scala:72)
at com.rackspace.spark.Sentiment$.main(Sentiment.scala:189)
at com.rackspace.spark.Sentiment.main(Sentiment.scala)
... 6 more
Caused by: java.lang.ClassNotFoundException: com.databricks.spark.csv.DefaultSource
at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
at org.apache.spark.sql.execution.datasources.ResolvedDataSource$$anonfun$4$$anonfun$apply$1.apply(ResolvedDataSource.scala:60)
at org.apache.spark.sql.execution.datasources.ResolvedDataSource$$anonfun$4$$anonfun$apply$1.apply(ResolvedDataSource.scala:60)
at scala.util.Try$.apply(Try.scala:161)
at org.apache.spark.sql.execution.datasources.ResolvedDataSource$$anonfun$4.apply(ResolvedDataSource.scala:60)
at org.apache.spark.sql.execution.datasources.ResolvedDataSource$$anonfun$4.apply(ResolvedDataSource.scala:60)
at scala.util.Try.orElse(Try.scala:82)
at org.apache.spark.sql.execution.datasources.ResolvedDataSource$.lookupDataSource(ResolvedDataSource.scala:60)
... 12 more
15/12/10 15:50:55 INFO SparkContext: Invoking stop() from shutdown hook
15/12/10 15:50:55 INFO BlockManagerMasterEndpoint: Registering block manager 10.223.112.199:42121 with 530.0 MB RAM, BlockManagerId(0, 10.223.112.199, 42121)
15/12/10 15:50:55 INFO SparkUI: Stopped Spark web UI at http://gateway-1.local:4040
15/12/10 15:50:55 INFO DAGScheduler: Stopping DAGScheduler
15/12/10 15:50:55 INFO SparkDeploySchedulerBackend: Shutting down all executors
15/12/10 15:50:55 INFO SparkDeploySchedulerBackend: Asking each executor to shut down
15/12/10 15:50:55 INFO SparkDeploySchedulerBackend: Registered executor: AkkaRpcEndpointRef(Actor[akka.tcp://sparkExecutor@10.223.112.209:46567/user/Executor#1038335329]) with ID 1
15/12/10 15:50:55 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
15/12/10 15:50:55 INFO MemoryStore: MemoryStore cleared
15/12/10 15:50:55 INFO BlockManager: BlockManager stopped
15/12/10 15:50:55 INFO BlockManagerMaster: BlockManagerMaster stopped
15/12/10 15:50:55 INFO SparkContext: Successfully stopped SparkContext
15/12/10 15:50:55 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
15/12/10 15:50:55 INFO ShutdownHookManager: Shutdown hook called
15/12/10 15:50:55 INFO ShutdownHookManager: Deleting directory /data1/spark/spark-ee667b90-7be7-49f0-b542-9349c6f06224
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment