반응형
Scala spark-submit 이용해서 실행 시키기
1. jar 파일 만들기
1) sbt가 설치되어 있어야 함
echo "deb https://dl.bintray.com/sbt/debian /" | sudo tee -a /etc/apt/sources.list.d/sbt.list
sudo apt-key adv --keyserver hkp://keyserver.ubuntu.com:80 --recv 2EE0EA64E40A89B84B2DF73499E82A75642AC823
sudo apt-get update
sudo apt-get install sbt
2) 컴파일 (자세한 것은 여기서 : 링크)
/* SimpleApp.scala */
import org.apache.spark.SparkContext
import org.apache.spark.SparkContext._
import org.apache.spark.SparkConf
object SimpleApp {
def main(args: Array[String]) {
val logFile = "YOUR_SPARK_HOME/README.md" // Should be some file on your system
val conf = new SparkConf().setAppName("Simple Application")
val sc = new SparkContext(conf)
val logData = sc.textFile(logFile, 2).cache()
val numAs = logData.filter(line => line.contains("a")).count()
val numBs = logData.filter(line => line.contains("b")).count()
println("Lines with a: %s, Lines with b: %s".format(numAs, numBs))
}
}
name := "Simple Project"
version := "1.0"
scalaVersion := "2.11.7"
libraryDependencies += "org.apache.spark" %% "spark-core" % "2.0.1"
# Your directory layout should look like this
$ find .
.
./simple.sbt
./src
./src/main
./src/main/scala
./src/main/scala/SimpleApp.scala
# Package a jar containing your application
$ sbt package
...
[info] Packaging {..}/{..}/target/scala-2.11/simple-project_2.11-1.0.jar
# Use spark-submit to run your application
$ YOUR_SPARK_HOME/bin/spark-submit \
--class "SimpleApp" \
--master local[4] \
target/scala-2.11/simple-project_2.11-1.0.jar
...
Lines with a: 46, Lines with b: 23
2. 실행 시키기
spark-submit --class "SimpleApp" --master yarn target/scala-2.10/selector_2.10-1.0.jar test_argument
반응형
'Programming > Spark, Scala' 카테고리의 다른 글
HDP 2.5 설치 (0) | 2016.11.15 |
---|---|
XGBoost build (0) | 2016.11.02 |
Scala 실행 방법, python 프로그램 spark에서 실행 (0) | 2016.11.01 |
Scala Spark - error : org.apache.spark.sql.SQLContext.sql (0) | 2016.11.01 |
Spark 1.6 Feature Importances (0) | 2016.10.31 |