aboutsummaryrefslogtreecommitdiff
path: root/dev/audit-release/sbt_app_sql
diff options
context:
space:
mode:
authorTathagata Das <tathagata.das1565@gmail.com>2014-05-22 20:48:55 -0700
committerTathagata Das <tathagata.das1565@gmail.com>2014-05-22 20:48:55 -0700
commitb2bdd0e505f1ae3d39c46139f17bd43779ece635 (patch)
treefd1f6274986c3bf259c0dd3a0adaf6c2cfddc1a1 /dev/audit-release/sbt_app_sql
parentcce77457e00aa5f1f4db3d50454cf257efb156ed (diff)
downloadspark-b2bdd0e505f1ae3d39c46139f17bd43779ece635.tar.gz
spark-b2bdd0e505f1ae3d39c46139f17bd43779ece635.tar.bz2
spark-b2bdd0e505f1ae3d39c46139f17bd43779ece635.zip
Updated scripts for auditing releases
- Added script to automatically generate change list CHANGES.txt - Added test for verifying linking against maven distributions of `spark-sql` and `spark-hive` - Added SBT projects for testing functionality of `spark-sql` and `spark-hive` - Fixed issues in existing tests that might have come up because of changes in Spark 1.0 Author: Tathagata Das <tathagata.das1565@gmail.com> Closes #844 from tdas/update-dev-scripts and squashes the following commits: 25090ba [Tathagata Das] Added missing license e2e20b3 [Tathagata Das] Updated tests for auditing releases.
Diffstat (limited to 'dev/audit-release/sbt_app_sql')
-rw-r--r--dev/audit-release/sbt_app_sql/build.sbt29
-rw-r--r--dev/audit-release/sbt_app_sql/src/main/scala/SqlApp.scala57
2 files changed, 86 insertions, 0 deletions
diff --git a/dev/audit-release/sbt_app_sql/build.sbt b/dev/audit-release/sbt_app_sql/build.sbt
new file mode 100644
index 0000000000..6e0ad3b4b2
--- /dev/null
+++ b/dev/audit-release/sbt_app_sql/build.sbt
@@ -0,0 +1,29 @@
+//
+// Licensed to the Apache Software Foundation (ASF) under one or more
+// contributor license agreements. See the NOTICE file distributed with
+// this work for additional information regarding copyright ownership.
+// The ASF licenses this file to You under the Apache License, Version 2.0
+// (the "License"); you may not use this file except in compliance with
+// the License. You may obtain a copy of the License at
+//
+// http://www.apache.org/licenses/LICENSE-2.0
+//
+// Unless required by applicable law or agreed to in writing, software
+// distributed under the License is distributed on an "AS IS" BASIS,
+// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+// See the License for the specific language governing permissions and
+// limitations under the License.
+//
+
+name := "Simple Project"
+
+version := "1.0"
+
+scalaVersion := System.getenv.get("SCALA_VERSION")
+
+libraryDependencies += "org.apache.spark" %% "spark-sql" % System.getenv.get("SPARK_VERSION")
+
+resolvers ++= Seq(
+ "Spark Release Repository" at System.getenv.get("SPARK_RELEASE_REPOSITORY"),
+ "Akka Repository" at "http://repo.akka.io/releases/",
+ "Spray Repository" at "http://repo.spray.cc/")
diff --git a/dev/audit-release/sbt_app_sql/src/main/scala/SqlApp.scala b/dev/audit-release/sbt_app_sql/src/main/scala/SqlApp.scala
new file mode 100644
index 0000000000..50af90c213
--- /dev/null
+++ b/dev/audit-release/sbt_app_sql/src/main/scala/SqlApp.scala
@@ -0,0 +1,57 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package main.scala
+
+import scala.collection.mutable.{ListBuffer, Queue}
+
+import org.apache.spark.SparkConf
+import org.apache.spark.SparkContext
+import org.apache.spark.rdd.RDD
+import org.apache.spark.sql.SQLContext
+
+case class Person(name: String, age: Int)
+
+object SparkSqlExample {
+
+ def main(args: Array[String]) {
+ val conf = sys.env.get("SPARK_AUDIT_MASTER") match {
+ case Some(master) => new SparkConf().setAppName("Simple Sql App").setMaster(master)
+ case None => new SparkConf().setAppName("Simple Sql App")
+ }
+ val sc = new SparkContext(conf)
+ val sqlContext = new SQLContext(sc)
+
+ import sqlContext._
+ val people = sc.makeRDD(1 to 100, 10).map(x => Person(s"Name$x", x))
+ people.registerAsTable("people")
+ val teenagers = sql("SELECT name FROM people WHERE age >= 13 AND age <= 19")
+ val teenagerNames = teenagers.map(t => "Name: " + t(0)).collect()
+ teenagerNames.foreach(println)
+
+ def test(f: => Boolean, failureMsg: String) = {
+ if (!f) {
+ println(failureMsg)
+ System.exit(-1)
+ }
+ }
+
+ test(teenagerNames.size == 7, "Unexpected number of selected elements: " + teenagerNames)
+ println("Test succeeded")
+ sc.stop()
+ }
+}