Skip to content
Snippets Groups Projects
Commit db9b90fd authored by Matei Zaharia's avatar Matei Zaharia
Browse files

Change version to 0.7.1-SNAPSHOT for development branch

parent 4f840f4e
No related branches found
No related tags found
No related merge requests found
...@@ -3,8 +3,8 @@ markdown: kramdown ...@@ -3,8 +3,8 @@ markdown: kramdown
# These allow the documentation to be updated with nerw releases # These allow the documentation to be updated with nerw releases
# of Spark, Scala, and Mesos. # of Spark, Scala, and Mesos.
SPARK_VERSION: 0.7.0 SPARK_VERSION: 0.7.1-SNAPSHOT
SPARK_VERSION_SHORT: 0.7.0 SPARK_VERSION_SHORT: 0.7.1
SCALA_VERSION: 2.9.2 SCALA_VERSION: 2.9.2
MESOS_VERSION: 0.9.0-incubating MESOS_VERSION: 0.9.0-incubating
SPARK_ISSUE_TRACKER_URL: https://spark-project.atlassian.net SPARK_ISSUE_TRACKER_URL: https://spark-project.atlassian.net
...@@ -3,7 +3,7 @@ ...@@ -3,7 +3,7 @@
<modelVersion>4.0.0</modelVersion> <modelVersion>4.0.0</modelVersion>
<groupId>org.spark-project</groupId> <groupId>org.spark-project</groupId>
<artifactId>parent</artifactId> <artifactId>parent</artifactId>
<version>0.7.0</version> <version>0.7.1-SNAPSHOT</version>
<packaging>pom</packaging> <packaging>pom</packaging>
<name>Spark Project Parent POM</name> <name>Spark Project Parent POM</name>
<url>http://spark-project.org/</url> <url>http://spark-project.org/</url>
......
...@@ -35,7 +35,7 @@ object SparkBuild extends Build { ...@@ -35,7 +35,7 @@ object SparkBuild extends Build {
def sharedSettings = Defaults.defaultSettings ++ Seq( def sharedSettings = Defaults.defaultSettings ++ Seq(
organization := "org.spark-project", organization := "org.spark-project",
version := "0.7.0", version := "0.7.1-SNAPSHOT",
scalaVersion := "2.9.2", scalaVersion := "2.9.2",
scalacOptions := Seq(/*"-deprecation",*/ "-unchecked", "-optimize"), // -deprecation is too noisy due to usage of old Hadoop API, enable it once that's no longer an issue scalacOptions := Seq(/*"-deprecation",*/ "-unchecked", "-optimize"), // -deprecation is too noisy due to usage of old Hadoop API, enable it once that's no longer an issue
unmanagedJars in Compile <<= baseDirectory map { base => (base / "lib" ** "*.jar").classpath }, unmanagedJars in Compile <<= baseDirectory map { base => (base / "lib" ** "*.jar").classpath },
......
...@@ -200,7 +200,7 @@ class SparkILoop(in0: Option[BufferedReader], val out: PrintWriter, val master: ...@@ -200,7 +200,7 @@ class SparkILoop(in0: Option[BufferedReader], val out: PrintWriter, val master:
____ __ ____ __
/ __/__ ___ _____/ /__ / __/__ ___ _____/ /__
_\ \/ _ \/ _ `/ __/ '_/ _\ \/ _ \/ _ `/ __/ '_/
/___/ .__/\_,_/_/ /_/\_\ version 0.7.0 /___/ .__/\_,_/_/ /_/\_\ version 0.7.1
/_/ /_/
""") """)
import Properties._ import Properties._
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment