diff --git a/bin/compute-classpath.cmd b/bin/compute-classpath.cmd
index 1dff8fea220ed7c037f88cc1bcf6dbed5437ff29..6e7efbd3349870bfc47daf99b578d2261b1f3a36 100644
--- a/bin/compute-classpath.cmd
+++ b/bin/compute-classpath.cmd
@@ -6,7 +6,7 @@ rem script and the ExecutorRunner in standalone cluster mode.
 set SCALA_VERSION=2.9.3
 
 rem Figure out where the Spark framework is installed
-set FWDIR=%~dp0\..
+set FWDIR=%~dp0..\
 
 rem Load environment variables from conf\spark-env.cmd, if it exists
 if exist "%FWDIR%conf\spark-env.cmd" call "%FWDIR%conf\spark-env.cmd"
@@ -45,7 +45,7 @@ rem Add Scala standard library
 set CLASSPATH=%CLASSPATH%;%SCALA_HOME%\lib\scala-library.jar;%SCALA_HOME%\lib\scala-compiler.jar;%SCALA_HOME%\lib\jline.jar
 
 rem A bit of a hack to allow calling this script within run2.cmd without seeing output
-if "x%DONT_PRINT_CLASSPATH%"=="x1" goto exit
+if "%DONT_PRINT_CLASSPATH%"=="1" goto exit
 
 echo %CLASSPATH%
 
diff --git a/run2.cmd b/run2.cmd
index 25e4f3b57c1b58686289fabf3d6b3b4b8c5eed6f..a9c4df180ff59db12a4530b5fd4ff5125733c51e 100644
--- a/run2.cmd
+++ b/run2.cmd
@@ -48,6 +48,7 @@ set REPL_DIR=%FWDIR%repl
 rem Compute classpath using external script
 set DONT_PRINT_CLASSPATH=1
 call "%FWDIR%bin\compute-classpath.cmd"
+set DONT_PRINT_CLASSPATH=0
 
 rem Figure out the JAR file that our examples were packaged into.
 rem First search in the build path from SBT: