- Apr 15, 2014
-
-
Sandeep authored
Currently it requires NumPy 1.7 due to using the copyto method (http://docs.scipy.org/doc/numpy/reference/generated/numpy.copyto.html) for extracting data out of an array. Replace it with a fallback Author: Sandeep <sandeep@techaddict.me> Closes #391 from techaddict/1426 and squashes the following commits: d365962 [Sandeep] SPARK-1426: Make MLlib work with NumPy versions older than 1.7 Currently it requires NumPy 1.7 due to using the copyto method (http://docs.scipy.org/doc/numpy/reference/generated/numpy.copyto.html) for extracting data out of an array. Replace it with a fallback
-
- Apr 07, 2014
-
-
Aaron Davidson authored
This is the default mode for running spark-shell and pyspark, intended to allow users running spark for the first time to see the performance benefits of using multiple cores, while not breaking backwards compatibility for users who use "local" mode and expect exactly 1 core. Author: Aaron Davidson <aaron@databricks.com> Closes #182 from aarondav/110 and squashes the following commits: a88294c [Aaron Davidson] Rebased changes for new spark-shell a9f393e [Aaron Davidson] SPARK-1099: Introduce local[*] mode to infer number of cores
-
- Apr 05, 2014
-
-
Matei Zaharia authored
The reason it wasn't working was passing a bytearray to stream.write(), which is not supported in Python 2.6 but is in 2.7. (This array came from NumPy when we converted data to send it over to Java). Now we just convert those bytearrays to strings of bytes, which preserves nonprintable characters as well. Author: Matei Zaharia <matei@databricks.com> Closes #335 from mateiz/mllib-python-2.6 and squashes the following commits: f26c59f [Matei Zaharia] Update docs to no longer say we need Python 2.7 a84d6af [Matei Zaharia] SPARK-1421. Make MLlib work on Python 2.6
-
- Mar 13, 2014
-
-
Sandy Ryza authored
Author: Sandy Ryza <sandy@cloudera.com> Closes #120 from sryza/sandy-spark-1183 and squashes the following commits: 5066a4a [Sandy Ryza] Remove "worker" in a couple comments 0bd1e46 [Sandy Ryza] Remove --am-class from usage bfc8fe0 [Sandy Ryza] Remove am-class from doc and fix yarn-alpha 607539f [Sandy Ryza] Address review comments 74d087a [Sandy Ryza] SPARK-1183. Don't use "worker" to mean executor
-
- Feb 26, 2014
-
-
Jyotiska NK authored
Author: Jyotiska NK <jyotiska123@gmail.com> Closes #22 from jyotiska/pyspark_docs and squashes the following commits: 426136c [Jyotiska NK] Updated link for pyspark examples
-
- Jan 15, 2014
-
-
Matei Zaharia authored
-
- Jan 12, 2014
-
-
Matei Zaharia authored
-
- Jan 07, 2014
-
-
Patrick Wendell authored
This patch removes compatibility for IPython < 1.0 but fixes the launch script and makes it much simpler. I tested this using the three commands in the PySpark documentation page: 1. IPYTHON=1 ./pyspark 2. IPYTHON_OPTS="notebook" ./pyspark 3. IPYTHON_OPTS="notebook --pylab inline" ./pyspark There are two changes: - We rely on PYTHONSTARTUP env var to start PySpark - Removed the quotes around $IPYTHON_OPTS... having quotes gloms them together as a single argument passed to `exec` which seemed to cause ipython to fail (it instead expects them as multiple arguments).
-
- Jan 06, 2014
-
-
Holden Karau authored
-
- Jan 02, 2014
-
-
Prashant Sharma authored
-
Prashant Sharma authored
-
- Dec 30, 2013
-
-
Matei Zaharia authored
-
- Oct 22, 2013
-
-
Ewen Cheslack-Postava authored
-
- Oct 09, 2013
-
-
Matei Zaharia authored
-
- Sep 10, 2013
-
-
Matei Zaharia authored
-
- Sep 08, 2013
-
-
Matei Zaharia authored
Also changed uses of "job" terminology to "application" when they referred to an entire Spark program, to avoid confusion.
-
- Sep 02, 2013
-
-
Matei Zaharia authored
-
- Sep 01, 2013
-
-
Matei Zaharia authored
-
- Aug 31, 2013
-
-
Matei Zaharia authored
and new Python stuff
-
- Aug 29, 2013
-
-
Matei Zaharia authored
are now needed
-
- Jul 29, 2013
-
-
Matei Zaharia authored
-
- Jul 01, 2013
-
-
root authored
-
- Jun 26, 2013
-
-
Matei Zaharia authored
-
- Feb 25, 2013
-
-
Matei Zaharia authored
-
- Feb 18, 2013
-
-
Tathagata Das authored
Added checkpointing and fault-tolerance semantics to the programming guide. Fixed default checkpoint interval to being a multiple of slide duration. Fixed visibility of some classes and objects to clean up docs.
-
- Jan 30, 2013
-
-
Patrick Wendell authored
Also, adds a line in doc explaining how to use.
-
Patrick Wendell authored
It's nicer if all the commands you need are made explicit.
-
- Jan 20, 2013
-
-
Matei Zaharia authored
-
- Jan 08, 2013
-
-
Josh Rosen authored
-
- Jan 01, 2013
-
-
Josh Rosen authored
Expand the PySpark programming guide.
-
Josh Rosen authored
-
- Dec 29, 2012
-
-
Josh Rosen authored
-