Skip to content
Snippets Groups Projects
  1. Mar 09, 2014
    • Patrick Wendell's avatar
      SPARK-782 Clean up for ASM dependency. · b9be1609
      Patrick Wendell authored
      This makes two changes.
      
      1) Spark uses the shaded version of asm that is (conveniently) published
         with Kryo.
      2) Existing exclude rules around asm are updated to reflect the new groupId
         of `org.ow2.asm`. This made all of the old rules not work with newer Hadoop
         versions that pull in new asm versions.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #100 from pwendell/asm and squashes the following commits:
      
      9235f3f [Patrick Wendell] SPARK-782 Clean up for ASM dependency.
      b9be1609
  2. Mar 08, 2014
    • Sandy Ryza's avatar
      SPARK-1193. Fix indentation in pom.xmls · a99fb374
      Sandy Ryza authored
      Author: Sandy Ryza <sandy@cloudera.com>
      
      Closes #91 from sryza/sandy-spark-1193 and squashes the following commits:
      
      a878124 [Sandy Ryza] SPARK-1193. Fix indentation in pom.xmls
      a99fb374
  3. Mar 06, 2014
    • Thomas Graves's avatar
      SPARK-1189: Add Security to Spark - Akka, Http, ConnectionManager, UI use servlets · 7edbea41
      Thomas Graves authored
      resubmit pull request.  was https://github.com/apache/incubator-spark/pull/332.
      
      Author: Thomas Graves <tgraves@apache.org>
      
      Closes #33 from tgravescs/security-branch-0.9-with-client-rebase and squashes the following commits:
      
      dfe3918 [Thomas Graves] Fix merge conflict since startUserClass now using runAsUser
      05eebed [Thomas Graves] Fix dependency lost in upmerge
      d1040ec [Thomas Graves] Fix up various imports
      05ff5e0 [Thomas Graves] Fix up imports after upmerging to master
      ac046b3 [Thomas Graves] Merge remote-tracking branch 'upstream/master' into security-branch-0.9-with-client-rebase
      13733e1 [Thomas Graves] Pass securityManager and SparkConf around where we can. Switch to use sparkConf for reading config whereever possible. Added ConnectionManagerSuite unit tests.
      4a57acc [Thomas Graves] Change UI createHandler routines to createServlet since they now return servlets
      2f77147 [Thomas Graves] Rework from comments
      50dd9f2 [Thomas Graves] fix header in SecurityManager
      ecbfb65 [Thomas Graves] Fix spacing and formatting
      b514bec [Thomas Graves] Fix reference to config
      ed3d1c1 [Thomas Graves] Add security.md
      6f7ddf3 [Thomas Graves] Convert SaslClient and SaslServer to scala, change spark.authenticate.ui to spark.ui.acls.enable, and fix up various other things from review comments
      2d9e23e [Thomas Graves] Merge remote-tracking branch 'upstream/master' into security-branch-0.9-with-client-rebase_rework
      5721c5a [Thomas Graves] update AkkaUtilsSuite test for the actorSelection changes, fix typos based on comments, and remove extra lines I missed in rebase from AkkaUtils
      f351763 [Thomas Graves] Add Security to Spark - Akka, Http, ConnectionManager, UI to use servlets
      7edbea41
  4. Mar 02, 2014
    • Patrick Wendell's avatar
      SPARK-1121: Include avro for yarn-alpha builds · c3f5e075
      Patrick Wendell authored
      This lets us explicitly include Avro based on a profile for 0.23.X
      builds. It makes me sad how convoluted it is to express this logic
      in Maven. @tgraves and @sryza curious if this works for you.
      
      I'm also considering just reverting to how it was before. The only
      real problem was that Spark advertised a dependency on Avro
      even though it only really depends transitively on Avro through
      other deps.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #49 from pwendell/avro-build-fix and squashes the following commits:
      
      8d6ee92 [Patrick Wendell] SPARK-1121: Add avro to yarn-alpha profile
      c3f5e075
    • Patrick Wendell's avatar
      Remove remaining references to incubation · 1fd2bfd3
      Patrick Wendell authored
      This removes some loose ends not caught by the other (incubating -> tlp) patches. @markhamstra this updates the version as you mentioned earlier.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #51 from pwendell/tlp and squashes the following commits:
      
      d553b1b [Patrick Wendell] Remove remaining references to incubation
      1fd2bfd3
  5. Feb 27, 2014
  6. Feb 23, 2014
    • Sean Owen's avatar
      SPARK-1071: Tidy logging strategy and use of log4j · c0ef3afa
      Sean Owen authored
      Prompted by a recent thread on the mailing list, I tried and failed to see if Spark can be made independent of log4j. There are a few cases where control of the underlying logging is pretty useful, and to do that, you have to bind to a specific logger.
      
      Instead I propose some tidying that leaves Spark's use of log4j, but gets rid of warnings and should still enable downstream users to switch. The idea is to pipe everything (except log4j) through SLF4J, and have Spark use SLF4J directly when logging, and where Spark needs to output info (REPL and tests), bind from SLF4J to log4j.
      
      This leaves the same behavior in Spark. It means that downstream users who want to use something except log4j should:
      
      - Exclude dependencies on log4j, slf4j-log4j12 from Spark
      - Include dependency on log4j-over-slf4j
      - Include dependency on another logger X, and another slf4j-X
      - Recreate any log config that Spark does, that is needed, in the other logger's config
      
      That sounds about right.
      
      Here are the key changes:
      
      - Include the jcl-over-slf4j shim everywhere by depending on it in core.
      - Exclude dependencies on commons-logging from third-party libraries.
      - Include the jul-to-slf4j shim everywhere by depending on it in core.
      - Exclude slf4j-* dependencies from third-party libraries to prevent collision or warnings
      - Added missing slf4j-log4j12 binding to GraphX, Bagel module tests
      
      And minor/incidental changes:
      
      - Update to SLF4J 1.7.5, which happily matches Hadoop 2’s version and is a recommended update over 1.7.2
      - (Remove a duplicate HBase dependency declaration in SparkBuild.scala)
      - (Remove a duplicate mockito dependency declaration that was causing warnings and bugging me)
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #570 from srowen/SPARK-1071 and squashes the following commits:
      
      52eac9f [Sean Owen] Add slf4j-over-log4j12 dependency to core (non-test) and remove it from things that depend on core.
      77a7fa9 [Sean Owen] SPARK-1071: Tidy logging strategy and use of log4j
      c0ef3afa
  7. Feb 17, 2014
    • CodingCat's avatar
      [SPARK-1090] improvement on spark_shell (help information, configure memory) · e0d49ad2
      CodingCat authored
      https://spark-project.atlassian.net/browse/SPARK-1090
      
      spark-shell should print help information about parameters and should allow user to configure exe memory
      there is no document about hot to set --cores/-c in spark-shell
      
      and also
      
      users should be able to set executor memory through command line options
      
      In this PR I also check the format of the options passed by the user
      
      Author: CodingCat <zhunansjtu@gmail.com>
      
      Closes #599 from CodingCat/spark_shell_improve and squashes the following commits:
      
      de5aa38 [CodingCat] add parameter to set driver memory
      915cbf8 [CodingCat] improvement on spark_shell (help information, configure memory)
      e0d49ad2
  8. Feb 09, 2014
    • Patrick Wendell's avatar
      Merge pull request #557 from ScrapCodes/style. Closes #557. · b69f8b2a
      Patrick Wendell authored
      SPARK-1058, Fix Style Errors and Add Scala Style to Spark Build.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      Author: Prashant Sharma <scrapcodes@gmail.com>
      
      == Merge branch commits ==
      
      commit 1a8bd1c059b842cb95cc246aaea74a79fec684f4
      Author: Prashant Sharma <scrapcodes@gmail.com>
      Date:   Sun Feb 9 17:39:07 2014 +0530
      
          scala style fixes
      
      commit f91709887a8e0b608c5c2b282db19b8a44d53a43
      Author: Patrick Wendell <pwendell@gmail.com>
      Date:   Fri Jan 24 11:22:53 2014 -0800
      
          Adding scalastyle snapshot
      b69f8b2a
  9. Feb 08, 2014
    • Mark Hamstra's avatar
      Merge pull request #542 from markhamstra/versionBump. Closes #542. · c2341c92
      Mark Hamstra authored
      Version number to 1.0.0-SNAPSHOT
      
      Since 0.9.0-incubating is done and out the door, we shouldn't be building 0.9.0-incubating-SNAPSHOT anymore.
      
      @pwendell
      
      Author: Mark Hamstra <markhamstra@gmail.com>
      
      == Merge branch commits ==
      
      commit 1b00a8a7c1a7f251b4bb3774b84b9e64758eaa71
      Author: Mark Hamstra <markhamstra@gmail.com>
      Date:   Wed Feb 5 09:30:32 2014 -0800
      
          Version number to 1.0.0-SNAPSHOT
      c2341c92
  10. Jan 14, 2014
  11. Jan 12, 2014
  12. Jan 10, 2014
    • Ankur Dave's avatar
      Revert GraphX changes to SparkILoopInit · 0ca18b8b
      Ankur Dave authored
      The changes were to support a custom banner in spark-shell for use by
      graphx-shell, but once GraphX is merged into Spark, a separate shell
      will be unnecessary.
      0ca18b8b
  13. Jan 07, 2014
  14. Jan 03, 2014
    • Luca Rosellini's avatar
      Added ‘-i’ command line option to spark REPL. · 0b6db8c1
      Luca Rosellini authored
      We had to create a new implementation of both scala.tools.nsc.CompilerCommand and scala.tools.nsc.Settings, because using scala.tools.nsc.GenericRunnerSettings would bring in other options (-howtorun, -save and -execute) which don’t make sense in Spark.
      Any new Spark specific command line option could now be added to org.apache.spark.repl.SparkRunnerSettings class.
      
      Since the behavior of loading a script from the command line should be the same as loading it using the “:load” command inside the shell, the script should be loaded when the SparkContext is available, that’s why we had to move the call to ‘loadfiles(settings)’ _after_ the call to postInitialization(). This still doesn’t work if ‘isAsync = true’.
      0b6db8c1
    • Prashant Sharma's avatar
      fixed review comments · 94f2fffa
      Prashant Sharma authored
      94f2fffa
  15. Jan 01, 2014
  16. Dec 31, 2013
  17. Dec 30, 2013
    • Patrick Wendell's avatar
      SPARK-1008: Logging improvments · cffe1c1d
      Patrick Wendell authored
      1. Adds a default log4j file that gets loaded if users haven't specified a log4j file.
      2. Isolates use of the tools assembly jar. I found this produced SLF4J warnings
         after building with SBT (and I've seen similar warnings on the mailing list).
      cffe1c1d
  18. Dec 28, 2013
    • Matei Zaharia's avatar
      Various fixes to configuration code · 642029e7
      Matei Zaharia authored
      - Got rid of global SparkContext.globalConf
      - Pass SparkConf to serializers and compression codecs
      - Made SparkConf public instead of private[spark]
      - Improved API of SparkContext and SparkConf
      - Switched executor environment vars to be passed through SparkConf
      - Fixed some places that were still using system properties
      - Fixed some tests, though others are still failing
      
      This still fails several tests in core, repl and streaming, likely due
      to properties not being set or cleared correctly (some of the tests run
      fine in isolation).
      642029e7
  19. Dec 24, 2013
  20. Dec 15, 2013
  21. Dec 13, 2013
  22. Dec 10, 2013
  23. Dec 07, 2013
  24. Nov 26, 2013
  25. Nov 15, 2013
    • Aaron Davidson's avatar
      Various merge corrections · f629ba95
      Aaron Davidson authored
      I've diff'd this patch against my own -- since they were both created
      independently, this means that two sets of eyes have gone over all the
      merge conflicts that were created, so I'm feeling significantly more
      confident in the resulting PR.
      
      @rxin has looked at the changes to the repl and is resoundingly
      confident that they are correct.
      f629ba95
  26. Nov 09, 2013
  27. Nov 04, 2013
    • Joseph E. Gonzalez's avatar
      This commit adds a new graphx-shell which is essentially the same as · 3c37928f
      Joseph E. Gonzalez authored
      the spark shell but with GraphX packages automatically imported
      and with Kryo serialization enabled for GraphX types.
      
      In addition the graphx-shell has a nifty new logo.
      
      To make these changes minimally invasive in the SparkILoop.scala
      I added some additional environment variables:
      
         SPARK_BANNER_TEXT: If set this string is displayed instead
         of the spark logo
      
         SPARK_SHELL_INIT_BLOCK: if set this expression is evaluated in the
         spark shell after the spark context is created.
      3c37928f
  28. Oct 24, 2013
  29. Oct 17, 2013
    • Aaron Davidson's avatar
      Spark shell exits if it cannot create SparkContext · 74737264
      Aaron Davidson authored
      Mainly, this occurs if you provide a messed up MASTER url (one that doesn't match one
      of our regexes). Previously, we would default to Mesos, fail, and then start the shell
      anyway, except that any Spark command would fail.
      74737264
  30. Oct 12, 2013
  31. Oct 06, 2013
  32. Sep 26, 2013
  33. Sep 24, 2013
  34. Sep 15, 2013
  35. Sep 10, 2013
Loading