diff --git a/.github/PULL_REQUEST_TEMPLATE b/.github/PULL_REQUEST_TEMPLATE index 0e41cf18264533842c865982539fbe2911b8ee78..5af45d6fa7988488989dba428d3d469e3d7df50c 100644 --- a/.github/PULL_REQUEST_TEMPLATE +++ b/.github/PULL_REQUEST_TEMPLATE @@ -7,4 +7,4 @@ (Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests) (If this patch involves UI changes, please attach a screenshot; otherwise, remove this) -Please review https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark before opening a pull request. +Please review http://spark.apache.org/contributing.html before opening a pull request. diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 1a8206abe38387d6069e31b8019fe1707199d4f0..8fdd5aa9e7dfbcb30ee62d4957080a7505ac5826 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -1,12 +1,12 @@ ## Contributing to Spark *Before opening a pull request*, review the -[Contributing to Spark wiki](https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark). +[Contributing to Spark guide](http://spark.apache.org/contributing.html). It lists steps that are required before creating a PR. In particular, consider: - Is the change important and ready enough to ask the community to spend time reviewing? - Have you searched for existing, related JIRAs and pull requests? -- Is this a new feature that can stand alone as a [third party project](https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects) ? +- Is this a new feature that can stand alone as a [third party project](http://spark.apache.org/third-party-projects.html) ? - Is the change being proposed clearly explained and motivated? When you contribute code, you affirm that the contribution is your original work and that you diff --git a/R/README.md b/R/README.md index 47f9a86dfde114edfa57cd2072497cac254c7d83..4c40c5963db708a25d05441cafbaa9836d5463f5 100644 --- a/R/README.md +++ b/R/README.md @@ -51,7 +51,7 @@ sparkR.session() #### Making changes to SparkR -The [instructions](https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark) for making contributions to Spark also apply to SparkR. +The [instructions](http://spark.apache.org/contributing.html) for making contributions to Spark also apply to SparkR. If you only make R file changes (i.e. no Scala changes) then you can just re-install the R package using `R/install-dev.sh` and test your changes. Once you have made your changes, please include unit tests for them and run existing unit tests using the `R/run-tests.sh` script as described below. diff --git a/R/pkg/DESCRIPTION b/R/pkg/DESCRIPTION index fe41a9e7dabbdafa78560a025f26cbdf556649f6..981ae1246476b8a811a57a1ef4361540b900173d 100644 --- a/R/pkg/DESCRIPTION +++ b/R/pkg/DESCRIPTION @@ -11,7 +11,7 @@ Authors@R: c(person("Shivaram", "Venkataraman", role = c("aut", "cre"), email = "felixcheung@apache.org"), person(family = "The Apache Software Foundation", role = c("aut", "cph"))) URL: http://www.apache.org/ http://spark.apache.org/ -BugReports: https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark#ContributingtoSpark-ContributingBugReports +BugReports: http://spark.apache.org/contributing.html Depends: R (>= 3.0), methods diff --git a/README.md b/README.md index dd7d0e22495b3361cc2ed06899989157874f62b3..853f7f5ded3cb63feabd492a48a90b2716aee044 100644 --- a/README.md +++ b/README.md @@ -29,8 +29,9 @@ To build Spark and its example programs, run: You can build Spark using more than one thread by using the -T option with Maven, see ["Parallel builds in Maven 3"](https://cwiki.apache.org/confluence/display/MAVEN/Parallel+builds+in+Maven+3). More detailed documentation is available from the project site, at ["Building Spark"](http://spark.apache.org/docs/latest/building-spark.html). -For developing Spark using an IDE, see [Eclipse](https://cwiki.apache.org/confluence/display/SPARK/Useful+Developer+Tools#UsefulDeveloperTools-Eclipse) -and [IntelliJ](https://cwiki.apache.org/confluence/display/SPARK/Useful+Developer+Tools#UsefulDeveloperTools-IntelliJ). + +For general development tips, including info on developing Spark using an IDE, see +[http://spark.apache.org/developer-tools.html](the Useful Developer Tools page). ## Interactive Scala Shell @@ -80,7 +81,7 @@ can be run using: ./dev/run-tests Please see the guidance on how to -[run tests for a module, or individual tests](https://cwiki.apache.org/confluence/display/SPARK/Useful+Developer+Tools). +[run tests for a module, or individual tests](http://spark.apache.org/developer-tools.html#individual-tests). ## A Note About Hadoop Versions @@ -100,5 +101,5 @@ in the online documentation for an overview on how to configure Spark. ##Â Contributing -Please review the [Contribution to Spark](https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark) -wiki for information on how to get started contributing to the project. +Please review the [Contribution to Spark guide](http://spark.apache.org/contributing.html) +for information on how to get started contributing to the project. diff --git a/dev/checkstyle.xml b/dev/checkstyle.xml index 92c5251c8503713b1fbf4f6b96bc2cd5a799f393..fd73ca73ee7ef396d46b1cfe2a169c8f3c5d60f8 100644 --- a/dev/checkstyle.xml +++ b/dev/checkstyle.xml @@ -28,7 +28,7 @@ with Spark-specific changes from: - https://cwiki.apache.org/confluence/display/SPARK/Spark+Code+Style+Guide + http://spark.apache.org/contributing.html#code-style-guide Checkstyle is very configurable. Be sure to read the documentation at http://checkstyle.sf.net (or in your downloaded distribution). diff --git a/docs/_layouts/global.html b/docs/_layouts/global.html index ad5b5c9adfac878e2b9958db5891572caf20bd96..c00d0db63cd103d3b058bd0da6fbaa33955b018e 100755 --- a/docs/_layouts/global.html +++ b/docs/_layouts/global.html @@ -113,8 +113,8 @@ <li><a href="hardware-provisioning.html">Hardware Provisioning</a></li> <li class="divider"></li> <li><a href="building-spark.html">Building Spark</a></li> - <li><a href="https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark">Contributing to Spark</a></li> - <li><a href="https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects">Third Party Projects</a></li> + <li><a href="http://spark.apache.org/contributing.html">Contributing to Spark</a></li> + <li><a href="http://spark.apache.org/third-party-projects.html">Third Party Projects</a></li> </ul> </li> </ul> diff --git a/docs/building-spark.md b/docs/building-spark.md index 88da0cc9c3bbfbb01ebfc72dbb97c63b11d8cedc..65c2895b29b10cad8c67cadd263984eede0be6b6 100644 --- a/docs/building-spark.md +++ b/docs/building-spark.md @@ -197,7 +197,7 @@ can be set to control the SBT build. For example: To avoid the overhead of launching sbt each time you need to re-compile, you can launch sbt in interactive mode by running `build/sbt`, and then run all build commands at the command prompt. For more recommendations on reducing build time, refer to the -[wiki page](https://cwiki.apache.org/confluence/display/SPARK/Useful+Developer+Tools#UsefulDeveloperTools-ReducingBuildTimes). +[Useful Developer Tools page](http://spark.apache.org/developer-tools.html). ##Â Encrypted Filesystems @@ -215,7 +215,7 @@ to the `sharedSettings` val. See also [this PR](https://github.com/apache/spark/ ## IntelliJ IDEA or Eclipse For help in setting up IntelliJ IDEA or Eclipse for Spark development, and troubleshooting, refer to the -[wiki page for IDE setup](https://cwiki.apache.org/confluence/display/SPARK/Useful+Developer+Tools#UsefulDeveloperTools-IDESetup). +[Useful Developer Tools page](http://spark.apache.org/developer-tools.html). # Running Tests diff --git a/docs/contributing-to-spark.md b/docs/contributing-to-spark.md index ef1b3ad6da57a27da4ae0face2e9237a58cf0d1e..9252545e4a129f381058de7618f9dcf7a6d3eb95 100644 --- a/docs/contributing-to-spark.md +++ b/docs/contributing-to-spark.md @@ -5,4 +5,4 @@ title: Contributing to Spark The Spark team welcomes all forms of contributions, including bug reports, documentation or patches. For the newest information on how to contribute to the project, please read the -[wiki page on contributing to Spark](https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark). +[Contributing to Spark guide](http://spark.apache.org/contributing.html). diff --git a/docs/index.md b/docs/index.md index 39de11de854a7aca568e6b23ac7a7e11c015ef8c..c5d34cb5c4e73f28d55296c660904681d4cdfff4 100644 --- a/docs/index.md +++ b/docs/index.md @@ -125,8 +125,8 @@ options for deployment: * Integration with other storage systems: * [OpenStack Swift](storage-openstack-swift.html) * [Building Spark](building-spark.html): build Spark using the Maven system -* [Contributing to Spark](https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark) -* [Third Party Projects](https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects): related third party Spark projects +* [Contributing to Spark](http://spark.apache.org/contributing.html) +* [Third Party Projects](http://spark.apache.org/third-party-projects.html): related third party Spark projects **External Resources:** diff --git a/docs/sparkr.md b/docs/sparkr.md index f30bd4026fed302fcc5f73203edde630b24c44a1..d26949226b11715d079fe9b999405b53d66e0590 100644 --- a/docs/sparkr.md +++ b/docs/sparkr.md @@ -126,7 +126,7 @@ head(df) SparkR supports operating on a variety of data sources through the `SparkDataFrame` interface. This section describes the general methods for loading and saving data using Data Sources. You can check the Spark SQL programming guide for more [specific options](sql-programming-guide.html#manually-specifying-options) that are available for the built-in data sources. The general method for creating SparkDataFrames from data sources is `read.df`. This method takes in the path for the file to load and the type of data source, and the currently active SparkSession will be used automatically. -SparkR supports reading JSON, CSV and Parquet files natively, and through packages available from sources like [Third Party Projects](https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects), you can find data source connectors for popular file formats like Avro. These packages can either be added by +SparkR supports reading JSON, CSV and Parquet files natively, and through packages available from sources like [Third Party Projects](http://spark.apache.org/third-party-projects.html), you can find data source connectors for popular file formats like Avro. These packages can either be added by specifying `--packages` with `spark-submit` or `sparkR` commands, or if initializing SparkSession with `sparkPackages` parameter when in an interactive R shell or from RStudio. <div data-lang="r" markdown="1"> diff --git a/docs/streaming-programming-guide.md b/docs/streaming-programming-guide.md index 18fc1cd9348268b6a68645ffd24e6359b6fb3c57..1fcd198685a5154f797d0ce22c820f4d4b09e27b 100644 --- a/docs/streaming-programming-guide.md +++ b/docs/streaming-programming-guide.md @@ -2382,7 +2382,7 @@ additional effort may be necessary to achieve exactly-once semantics. There are - [Kafka Integration Guide](streaming-kafka-integration.html) - [Kinesis Integration Guide](streaming-kinesis-integration.html) - [Custom Receiver Guide](streaming-custom-receivers.html) -* Third-party DStream data sources can be found in [Third Party Projects](https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects) +* Third-party DStream data sources can be found in [Third Party Projects](http://spark.apache.org/third-party-projects.html) * API documentation - Scala docs * [StreamingContext](api/scala/index.html#org.apache.spark.streaming.StreamingContext) and diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala index cfee7be1e3f07aa4b57a66329faf7487ce234c0a..84fde0bbf9268f26831cc0626d97e9137eaee604 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala @@ -505,12 +505,11 @@ object DataSource { provider1 == "com.databricks.spark.avro") { throw new AnalysisException( s"Failed to find data source: ${provider1.toLowerCase}. Please find an Avro " + - "package at " + - "https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects") + "package at http://spark.apache.org/third-party-projects.html") } else { throw new ClassNotFoundException( s"Failed to find data source: $provider1. Please find packages at " + - "https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects", + "http://spark.apache.org/third-party-projects.html", error) } }