Skip to content

Commit

Permalink
Merge pull request apache#175 from falaki/docfix
Browse files Browse the repository at this point in the history
Minor documentation cleanup
  • Loading branch information
shivaram committed Feb 19, 2015
2 parents 0981dff + ba2b72b commit 17eda4c
Showing 1 changed file with 5 additions and 10 deletions.
15 changes: 5 additions & 10 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -13,12 +13,6 @@ SparkR requires Scala 2.10 and Spark version >= 0.9.0. Current build by default
Apache Spark 1.1.0. You can also build SparkR against a
different Spark version (>= 0.9.0) by modifying `pkg/src/build.sbt`.

SparkR also requires the R package `rJava` to be installed. To install `rJava`,
you can run the following command in R:

install.packages("rJava")


### Package installation
To develop SparkR, you can build the scala package and the R package using

Expand All @@ -31,9 +25,9 @@ If you wish to try out the package directly from github, you can use [`install_g

SparkR by default uses Apache Spark 1.1.0. You can switch to a different Spark
version by setting the environment variable `SPARK_VERSION`. For example, to
use Apache Spark 1.2.0, you can run
use Apache Spark 1.3.0, you can run

SPARK_VERSION=1.2.0 ./install-dev.sh
SPARK_VERSION=1.3.0 ./install-dev.sh

SparkR by default links to Hadoop 1.0.4. To use SparkR with other Hadoop
versions, you will need to rebuild SparkR with the same version that [Spark is
Expand Down Expand Up @@ -97,8 +91,9 @@ To run one of them, use `./sparkR <filename> <args>`. For example:

./sparkR examples/pi.R local[2]

You can also run the unit-tests for SparkR by running
You can also run the unit-tests for SparkR by running (you need to install the [testthat](http://cran.r-project.org/web/packages/testthat/index.html) package first):

R -e 'install.packages("testthat", repos="http://cran.us.r-project.org")'
./run-tests.sh

## Running on EC2
Expand All @@ -110,7 +105,7 @@ Instructions for running SparkR on EC2 can be found in the
Currently, SparkR supports running on YARN with the `yarn-client` mode. These steps show how to build SparkR with YARN support and run SparkR programs on a YARN cluster:

```
# assumes Java, R, rJava, yarn, spark etc. are installed on the whole cluster.
# assumes Java, R, yarn, spark etc. are installed on the whole cluster.
cd SparkR-pkg/
USE_YARN=1 SPARK_YARN_VERSION=2.4.0 SPARK_HADOOP_VERSION=2.4.0 ./install-dev.sh
```
Expand Down

0 comments on commit 17eda4c

Please sign in to comment.