Integrate Spark with R
You integrate Spark with R when you want to run R programs as Spark jobs.
About this task
Procedure
-
On each node that will submit Spark jobs, install R 3.2.2 or greater:
- On Ubuntu:
apt-get install r-base-dev
- On CentOS/RedHat:
yum install R
For more information about installing R, see the R documentation.
- On Ubuntu:
-
To verify the integration, run the following commands as the mapr user or as a
user that mapr impersonates: