
R STUDIO FOR LOOP HOW TO
Specific documentation for your environment will point out how to retrieve these logs. It is worth mentioning that different cluster providers and platforms expose worker logs in different ways. Spark_apply(iris_tbl, function(e) stop( "Make this fail")) For instance, the following R code causes the distributed execution to fail and suggests you check the logs for details. It can be more difficult to troubleshoot R issues in a cluster than in local mode. Note: packages are not copied in local mode ( master="local") because the packages already exist on the system. You can disable package distribution by setting packages = FALSE. It’s not uncommon for R libraries to be several gigabytes in size, so be prepared for a one-time tax while the R packages are copied over to your Spark cluster. Packages will only be copied once and will persist as long as the connection remains open. libPaths() will be copied into each Spark worker node via the SparkConf.addFile() function.

The first time you call spark_apply() all of the contents in your local. To use R packages inside Spark, your packages must be installed on the worker nodes.
