(Deprecated) Initialize a new Spark Context
sparkR.init-deprecated.RdThis function initializes a new SparkContext.
Usage
sparkR.init(
  master = "",
  appName = "SparkR",
  sparkHome = Sys.getenv("SPARK_HOME"),
  sparkEnvir = list(),
  sparkExecutorEnv = list(),
  sparkJars = "",
  sparkPackages = ""
)Arguments
- master
- The Spark master URL 
- appName
- Application name to register with cluster manager 
- sparkHome
- Spark Home directory 
- sparkEnvir
- Named list of environment variables to set on worker nodes 
- sparkExecutorEnv
- Named list of environment variables to be used when launching executors 
- sparkJars
- Character vector of jar files to pass to the worker nodes 
- sparkPackages
- Character vector of package coordinates 
Examples
if (FALSE) {
sc <- sparkR.init("local[2]", "SparkR", "/home/spark")
sc <- sparkR.init("local[2]", "SparkR", "/home/spark",
                 list(spark.executor.memory="1g"))
sc <- sparkR.init("yarn-client", "SparkR", "/home/spark",
                 list(spark.executor.memory="4g"),
                 list(LD_LIBRARY_PATH="/directory of JVM libraries (libjvm.so) on workers/"),
                 c("one.jar", "two.jar", "three.jar"),
                 c("com.databricks:spark-avro_2.11:2.0.1"))
}