A B C D E F G H I J K L M N O P Q R S T U V W Y misc
| abs | abs | 
| abs-method | abs | 
| acos | acos | 
| acos-method | acos | 
| add_months | add_months | 
| add_months-method | add_months | 
| AFTSurvivalRegressionModel-class | S4 class that represents a AFTSurvivalRegressionModel | 
| agg | summarize | 
| agg-method | summarize | 
| alias | alias | 
| alias-method | alias | 
| ALSModel-class | S4 class that represents an ALSModel | 
| approxCountDistinct | Returns the approximate number of distinct items in a group | 
| approxCountDistinct-method | Returns the approximate number of distinct items in a group | 
| approxQuantile | Calculates the approximate quantiles of a numerical column of a SparkDataFrame | 
| approxQuantile-method | Calculates the approximate quantiles of a numerical column of a SparkDataFrame | 
| arrange | Arrange Rows by Variables | 
| arrange-method | Arrange Rows by Variables | 
| array_contains | array_contains | 
| array_contains-method | array_contains | 
| as.data.frame | Download data from a SparkDataFrame into a R data.frame | 
| as.data.frame-method | Download data from a SparkDataFrame into a R data.frame | 
| as.DataFrame | Create a SparkDataFrame | 
| as.DataFrame.default | Create a SparkDataFrame | 
| asc | A set of operations working with SparkDataFrame columns | 
| ascii | ascii | 
| ascii-method | ascii | 
| asin | asin | 
| asin-method | asin | 
| atan | atan | 
| atan-method | atan | 
| atan2 | atan2 | 
| atan2-method | atan2 | 
| attach | Attach SparkDataFrame to R search path | 
| attach-method | Attach SparkDataFrame to R search path | 
| avg | avg | 
| avg-method | avg | 
| base64 | base64 | 
| base64-method | base64 | 
| between | between | 
| between-method | between | 
| bin | bin | 
| bin-method | bin | 
| bitwiseNOT | bitwiseNOT | 
| bitwiseNOT-method | bitwiseNOT | 
| bround | bround | 
| bround-method | bround | 
| cache | Cache | 
| cache-method | Cache | 
| cacheTable | Cache Table | 
| cacheTable.default | Cache Table | 
| cancelJobGroup | Cancel active jobs for the specified group | 
| cancelJobGroup.default | Cancel active jobs for the specified group | 
| cast | Casts the column to a different data type. | 
| cast-method | Casts the column to a different data type. | 
| cbrt | cbrt | 
| cbrt-method | cbrt | 
| ceil | Computes the ceiling of the given value | 
| ceil-method | Computes the ceiling of the given value | 
| ceiling | Computes the ceiling of the given value | 
| ceiling-method | Computes the ceiling of the given value | 
| clearCache | Clear Cache | 
| clearCache.default | Clear Cache | 
| clearJobGroup | Clear current job group ID and its description | 
| clearJobGroup.default | Clear current job group ID and its description | 
| coalesce | Coalesce | 
| coalesce-method | Coalesce | 
| collect | Collects all the elements of a SparkDataFrame and coerces them into an R data.frame. | 
| collect-method | Collects all the elements of a SparkDataFrame and coerces them into an R data.frame. | 
| colnames | Column Names of SparkDataFrame | 
| colnames-method | Column Names of SparkDataFrame | 
| colnames<- | Column Names of SparkDataFrame | 
| colnames<--method | Column Names of SparkDataFrame | 
| coltypes | coltypes | 
| coltypes-method | coltypes | 
| coltypes<- | coltypes | 
| coltypes<--method | coltypes | 
| column | S4 class that represents a SparkDataFrame column | 
| Column-class | S4 class that represents a SparkDataFrame column | 
| column-method | S4 class that represents a SparkDataFrame column | 
| columnfunctions | A set of operations working with SparkDataFrame columns | 
| columns | Column Names of SparkDataFrame | 
| columns-method | Column Names of SparkDataFrame | 
| concat | concat | 
| concat-method | concat | 
| concat_ws | concat_ws | 
| concat_ws-method | concat_ws | 
| contains | A set of operations working with SparkDataFrame columns | 
| conv | conv | 
| conv-method | conv | 
| corr | corr | 
| corr-method | corr | 
| cos | cos | 
| cos-method | cos | 
| cosh | cosh | 
| cosh-method | cosh | 
| count | Count | 
| count-method | Count | 
| count-method | Returns the number of rows in a SparkDataFrame | 
| countDistinct | Count Distinct Values | 
| countDistinct-method | Count Distinct Values | 
| cov | cov | 
| cov-method | cov | 
| covar_pop | covar_pop | 
| covar_pop-method | covar_pop | 
| covar_samp | cov | 
| covar_samp-method | cov | 
| crc32 | crc32 | 
| crc32-method | crc32 | 
| createDataFrame | Create a SparkDataFrame | 
| createDataFrame.default | Create a SparkDataFrame | 
| createExternalTable | Create an external table | 
| createExternalTable.default | Create an external table | 
| createOrReplaceTempView | Creates a temporary view using the given name. | 
| createOrReplaceTempView-method | Creates a temporary view using the given name. | 
| crossJoin | CrossJoin | 
| crossJoin-method | CrossJoin | 
| crosstab | Computes a pair-wise frequency table of the given columns | 
| crosstab-method | Computes a pair-wise frequency table of the given columns | 
| cume_dist | cume_dist | 
| cume_dist-method | cume_dist | 
| dapply | dapply | 
| dapply-method | dapply | 
| dapplyCollect | dapplyCollect | 
| dapplyCollect-method | dapplyCollect | 
| datediff | datediff | 
| datediff-method | datediff | 
| date_add | date_add | 
| date_add-method | date_add | 
| date_format | date_format | 
| date_format-method | date_format | 
| date_sub | date_sub | 
| date_sub-method | date_sub | 
| dayofmonth | dayofmonth | 
| dayofmonth-method | dayofmonth | 
| dayofyear | dayofyear | 
| dayofyear-method | dayofyear | 
| decode | decode | 
| decode-method | decode | 
| dense_rank | dense_rank | 
| dense_rank-method | dense_rank | 
| desc | A set of operations working with SparkDataFrame columns | 
| describe | summary | 
| describe-method | summary | 
| dim | Returns the dimensions of SparkDataFrame | 
| dim-method | Returns the dimensions of SparkDataFrame | 
| distinct | Distinct | 
| distinct-method | Distinct | 
| drop | drop | 
| drop-method | drop | 
| dropDuplicates | dropDuplicates | 
| dropDuplicates-method | dropDuplicates | 
| dropna | A set of SparkDataFrame functions working with NA values | 
| dropna-method | A set of SparkDataFrame functions working with NA values | 
| dropTempTable | (Deprecated) Drop Temporary Table | 
| dropTempTable.default | (Deprecated) Drop Temporary Table | 
| dropTempView | Drops the temporary view with the given view name in the catalog. | 
| dtypes | DataTypes | 
| dtypes-method | DataTypes | 
| encode | encode | 
| encode-method | encode | 
| endsWith | endsWith | 
| endsWith-method | endsWith | 
| except | except | 
| except-method | except | 
| exp | exp | 
| exp-method | exp | 
| explain | Explain | 
| explain-method | Explain | 
| explode | explode | 
| explode-method | explode | 
| expm1 | expm1 | 
| expm1-method | expm1 | 
| expr | expr | 
| expr-method | expr | 
| factorial | factorial | 
| factorial-method | factorial | 
| fillna | A set of SparkDataFrame functions working with NA values | 
| fillna-method | A set of SparkDataFrame functions working with NA values | 
| filter | Filter | 
| filter-method | Filter | 
| first | Return the first row of a SparkDataFrame | 
| first-method | Return the first row of a SparkDataFrame | 
| fitted | Get fitted result from a k-means model | 
| fitted-method | Get fitted result from a k-means model | 
| floor | floor | 
| floor-method | floor | 
| format_number | format_number | 
| format_number-method | format_number | 
| format_string | format_string | 
| format_string-method | format_string | 
| freqItems | Finding frequent items for columns, possibly with false positives | 
| freqItems-method | Finding frequent items for columns, possibly with false positives | 
| from_unixtime | from_unixtime | 
| from_unixtime-method | from_unixtime | 
| from_utc_timestamp | from_utc_timestamp | 
| from_utc_timestamp-method | from_utc_timestamp | 
| gapply | gapply | 
| gapply-method | gapply | 
| gapplyCollect | gapplyCollect | 
| gapplyCollect-method | gapplyCollect | 
| GaussianMixtureModel-class | S4 class that represents a GaussianMixtureModel | 
| GBTClassificationModel-class | S4 class that represents a GBTClassificationModel | 
| GBTRegressionModel-class | S4 class that represents a GBTRegressionModel | 
| GeneralizedLinearRegressionModel-class | S4 class that represents a generalized linear model | 
| generateAliasesForIntersectedCols | Creates a list of columns by replacing the intersected ones with aliases | 
| getField | A set of operations working with SparkDataFrame columns | 
| getItem | A set of operations working with SparkDataFrame columns | 
| getNumPartitions | getNumPartitions | 
| getNumPartitions-method | getNumPartitions | 
| glm | Generalized Linear Models (R-compliant) | 
| glm-method | Generalized Linear Models (R-compliant) | 
| greatest | greatest | 
| greatest-method | greatest | 
| groupBy | GroupBy | 
| groupBy-method | GroupBy | 
| groupedData | S4 class that represents a GroupedData | 
| GroupedData-class | S4 class that represents a GroupedData | 
| group_by | GroupBy | 
| group_by-method | GroupBy | 
| hash | hash | 
| hash-method | hash | 
| hashCode | Compute the hashCode of an object | 
| head | Head | 
| head-method | Head | 
| hex | hex | 
| hex-method | hex | 
| histogram | Compute histogram statistics for given column | 
| histogram-method | Compute histogram statistics for given column | 
| hour | hour | 
| hour-method | hour | 
| hypot | hypot | 
| hypot-method | hypot | 
| ifelse | ifelse | 
| ifelse-method | ifelse | 
| initcap | initcap | 
| initcap-method | initcap | 
| insertInto | insertInto | 
| insertInto-method | insertInto | 
| install.spark | Download and Install Apache Spark to a Local Directory | 
| instr | instr | 
| instr-method | instr | 
| intersect | Intersect | 
| intersect-method | Intersect | 
| is.nan | is.nan | 
| is.nan-method | is.nan | 
| isLocal | isLocal | 
| isLocal-method | isLocal | 
| isNaN | A set of operations working with SparkDataFrame columns | 
| isnan | is.nan | 
| isnan-method | is.nan | 
| isNotNull | A set of operations working with SparkDataFrame columns | 
| isNull | A set of operations working with SparkDataFrame columns | 
| IsotonicRegressionModel-class | S4 class that represents an IsotonicRegressionModel | 
| join | Join | 
| join-method | Join | 
| jsonFile | Create a SparkDataFrame from a JSON file. | 
| jsonFile.default | Create a SparkDataFrame from a JSON file. | 
| KMeansModel-class | S4 class that represents a KMeansModel | 
| KSTest-class | S4 class that represents an KSTest | 
| kurtosis | kurtosis | 
| kurtosis-method | kurtosis | 
| lag | lag | 
| lag-method | lag | 
| last | last | 
| last-method | last | 
| last_day | last_day | 
| last_day-method | last_day | 
| LDAModel-class | S4 class that represents an LDAModel | 
| lead | lead | 
| lead-method | lead | 
| least | least | 
| least-method | least | 
| length | length | 
| length-method | length | 
| levenshtein | levenshtein | 
| levenshtein-method | levenshtein | 
| like | A set of operations working with SparkDataFrame columns | 
| limit | Limit | 
| limit-method | Limit | 
| lit | lit | 
| lit-method | lit | 
| loadDF | Load a SparkDataFrame | 
| loadDF.default | Load a SparkDataFrame | 
| locate | locate | 
| locate-method | locate | 
| log | log | 
| log-method | log | 
| log10 | log10 | 
| log10-method | log10 | 
| log1p | log1p | 
| log1p-method | log1p | 
| log2 | log2 | 
| log2-method | log2 | 
| LogisticRegressionModel-class | S4 class that represents an LogisticRegressionModel | 
| lower | lower | 
| lower-method | lower | 
| lpad | lpad | 
| lpad-method | lpad | 
| ltrim | ltrim | 
| ltrim-method | ltrim | 
| max | max | 
| max-method | max | 
| md5 | md5 | 
| md5-method | md5 | 
| mean | mean | 
| mean-method | mean | 
| merge | Merges two data frames | 
| merge-method | Merges two data frames | 
| min | min | 
| min-method | min | 
| minute | minute | 
| minute-method | minute | 
| monotonically_increasing_id | monotonically_increasing_id | 
| monotonically_increasing_id-method | monotonically_increasing_id | 
| month | month | 
| month-method | month | 
| months_between | months_between | 
| months_between-method | months_between | 
| MultilayerPerceptronClassificationModel-class | S4 class that represents a MultilayerPerceptronClassificationModel | 
| mutate | Mutate | 
| mutate-method | Mutate | 
| n | Count | 
| n-method | Count | 
| na.omit | A set of SparkDataFrame functions working with NA values | 
| na.omit-method | A set of SparkDataFrame functions working with NA values | 
| NaiveBayesModel-class | S4 class that represents a NaiveBayesModel | 
| names | Column Names of SparkDataFrame | 
| names-method | Column Names of SparkDataFrame | 
| names<- | Column Names of SparkDataFrame | 
| names<--method | Column Names of SparkDataFrame | 
| nanvl | nanvl | 
| nanvl-method | nanvl | 
| ncol | Returns the number of columns in a SparkDataFrame | 
| ncol-method | Returns the number of columns in a SparkDataFrame | 
| negate | negate | 
| negate-method | negate | 
| next_day | next_day | 
| next_day-method | next_day | 
| nrow | Returns the number of rows in a SparkDataFrame | 
| nrow-method | Returns the number of rows in a SparkDataFrame | 
| ntile | ntile | 
| ntile-method | ntile | 
| n_distinct | Count Distinct Values | 
| n_distinct-method | Count Distinct Values | 
| orderBy | Ordering Columns in a WindowSpec | 
| orderBy-method | Arrange Rows by Variables | 
| orderBy-method | Ordering Columns in a WindowSpec | 
| otherwise | otherwise | 
| otherwise-method | otherwise | 
| over | over | 
| over-method | over | 
| parquetFile | Create a SparkDataFrame from a Parquet file. | 
| parquetFile.default | Create a SparkDataFrame from a Parquet file. | 
| partitionBy | partitionBy | 
| partitionBy-method | partitionBy | 
| percent_rank | percent_rank | 
| percent_rank-method | percent_rank | 
| persist | Persist | 
| persist-method | Persist | 
| pivot | Pivot a column of the GroupedData and perform the specified aggregation. | 
| pivot-method | Pivot a column of the GroupedData and perform the specified aggregation. | 
| pmod | pmod | 
| pmod-method | pmod | 
| posexplode | posexplode | 
| posexplode-method | posexplode | 
| predict | Makes predictions from a MLlib model | 
| predict-method | Alternating Least Squares (ALS) for Collaborative Filtering | 
| predict-method | Multivariate Gaussian Mixture Model (GMM) | 
| predict-method | Gradient Boosted Tree Model for Regression and Classification | 
| predict-method | Generalized Linear Models | 
| predict-method | Isotonic Regression Model | 
| predict-method | K-Means Clustering Model | 
| predict-method | Logistic Regression Model | 
| predict-method | Multilayer Perceptron Classification Model | 
| predict-method | Naive Bayes Models | 
| predict-method | Random Forest Model for Regression and Classification | 
| predict-method | Accelerated Failure Time (AFT) Survival Regression Model | 
| print.jobj | Print a JVM object reference. | 
| print.structField | Print a Spark StructField. | 
| print.structType | Print a Spark StructType. | 
| print.summary.GBTClassificationModel | Gradient Boosted Tree Model for Regression and Classification | 
| print.summary.GBTRegressionModel | Gradient Boosted Tree Model for Regression and Classification | 
| print.summary.GeneralizedLinearRegressionModel | Generalized Linear Models | 
| print.summary.KSTest | (One-Sample) Kolmogorov-Smirnov Test | 
| print.summary.RandomForestClassificationModel | Random Forest Model for Regression and Classification | 
| print.summary.RandomForestRegressionModel | Random Forest Model for Regression and Classification | 
| printSchema | Print Schema of a SparkDataFrame | 
| printSchema-method | Print Schema of a SparkDataFrame | 
| quarter | quarter | 
| quarter-method | quarter | 
| rand | rand | 
| rand-method | rand | 
| randn | randn | 
| randn-method | randn | 
| RandomForestClassificationModel-class | S4 class that represents a RandomForestClassificationModel | 
| RandomForestRegressionModel-class | S4 class that represents a RandomForestRegressionModel | 
| randomSplit | randomSplit | 
| randomSplit-method | randomSplit | 
| rangeBetween | rangeBetween | 
| rangeBetween-method | rangeBetween | 
| rank | rank | 
| rank-method | rank | 
| rbind | Union two or more SparkDataFrames | 
| rbind-method | Union two or more SparkDataFrames | 
| read.df | Load a SparkDataFrame | 
| read.df.default | Load a SparkDataFrame | 
| read.jdbc | Create a SparkDataFrame representing the database table accessible via JDBC URL | 
| read.json | Create a SparkDataFrame from a JSON file. | 
| read.json.default | Create a SparkDataFrame from a JSON file. | 
| read.ml | Load a fitted MLlib model from the input path. | 
| read.orc | Create a SparkDataFrame from an ORC file. | 
| read.parquet | Create a SparkDataFrame from a Parquet file. | 
| read.parquet.default | Create a SparkDataFrame from a Parquet file. | 
| read.text | Create a SparkDataFrame from a text file. | 
| read.text.default | Create a SparkDataFrame from a text file. | 
| regexp_extract | regexp_extract | 
| regexp_extract-method | regexp_extract | 
| regexp_replace | regexp_replace | 
| regexp_replace-method | regexp_replace | 
| registerTempTable | (Deprecated) Register Temporary Table | 
| registerTempTable-method | (Deprecated) Register Temporary Table | 
| rename | rename | 
| rename-method | rename | 
| repartition | Repartition | 
| repartition-method | Repartition | 
| reverse | reverse | 
| reverse-method | reverse | 
| rint | rint | 
| rint-method | rint | 
| rlike | A set of operations working with SparkDataFrame columns | 
| round | round | 
| round-method | round | 
| rowsBetween | rowsBetween | 
| rowsBetween-method | rowsBetween | 
| row_number | row_number | 
| row_number-method | row_number | 
| rpad | rpad | 
| rpad-method | rpad | 
| rtrim | rtrim | 
| rtrim-method | rtrim | 
| sample | Sample | 
| sample-method | Sample | 
| sampleBy | Returns a stratified sample without replacement | 
| sampleBy-method | Returns a stratified sample without replacement | 
| sample_frac | Sample | 
| sample_frac-method | Sample | 
| saveAsParquetFile | Save the contents of SparkDataFrame as a Parquet file, preserving the schema. | 
| saveAsParquetFile-method | Save the contents of SparkDataFrame as a Parquet file, preserving the schema. | 
| saveAsTable | Save the contents of the SparkDataFrame to a data source as a table | 
| saveAsTable-method | Save the contents of the SparkDataFrame to a data source as a table | 
| saveDF | Save the contents of SparkDataFrame to a data source. | 
| saveDF-method | Save the contents of SparkDataFrame to a data source. | 
| schema | Get schema object | 
| schema-method | Get schema object | 
| sd | sd | 
| sd-method | sd | 
| second | second | 
| second-method | second | 
| select | Select | 
| select-method | Select | 
| selectExpr | SelectExpr | 
| selectExpr-method | SelectExpr | 
| setJobGroup | Assigns a group ID to all the jobs started by this thread until the group ID is set to a different value or cleared. | 
| setJobGroup.default | Assigns a group ID to all the jobs started by this thread until the group ID is set to a different value or cleared. | 
| setLogLevel | Set new log level | 
| sha1 | sha1 | 
| sha1-method | sha1 | 
| sha2 | sha2 | 
| sha2-method | sha2 | 
| shiftLeft | shiftLeft | 
| shiftLeft-method | shiftLeft | 
| shiftRight | shiftRight | 
| shiftRight-method | shiftRight | 
| shiftRightUnsigned | shiftRightUnsigned | 
| shiftRightUnsigned-method | shiftRightUnsigned | 
| show | show | 
| show-method | show | 
| showDF | showDF | 
| showDF-method | showDF | 
| sign | signum | 
| sign-method | signum | 
| signum | signum | 
| signum-method | signum | 
| sin | sin | 
| sin-method | sin | 
| sinh | sinh | 
| sinh-method | sinh | 
| size | size | 
| size-method | size | 
| skewness | skewness | 
| skewness-method | skewness | 
| sort_array | sort_array | 
| sort_array-method | sort_array | 
| soundex | soundex | 
| soundex-method | soundex | 
| spark.addFile | Add a file or directory to be downloaded with this Spark job on every node. | 
| spark.als | Alternating Least Squares (ALS) for Collaborative Filtering | 
| spark.als-method | Alternating Least Squares (ALS) for Collaborative Filtering | 
| spark.gaussianMixture | Multivariate Gaussian Mixture Model (GMM) | 
| spark.gaussianMixture-method | Multivariate Gaussian Mixture Model (GMM) | 
| spark.gbt | Gradient Boosted Tree Model for Regression and Classification | 
| spark.gbt-method | Gradient Boosted Tree Model for Regression and Classification | 
| spark.getSparkFiles | Get the absolute path of a file added through spark.addFile. | 
| spark.getSparkFilesRootDirectory | Get the root directory that contains files added through spark.addFile. | 
| spark.glm | Generalized Linear Models | 
| spark.glm-method | Generalized Linear Models | 
| spark.isoreg | Isotonic Regression Model | 
| spark.isoreg-method | Isotonic Regression Model | 
| spark.kmeans | K-Means Clustering Model | 
| spark.kmeans-method | K-Means Clustering Model | 
| spark.kstest | (One-Sample) Kolmogorov-Smirnov Test | 
| spark.kstest-method | (One-Sample) Kolmogorov-Smirnov Test | 
| spark.lapply | Run a function over a list of elements, distributing the computations with Spark | 
| spark.lda | Latent Dirichlet Allocation | 
| spark.lda-method | Latent Dirichlet Allocation | 
| spark.logit | Logistic Regression Model | 
| spark.logit-method | Logistic Regression Model | 
| spark.mlp | Multilayer Perceptron Classification Model | 
| spark.mlp-method | Multilayer Perceptron Classification Model | 
| spark.naiveBayes | Naive Bayes Models | 
| spark.naiveBayes-method | Naive Bayes Models | 
| spark.perplexity | Latent Dirichlet Allocation | 
| spark.perplexity-method | Latent Dirichlet Allocation | 
| spark.posterior | Latent Dirichlet Allocation | 
| spark.posterior-method | Latent Dirichlet Allocation | 
| spark.randomForest | Random Forest Model for Regression and Classification | 
| spark.randomForest-method | Random Forest Model for Regression and Classification | 
| spark.survreg | Accelerated Failure Time (AFT) Survival Regression Model | 
| spark.survreg-method | Accelerated Failure Time (AFT) Survival Regression Model | 
| SparkDataFrame-class | S4 class that represents a SparkDataFrame | 
| sparkR.callJMethod | Call Java Methods | 
| sparkR.callJStatic | Call Static Java Methods | 
| sparkR.conf | Get Runtime Config from the current active SparkSession | 
| sparkR.init | (Deprecated) Initialize a new Spark Context | 
| sparkR.newJObject | Create Java Objects | 
| sparkR.session | Get the existing SparkSession or initialize a new SparkSession. | 
| sparkR.session.stop | Stop the Spark Session and Spark Context | 
| sparkR.stop | Stop the Spark Session and Spark Context | 
| sparkR.uiWebUrl | Get the URL of the SparkUI instance for the current active SparkSession | 
| sparkR.version | Get version of Spark on which this application is running | 
| sparkRHive.init | (Deprecated) Initialize a new HiveContext | 
| sparkRSQL.init | (Deprecated) Initialize a new SQLContext | 
| spark_partition_id | Return the partition ID as a column | 
| spark_partition_id-method | Return the partition ID as a column | 
| sql | SQL Query | 
| sql.default | SQL Query | 
| sqrt | sqrt | 
| sqrt-method | sqrt | 
| startsWith | startsWith | 
| startsWith-method | startsWith | 
| stddev | sd | 
| stddev-method | sd | 
| stddev_pop | stddev_pop | 
| stddev_pop-method | stddev_pop | 
| stddev_samp | stddev_samp | 
| stddev_samp-method | stddev_samp | 
| storageLevel | StorageLevel | 
| storageLevel-method | StorageLevel | 
| str | Compactly display the structure of a dataset | 
| str-method | Compactly display the structure of a dataset | 
| struct | struct | 
| struct-method | struct | 
| structField | structField | 
| structField.character | structField | 
| structField.jobj | structField | 
| structType | structType | 
| structType.jobj | structType | 
| structType.structField | structType | 
| subset | Subset | 
| subset-method | Subset | 
| substr | substr | 
| substr-method | substr | 
| substring_index | substring_index | 
| substring_index-method | substring_index | 
| sum | sum | 
| sum-method | sum | 
| sumDistinct | sumDistinct | 
| sumDistinct-method | sumDistinct | 
| summarize | summarize | 
| summarize-method | summarize | 
| summary | summary | 
| summary-method | Alternating Least Squares (ALS) for Collaborative Filtering | 
| summary-method | Multivariate Gaussian Mixture Model (GMM) | 
| summary-method | Gradient Boosted Tree Model for Regression and Classification | 
| summary-method | Generalized Linear Models | 
| summary-method | Isotonic Regression Model | 
| summary-method | K-Means Clustering Model | 
| summary-method | (One-Sample) Kolmogorov-Smirnov Test | 
| summary-method | Latent Dirichlet Allocation | 
| summary-method | Logistic Regression Model | 
| summary-method | Multilayer Perceptron Classification Model | 
| summary-method | Naive Bayes Models | 
| summary-method | Random Forest Model for Regression and Classification | 
| summary-method | Accelerated Failure Time (AFT) Survival Regression Model | 
| summary-method | summary | 
| tableNames | Table Names | 
| tableNames.default | Table Names | 
| tables | Tables | 
| tables.default | Tables | 
| tableToDF | Create a SparkDataFrame from a SparkSQL Table | 
| take | Take the first NUM rows of a SparkDataFrame and return the results as a R data.frame | 
| take-method | Take the first NUM rows of a SparkDataFrame and return the results as a R data.frame | 
| tan | tan | 
| tan-method | tan | 
| tanh | tanh | 
| tanh-method | tanh | 
| toDegrees | toDegrees | 
| toDegrees-method | toDegrees | 
| toRadians | toRadians | 
| toRadians-method | toRadians | 
| to_date | to_date | 
| to_date-method | to_date | 
| to_utc_timestamp | to_utc_timestamp | 
| to_utc_timestamp-method | to_utc_timestamp | 
| transform | Mutate | 
| transform-method | Mutate | 
| translate | translate | 
| translate-method | translate | 
| trim | trim | 
| trim-method | trim | 
| unbase64 | unbase64 | 
| unbase64-method | unbase64 | 
| uncacheTable | Uncache Table | 
| uncacheTable.default | Uncache Table | 
| unhex | unhex | 
| unhex-method | unhex | 
| union | Return a new SparkDataFrame containing the union of rows | 
| union-method | Return a new SparkDataFrame containing the union of rows | 
| unionAll | Return a new SparkDataFrame containing the union of rows | 
| unionAll-method | Return a new SparkDataFrame containing the union of rows | 
| unique | Distinct | 
| unique-method | Distinct | 
| unix_timestamp | unix_timestamp | 
| unix_timestamp-method | unix_timestamp | 
| unpersist | Unpersist | 
| unpersist-method | Unpersist | 
| upper | upper | 
| upper-method | upper | 
| var | var | 
| var-method | var | 
| variance | var | 
| variance-method | var | 
| var_pop | var_pop | 
| var_pop-method | var_pop | 
| var_samp | var_samp | 
| var_samp-method | var_samp | 
| weekofyear | weekofyear | 
| weekofyear-method | weekofyear | 
| when | when | 
| when-method | when | 
| where | Filter | 
| where-method | Filter | 
| window | window | 
| window-method | window | 
| windowOrderBy | windowOrderBy | 
| windowOrderBy-method | windowOrderBy | 
| windowPartitionBy | windowPartitionBy | 
| windowPartitionBy-method | windowPartitionBy | 
| WindowSpec-class | S4 class that represents a WindowSpec | 
| with | Evaluate a R expression in an environment constructed from a SparkDataFrame | 
| with-method | Evaluate a R expression in an environment constructed from a SparkDataFrame | 
| withColumn | WithColumn | 
| withColumn-method | WithColumn | 
| withColumnRenamed | rename | 
| withColumnRenamed-method | rename | 
| write.df | Save the contents of SparkDataFrame to a data source. | 
| write.df-method | Save the contents of SparkDataFrame to a data source. | 
| write.jdbc | Save the content of SparkDataFrame to an external database table via JDBC. | 
| write.jdbc-method | Save the content of SparkDataFrame to an external database table via JDBC. | 
| write.json | Save the contents of SparkDataFrame as a JSON file | 
| write.json-method | Save the contents of SparkDataFrame as a JSON file | 
| write.ml | Saves the MLlib model to the input path | 
| write.ml-method | Alternating Least Squares (ALS) for Collaborative Filtering | 
| write.ml-method | Multivariate Gaussian Mixture Model (GMM) | 
| write.ml-method | Gradient Boosted Tree Model for Regression and Classification | 
| write.ml-method | Generalized Linear Models | 
| write.ml-method | Isotonic Regression Model | 
| write.ml-method | K-Means Clustering Model | 
| write.ml-method | Latent Dirichlet Allocation | 
| write.ml-method | Logistic Regression Model | 
| write.ml-method | Multilayer Perceptron Classification Model | 
| write.ml-method | Naive Bayes Models | 
| write.ml-method | Random Forest Model for Regression and Classification | 
| write.ml-method | Accelerated Failure Time (AFT) Survival Regression Model | 
| write.orc | Save the contents of SparkDataFrame as an ORC file, preserving the schema. | 
| write.orc-method | Save the contents of SparkDataFrame as an ORC file, preserving the schema. | 
| write.parquet | Save the contents of SparkDataFrame as a Parquet file, preserving the schema. | 
| write.parquet-method | Save the contents of SparkDataFrame as a Parquet file, preserving the schema. | 
| write.text | Save the content of SparkDataFrame in a text file at the specified path. | 
| write.text-method | Save the content of SparkDataFrame in a text file at the specified path. | 
| year | year | 
| year-method | year | 
| $ | Select | 
| $-method | Select | 
| $<- | Select | 
| $<--method | Select | 
| %in% | Match a column with given values. | 
| %in%-method | Match a column with given values. | 
| [ | Subset | 
| [-method | Subset | 
| [[ | Subset | 
| [[-method | Subset | 
| [[<- | Subset | 
| [[<--method | Subset |