copy_to.spark_connection {sparklyr} | R Documentation |
Copy an R Data Frame to Spark
Description
Copy an R data.frame
to Spark, and return a reference to the
generated Spark DataFrame as a tbl_spark
. The returned object will
act as a dplyr
-compatible interface to the underlying Spark table.
Usage
## S3 method for class 'spark_connection'
copy_to(
dest,
df,
name = spark_table_name(substitute(df)),
overwrite = FALSE,
memory = TRUE,
repartition = 0L,
...
)
Arguments
dest |
A |
df |
An R |
name |
The name to assign to the copied table in Spark. |
overwrite |
Boolean; overwrite a pre-existing table with the name |
memory |
Boolean; should the table be cached into memory? |
repartition |
The number of partitions to use when distributing the table across the Spark cluster. The default (0) can be used to avoid partitioning. |
... |
Optional arguments; currently unused. |
Value
A tbl_spark
, representing a dplyr
-compatible interface
to a Spark DataFrame.
[Package sparklyr version 1.8.6 Index]