SparkDataFrames can be created using functions like createDataFrame, read.json, table etc.
Slotsenv
An R environment that stores bookkeeping states of the SparkDataFrame
sdf
A Java object reference to the backing Scala DataFrame
SparkDataFrame since 2.0.0
See alsocreateDataFrame, read.json, table
https://spark.apache.org/docs/latest/sparkr.html#sparkr-dataframes
Other SparkDataFrame functions: agg()
, alias()
, arrange()
, as.data.frame()
, attach,SparkDataFrame-method
, broadcast()
, cache()
, checkpoint()
, coalesce()
, collect()
, colnames()
, coltypes()
, createOrReplaceTempView()
, crossJoin()
, cube()
, dapplyCollect()
, dapply()
, describe()
, dim()
, distinct()
, dropDuplicates()
, dropna()
, drop()
, dtypes()
, exceptAll()
, except()
, explain()
, filter()
, first()
, gapplyCollect()
, gapply()
, getNumPartitions()
, group_by()
, head()
, hint()
, histogram()
, insertInto()
, intersectAll()
, intersect()
, isLocal()
, isStreaming()
, join()
, limit()
, localCheckpoint()
, merge()
, mutate()
, ncol()
, nrow()
, persist()
, printSchema()
, randomSplit()
, rbind()
, rename()
, repartitionByRange()
, repartition()
, rollup()
, sample()
, saveAsTable()
, schema()
, selectExpr()
, select()
, showDF()
, show()
, storageLevel()
, str()
, subset()
, summary()
, take()
, toJSON()
, unionAll()
, unionByName()
, union()
, unpersist()
, unpivot()
, withColumn()
, withWatermark()
, with()
, write.df()
, write.jdbc()
, write.json()
, write.orc()
, write.parquet()
, write.stream()
, write.text()
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4