Sort a SparkDataFrame by the specified column(s).
Usagearrange(x, col, ...)
# S4 method for class 'SparkDataFrame,Column'
arrange(x, col, ..., withinPartitions = FALSE)
# S4 method for class 'SparkDataFrame,character'
arrange(x, col, ..., decreasing = FALSE, withinPartitions = FALSE)
# S4 method for class 'SparkDataFrame,characterOrColumn'
orderBy(x, col, ...)
Arguments
a SparkDataFrame to be sorted.
a character or Column object indicating the fields to sort on
additional sorting fields
a logical argument indicating whether to sort only within each partition
a logical argument indicating sorting order for columns when a character vector is specified for col
A SparkDataFrame where all elements are sorted.
Notearrange(SparkDataFrame, Column) since 1.4.0
arrange(SparkDataFrame, character) since 1.4.0
orderBy(SparkDataFrame, characterOrColumn) since 1.4.0
See alsoOther SparkDataFrame functions: SparkDataFrame-class
, agg()
, alias()
, as.data.frame()
, attach,SparkDataFrame-method
, broadcast()
, cache()
, checkpoint()
, coalesce()
, collect()
, colnames()
, coltypes()
, createOrReplaceTempView()
, crossJoin()
, cube()
, dapplyCollect()
, dapply()
, describe()
, dim()
, distinct()
, dropDuplicates()
, dropna()
, drop()
, dtypes()
, exceptAll()
, except()
, explain()
, filter()
, first()
, gapplyCollect()
, gapply()
, getNumPartitions()
, group_by()
, head()
, hint()
, histogram()
, insertInto()
, intersectAll()
, intersect()
, isLocal()
, isStreaming()
, join()
, limit()
, localCheckpoint()
, merge()
, mutate()
, ncol()
, nrow()
, persist()
, printSchema()
, randomSplit()
, rbind()
, rename()
, repartitionByRange()
, repartition()
, rollup()
, sample()
, saveAsTable()
, schema()
, selectExpr()
, select()
, showDF()
, show()
, storageLevel()
, str()
, subset()
, summary()
, take()
, toJSON()
, unionAll()
, unionByName()
, union()
, unpersist()
, unpivot()
, withColumn()
, withWatermark()
, with()
, write.df()
, write.jdbc()
, write.json()
, write.orc()
, write.parquet()
, write.stream()
, write.text()
if (FALSE) { # \dontrun{
sparkR.session()
path <- "path/to/file.json"
df <- read.json(path)
arrange(df, df$col1)
arrange(df, asc(df$col1), desc(abs(df$col2)))
arrange(df, "col1", decreasing = TRUE)
arrange(df, "col1", "col2", decreasing = c(TRUE, FALSE))
arrange(df, "col1", "col2", withinPartitions = TRUE)
} # }
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4