Return a new SparkDataFrame containing rows in this SparkDataFrame but not in another SparkDataFrame while preserving the duplicates. This is equivalent to EXCEPT ALL
in SQL. Also as standard in SQL, this function resolves columns by position (not by name).
exceptAll(x, y)
# S4 method for class 'SparkDataFrame,SparkDataFrame'
exceptAll(x, y)
Arguments
a SparkDataFrame.
a SparkDataFrame.
A SparkDataFrame containing the result of the except all operation.
NoteexceptAll since 2.4.0
See alsoOther SparkDataFrame functions: SparkDataFrame-class
, agg()
, alias()
, arrange()
, as.data.frame()
, attach,SparkDataFrame-method
, broadcast()
, cache()
, checkpoint()
, coalesce()
, collect()
, colnames()
, coltypes()
, createOrReplaceTempView()
, crossJoin()
, cube()
, dapplyCollect()
, dapply()
, describe()
, dim()
, distinct()
, dropDuplicates()
, dropna()
, drop()
, dtypes()
, except()
, explain()
, filter()
, first()
, gapplyCollect()
, gapply()
, getNumPartitions()
, group_by()
, head()
, hint()
, histogram()
, insertInto()
, intersectAll()
, intersect()
, isLocal()
, isStreaming()
, join()
, limit()
, localCheckpoint()
, merge()
, mutate()
, ncol()
, nrow()
, persist()
, printSchema()
, randomSplit()
, rbind()
, rename()
, repartitionByRange()
, repartition()
, rollup()
, sample()
, saveAsTable()
, schema()
, selectExpr()
, select()
, showDF()
, show()
, storageLevel()
, str()
, subset()
, summary()
, take()
, toJSON()
, unionAll()
, unionByName()
, union()
, unpersist()
, unpivot()
, withColumn()
, withWatermark()
, with()
, write.df()
, write.jdbc()
, write.json()
, write.orc()
, write.parquet()
, write.stream()
, write.text()
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4