pyspark.sql.DataFrame.exceptAll#
- DataFrame.exceptAll(other)[source]#
- Return a new - DataFramecontaining rows in this- DataFramebut not in another- DataFramewhile preserving duplicates.- This is equivalent to EXCEPT ALL in SQL. As standard in SQL, this function resolves columns by position (not by name). - New in version 2.4.0. - Changed in version 3.4.0: Supports Spark Connect. - See also - DataFrame.subtract
- Similar to exceptAll, but eliminates duplicates. 
 - Examples - >>> df1 = spark.createDataFrame( ... [("a", 1), ("a", 1), ("a", 1), ("a", 2), ("b", 3), ("c", 4)], ["C1", "C2"]) >>> df2 = spark.createDataFrame([("a", 1), ("b", 3)], ["C1", "C2"]) >>> df1.exceptAll(df2).show() +---+---+ | C1| C2| +---+---+ | a| 1| | a| 1| | a| 2| | c| 4| +---+---+