Things on this page are fragmentary and immature notes/thoughts of the author. Please read with your own judgement!
Besides using the
colfunction to reference a column, Spark/Scala DataFrame supports using$"col_name"(based on implicit conversion and must haveimport spark.implicit._) while PySpark DataFrame support usingdf.col_name(similar to what you can do with a pandas DataFrame).Spark/Scala PySpark col(“col_name”) col(“col_name”) Implicit Conversion $“col_name” X Dot reference X df.col_name ===(null safe equality comparison) is supported in Spark/Scala but not available in PySpark.
References¶
https://
https://
https://
https://
https://
https://
https://