About 9,600,000 results
Open links in new tab
  1. python - Spark Equivalent of IF Then ELSE - Stack Overflow

    python apache-spark pyspark apache-spark-sql edited Dec 10, 2017 at 1:43 Community Bot 1 1

  2. pyspark - How to use AND or OR condition in when in Spark - Stack …

    107 pyspark.sql.functions.when takes a Boolean Column as its condition. When using PySpark, it's often useful to think "Column Expression" when you read "Column". Logical operations on PySpark …

  3. Comparison operator in PySpark (not equal/ !=) - Stack Overflow

    Aug 24, 2016 · The selected correct answer does not address the question, and the other answers are all wrong for pyspark. There is no "!=" operator equivalent in pyspark for this solution.

  4. Rename more than one column using withColumnRenamed

    Since pyspark 3.4.0, you can use the withColumnsRenamed() method to rename multiple columns at once. It takes as an input a map of existing column names and the corresponding desired column …

  5. PySpark: multiple conditions in when clause - Stack Overflow

    Jun 8, 2016 · Very helpful observation when in pyspark multiple conditions can be built using & (for and) and | (for or). Note:In pyspark t is important to enclose every expressions within parenthesis () that …

  6. Best way to get the max value in a Spark dataframe column

    Remark: Spark is intended to work on Big Data - distributed computing. The size of the example DataFrame is very small, so the order of real-life examples can be altered with respect to the small …

  7. spark dataframe drop duplicates and keep first - Stack Overflow

    Aug 1, 2016 · 2 I just did something perhaps similar to what you guys need, using drop_duplicates pyspark. Situation is this. I have 2 dataframes (coming from 2 files) which are exactly same except 2 …

  8. Filtering a Pyspark DataFrame with SQL-like IN clause

    Mar 8, 2016 · Filtering a Pyspark DataFrame with SQL-like IN clause Asked 9 years, 9 months ago Modified 3 years, 8 months ago Viewed 123k times

  9. Pyspark: display a spark data frame in a table format

    Pyspark: display a spark data frame in a table format Asked 9 years, 3 months ago Modified 2 years, 4 months ago Viewed 413k times

  10. Join two data frames, select all columns from one and some columns …

    Mar 21, 2016 · 0 If you need multiple columns from other pyspark dataframe then you can use this based on single join condition