From b131ad3c3e107e1b09ddcf7a056f10afc602c6f4 Mon Sep 17 00:00:00 2001 From: "R. RyanTM" Date: Tue, 19 Oct 2021 03:54:11 +0000 Subject: [PATCH 1/2] python38Packages.pyspark: 3.1.2 -> 3.2.0 --- pkgs/development/python-modules/pyspark/default.nix | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/pkgs/development/python-modules/pyspark/default.nix b/pkgs/development/python-modules/pyspark/default.nix index 625041fd81b2..68aa78db462d 100644 --- a/pkgs/development/python-modules/pyspark/default.nix +++ b/pkgs/development/python-modules/pyspark/default.nix @@ -2,11 +2,11 @@ buildPythonPackage rec { pname = "pyspark"; - version = "3.1.2"; + version = "3.2.0"; src = fetchPypi { inherit pname version; - sha256 = "5e25ebb18756e9715f4d26848cc7e558035025da74b4fc325a0ebc05ff538e65"; + sha256 = "bfea06179edbfb4bc76a0f470bd3c38e12f00e1023e3ad0373558d07cff102ab"; }; # pypandoc is broken with pandoc2, so we just lose docs. From 06c825a9c23e62870b2d2f89fe571f4341ca514c Mon Sep 17 00:00:00 2001 From: Fabian Affolter Date: Tue, 19 Oct 2021 08:52:32 +0200 Subject: [PATCH 2/2] python3Packages.pyspark: update postPatch and add pythonImportsCheck --- .../python-modules/pyspark/default.nix | 21 ++++++++++++++----- 1 file changed, 16 insertions(+), 5 deletions(-) diff --git a/pkgs/development/python-modules/pyspark/default.nix b/pkgs/development/python-modules/pyspark/default.nix index 68aa78db462d..2e6f41aa2332 100644 --- a/pkgs/development/python-modules/pyspark/default.nix +++ b/pkgs/development/python-modules/pyspark/default.nix @@ -1,4 +1,8 @@ -{ buildPythonPackage, fetchPypi, lib, py4j }: +{ lib +, buildPythonPackage +, fetchPypi +, py4j +}: buildPythonPackage rec { pname = "pyspark"; @@ -13,16 +17,23 @@ buildPythonPackage rec { postPatch = '' sed -i "s/'pypandoc'//" setup.py - substituteInPlace setup.py --replace py4j==0.10.9 'py4j>=0.10.9,<0.11' + substituteInPlace setup.py \ + --replace py4j==0.10.9.2 'py4j>=0.10.9,<0.11' ''; - propagatedBuildInputs = [ py4j ]; + propagatedBuildInputs = [ + py4j + ]; - # Tests assume running spark... + # Tests assume running spark instance doCheck = false; + pythonImportsCheck = [ + "pyspark" + ]; + meta = with lib; { - description = "Apache Spark"; + description = "Python bindings for Apache Spark"; homepage = "https://github.com/apache/spark/tree/master/python"; license = licenses.asl20; maintainers = [ maintainers.shlevy ];