Version 3.5.3

pyspark

Apache Spark - A unified analytics engine for large-scale data processing

Install Instructions

pip install pyspark
Current Version Release Date September 24, 2024
Language Python
Package URL (purl) pkg:pip/pyspark@3.5.3

Find pyspark vulnerabilities in your supply chain.

Scan for Free

pyspark Vulnerabilities

Sort by
icon CVVS Score
  • icon CVVS Score
  • icon CVVS Score
  • icon CVE
  • icon CVE
CVSS Score question mark icon CVE question mark icon CWE(s) question mark icon EPSS Score question mark icon EPSS % question mark icon Impacted Versions
Medium 5.4 CVE-2022-31777 CWE-74 0.00069 0.3119
  • 3.0.0–3.2.1
  • 2.1.1–2.4.8
High 8.8 CVE-2022-33891 CWE-77, CWE-78 0.97286 0.99893
  • 3.0.0–3.2.1
  • 2.1.1–2.4.8
High 9.9 CVE-2023-22946 CWE-269 0.00089 0.38675
  • 3.0.0–3.3.4
  • 2.1.1–2.4.8
High 8.8 CVE-2023-32007 CWE-77 0.01269 0.85948
  • 3.1.1–3.2.1
Medium 5.5 CVE-2018-11760 0.00042 0.05089
  • 2.1.1–2.3.1
Medium 4.7 CVE-2018-1334 CWE-200 0.00042 0.05089
  • 2.1.1–2.2.1
High 7.5 CVE-2019-10099 CWE-310, CWE-312 0.00123 0.4745
  • 2.1.1–2.3.2
High 9.8 CVE-2020-9480 CWE-306 0.02985 0.91063
  • 2.1.1–2.4.5
High 7.5 CVE-2021-38296 CWE-294 0.0005 0.19828
  • 3.0.0–3.1.2
  • 2.1.1–2.4.8

pyspark Vulnerability Remediation Guidance

CVE Description Full list of Impacted Versions Fix
CVE-2022-31777 A stored cross-site scripting (XSS) vulnerability in Apache Spark 3.2.1 and earlier, and 3.3.0, allows remote attackers to execute arbitrary JavaScript in the web browser of a user, by including a malicious payload into the logs which would be returned in logs rendered in the UI. 3.1.3, 2.2.1, 2.1.2, 2.1.3, 2.2.0, 3.2.1, 3.2.0, 3.1.1 (Show all) Minor → 3.4.0
CVE-2022-33891 The Apache Spark UI offers the possibility to enable ACLs via the configuration option spark.acls.enable. With an authentication filter, this checks whether a user has access permissions to view or modify the application. If ACLs are enabled, a code path in HttpSecurityFilter can allow someone to perform impersonation by providing an arbitrary user name. A malicious user might then be able to reach a permission check function that will ultimately build a Unix shell command based on their input, and execute it. This will result in arbitrary shell command execution as the user Spark is currently running as. This affects Apache Spark versions 3.0.3 and earlier, versions 3.1.1 to 3.1.2, and versions 3.2.0 to 3.2.1. 3.1.3, 2.2.1, 2.1.2, 2.1.3, 2.2.0, 3.2.1, 3.2.0, 3.1.1 (Show all) Minor → 3.4.0
CVE-2023-22946 In Apache Spark versions prior to 3.4.0, applications using spark-submit can specify a 'proxy-user' to run as, limiting privileges. The application can execute code with the privileges of the submitting user, however, by providing malicious configuration-related classes on the classpath. This affects architectures relying on proxy-user, for example those using Apache Livy to manage submitted applications. Update to Apache Spark 3.4.0 or later, and ensure that spark.submit.proxyUser.allowCustomClasspathInClusterMode is set to its default of "false", and is not overridden by submitted applications. 3.1.3, 2.2.1, 2.1.2, 3.2.2, 3.2.3, 3.3.4, 2.1.3, 2.2.0 (Show all) Minor → 3.4.0
CVE-2023-32007 ** UNSUPPORTED WHEN ASSIGNED ** The Apache Spark UI offers the possibility to enable ACLs via the configuration option spark.acls.enable. With an authentication filter, this checks whether a user has access permissions to view or modify the application. If ACLs are enabled, a code path in HttpSecurityFilter can allow someone to perform impersonation by providing an arbitrary user name. A malicious user might then be able to reach a permission check function that will ultimately build a Unix shell command based on their input, and execute it. This will result in arbitrary shell command execution as the user Spark is currently running as. This issue was disclosed earlier as CVE-2022-33891, but incorrectly claimed version 3.1.3 (which has since gone EOL) would not be affected. NOTE: This vulnerability only affects products that are no longer supported by the maintainer. Users are recommended to upgrade to a supported version of Apache Spark, such as version 3.4.0. 3.1.3, 3.2.1, 3.2.0, 3.1.1, 3.1.2 Minor → 3.4.0
CVE-2018-11760 When using PySpark , it's possible for a different local user to connect to the Spark application and impersonate the user running the Spark application. This affects versions 1.x, 2.0.x, 2.1.x, 2.2.0 to 2.2.2, and 2.3.0 to 2.3.1. 2.2.1, 2.1.2, 2.1.3, 2.2.0, 2.3.1, 2.3.0, 2.1.1 Major → 3.4.0
CVE-2018-1334 In Apache Spark 1.0.0 to 2.1.2, 2.2.0 to 2.2.1, and 2.3.0, when using PySpark or SparkR, it's possible for a different local user to connect to the Spark application and impersonate the user running the Spark application. 2.2.1, 2.1.2, 2.2.0, 2.1.1 Major → 3.4.0
CVE-2019-10099 Prior to Spark 2.3.3, in certain situations Spark would write user data to local disk unencrypted, even if spark.io.encryption.enabled=true. This includes cached blocks that are fetched to disk (controlled by spark.maxRemoteBlockSizeFetchToMem); in SparkR, using parallelize; in Pyspark, using broadcast and parallelize; and use of python udfs. 2.2.1, 2.1.2, 2.1.3, 2.2.0, 2.3.1, 2.3.2, 2.3.0, 2.1.1 (Show all) Major → 3.4.0
CVE-2020-9480 In Apache Spark 2.4.5 and earlier, a standalone resource manager's master may be configured to require authentication (spark.authenticate) via a shared secret. When enabled, however, a specially-crafted RPC to the master can succeed in starting an application's resources on the Spark cluster, even without the shared key. This can be leveraged to execute shell commands on the host machine. This does not affect Spark clusters using other resource managers (YARN, Mesos, etc). 2.2.1, 2.1.2, 2.1.3, 2.2.0, 2.4.5, 2.4.4, 2.4.1, 2.3.1 (Show all) Major → 3.4.0
CVE-2021-38296 Apache Spark supports end-to-end encryption of RPC connections via "spark.authenticate" and "spark.network.crypto.enabled". In versions 3.1.2 and earlier, it uses a bespoke mutual authentication protocol that allows for full encryption key recovery. After an initial interactive attack, this would allow someone to decrypt plaintext traffic offline. Note that this does not affect security mechanisms controlled by "spark.authenticate.enableSaslEncryption", "spark.io.encryption.enabled", "spark.ssl", "spark.ui.strictTransportSecurity". Update to Apache Spark 3.1.3 or later 2.2.1, 2.1.2, 2.1.3, 2.2.0, 3.1.1, 3.1.2, 3.0.3, 3.0.0 (Show all) Major → 3.4.0

Instantly see if these pyspark vulnerabilities affect your code.

Scan for Free