Spark 3.2.0.200 - 2405 (EEP 8.1.2) Release Notes

This section provides reference information, including new features, patches, and known issues for Spark 3.2.0.200.

The notes below relate specifically to the Hewlett Packard Enterprise Distribution for Apache Hadoop. For more information, you may also want to consult the open-source Spark 3.2.0 Release Notes.

These release notes contain only Hewlett Packard Enterprise specific information and are not necessarily cumulative in nature. For information about how to use the release notes, see Ecosystem Component Release Notes.

Spark Version 3.2.0.200
Release Date May 2024
HPE Version Interoperability See Component Versions for Released EEPs and EEP Components and OS Support.
Source on GitHub https://github.com/mapr/spark
GitHub Release Tag 3.2.0.200-eep-812
Maven Artifacts https://repository.mapr.com/maven/
Package Names Navigate to https://package.ezmeral.hpe.com/releases/MEP/ and select your EEP and OS to view the list of package names.
IMPORTANT
  • Beginning with EEP 6.0.0, the KeyStore and TrustStore password can be removed from spark-defaults.conf and set in /opt/mapr/conf/ssl-client.xml.
  • Beginning with Core 6.2 and EEP 7.0, Spark supports SSL for WebUI.

Hive Support

  • Starting from Spark 3.1.2, Spark supports Hive 2.3.

New in This Release

For a complete list of new features, see the open-source Spark 3.2.0 Release Notes.
  • Bug fixes.

Fixes

This HPE release includes the following new fixes since the latest Spark release. For details, refer to the commit log for this project in GitHub.

GitHub Commit Date (YYYY-MM-DD) Comment
eb80ac4 19/06/2023 MapR [SPARK-1190] Address CVE-2022-37865
cf3a39e

18/04/2024

[SPARK-37628][BUILD] Upgrade Netty from 4.1.68 to 4.1.72
ba46b34 18/04/2024 MapR [SPARK-1280] Prepare EEP-8.1.2 Spark release
7187d5c 03/05/2024 MapR [SPARK-1286] Pyspark + Panda integration doesn't work in Spark 3.2.0
10f677b 03/05/2024 MapR [SPARK-1195] Fix Spark scripts as htrace-core*.jar was removed from HBase
ee9bcf9 03/05/2024 MapR [SPARK-1108] Parallel jobs running causes errors with manageSSLKeys.sh

Known Issues and Limitations

  • When you enable the SSL in a mixed (FIPS and non-FIPS) configuration, Spark application run fails. To run Spark applications, set spark.ssl.ui.enabled option to false in spark-defaults.conf configuration file.

  • If you are using Spark SQL with Derby database without Hive or Hive Metastore installation, you will see the Java Runtime Exception. See Apache Spark Feature Support for workaround. Spark 3.1.2 does not support log4j1.2 logging on HPE Ezmeral Data Fabric.

Resolved Issues

  • None.