Skip to content

Excel

Bases: ReadWriteFileFormat

Excel file format. |support_hooks|

Based on Spark Excel <https://github.com/crealytics/spark-excel>_ file format.

Supports reading/writing files with .xlsx (read/write) and .xls (read only) extensions.

.. dropdown:: Version compatibility

* Spark versions: 3.2.x - 4.0.x

    .. warning::

        Not all combinations of Spark version and package version are supported.
        See `Maven index <https://mvnrepository.com/artifact/dev.mauch/spark-excel>`_
        and `official documentation <https://github.com/crealytics/spark-excel>`_.

* Java versions: 8 - 22

See documentation from link above.

.. versionadded:: 0.9.4

Examples

.. note ::

You can pass any option mentioned in
`official documentation <https://github.com/crealytics/spark-excel>`_.
**Option names should be in** ``camelCase``!

The set of supported options depends on ``spark-excel`` package version.

.. tabs::

.. code-tab:: py Reading files

    from pyspark.sql import SparkSession
    from onetl.file.format import Excel

    # Create Spark session with Excel package loaded
    maven_packages = Excel.get_packages(
        package_version="0.31.2",
        spark_version="3.5.7",
    )
    spark = (
        SparkSession.builder.appName("spark-app-name")
        .config("spark.jars.packages", ",".join(maven_packages))
        .getOrCreate()
    )

    excel = Excel(header=True, inferSchema=True)

.. code-tab:: py Writing files

    # Create Spark session with Excel package loaded
    spark = ...

    from onetl.file.format import XML

    excel = Excel(header=True, dataAddress="'Sheet1'!A1")
Source code in onetl/file/format/excel.py
 27
 28
 29
 30
 31
 32
 33
 34
 35
 36
 37
 38
 39
 40
 41
 42
 43
 44
 45
 46
 47
 48
 49
 50
 51
 52
 53
 54
 55
 56
 57
 58
 59
 60
 61
 62
 63
 64
 65
 66
 67
 68
 69
 70
 71
 72
 73
 74
 75
 76
 77
 78
 79
 80
 81
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
@support_hooks
class Excel(ReadWriteFileFormat):
    """
    Excel file format. |support_hooks|

    Based on `Spark Excel <https://github.com/crealytics/spark-excel>`_ file format.

    Supports reading/writing files with ``.xlsx`` (read/write) and ``.xls`` (read only) extensions.

    .. dropdown:: Version compatibility

        * Spark versions: 3.2.x - 4.0.x

            .. warning::

                Not all combinations of Spark version and package version are supported.
                See `Maven index <https://mvnrepository.com/artifact/dev.mauch/spark-excel>`_
                and `official documentation <https://github.com/crealytics/spark-excel>`_.

        * Java versions: 8 - 22

        See documentation from link above.

    .. versionadded:: 0.9.4

    Examples
    --------

    .. note ::

        You can pass any option mentioned in
        `official documentation <https://github.com/crealytics/spark-excel>`_.
        **Option names should be in** ``camelCase``!

        The set of supported options depends on ``spark-excel`` package version.

    .. tabs::

        .. code-tab:: py Reading files

            from pyspark.sql import SparkSession
            from onetl.file.format import Excel

            # Create Spark session with Excel package loaded
            maven_packages = Excel.get_packages(
                package_version="0.31.2",
                spark_version="3.5.7",
            )
            spark = (
                SparkSession.builder.appName("spark-app-name")
                .config("spark.jars.packages", ",".join(maven_packages))
                .getOrCreate()
            )

            excel = Excel(header=True, inferSchema=True)

        .. code-tab:: py Writing files

            # Create Spark session with Excel package loaded
            spark = ...

            from onetl.file.format import XML

            excel = Excel(header=True, dataAddress="'Sheet1'!A1")
    """

    name: ClassVar[str] = "excel"

    header: bool = False
    """
    If ``True``, the first row in file is conditioned as a header.
    Default ``False``.
    """

    dataAddress: Optional[str] = None
    """
    Cell address used as starting point.
    For example: ``'A1'`` or ``'Sheet1'!A1``
    """

    timestampFormat: Optional[str] = None
    """
    Format string used for parsing or serializing timestamp values.
    Default ``yyyy-mm-dd hh:mm:ss[.fffffffff]``.
    """

    dateFormat: Optional[str] = None
    """
    Format string used for parsing or serializing date values.
    Default ``yyyy-MM-dd``.
    """

    treatEmptyValuesAsNulls: Optional[bool] = None
    """
    If ``True``, empty cells are parsed as ``null`` values.
    If ``False``, empty cells are parsed as empty strings.
    Default ``True``.

    .. note::

        Used only for reading files.
    """

    setErrorCellsToFallbackValues: Optional[bool] = None
    """
    If ``True``, cells containing ``#N/A`` value are replaced with default value for column type,
    e.g. 0 for ``IntegerType()``. If ``False``, ``#N/A`` values are replaced with ``null``.
    Default ``False``.

    .. note::

        Used only for reading files.
    """

    usePlainNumberFormat: Optional[bool] = None
    """
    If ``True``, read or write numeric values with plain format, without using scientific notation or rounding.
    Default ``False``.
    """

    inferSchema: Optional[bool] = None
    """
    If ``True``, infer DataFrame schema based on cell content.
    If ``False`` and no explicit DataFrame schema is passed, all columns are ``StringType()``.

    .. note::

        Used only for reading files.
    """

    workbookPassword: Optional[SecretStr] = None
    """
    If Excel file is encrypted, provide password to open it.

    .. note::

        Used only for reading files. Cannot be used to write files.
    """

    maxRowsInMemory: Optional[int] = None
    """
    If set, use streaming reader and fetch only specified number of rows per iteration.
    This reduces memory usage for large files.
    Default ``None``, which means reading the entire file content to memory.

    .. warning::

        Can be used only with ``.xlsx`` files, but fails on ``.xls``.

    .. note::

        Used only for reading files.
    """

    maxByteArraySize: Optional[ByteSize] = None
    """
    If set, overrides memory limit (in bytes) of byte array size used for reading rows from input file.
    Default ``0``, which means using default limit.

    See `IOUtils.setByteArrayMaxOverride <https://poi.apache.org/apidocs/5.0/org/apache/poi/util/IOUtils.html#setByteArrayMaxOverride-int->`_
    documentation.

    .. note::

        Used only for reading files.
    """

    tempFileThreshold: Optional[ByteSize] = None
    """
    If value is greater than 0, large zip entries will be written to temporary files after reaching this threshold.
    If value is 0, all zip entries will be written to temporary files.
    If value is -1, no temp files will be created, which may cause errors if zip entry is larger than 2GiB.

    .. note::

        Used only for reading files.
    """

    excerptSize: Optional[int] = None
    """
    If ``inferSchema=True``, set number of rows to infer schema from.
    Default ``10``.

    .. note::

        Used only for reading files.
    """

    class Config:
        known_options = frozenset()
        extra = "allow"

    @slot
    @classmethod
    def get_packages(
        cls,
        package_version: str,
        spark_version: str,
        scala_version: str | None = None,
    ) -> list[str]:
        """
        Get package names to be downloaded by Spark. |support_hooks|

        .. warning::

            Not all combinations of Spark version and package version are supported.
            See `Maven index <https://mvnrepository.com/artifact/dev.mauch/spark-excel>`_
            and `official documentation <https://github.com/crealytics/spark-excel>`_.

        .. versionadded:: 0.9.4
        .. versionchanged:: 0.14.0
            Maven package ``com.crealytics:spark-excel`` was renamed to ``dev.mauch:spark-excel``.

        Parameters
        ----------
        package_version : str
            Package version in format ``major.minor.patch``.

            .. versionchanged:: 0.14.0
                This parameter is now mandatory.

        spark_version : str
            Spark version in format ``major.minor.patch``.

        scala_version : str, optional
            Scala version in format ``major.minor``.

            If ``None``, ``spark_version`` is used to determine Scala version.

        Examples
        --------

        .. code:: python

            from onetl.file.format import Excel

            Excel.get_packages(
                package_version="0.31.2",
                spark_version="3.5.7",
            )
            Excel.get_packages(
                package_version="0.31.2",
                spark_version="3.5.7",
                scala_version="2.12",
            )

        """

        version = Version(package_version)
        if version < Version("0.30"):
            raise ValueError(f"Package version should be at least 0.30, got {package_version}")

        spark_ver = Version(spark_version).min_digits(3)
        scala_ver = Version(scala_version).min_digits(2) if scala_version else get_default_scala_version(spark_ver)
        return [
            f"dev.mauch:spark-excel_{scala_ver.format('{0}.{1}')}:{spark_ver.format('{0}.{1}.{2}')}_{version}",
        ]

    @slot
    def check_if_supported(self, spark: SparkSession) -> None:
        java_class = "dev.mauch.spark.excel.v2.ExcelDataSource"

        try:
            try_import_java_class(spark, java_class)
        except Exception as e:
            spark_version = get_spark_version(spark).format("{0}.{1}.{2}")
            msg = MISSING_JVM_CLASS_MSG.format(
                java_class=java_class,
                package_source=self.__class__.__name__,
                args=f"spark_version='{spark_version}'",
            )
            raise ValueError(msg) from e

    @slot
    def apply_to_reader(self, reader: DataFrameReader) -> DataFrameReader:
        options = self.dict(by_alias=True, exclude_none=True)
        if self.workbookPassword:
            options["workbookPassword"] = self.workbookPassword.get_secret_value()
        return reader.format(self.name).options(**options)

    def __repr__(self):
        options_dict = self.dict(by_alias=True, exclude_none=True)
        options_dict = dict(sorted(options_dict.items()))
        options_kwargs = ", ".join(f"{k}={v!r}" for k, v in options_dict.items())
        return f"{self.__class__.__name__}({options_kwargs})"

header = False class-attribute instance-attribute

If True, the first row in file is conditioned as a header. Default False.

dataAddress = None class-attribute instance-attribute

Cell address used as starting point. For example: 'A1' or 'Sheet1'!A1

treatEmptyValuesAsNulls = None class-attribute instance-attribute

If True, empty cells are parsed as null values. If False, empty cells are parsed as empty strings. Default True.

.. note::

Used only for reading files.

setErrorCellsToFallbackValues = None class-attribute instance-attribute

If True, cells containing #N/A value are replaced with default value for column type, e.g. 0 for IntegerType(). If False, #N/A values are replaced with null. Default False.

.. note::

Used only for reading files.

usePlainNumberFormat = None class-attribute instance-attribute

If True, read or write numeric values with plain format, without using scientific notation or rounding. Default False.

inferSchema = None class-attribute instance-attribute

If True, infer DataFrame schema based on cell content. If False and no explicit DataFrame schema is passed, all columns are StringType().

.. note::

Used only for reading files.

timestampFormat = None class-attribute instance-attribute

Format string used for parsing or serializing timestamp values. Default yyyy-mm-dd hh:mm:ss[.fffffffff].

maxRowsInMemory = None class-attribute instance-attribute

If set, use streaming reader and fetch only specified number of rows per iteration. This reduces memory usage for large files. Default None, which means reading the entire file content to memory.

.. warning::

Can be used only with ``.xlsx`` files, but fails on ``.xls``.

.. note::

Used only for reading files.

maxByteArraySize = None class-attribute instance-attribute

If set, overrides memory limit (in bytes) of byte array size used for reading rows from input file. Default 0, which means using default limit.

See IOUtils.setByteArrayMaxOverride <https://poi.apache.org/apidocs/5.0/org/apache/poi/util/IOUtils.html#setByteArrayMaxOverride-int->_ documentation.

.. note::

Used only for reading files.

tempFileThreshold = None class-attribute instance-attribute

If value is greater than 0, large zip entries will be written to temporary files after reaching this threshold. If value is 0, all zip entries will be written to temporary files. If value is -1, no temp files will be created, which may cause errors if zip entry is larger than 2GiB.

.. note::

Used only for reading files.

excerptSize = None class-attribute instance-attribute

If inferSchema=True, set number of rows to infer schema from. Default 10.

.. note::

Used only for reading files.

workbookPassword = None class-attribute instance-attribute

If Excel file is encrypted, provide password to open it.

.. note::

Used only for reading files. Cannot be used to write files.

get_packages(package_version, spark_version, scala_version=None) classmethod

Get package names to be downloaded by Spark. |support_hooks|

.. warning::

Not all combinations of Spark version and package version are supported.
See `Maven index <https://mvnrepository.com/artifact/dev.mauch/spark-excel>`_
and `official documentation <https://github.com/crealytics/spark-excel>`_.

.. versionadded:: 0.9.4 .. versionchanged:: 0.14.0 Maven package com.crealytics:spark-excel was renamed to dev.mauch:spark-excel.

Parameters

package_version : str Package version in format major.minor.patch.

.. versionchanged:: 0.14.0
    This parameter is now mandatory.
str

Spark version in format major.minor.patch.

str, optional

Scala version in format major.minor.

If None, spark_version is used to determine Scala version.

Examples

.. code:: python

from onetl.file.format import Excel

Excel.get_packages(
    package_version="0.31.2",
    spark_version="3.5.7",
)
Excel.get_packages(
    package_version="0.31.2",
    spark_version="3.5.7",
    scala_version="2.12",
)
Source code in onetl/file/format/excel.py
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
@slot
@classmethod
def get_packages(
    cls,
    package_version: str,
    spark_version: str,
    scala_version: str | None = None,
) -> list[str]:
    """
    Get package names to be downloaded by Spark. |support_hooks|

    .. warning::

        Not all combinations of Spark version and package version are supported.
        See `Maven index <https://mvnrepository.com/artifact/dev.mauch/spark-excel>`_
        and `official documentation <https://github.com/crealytics/spark-excel>`_.

    .. versionadded:: 0.9.4
    .. versionchanged:: 0.14.0
        Maven package ``com.crealytics:spark-excel`` was renamed to ``dev.mauch:spark-excel``.

    Parameters
    ----------
    package_version : str
        Package version in format ``major.minor.patch``.

        .. versionchanged:: 0.14.0
            This parameter is now mandatory.

    spark_version : str
        Spark version in format ``major.minor.patch``.

    scala_version : str, optional
        Scala version in format ``major.minor``.

        If ``None``, ``spark_version`` is used to determine Scala version.

    Examples
    --------

    .. code:: python

        from onetl.file.format import Excel

        Excel.get_packages(
            package_version="0.31.2",
            spark_version="3.5.7",
        )
        Excel.get_packages(
            package_version="0.31.2",
            spark_version="3.5.7",
            scala_version="2.12",
        )

    """

    version = Version(package_version)
    if version < Version("0.30"):
        raise ValueError(f"Package version should be at least 0.30, got {package_version}")

    spark_ver = Version(spark_version).min_digits(3)
    scala_ver = Version(scala_version).min_digits(2) if scala_version else get_default_scala_version(spark_ver)
    return [
        f"dev.mauch:spark-excel_{scala_ver.format('{0}.{1}')}:{spark_ver.format('{0}.{1}.{2}')}_{version}",
    ]