.. etl_toolkit documentation master file, created by sphinx-quickstart on Sat Jul 13 13:40:31 2024. You can adapt this file completely to your liking, but it should at least contain the root `toctree` directive. Welcome to etl_toolkit's documentation! ======================================= The ``etl_toolkit`` contains many utilities to write better pyspark to simplify pipelines on Databricks. Navigate to each module of the ``etl_toolkit`` to learn about the functions provided and how to use them. - **expressions**: Contains functions for deriving complex ``pyspark.Columns`` - **analyses**: Contains functions for deriving complex ``pyspark.DataFrames`` that apply many data transformations. .. toctree:: :maxdepth: 3 :caption: Contents: expressions analyses table_creation mutable_table Indices and tables ================== * :ref:`genindex` * :ref:`modindex` * :ref:`search` .. meta:: :description: The ETL toolkit contains many pyspark-based functions to build high quality data pipelines while following spark best practices. The functions conveniently encapsulate common cleaning and analytical operations teams need to do on incoming sources of data.