Skip to main content

Local filesystem

Sample setup

If running the samples on MacOS or Linux, you may skip this section.

To run the samples locally on Windows, you must first create the docker image:

$ docker build --build-arg SL_VERSION=0.7.2.2 -t starlake .

One the docker image is built locally, run it:

$ docker run -it starlakeai/starlake:VERSION bash

Running the samples

Inside the docker container, make sure you are in the samples/local folder

The userguide-template is first duplicated into the samples/local folder to create a startup project

$ ./0.data-init.sh

Then you need to import he files located in userguide/incoming into the correct pending folder depending on the domain they belong to:

$ ./1.data-import.sh

To start the ingestion process, run the load command. The resulting tables should be available in the userguide/datasets/accepted folder:

$ ./2.data-load.sh

To join multiple datasets using the KPI job example located in userguide/metadata/jobs/kpi.sql, run the corresponding transformation:

$ ./3.data-transform.sh

To view the data ingested and stored as parquet files:

$ ./4.data-view-results.sh

To exit the spark shell above type :quit

To view the log produced:

$ ./4.data-view-audit.sh

To exit the spark shell above type :quit

Optional

You may view the relationship between your tables by generating a graphviz diagram using the command below:

$ ./1.data-visualization.sh