Iris dataset example project¶
In this chapter we describe the project directory structure for a typical Kedro project.
This example uses the simple and familiar Iris dataset.
The Iris dataset was generated in 1936 by the British statistician and biologist Ronald Fisher. It contains 150 samples in total, comprising 50 samples of 3 different species of Iris plant (Iris Setosa, Iris Versicolour and Iris Virginica). For each sample, the flower measurements are recorded for the sepal length, sepal width, petal length and petal width, as illustrated in the following graphic.
The Iris dataset can be used by a machine learning model to illustrate classification (a method used to determine the type of an object by comparison with similar objects that have previously been categorised). Once trained on known data, the machine learning model can make a predictive classification by comparing a test object to the output of its training data.
Create the example project¶
You must first create a project. Feel free to name your project as you like, but here we will assume the project’s repository name is
Be sure to enter
Yat the final prompt, so your new project is pre-populated with the Iris dataset example.
If you prefer, clone the example project from the
kedro-examples repository on Github
Project directory structure¶
This example project illustrates a convenient starting point and some best-practices. It follows the default Kedro project template and uses folders to store datasets, notebooks, configuration and source code. When you create your own projects, you can adapt the folder structure if you need to.
The example project directory is set out as follows:
getting-started # Parent directory of the template ├── conf # Project configuration files ├── data # Local project data (not committed to version control) ├── docs # Project documentation ├── kedro_cli.py # A collection of Kedro command line interface (CLI) commands ├── logs # Project output logs (not committed to version control) ├── notebooks # Project related Jupyter notebooks (can be used for experimental code before moving the code to src) ├── README.md # Project README ├── setup.cfg # Configuration options for `pytest` when doing `kedro test` └── src # Project source code
Kedro also creates the following hidden files and folders:
getting-started ├── .coveragerc # Configuration file for the coverage reporting when doing `kedro test` ├── .gitignore # Prevent staging of unnecessary files to `git` ├── .ipython # IPython startup scripts ├── .isort.cfg # Configuration file for the `isort` utility when doing `kedro lint` └── .kedro.yml # Identifies the project root and [contains configuration information](https://kedro.readthedocs.io/en/latest/06_resources/02_architecture_overview.html?#kedro-yml)
conf folder, there are two subfolders for storing configuration information:
For project-specific settings to share across different installations (for example, with different users) you should use the
base subfolder of
The folder contains three files for the example, but you can add others as you require:
catalog.yml- Configures the Data Catalog with the file paths and load/save configuration required for different datasets
logging.yml- Uses Python’s default
logginglibrary to set up logging
parameters.yml- Allows you to define parameters for machine learning experiments e.g. train / test split and number of iterations
local subfolder of
conf is used for settings that should not be shared, such as access credentials, custom editor configuration, personal IDE configuration and other sensitive or personal content. It is specific to user and installation. The contents of
conf/local/ is ignored by
git (through inclusion in
By default, when Kedro creates
conf/local folder, it is empty. However, Kedro creates
conf/base/ for use as an example. Before you populate and use the file, you should first move it to
data folder contains a number of subfolders to store project data. We recommend that you put raw data into
raw and move processed data to other subfolders according to data engineering convention.
The example project has a single file,
iris.csv, that contains the Iris dataset. The subfolders of
data are ignored by
git through inclusion in
.gitignore since data is more frequently stored elsewhere, such as a in an S3 bucket. However, if you are familiar with
.gitignore you can edit it, if you are confident that you need to manage your data in
This subfolder contains the project’s source code. It contains 2 subfolders:
getting_started/This is the Python package for your project
tests/The subfolder for unit tests for your project. Projects are preconfigured to run tests using
pytestwhen you call
kedro testfrom the project’s root directory
What best practice should I follow to avoid leaking confidential data?¶
- Avoid committing data to version control.
- Avoid committing data to notebook output cells (data can easily sneak into notebooks when you don’t delete output cells).
- Don’t commit sensitive results or plots to version control (in notebooks or otherwise).
- Don’t commit credentials in
conf/. Only the
conf/local/folder should be used for sensitive information like access credentials. To add credentials, please refer to the
conf/base/credentials.ymlfile in the project template.
- By default any file inside the
conf/folder (and its subfolders) containing
credentialsin its name will be ignored via
.gitignoreand not committed to your git repository.
- To describe where your colleagues can access the credentials, you may edit the
README.mdto provide instructions.
Run the example project¶
Navigate to the root directory of the example project, and call
cd getting-started kedro run
When it completes, you should see a log message similar to the following in your console:
2019-02-13 16:59:26,293 - kedro.runner.sequential_runner - INFO - Completed 4 out of 4 tasks 2019-02-13 16:59:26,293 - kedro.runner.sequential_runner - INFO - Pipeline execution completed successfully.
Under the hood: Pipelines and nodes¶
The example project contains two modular pipelines:
src/getting_started/pipelines/data_engineering/pipeline.py) responsible for splitting the data into training and testing samples
src/getting_started/pipelines/data_science/pipeline.py) responsible for model training, predictions and accuracy-reporting
Data engineering node
This is the data engineering node function within
|Node||Description||Node Function Name|
|Split data||Splits the example Iris dataset into train and test samples||
Data science node
These are the data science node functions within
|Node||Description||Node Function Name|
|Train model||Trains a simple multi-class logistic regression model||
|Predict||Makes class predictions given a pre-trained model and a test set||
|Report accuracy||Reports the accuracy of the predictions performed by the previous node||
src/pipeline.py creates and collates the project’s modular pipelines into a single pipeline, resolving node execution order from the input and output data dependencies between the nodes.