Tutorial | Prepare recipe#
The Prepare recipe allows you to define a series of steps, or actions, to take on the dataset.
Get started#
Objectives#
In this tutorial, you will:
Create a Prepare recipe.
Add various processor steps in different ways.
Process dates, numbers and text.
Use the Formula language.
Prerequisites#
To complete this tutorial, you’ll need the following:
Dataiku 12.0 or later.
Create the project#
From the Dataiku Design homepage, click + New Project.
Select Learning projects.
Search for and select Prepare Recipe.
Click Install.
From the project homepage, click Go to Flow (or
g
+f
).
From the Dataiku Design homepage, click + New Project.
Select DSS tutorials.
Filter by Core Designer.
Select Prepare Recipe.
From the project homepage, click Go to Flow (or
g
+f
).
Note
You can also download the starter project from this website and import it as a zip file.
You’ll next want to build the Flow.
Click Flow Actions at the bottom right of the Flow.
Click Build all.
Keep the default settings and click Build.
Use case summary#
Let’s say we’re a financial company that uses some credit card data to detect fraudulent transactions.
The project comes with three datasets, described in the table below.
Dataset |
Description |
---|---|
tx |
Each row is a unique credit card transaction with information such as the card that was used and the merchant where the transaction was made. It also indicates whether the transaction has either been:
|
merchants |
Each row is a unique merchant with information such as the merchant’s location and category. |
cards |
Each row is a unique credit card ID with information such as the card’s activation month or the cardholder’s FICO score (a common measure of creditworthiness in the US). |
Create a Prepare recipe#
Let’s get started preparing the dataset!
From the Flow, select the tx_joined dataset and click the Actions icon (+) from the right panel to open the Actions tab.
Under the Visual Recipes section, click on Prepare.
The New data preparation recipe window opens.
Keep tx_joined as the input dataset.
Name the output dataset
tx_prepared
.Click Create Recipe.
Add steps to the recipe#
Now that we have our Prepare recipe, it’s time to add some steps to it.
Process dates#
Let’s start with dates.
Only display date columns#
This step is optional but it shows you how to display only the columns you select.
Here, we want to focus our attention on the date columns, which are purchase_date and first_active_month. To do so:
At the top right of the dataset preview, click on Display > Select displayed columns.
In the Displayed columns dialog:
Set the Display option to Selected columns.
Unselect all columns using the checkbox at the top, to the left of the search field.
Select purchase_date and first_active_month.
Click OK. The dataset should now display two columns only.
Note
Recall that the data visible in the recipe is a preview of a sample of your data.
Click Save.
Parse dates#
In this section, we’ll parse the two date columns: purchase_date and first_active_month.
At this point, the storage type of these columns is a string, but their meanings inferred by Dataiku is an unparsed date. Let’s parse them so that we can treat them as a proper date.
Open the purchase_date column header dropdown, and select Parse date.
It opens the Smart Date dialog, which shows the most likely formats for our dates and what the dates would look like once parsed, with some sample values from the dataset. In our case, the dates appear to be in yyyy-MM-dd format.
With the yyyy-MM-dd format selected, click Use Date Format.
See that a Parse date step has been added to the script.
Repeat steps 1 to 3 on first_active_month, using the yyyy-MM format.
Click Save.
By default, these steps create two new columns named purchase_date_parsed and first_active_month_parsed. Note how both their storage types and meanings are a date.
Delete and rename columns#
In the Parse date script step, we could leave the name of the output column empty in order to parse the column in place. Instead of overriding the original date column, let’s delete it and rename the new column.
Click on the purchase_date column header dropdown, and choose Delete. It adds a Remove columns step to the Script panel.
In the Remove columns step, select multiple next to Column and add the first_active_month column.
Click on the purchase_date_parsed column header dropdown, and choose Rename.
Enter the name
purchase_date
, and click OK.Repeat steps 3 and 4 to rename first_active_month_parsed into first_active_month.
For the last step in the script, click on the eye icon to re-enable the step preview.
Note
Recall that the data visible in the recipe is a preview of what your output dataset will look like based on a sample of data.
Click Save.
Extract date components#
Now that the dates are parsed, let’s extract some date components from purchase_date, like the year, month, day and day of the week.
Click on the purchase_date column header dropdown, and choose Extract date components. It adds a step to the Script panel and three columns to the dataset.
In the script, set the following options as below:
In the Year column field, enter
purchase_year
.In the Month column field, enter
purchase_month
.In the Day column field, enter
purchase_day
.In the Day of week column field, enter
purchase_dow
.
Click Save.
Important
The values in the purchase_dow (day of the week) column ranges from 1
to 7
, where 1
corresponds to Monday.
Compute time difference#
Let’s compute the time from the first active month to the purchase date.
Click on the first_active_month column header dropdown, and choose Compute time since. It adds a step to the Script panel.
In the script, configure the Compute time difference step as follows:
Leave the Time since column option to
first_active_month
.Set the Until option to Another date column.
Set the Other column option to
purchase_date
.Ensure the Output time unit option is set to Days.
Enter
days_active
as the name of the Output column.
Click Save.
Show all columns from the dataset#
Now that we’ve finished processing the dates, let’s display all the columns in the dataset.
At the top right of the dataset preview, click on Display > Select displayed columns.
In the Displayed columns dialog, set the Display option to All columns.
Click OK. The dataset should now display all columns.
Click Save.
Process numbers#
Convert currencies#
One handy processor provides the ability to convert currencies based on a date column.
This time, we will not add the step by clicking on a column header, but instead search the processors library which includes about 100 data preparation processors.
At the bottom of the Script panel, click + Add a New Step.
Search for Convert currencies.
Enter
purchase_amount
as the column.Select USD as the input currency.
Select EUR as the output currency.
For input date source, select From Column (Date), and enter
purchase_date
as the date column.Name the output column
purchase_amount_eur
.Click Save.
Round numbers#
Let’s also round the original purchase_amount values for simplicity.
Click on the column header for purchase_amount.
Click on Round to integer.
Increase the Decimal places to
2
.Click Save.
Tip
In the Script panel, try to disable/enable this step. Dataiku processes the steps sequentially. So, rounding the amounts after the conversion step does not have any impact on the purchase_amount_eur col.
Process text#
Simplify text#
If we want to work with a text or natural language column, a good starting point is often to normalize the text. When you do so, Dataiku transforms all text to lowercase, removes punctuation and accents, and performs unicode normalization.
Click on the column header for product_title.
Click Simplify text.
Leave the default option to normalize the text.
Click Save.
Extract text with a regular expression#
We can manipulate string data with regular expressions in many places throughout Dataiku, including in data preparation. As an example, let’s use a regular expression to extract into a new column the name of the first match of common Apple products.
At the bottom of the Script panel, click + Add a New Step.
Search for Extract with regular expression.
Enter:
product_title
as the input column.apple
as the prefix for the output column.\b(apple|macbook|ipad|iphone|ipod)\b
as the regular expression.
Click Save.
As this regular expression does not impact all rows, to see the impact more easily, let’s filter the column:
From the apple1 column header dropdown, select Filter.
In the filter box, check all values except No value to exclude rows that are empty.
Once you’ve reviewed the apple1 column, delete the filter by clicking the bin icon next to it.
Note
Here, we’ve provided the regular expression for you, but you can explore how to use the smart pattern builder on your own.
Use a formula#
Finally, let’s use a Formula step to create a binary column named purchase_weekend. Dataiku formulas are a very powerful expression language to perform calculations, manipulate strings, and much more.
At the bottom of the Script panel, click + Add a New Step.
Search for Formula.
In the Script panel, enter
purchase_weekend
as the name of the new output column.Click the Open Editor Panel to use the advanced formula editor, which will autocomplete column names and show you a preview of the output.
In the Formula field, enter
if(purchase_dow>5,1,0)
.Click Apply then Save.
Important
The purchase_weekend column appears in the dataset only after clicking the Apply button.
As you can see, purchase_weekend is set to:
0
(false) when the day of the week stored in the purchase_dow column ranges from1
to5
, as1
corresponds to Monday and5
to Friday.1
(true) when the day of the week is either6
or7
(Saturday and Sunday, respectively).
Group steps together#
Dataiku lets you group steps to organize the Script panel. Here let’s group together the steps used to process the dates.
In the Script panel, select the first six steps (two Parse date steps, Remove, Rename, Extract date and Compute time difference).
On top of the step list, click Actions > Group.
Name the group
Process dates
.Click Save.
At any time, you can click on the group to expand all the steps it includes.
Run the recipe#
Once all the steps are defined in your script, it is time to apply the script on the whole dataset and explore the output dataset.
Click Run in the lower-left corner of the page (or use the keyboard shortcut
@+r+u+n
).When the job completes, click Explore dataset tx_prepared to view the output dataset. You can also return to the Flow and see your progress.
What’s next?#
Congratulations on completing your first Prepare recipe! This tutorial only shows a small subset of what can be done in a Prepare recipe.
See also
For more information on data preparation, see also:
The Processors reference article in the reference documentation.
The Formula cheat sheet.
The data preparation section in the Knowledge Base.
Now, there’s still more data exploration and preparation to be done using other visual recipes.
Let’s dive deeper by grouping some data on a dataset.