Data Science Workspace is no longer available for purchase.
This documentation is intended for existing customers with prior entitlements to Data Science Workspace.
This tutorial provides instructions on how you can package the provided Retail Sales sample source files into an archive file, which can be used to create a recipe in Adobe Experience Platform Data Science Workspace by following the recipe import workflow either in the UI or using the API.
Concepts to understand:
Recipe creation starts with packaging source files to build an archive file. Source files define the machine learning logic and algorithms used to solve a specific problem at hand, and are written in either Python, R, PySpark, or Scala. Built archive files take the form of a Docker image. Once built, the packaged archive file is imported into Data Science Workspace to create a recipe in the UI or using the API.
A Docker image allows a developer to package up an application with all the parts it needs, such as libraries and other dependencies, and ship it out as one package.
The built Docker image is pushed to the Azure Container Registry using credentials supplied to you during the recipe creation workflow.
To obtain your Azure Container Registry credentials, log into Adobe Experience Platform. On the left navigation column, navigate to Workflows. Select Import Recipe followed by selecting Launch. See the screen shot below for reference.
The Configure page opens. Provide an appropriate Recipe Name, for example, “Retail Sales recipe”, and optionally provide a description or documentation URL. Once complete, click Next.
Select the appropriate Runtime, then choose a Classification for Type. Your Azure Container Registry credentials are generated once complete.
Type is the class of machine learning problem the recipe is designed for and is used after training to help tailor evaluating the training run.
Note the values for Docker host, username, and password. These are used to build and push your Docker image in the workflows outlined below.
The Source URL is provided after completing the steps outlined below. The configuration file is explained in subsequent tutorials found in next steps.
Start by obtaining the sample codebase found in the Experience Platform Data Science Workspace Reference repository.
If you have not done so, clone the GitHub repository onto your local system with the following command:
git clone https://github.com/adobe/experience-platform-dsw-reference.git
Navigate to the directory experience-platform-dsw-reference/recipes/python/retail
. Here, you will find the scripts login.sh
and build.sh
used to login to Docker and to build the Python Docker image. If you have your Docker credentials ready, enter the following commands in order:
# for logging in to Docker
./login.sh
# for building Docker image
./build.sh
Note that when executing the login script, you need to provide the Docker host, username, and password. When building, you are required to provide the Docker host and a version tag for the build.
Once the build script is complete, you are given a Docker source file URL in your console output. For this specific example, it will look something like:
# URL format:
{DOCKER_HOST}/ml-retailsales-python:{VERSION_TAG}
Copy this URL and move on to the next steps.
If you have not done so, clone the GitHub repository onto your local system with the following command:
git clone https://github.com/adobe/experience-platform-dsw-reference.git
Navigate to the directory experience-platform-dsw-reference/recipes/R/Retail - GradientBoosting
inside your cloned repository. Here, you’ll find the files login.sh
and build.sh
which you will use to login to Docker and to build the R Docker image. If you have your Docker credentials ready, enter the following commands in order:
# for logging in to Docker
./login.sh
# for build Docker image
./build.sh
Note that when executing the login script, you need to provide the Docker host, username, and password. When building, you are required to provide the Docker host and a version tag for the build.
Once the build script is complete, you are given a Docker source file URL in your console output. For this specific example, it will look something like:
# URL format:
{DOCKER_HOST}/ml-retail-r:{VERSION_TAG}
Copy this URL and move on to the next steps.
Start by cloning the GitHub repository onto your local system with the following command:
git clone https://github.com/adobe/experience-platform-dsw-reference.git
Navigate to the directory experience-platform-dsw-reference/recipes/pyspark/retail
. The scripts login.sh
and build.sh
are located here and used to login to Docker and to build the Docker image. If you have your Docker credentials ready, enter the following commands in order:
# for logging in to Docker
./login.sh
# for building Docker image
./build.sh
Note that when executing the login script, you need to provide the Docker host, username, and password. When building, you are required to provide the Docker host and a version tag for the build.
Once the build script is complete, you are given a Docker source file URL in your console output. For this specific example, it will look something like:
# URL format:
{DOCKER_HOST}/ml-retailsales-pyspark:{VERSION_TAG}
Copy this URL and move on to the next steps.
Start by cloning the GitHub repository onto your local system with the following command in terminal:
git clone https://github.com/adobe/experience-platform-dsw-reference.git
Next, navigate to the directory experience-platform-dsw-reference/recipes/scala
where you can find the scripts login.sh
and build.sh
. These scripts are used to login to Docker and build the Docker image. If you have your Docker credentials ready, enter the following commands to terminal in order:
# for logging in to Docker
./login.sh
# for building Docker image
./build.sh
If you are receiving a permission error when trying to login to Docker using the login.sh
script, try using the command bash login.sh
.
When executing the login script, you need to provide the Docker host, username, and password. When building, you are required to provide the Docker host and a version tag for the build.
Once the build script is complete, you are given a Docker source file URL in your console output. For this specific example, it will look something like:
# URL format:
{DOCKER_HOST}/ml-retailsales-spark:{VERSION_TAG}
Copy this URL and move on to the next steps.
This tutorial went over packaging source files into a Recipe, the prerequisite step for importing a Recipe into Data Science Workspace. You should now have a Docker image in Azure Container Registry along with the corresponding image URL. You are now ready to begin the tutorial on importing a packaged recipe into Data Science Workspace. Select one of the tutorial links below to get started: