Table of Contents

Article

Connect to remote repositories through File data sets

You can now transfer data between cloud-based storage systems and your applications in the cloud by using data flows. Instead of creating and maintaining relational databases to transfer data to or from remote files, you can add a reference to a remote repository in a File data set. The files may contain information, such as customer data, purchase lists, or offer details, that you can apply to your applications by using parallel loads from CSV or JSON files, including compressed ZIP and GZIP archives.

The following figure presents the configuration of remote repositories:

Thumbnail

Data set configuration - remote repositories

Configuring a File data set with a remote repository connection allows you to automatically process that data in various scenarios. For example, by referencing that File data set from a data flow in a decisioning strategy for promotional offers, you can automatically download a detailed list of customers that is stored in a collection of GZIP archives in a Microsoft Azure cloud directory and let the strategy model select the clients who receive new discount offers.

By using the updated data set interface, you can map Pega Platform properties to the relevant columns in the remote file:

Thumbnail

Data set configuration - mapping properties

 

See the following video for a sample data set creation process:

Apart from sourcing data, you can also configure an automatic export of data to a set of remote files at the same time through wildcards.

Currently, Pega Platform supports connections with the following repositories:

  • JFrog Artifactory
  • Amazon S3
  • Microsoft Azure
  • local file system

For more information, see:

Published August 13, 2018 — Updated November 5, 2018

Related Content

Have a question? Get answers now.

Visit the Pega Support Community to ask questions, engage in discussions, and help others.