Tutorial: Configuring a remote repository as a source for a File data set
To use data stored in cloud-based storage systems such as JFrog Artifactory, Amazon S3, or Microsoft Azure in your Pega Platform™ applications, configure a remote repository as a data source for a File data set. You can apply this functionality to set an automated data transfer to Pega Platform in the cloud by enabling parallel loads from CSV and JSON files that are stored remotely, instead of creating and maintaining relational databases to transfer remote files.
You can then convert and use the transferred data in various scenarios, such as in a marketing strategy or for customized correspondence.
After you create a File data set with a remote repository, you add a reference to the File data set in a data flow. See Referencing remote repository data in a data flow.
The marketing team of the Xauto company wants to use their client data, which is stored in Amazon S3, in their decision strategy on Pega Platform in the cloud.
Creating a File data set with a remote repository
To use data that is stored in a remote repository, such as Amazon S3, create a File data set that references that directory.
- In Dev Studio, click .
- In the Data Set Record Configuration section, enter the data set parameters:
- In the Label field, enter a name for the new record, for example:
- In the Type field, select File.
- In the Context section, select the Apply to class and ruleset version.
- Confirm the settings by clicking .
- In the New tab, in the Data source section, click .
- In the Connection section, click the icon to the right of the Repository configuration field and configure the remote repository:
- In the Create Repository tab, enter a description in the Short description field, for example:
Xauto customer data directory (Amazon S3)
- Enter the Repository name, for example:
- In the Edit Repository tab, click and select S3 as your repository type.
- In the Configuration section, enter the parameters for your Amazon S3 repository:
- In the Bucket field, enter the S3 bucket location where artifacts are stored, for example:
- In the Authentication profile field, select or create an authentication profile to connect to the repository.
For more information, see Creating an authentication profile.
- To use an Amazon Key Management Service (KMS) keystore for storing keys and certificates, select the Server side data encryption with KMS managed keys check box and enter the KMS key ID.
For more information, see Keystores.
- In the Root path field, enter the location of the root folder in S3, for example:
- Verify the credentials by clicking .
- Click .
- In the File configuration section, in the File path field, enter the source file directory, for example:
To match multiple files in a folder, use an asterisk (*), for example:
Additional details about the selected file are displayed in the section.
- Optional: To preview the file, click .
- Optional: To update the settings for the selected file, in the Parser configuration section, enter new parameter values as in the following example:
Time properties in the selected file can be in a different time zone than the time zone that is used by Pega Platform. To avoid confusion, specify the time zone in the time properties of the file, and use the appropriate pattern in the settings.
- For CSV files, in the Mapping tab, modify the number of mapped columns:
- To add a new column, click Add mapping.
- To remove a column and the associated property mapping, click the Delete mapping icon for the applicable row.
- For CSV files, in the Mapping tab, check the mapping between the fields in the CSV file and the corresponding properties in Pega Platform:
- Confirm the data set settings by clicking Save.
Add a reference to your File data set in the Source shape of a data flow. See Referencing remote repository data in a data flow.
Published June 22, 2018 — Updated September 28, 2018