Making decisions in data flow runs
- Creating a batch run for data flows
Create batch runs for your data flows to make simultaneous decisions for large groups of customers. You can also create a batch run for data flows with a non-streamable primary input, for example, a Facebook data set.
- Creating a real-time run for data flows
Provide your decision strategies with the latest data by creating real-time runs for data flows with a streamable data set source, for example, a Kafka data set.
- Creating an external data flow run
You can specify where to run external data flows and manage and monitor running them on the External processing tab of the Data Flows landing page. External data flows run in an external environment (data set) that is referenced by a Hadoop record on the Pega Platform platform.
- Monitoring single case data flow runs
View and monitor statistics of data flow runs that are triggered in the single case mode from the DataFlow-Execute method. Check the number of invocations for single case data flow runs to evaluate the system usage for licensing purposes. Analyze run metrics to support performance investigation when Service Level Agreements (SLAs) are breached.