Table of Contents

Making decisions in data flow runs


Only available versions of this content are shown in the dropdown

  • Creating a batch run for data flows

    Create batch runs for your data flows to make simultaneous decisions for large groups of customers. You can also create a batch run for data flows with a non-streamable primary input, for example, a Facebook data set.

  • Creating a real-time run for data flows

    Provide your decision strategies with the latest data by creating real-time runs for data flows with a streamable data set source, for example, a Kafka data set.

  • Creating an external data flow run

    You can specify where to run external data flows and manage and monitor running them on the External processing tab of the Data Flows landing page. External data flows run in an external environment (data set) that is referenced by a Hadoop record on the Pega Platform platform.

  • Monitoring single case data flow runs

    View and monitor statistics of data flow runs that are triggered in the single case mode from the DataFlow-Execute method. Check the number of invocations for single case data flow runs to evaluate the system usage for licensing purposes. Analyze run metrics to support performance investigation when Service Level Agreements (SLAs) are breached.

Did you find this content helpful?

Have a question? Get answers now.

Visit the Collaboration Center to ask questions, engage in discussions, share ideas, and help others.