In Prediction Studio, you can now monitor the predictive performance of your models to validate that they make accurate predictions. Based on that information, you can re-create or adjust the models to provide better business results, such as higher accept rates or decreased customer churn.
In Kafka data sets, you can now create and receive messages in your custom formats, as well as in the default JSON format. To use custom logic and formats for serializing and deserializing ClipboardPage objects, create and implement a Java class. When you create a Kafka data set, you can choose to apply JSON or your custom format that uses a PegaSerde implementation.
Additional configuration options for File data sets
Valid from Pega Version 8.2
You can now create File data sets for more advanced scenarios by adding custom Java classes for data encryption and decryption, and by defining a file set in a manifest file.
Additionally, you can improve data management by viewing detailed information in the dedicated meta file for every file that is saved, or by automatically extending the filenames with the creation date and time.
Evaluate event strategies by creating test runs. During each run, you can enter a number of sample events with simulated property values, such as the event time, the event key, and so on. By testing a strategy against sample data, you can understand the strategy configuration better and troubleshoot potential issues.
You can now generate a report from the Run details section of a Data Flow rule that provides information about run events. The report includes reasons for specific events which you can analyze to troubleshoot and debug issues more quickly. You can export the report and share it with others, such as Global Customer Support.
Real-time and batch data flow runs now retry dataset connections that fail when the related service is temporarily unavailable, for example, when a connection to a Cassandra database times out. With the automatic retries, you no longer need to run data-heavy and CPU-intensive jobs multiple times and the maintenance of data flow runs diminishes significantly.
Strategies that are part of data flows are now automatically optimized to achieve the best performance. You can also choose the properties that a strategy outputs to further increase the efficiency without additional strain on your hardware.
Gain a better understanding of how strategy components influence decision outcomes by using decision funnel explanations. By running simulation tests with the new Decision funnel explanation purpose, you can now break down the results of a decision funnel into granular analyses to assess how certain components and expressions influence the overall outcome.