Skip to main content
circle-square
Outline Circle
square-square
Little Circle
square-square
Little Circle

Published Release Notes

Find release notes for the selected Pega Version and Capability.

Browse resolved issues for Platform releases.

Required Oracle optimization parameter

Valid from Pega Version 7.1.3

To optimize performance, set the Oracle parameter optimizer_index_cost_adj to a value between 20 and 25. If this value is not set, the system can run exceedingly slowly, potentially blocking users from login.

Split schema upgrade instructions missing properties

Valid from Pega Version 7.1.3

If you upgraded from 5.x, 6.x, or 7.x using the instructions in previous versions of the upgrade guide, you may have neglected to set the properties below in your migrateSystem.properties file when you migrated your upgraded schema to the source system:

 

pega.rules.objects.generate=true

pega.rules.objects.apply=true

 

If these properties were not set during an upgrade that splits the schema, your environment does not have the indexes, triggers, and primary keys on the rules tables.

 

To check for this issue, see if the pr4_base and pr4_rule rules tables in your existing rules schema are missing primary keys. If they are, use the SQL scripts in the ResourceKit\MigrationRecoveryScripts directory of the release to cleanup duplicate rules that were created due to this issue. Follow the steps below to run the scripts.

To run the scripts on Microsoft SQL, Oracle, or PostgreSLQ

  1. Take down any app servers using the affected schema.
  2. Backup your database.
  3. Replace all instances of @RULES_SCHEMA in <database>_cleanDups.sql with the name of the schema that contains the pr4_base table.
  4. Run the <database>_cleanDups.sql script on the database with vendor tools (sqlPlus, SQL Server Management Studio, etc).
  5. Replace all instances of @RULES_SCHEMA in <database>_fix_vw_table.sql with the name of the schema that contains the pr4_base table.
  6. Run the <database>_fix_vw_table.sql script on the database with vendor tools (sqlPlus, SQL Server Management Studio, etc).
  7. Generate and apply the ddl using the command line generateDDL command. Check the installation guide for your database or the upgrade guide for details about how to use the generateDDL command line script.
  8. Rebuild the indexes for the tables in your rules schema using vendor tools. This is necessary so that your system runs at an optimum speed.
  9. Optionally upgrade to the latest release, at this point your database is ready to be upgraded or used depending on your needs.

      

To run the scripts on DB2 for LUW or z/OS

  1. Take down any app servers using the affected schema.
  2. Backup your database.
  3. Run the <database>_cleanDups.sql script on the database with vendor tools (UDB CLP, Data Studio, etc) to create the CLEANSE_RULES_DUPS stored procedure.
  4. Run the query Call CLEANSE_RULES_DUPS(‘<rulesSchema>’); where <rulesSchema> is the name of schema that contains the pr4_base table.
  5. After the previous step is complete drop the CLEANSE_RULES_DUPS procedure.
  6. Replace all instances of @RULES_SCHEMA in <database>_fix_vw_table.sql with the name of the schema that contains the pr4_base table.
  7. Run the <database>_fix_vw_table.sql script on the database with vendor tools (UDB CLP, Data Studio, etc).
  8. Generate and apply the ddl using the command line generateDDL command. Check the installation guide for your database or the upgrade guide for details about how to use the generateDDL command line script.
  9. Rebuild the indexes for the tables in your rules schema using vendor tools. This is necessary so that your system runs at an optimum speed.
  10. Optionally upgrade to the latest release. At this point your database is ready to be upgraded or used depending on your needs.

Synchronized database and application server settings

Valid from Pega Version 7.1.3

Configure your database and application server to use the same time zone and character encoding to avoid conflicts.

Decision Strategy Manager

Valid from Pega Version 7.1.3

Fixes were made that improve the configuration and execution of Batch Decisions. In particular,  the capability to use Structured Data has been enhanced. Some notable improvements have also been made to the UI of Visual Business Director.

  • The structured data input configuration will now work even though there is no data.
  • VBD has been enhanced to work with Version 7.1.
  • The structured data input configuration will now work with nested structures.
  • Association rules from Version 6.3 will now work with 7.1 structured data.

System Management

Valid from Pega Version 7.1.3

The PRPC installer has been enhanced to handle additional error conditions related to partial installs and duplicate keys; additionally, the hotfix rollback feature has been improved, and the Rulebase compare tool now supports HTTP(S) environments.

  • Rollback feature has been enhanced.
  • RuleBase Compare will work for secure (HTTPS) environments as well as  non-secure (HTTP) environments.
  • Install will gracefully handle the presence of a duplicate key during import.
  • When install fails, system will set tables back to original state.
  • Remote MBean methods will now work on multi-servant load-balancing systems.
  • Enhancements were made to harnesses in Split Schema systems.

Predictive models monitoring

Valid from Pega Version 8.2

In Prediction Studio, you can now monitor the predictive performance of your models to validate that they make accurate predictions. Based on that information, you can re-create or adjust the models to provide better business results, such as higher accept rates or decreased customer churn.

For more information, see Monitoring predictive models.

Kafka custom serializer

Valid from Pega Version 8.2

In Kafka data sets, you can now create and receive messages in your custom formats, as well as in the default JSON format. To use custom logic and formats for serializing and deserializing ClipboardPage objects, create and implement a Java class. When you create a Kafka data set, you can choose to apply JSON or your custom format that uses a PegaSerde implementation.

For more information, see Creating a Kafka data set and Kafka custom serializer/deserialized implementation.

Additional configuration options for File data sets

Valid from Pega Version 8.2

You can now create File data sets for more advanced scenarios by adding custom Java classes for data encryption and decryption, and by defining a file set in a manifest file.

Additionally, you can improve data management by viewing detailed information in the dedicated meta file for every file that is saved, or by automatically extending the filenames with the creation date and time.

For more information, see Creating a File data set for files on repositories and Requirements for custom stream processing in File data sets.

Simplified testing of event strategies

Valid from Pega Version 8.2

Evaluate event strategies by creating test runs. During each run, you can enter a number of sample events with simulated property values, such as the event time, the event key, and so on. By testing a strategy against sample data, you can understand the strategy configuration better and troubleshoot potential issues.

For more information, see Evaluate event strategies through test runs.

Data flow life cycle monitoring

Valid from Pega Version 8.2

You can now generate a report from the Run details section of a Data Flow rule that provides information about run events. The report includes reasons for specific events which you can analyze to troubleshoot and debug issues more quickly. You can export the report and share it with others, such as Global Customer Support.

For more information about accessing event details, see Creating a real-time run for data flows and Creating a batch run for data flows.

All Pega product release notes can be found on their product pages.

Ready to crush complexity?

Experience the benefits of Pega Community when you log in.

We'd prefer it if you saw us at our best.

Pega Community has detected you are using a browser which may prevent you from experiencing the site as intended. To improve your experience, please update your browser.

Close Deprecation Notice
Contact us