Skip to main content

Resolved Issues

View the resolved issues for a specific Platform release.

Go to download resolved issues by patch release.

Browse release notes for a selected Pega Version.

NOTE: Enter just the Case ID number (SR or INC) in order to find the associated Support Request.

Please note: beginning with the Pega Platform 8.7.4 Patch, the Resolved Issues have moved to the Support Center.

INC-126129 · Issue 569668

PropertyToColumnMap made more robust

Resolved in Pega Version 8.4.2

The DF_ProcessEmails dataflow was intermittently failing with a StageException error. This was traced to schema changes being propagated asynchronously by system pulse, which seem to have caused PropertyToColumnMap to cache stale schema. To resolve this, if the property mapping is not found the first time, the system will make a second attempt to get the mapping. Additional logging has also been added for better diagnostics.

INC-126187 · Issue 564340

Report Definition replaced with a data set for training the data and build model

Resolved in Pega Version 8.4.2

An issue with training the data and build model via an email or chat channel was traced in an exception in logs indicating the inability to execute a Report Definition. This can happen when UDF are removed/disabled in the database or in environments where functions are not allowed to read the blob for performance reasons. To resolve this, the Report Definition has been replaced with a data set.

INC-126750 · Issue 564155

DSS added to configure Dataset-Execute page handling

Resolved in Pega Version 8.4.2

When Kafka Data Set pages were saved to a data set with the Dataset-Execute method, there was is feedback if any of the pages were not successfully saved. Instead, the step always completes as successful. In addition, if any properties are added or modified by the save operation itself, those changes are not visible. This is due to the data set execute save operation saving pages as DSM pages to the data set. Due to the conversion of the pages, copies of the pages are used which do not reflect back any changes on the input pages. DSM pages are used by default because they are more lightweight than regular clipboard pages and therefore have potentially better performance. In order to allow the use of DSM pages to be customized, a new Dynamic System Setting: dataset/execute/save/statusFailOnError has been added. This can be enabled by setting it to true; it is disabled by default for greater backwards compatibility. By removing the DSM page conversion in the generated save code, changes to input pages will be reflected back if any are performed by the data set save operation, and the system will report back which pages are saved or failed by adding messages to the pages that failed to save. Performance may be affected with this change as regular clipboard pages are in general slower than DSM pages, however, that may be offset by removing the conversion to DSM pages process and will depend on the site configuration.

INC-126796 · Issue 561536

Modifications to getFunctionalServiceNodes process

Resolved in Pega Version 8.4.2

The count of the Interaction History write related threads was increasing rapidly and a stack trace indicated "waiting on condition" and "java.lang.Thread.State: WAITING (parking)" errors. Investigation showed that this was due to getFunctionalServiceNodes using Hazelcast to determine node status by making a service request on an installation with a very large number of nodes, causing thread locking. To resolve this, the implementation has been updated to avoid calling getFunctionalServiceNodes on save of Interaction History, instead using Cassandra and only calling getFunctionalServiceNodes on the master node, not on all nodes.

INC-128385 · Issue 564522

Behavior made consistent between SSA and legacy engines

Resolved in Pega Version 8.4.2

There was a behavioral disparity between the legacy execution engine and the SSA engine where the latter was not creating a new page when the index was one above the size of the page list. This has now been corrected in order to make the SSA behavior fully backward compatible with the legacy engine, i.e. a new blank page will be added to the list if the index is one above the size of the list.

INC-128898 · Issue 564692

Updated precondition checks for Tumbling Time data in event strategies

Resolved in Pega Version 8.4.2

Tumbling Time data keys in event strategies were not being properly executed when certain window configurations were used. This was has been resolved by turning off the optimalization of Cassandra reads for small windows when window size is not known upfront and set dynamically (set size by property).

INC-130037 · Issue 570959

Corrected caching sync for proposition filter

Resolved in Pega Version 8.4.2

Unit testing or a manual run of the generated strategy through next best designer was not giving the result, but instead was throwing a stale thread error. Investigation showed that the PegaAPI reference was cached in proposition filter at first execution and the filter was trying to use same reference for other further executions. This has been resolved by ensuring the cached reference is properly synced.

INC-130471 · Issue 569670

Null handling added for SSA parsing

Resolved in Pega Version 8.4.2

Strategy save was failing for the FormatDateTime expressions such as FormatDateTime(@addToDate(@getCurrentTimeStamp(),3,0,0,0),"yyyy",null,null) FormatDateTime(@addToDate(@getCurrentTimeStamp(),3,0,0,0),"MMM",null,null) FormatDateTime(@addToDate(@getCurrentTimeStamp(),3,0,0,0),"yyyyMM",null,null) This was due to the SSA engine not considering NULL to be a valid type-safe optimizable expression. To resolve this, the parsing process has been updated so that unsupported IL nodes (such as the node representing literal null) will cause the expression to fall back to legacy instead of failing.

INC-130757 · Issue 571421

Migration data pipeline import activity updated

Resolved in Pega Version 8.4.2

The migration data pipeline was not working in the DevOps environment when moving data from Production to Simulation environments. This has been resolved by updating the activity used to generate file data set for configuring the datetime, date and time formats and inserting a check so that if a format for a type is empty, parsing properties of this type will be skipped.

INC-131260 · Issue 571786

Handling added for runs using Completed state

Resolved in Pega Version 8.4.2

After upgrade, creating a data flow rule and the batch Data flow run using the Dataflow-Execute method in the activity resulted in the exception "Could not start run DataFlowRunConfig" when executing. This was a missed use case in the new architecture of data flow engine introduced in 8.4, causing DataFlow-Execute to not work properly for existing runs in COMPLETED state. This has been resolved with the addition of a branch of execution in DataFlow-Execute.Start to correctly handle completed runs.

We'd prefer it if you saw us at our best.

Pega.com is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice
Contact us