Skip to main content

Resolved Issues

View the resolved issues for a specific Platform release.

Go to download resolved issues by patch release.

Browse release notes for a selected Pega Version.

NOTE: Enter just the Case ID number (SR or INC) in order to find the associated Support Request.

INC-167311 · Issue 646477

Updated upgrade handling for migrating work objects

Resolved in Pega Version 8.3.6

After upgrading from Pega 6.2 to 8.3, the work migrated work objects were missing SLAs due to missed entries in the assignment tables (PC_ASSIGN_WORKLIST/ PC_ASSIGN_WORKBASKET) . The SLA was firing, but the processing failed due to the fact the runtime could not resolve a 'AddHistoryPage' library function. In this case, multiple upgrades of the application dating back to Pega 4 resulted in the runtime context containing older ruleset versions in higher ruleset versions, hiding the underlying Pega 8 version of the rule. For releases prior to Pega 7.3, Rule-Application was stored in pr4_rule and will be migrated to pr4_rule_application during upgrades. However, since Context Upgrade is run before Optimize Newly Exposed Columns, the pyDependsOnName won't always be populated. To resolve this, the system will filter based on the value in the blob rather than the exposed column so there will be a value regardless of the upgrade-from version.

INC-172675 · Issue 649451

Configuration added for extending queue processor timeout

Resolved in Pega Version 8.3.6

Alerts for queue processor (QP) items which took more than 15 minutes to run could result in the system marking the node as 'unhealthy'. In environments with Pega Health Check enabled, this would shut down the node gracefully. It was not possible to change this default as it was hardcoded. In order to support systems that may have custom processes that run beyond 15 minutes, a a new setting has been exposed that allows configuration of the interval after which a node with long-running queue processor is marked as unhealthy and is restarted. By default this remains 900000 milliseconds / 900 seconds / 15 minutes, but it may be adjusted up to 24 hours to avoid premature node shutdown. The stale thread detection mechanism will take that setting into account and use the provided value or default to 15 minutes if the value was not provided. In addition, the threshold's units in the UI have been changed from ms to seconds.

SR-B35942 · Issue 295191

Grid filter column localization fixed

Resolved in Pega Version 7.3.1

The repeating grid's column filter was not displaying the localized pyStatusLabel value. This has been fixed.

SR-B64134 · Issue 313542

Workaround added for Oracle 4000 character column value max

Resolved in Pega Version 7.3.1

Due to an Oracle limitation, RD on a clob column fails if the column value is of length greater than 4000. To work around this, support has been added to retrieve data from a clob column of length greater than 4000 characters using report definition based on DASS setting "reporting/retrieveFullClobContent" defined on "Pega-RULES".

SR-B51989 · Issue 310393

TimeStamp column includes milliseconds in extract

Resolved in Pega Version 7.3.1

Extraction to database output format was not showing the millisecond part of the datetime column. This was a missed use case for BIX when the support of Datetime property mapping to timestamp column was introduced in Oracle, and has been fixed.

SR-B64125 · Issue 313855

Fixed SLA calculations with custom classes in Oracle

Resolved in Pega Version 7.3.1

For work classes, columns for pySLAGoal and pySLADeadLine are of type TIMESTAMP in Oracle DB. The Pega getContentItemsPastSLA activity uses the columns to find date differences with SYSDATE and then used a subtraction operation between columns of type TIMESTAMP and SYSDATE to calculate the SLA. However, an exception resulted when custom classes were used. To correct this, getContentItemsPastSLA has been updated to cast the columns of type TIMESTAMP to DATE type.

SR-B66297 · Issue 314861

Instance insertion fixed for PostgreSQL partitioning

Resolved in Pega Version 7.3.1

Obj-Save was failing when trying to insert an instance in a partitioned PostgreSQL table with 'only if new' enabled. This record insertion failed because of a system check that is present when inserting an instance into the child table, and the issue has been resolved by reworking the code that was checking (assuming) if the insert was successful based on the returned inserted count.

SR-B66297 · Issue 299892

Instance insertion fixed for PostgreSQL partitioning

Resolved in Pega Version 7.3.1

Obj-Save was failing when trying to insert an instance in a partitioned PostgreSQL table with 'only if new' enabled. This record insertion failed because of a system check that is present when inserting an instance into the child table, and the issue has been resolved by reworking the code that was checking (assuming) if the insert was successful based on the returned inserted count.

SR-B51825 · Issue 310042

Added handling for WO lock when OAuth2 token expires

Resolved in Pega Version 7.3.1

The expiration of the OAuth2 token committed the WOs in context and caused all WOs to lose their locks and eventually end up in the SLA broken queue. To fix this, code has been added to perform the commit on a new database thread so the commit operation will be independent from the actual thread context.

SR-B49520 · Issue 310724

Schema extraction over multiple DBs uses log table instead of class

Resolved in Pega Version 7.3.1

When running an extract, an Null Pointer Exception error was generated if the data being extracted was on a different schema / database than the Pega data schema. This has been resolved by reworking the sequence number generator to use a pr_log table connection of instead of using the class on which the extract was defined.

We'd prefer it if you saw us at our best.

Pega.com is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice
Contact us