Skip to main content

Published Release Notes

Find release notes for the selected Pega Version and Capability

Browse resolved issues for Platform releases.

This documentation is for non-current versions of Pega Platform. For current release notes, go here.

Decision Data Store data sets can be used only on DNodes

Valid from Pega Version 7.1.8

Data flows that contain Decision Data Store data sets as the primary or secondary source must be created and executed only on DNodes.

Data flows are not restarted automatically after application server restart

Valid from Pega Version 7.1.8

When you restart the application server or the Pega 7 server, you stop the execution of your data flows. The interrupted batch and real-time runs are marked as Failed

Recommendation:

  • Go to the Designer StudioDecisioning > Decisions > Data Flows > Batch processing tab and start the failed batch runs.
  • Go to the Designer StudioDecisioning > Decisions > Data Flows > Real-time processing tab and activate the failed real-time runs.

 

Value list and value group properties are not supported inside data flows

Valid from Pega Version 7.1.8

Value list and value group properties are not supported inside data flows, and you need to instead use other property types.

See the Pega 7 developer help to check all available property types.

Data flow preview size is fixed to 10

Valid from Pega Version 7.1.8

The Preview option for each shape in data flows returns the first 10 records. This value is fixed and currently cannot be changed.

Data Flow transformation shapes cannot be used in combination with the Compose or Merge shapes

Valid from Pega Version 7.1.8

When you reference another data flow from a data flow that contains the Compose or Merge shape, the referenced data flow cannot contain transformation shapes (EventStrategy, Decision strategy, Convert).

Data flow validation does not currently prevent you from designing a data flow that goes against this design pattern. Make sure that your data flows follow this design pattern by checking the referenced and referencing data flows.

 

Data flow destination is resolved at assembly time

Valid from Pega Version 7.1.8

Data flow execution always uses the design-time destination data sets or destination data flows, regardless of the input records at run time.

Example:

You create a data flow and set the destination to a specific data set. In the data flow, you use a strategy that produces some results. Regardless of these results, the execution of your data flow always uses the destination data set that you specified when you designed the data flow.    

The same applies when you set the destination to a data flow.

 

 

Schema not found error when importing a Rule-Admin-Product file

Valid from Pega Version 7.1.8

If you import a Rule-Admin-Product file (RAP) file that includes a schema that has a view, grammar parsing changes can mean that the Pega 7 Platform is unable to find the schema.

To enable the Pega 7 Platform to find the schema:

  1. Review the error message that indicates that the schema was not found and identify the following items:
    • The name of the schema that is causing the error
    • The table name in the format: wrong_schema_name.table_name...
    • The DDL that the Pega 7 Platform is trying to apply
  2. Use the table name to identify the correct schema name in your system. This schema is usually your data schema, but might also be an external database.
  3. Download the DDL.
  4. Correct the schema name and save the DDL.
  5. Manually apply the DDL.

WebSphere log error with Java 8 for Enterprise JavaBeans

Valid from Pega Version 7.1.8

For environments that are deployed by using Java 8 on IBM WebSphere Application Server version 8.5.5.9 through the last 8.5.5.x fix pack, the log might report that a class was not found when the application server attempts to start Enterprise JavaBeans. If Java EE functionality is not needed, no further action is required. However, to use Enterprise JavaBeans, see WebSphere log error with Java 8 for Enterprise JavaBeans.

Extension attributes are not supported in PMML models

Valid from Pega Version 7.3.1

Models in the Predictive Model Markup Language (PMML) format version 4.3 that contain extension attributes with the x- prefix are not valid. These extension attributes are deprecated; you must use extension elements instead. In addition, if the output type of any output field in the model is set to FLOAT, change it to DOUBLE.

For more information, see PMML 4.3 - General Structure in the Data Mining Group documentation.

The Upload responses action is not supported for adaptive models with customized context

Valid from Pega Version 7.3.1

A default instance of the Adaptive Model rule contains five model identifiers (.pyIssue, .pyGroup, .pyName, .pyDirection, .pyChannel) that are used to partition adaptive models. If you add other identifiers in your Adaptive Model rule instance, you cannot upload responses to this instance with the Upload Responses wizard and the following error is displayed: The Flow Action post-processing activity pzUploadCSVFile failed: Cannot parse csv file.You can still train such adaptive models with data flows.

For more information, see Training adaptive models in bulk with data flows, Model context, and Uploading customer responses.

We'd prefer it if you saw us at our best.

Pega.com is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice
Contact us