Symptom
- The duration of the pipeline's T&L refresh is increasing over time.
- Concerns regarding performance impact when adding data to the same table in the data source.
Environment
- Data Source (CSV, Connector)
Reproducing the Issue
- Store historical BSEG data in the data source table.
- Continuously add new data to the same table over time.
- Observe the increasing duration of the pipeline's T&L refresh.
Resolution
Current behavior (summary)
- The Transform step processes extracted data into a single event log file.
- The Load step always overwrites the entire Process data with that new event log.
- Because the Process data (event log) is replaced every run, the Transform must produce whatever the Process requires each time.
Why T&L time increases
- Transform work grows with the volume of input rows.
- As more data is extracted, the Transform must process more rows, so run time increases.
- Because Load overwrites Process data every T&L run must produce a complete (or at least required) event log.
An option that may be relevant for you is to filter during Transformation:
- Add WHERE clauses or other filters in Transform scripts to limit processed rows (examples: date ranges).
- Benefit: less data processed → shorter Transform and Load times.
You may need to consult with your internal expert business users to determine what filters to put in place here as you will need to decide on what data is required for your investigations
See Also
Refer to: SAP Signavio Process Intelligence Documentation
Refer to: Data Extraction Options User Guide
Keywords
data source, process intelligence, historical data, pipeline refresh, T&L refresh duration, performance impact, BSEG data, data table, data management, data pipeline , KBA , BPI-SIG-PI-DS , Data Source (CSV, Connector) , Problem
Product
SAP Signavio Process Intelligence all versions
SAP Knowledge Base Article - Public