Expand my Community achievements bar.

Join expert-led sessions on Real-Time CDP & Journey Optimizer designed to boost your impact.
SOLVED

Calculated column can impact system performance when data load happens for 1lakhs records ?

Avatar

Level 1

Hi Community,

I am ingesting data from external sources into AEP dataset. Each table has 4-5 calculated fields. When size of data will be huge around like 1lakh records in each file with calculated fields, will it going to impact system performance. I am creating calculated fields to convert datatime into UTC format and some fields are using split function. Do we have any best practices on using calculated fields in data

mapping.

 

Thanks in advance

 

 

1 Accepted Solution

Avatar

Correct answer by
Community Advisor

@SantoshRa5  Short Answer -  Yes, calculated fields are super handy but using calculated fields in your dataflow mappings mainly when you're working with large files, can impact ingestion performance depending on how complex the logic is.  However, AEP is designed to handle such workloads.

A few best practices:

  • Keep calculations simple, example date formatting, basic string splits
  • Avoid chaining too many operations like nested split and substring 
  • If possible, pre-process heavy transformations before ingestion in your ETL and source layer

If you're just doing things like converting to UTC, splitting a timestamp, you're likely fine, just be mindful of stacking 4, 5 calculated fields per record on very large files.

View solution in original post

2 Replies

Avatar

Correct answer by
Community Advisor

@SantoshRa5  Short Answer -  Yes, calculated fields are super handy but using calculated fields in your dataflow mappings mainly when you're working with large files, can impact ingestion performance depending on how complex the logic is.  However, AEP is designed to handle such workloads.

A few best practices:

  • Keep calculations simple, example date formatting, basic string splits
  • Avoid chaining too many operations like nested split and substring 
  • If possible, pre-process heavy transformations before ingestion in your ETL and source layer

If you're just doing things like converting to UTC, splitting a timestamp, you're likely fine, just be mindful of stacking 4, 5 calculated fields per record on very large files.

Avatar

Administrator

@SantoshRa5 Just checking in — were you able to resolve your issue?
We’d love to hear how things worked out. If the suggestions above helped, marking a response as correct can guide others with similar questions. And if you found another solution, feel free to share it — your insights could really benefit the community. Thanks again for being part of the conversation!



Kautuk Sahni