Is it possible to stream and batch ingest data simultaneously from different sources into the same profile dataset in Adobe Experience Platform (AEP)? | Community
Skip to main content
May 30, 2025
Solved

Is it possible to stream and batch ingest data simultaneously from different sources into the same profile dataset in Adobe Experience Platform (AEP)?

  • May 30, 2025
  • 2 replies
  • 517 views

I’m curious about the technical feasibility and best practices. Is combining streaming and batch ingestion into a single profile dataset advisable? What are the key considerations or potential pitfalls to be aware of when setting this up?

Thanks in advance for any insights!

This post is no longer active and is closed to new replies. Need help? Start a new post to ask your question.
Best answer by Asheesh_Pandey

@yatinsh It is technically feasible to stream and batch ingest data simultaneously into the same profile dataset. Key Considerations: Define identity strategy, XDM schema, and source mappings. Test in sandbox to identify and resolve issues like schema mismatch, data duplication, conflict merge rules, profile fragments etc.

2 replies

Asheesh_Pandey
Community Advisor
Asheesh_PandeyCommunity AdvisorAccepted solution
Community Advisor
May 31, 2025

@yatinsh It is technically feasible to stream and batch ingest data simultaneously into the same profile dataset. Key Considerations: Define identity strategy, XDM schema, and source mappings. Test in sandbox to identify and resolve issues like schema mismatch, data duplication, conflict merge rules, profile fragments etc.

- Asheesh
Parvesh_Parmar
Community Advisor
Community Advisor
June 2, 2025

Hi @yatinsh ,

As @Asheesh_Pandey mentioned, yes — it is technically possible.

However, from a future-proofing and best practices perspective, it's not the ideal approach. If there's an issue with one data source, it can become difficult to debug or analyze the root cause.

Also, if incorrect data is sent from one source and you need to delete it, there's a risk of having to clean the entire dataset — which could result in loss of data from both sources.

A better solution would be to create two separate datasets, one for each source. You can then stitch the data together based on identity namespaces and control how the data is prioritized using merge policies.

Hope this helps!

Kind regards,
Parvesh

Parvesh Parmar – Adobe Community Advisor https://www.linkedin.com/in/parvesh-parmar/