What AEP is doing if we send for example more than 20GB batch data in one hour? How is it queued? Or if we exceed number of requests?
I went through this guardrails but customer has additional pre-sales questions - https://experienceleague.adobe.com/en/docs/experience-platform/ingestion/guardrails
Solved! Go to Solution.
Topics help categorize Community content and increase your ability to discover relevant content.
if the guardrails are mentioned as hard limit, then the platform will not allow to exceed that guardrail. if not, platform will continue to accept but there will be delay in processing or/and there will be performance degradation. but if we exceed the guardrails in extremely high volume then there will be error as well and the ingestion will timeout.
In case of streaming data exceeding the guardrails continuously and platform is taking lot of time in evaluating the streaming segment for each streaming ingestion, then you will see the impact where the streaming segment evaluation will be stopped as well temporarily.
--ssj
Hi @Michael_Soprano , While it is not explicitly documented, it's possible that AEP internally queues excess data for later processing when the system load decreases, This has to be checked with product team. Otherwise,
When you exceed the 20GB hourly batch ingestion limit in AEP, the platform will either return an error indicating that the ingestion limit has been exceeded or may throttle your requests, slowing down ingestion or temporarily blocking further uploads.
you'll encounter similar behavior if the number of concurrent requests and the overall request rate limit exceeds.
if the guardrails are mentioned as hard limit, then the platform will not allow to exceed that guardrail. if not, platform will continue to accept but there will be delay in processing or/and there will be performance degradation. but if we exceed the guardrails in extremely high volume then there will be error as well and the ingestion will timeout.
In case of streaming data exceeding the guardrails continuously and platform is taking lot of time in evaluating the streaming segment for each streaming ingestion, then you will see the impact where the streaming segment evaluation will be stopped as well temporarily.
--ssj
@Michael_Soprano
One thing to note there is that the 20GB limit per hour is regarding Batch Ingestion API. Ingestion from batch source can handle up to 200GB per hour.