Guardrails for batch and streaming ingestion are calculated at the organization level and not the sandbox level. This means that your data usage per sandbox is bound to the total license usage entitlement that corresponds with your entire organization. Additionally, data usage in development sandboxes are limited to 10% of your total profiles. For more information about license usage entitlement, read the data management best practices guide.
Guardrails are thresholds that provide guidance for data and system usage, performance optimization, and avoidance of errors or unexpected results in Adobe Experience Platform. Guardrails can refer to your usage or consumption of data and processing in relation to your licensing entitlements.
Check your license entitlements in your Sales Order and corresponding Product Description on actual usage limits in addition to this guardrails page.
This document provides guidance on guardrails for data ingestion in Adobe Experience Platform.
The following table outlines guardrails to consider when using the batch ingestion API or sources:
Type of ingestion | Guidelines | Notes |
---|---|---|
Data lake ingestion using the batch ingestion API |
|
|
Data lake ingestion using batch sources |
|
Read the sources overview for a catalog of sources you can use for data ingestion. |
Batch ingestion to Profile |
|
|
Number of Profile or ExperienceEvent batches ingested per day | The maximum number of Profile or ExperienceEvent batches ingested per day is 90. This means that the combined total of Profile and ExperienceEvent batches ingested each day cannot exceed 90. Ingesting additional batches will affect system performance. | This is a soft limit. It is possible to go beyond a soft limit, however, soft limits provide a recommended guideline for system performance. |
Encrypted data ingestion | The maximum supported size of a single encrypted file is 1 GB. For example, while you can ingest 2 or more GBs worth of data in a single dataflow run, no individual file in the dataflow run can exceed 1 GB. | The process of ingesting encrypted data may take longer than that of a regular data ingestion. Read the encrypted data ingestion API guide for more information. |
Upsert batch ingestion | Ingestion of upsert batches can be up to 10x slower than regular batches, therefore, you should keep your upsert batches under two million records in order to ensure an efficient runtime and to avoid blocking other batches from being processed in the sandbox. | While you can undoubtedly ingest batches that exceed two million records, the time of your ingestion will be significantly longer due to the limitations of small sandboxes. |
Read the streaming ingestion overview for information on guardrails for streaming ingestion.
The following table outlines guardrails to consider when using the streaming sources:
Type of ingestion | Guidelines | Notes |
---|---|---|
Streaming sources |
|
Streaming sources such as Kafka, Azure Event Hubs, and Amazon Kinesis do not use the Data Collection Core Service (DCCS) route and can have different throughput limits. See the sources overview for a catalog of sources you can use for data ingestion. |
See the following documentation for more information on other Experience Platform services guardrails, on end-to-end latency information, and licensing information from Real-Time CDP Product Description documents: