ποΈ Architecture & Performance
Data Designer is an orchestration framework that coordinates synthetic data generation workflows. It is a client of LLM inference serversβit does not host models itself.
This guide explains the architecture, execution model, and how to tune performance for your specific use case.
Separation of Concerns
βββββββββββββββββββββββββββββββββββββββ βββββββββββββββββββββββββββββββββββββββ
β Data Designer β β Inference Server(s) β
β (Orchestration) β HTTP β (LLM Hosting) β
β β ββββββΊ β β
β β’ Dataset workflow management β β β’ Model weights and execution β
β β’ Column dependency resolution β β β’ GPU allocation and scheduling β
β β’ Batching and parallelism β β β’ Request queuing β
β β’ Retry and error handling β β β’ Token generation β
β β’ Adaptive concurrency (AIMD) β β β’ Rate limiting (optional) β
β β’ Data validation and quality β β β
βββββββββββββββββββββββββββββββββββββββ βββββββββββββββββββββββββββββββββββββββ
β² β²
β β
Your workflow Your infrastructure
configuration (or cloud API)
What Data Designer Does
- Orchestrates the generation workflow across multiple columns
- Resolves dependencies between columns (DAG-based execution)
- Batches work into manageable chunks (
buffer_size) - Parallelizes LLM calls within batches (
max_parallel_requests) - Adapts to rate limits automatically via AIMD concurrency control
- Handles errors with retries and early shutdown logic
- Validates generated data against schemas and constraints
What Data Designer Does NOT Do
- Host models: You must provide LLM endpoints
- Manage GPUs: Your inference server handles GPU allocation
- Scale inference: You must provision sufficient capacity
- Impose rate limits: Your server or API gateway sets rate limits (Data Designer reacts to them automatically)
Execution Model
Dataset Builder
This describes Data Designer's current DatasetBuilder, which generates columns sequentially within batches. Other dataset generation strategies are in development.
Data Designer processes datasets in batches, with parallel operations within each batch.
How It Works
Step 1: Split into batches
Your dataset is divided into batches of buffer_size records. Each batch is processed completely before moving to the next.
Step 2: Process columns sequentially
Within a batch, columns are generated one at a time following the dependency graph. The order depends on column dependenciesβexpression columns may come before LLM columns if the LLM columns depend on them.
Example workflow:
Batch 1 (100 records)
β
βββΊ Column 1: category (Sampler) ββββ All 100 values generated
βββΊ Column 2: prompt (LLM Text) ββββ All 100 values generated
βββΊ Column 3: response (LLM Text) ββββ All 100 values generated
βββΊ Column 4: score (Expression) ββββ All 100 values computed
β
βββΊ Write batch to disk
β
βΌ
Batch 2 (100 records)
...repeat...
Step 3: Generate cells in parallel
Within each column, cells are processed in parallel up to the configured limit:
| Column Type | Parallelism Control |
|---|---|
| Sampler | non_inference_max_parallel_workers |
| LLM (Text, Code, Structured, Judge) | max_parallel_requests |
| Expression | Sequential (fast, CPU-bound) |
Key Concepts
| Concept | Description |
|---|---|
| Batching | Records are split into batches of buffer_size. Each batch completes entirely before the next begins. |
| Sequential columns | Within a batch, columns are generated one at a time, respecting the dependency graph. |
| Parallel cells | Within a column, individual cells (records) are generated in parallel up to the configured limit. |
Concurrency Formula
At any moment, the number of concurrent LLM requests is:
concurrent_requests = min(
buffer_size, # Records in current batch
current_throttle_limit, # AIMD-managed limit (β€ max_parallel_requests)
remaining_cells_in_column # Cells left to generate
)
max_parallel_requests sets the ceiling. The actual limit (current_throttle_limit) is managed at runtime by an AIMD (Additive Increase / Multiplicative Decrease) controller that reacts to rate-limit signals from the inference server:
- On the first 429 in a burst: the limit is reduced by a configurable factor (default: 25% reduction) and a cooldown is applied. Further 429s from already in-flight requests in the same burst do not reduce the limit again β they release their permits and hold the limit steady.
- After consecutive successes: the limit increases by 1 (by default) until it reaches the ceiling or a stabilized rate-limit threshold.
This means Data Designer automatically finds the right concurrency level for your server without manual tuning.
Sync engine caveat
AIMD adaptive concurrency is fully active on the async engine path. On the current sync engine path, 429 responses are first retried at the HTTP transport layer; AIMD only engages as a fallback if transport retries are exhausted. In practice the concurrency limit stays near max_parallel_requests for most sync workloads. The async engine is landing soon and will be the recommended path for production workloads.
Example: With buffer_size=100 and max_parallel_requests=32, Data Designer starts sending up to 32 requests in parallel. If the server returns 429s, concurrency drops automatically (e.g., to 24, then 18) and recovers once the server catches up.
Configuration Parameters
buffer_size (RunConfig)
Controls how many records are processed per batch.
import data_designer.config as dd
from data_designer.interface import DataDesigner
run_config = dd.RunConfig(buffer_size=2000)
designer = DataDesigner()
designer.set_run_config(run_config)
| Value | Memory Usage | Throughput | Error Feedback |
|---|---|---|---|
| Low (100-500) | Lower | May not saturate inference | Fast |
| Default (1000) | Moderate | Good for most cases | Moderate |
| High (2000-5000) | Higher | Better for deep pipelines | Slower |
When to increase: High-capacity inference server, single-model workflows, memory not constrained
When to decrease: Memory-constrained environments, development/debugging, complex multi-model pipelines
max_parallel_requests (InferenceParams)
Sets the maximum concurrent LLM API calls per model. This is the ceiling that the AIMD throttle controller can ramp up to β the actual concurrency at runtime may be lower if the server signals rate limits.
import data_designer.config as dd
model = dd.ModelConfig(
alias="my-model",
model="nvidia/nemotron-3-nano-30b-a3b",
inference_parameters=dd.ChatCompletionInferenceParams(
max_parallel_requests=8,
),
)
Default: 4
When to increase: Your inference backend has high throughput capacity, you're using a cloud API with generous rate limits, or you're running vLLM/TensorRT-LLM with multiple GPUs. With AIMD, setting an aggressively high value is safer than before β the system will self-correct downward if the server can't keep up. (On the async engine the salvage queue reclaims failed rows; on the sync engine the initial burst of 429s before AIMD stabilizes can drop rows, so start with a more conservative ceiling if you're using the sync path.)
When to decrease: You want to cap resource usage to a known safe level, or you want more predictable/debuggable execution.
Finding the optimal value
The right value depends on your inference stack and model. Self-hosted vLLM servers can often handle values as high as 256, 512, or even 1024 depending on your hardware.
With AIMD, a practical approach is to set max_parallel_requests to the upper bound you're comfortable with and let the throttle controller find the sustainable level automatically. If you see frequent 429 β recovery cycles in the logs, your ceiling is above the server's true capacity but the system is handling it. If you never see any throttle activity, you may have room to increase the ceiling further.
Benchmark approach: Run a small dataset (e.g., 100 records) with increasing max_parallel_requests values (4 β 8 β 16 β 32 β ...) and measure generation time. Stop increasing when the runtime stops decreasingβthat's when your inference server is saturated.
non_inference_max_parallel_workers (RunConfig)
Controls thread pool size for non-LLM operations (samplers, expressions, validators).
run_config = dd.RunConfig(non_inference_max_parallel_workers=8)
designer.set_run_config(run_config)
Default: 4
When to increase: Many CPU-bound columns (complex expressions, heavy sampling)
Adaptive Throttling (RunConfig)
Data Designer uses an AIMD (Additive Increase / Multiplicative Decrease) controller to automatically adjust concurrency per model based on rate-limit feedback from the inference server. The defaults work well for most workloads. Override them via ThrottleConfig only when you understand the trade-offs.
Sync engine caveat
Adaptive throttling is fully active on the async engine path, where 429 responses propagate directly to the AIMD controller. On the sync engine path, 429s are first retried at the HTTP transport layer; ThrottleConfig settings only take effect as a fallback if transport retries are exhausted. The async engine is landing soon and will be the recommended path for production workloads.
import data_designer.config as dd
from data_designer.interface import DataDesigner
run_config = dd.RunConfig(
throttle=dd.ThrottleConfig(
reduce_factor=0.75, # Multiply limit by this on a 429 (default: 0.75)
additive_increase=1, # Add this many slots after success_window successes (default: 1)
success_window=25, # Consecutive successes before increasing (default: 25)
cooldown_seconds=2.0, # Pause after a 429 when no Retry-After header (default: 2.0)
ceiling_overshoot=0.10, # Probe 10% above observed server limit (default: 0.10)
),
)
designer = DataDesigner()
designer.set_run_config(run_config)
| Parameter | Default | Effect |
|---|---|---|
reduce_factor |
0.75 | How aggressively to cut concurrency on a 429. Lower = more aggressive. |
additive_increase |
1 | Slots added per recovery step. Higher = faster ramp-up, but riskier. |
success_window |
25 | Consecutive successes required before each increase step. |
cooldown_seconds |
2.0 | Pause duration after a 429 (used when the server doesn't send Retry-After). |
ceiling_overshoot |
0.10 | Fraction above the observed rate-limit ceiling the controller is allowed to probe. |
How it works in practice
When a model endpoint returns HTTP 429, the controller reduces the concurrency limit for that model and pauses briefly. After enough consecutive successes, it begins ramping back up. If the server rate-limits again, the controller records that level as a ceiling and stabilizes just below it, with a small overshoot band to detect when the server can handle more load.
You can observe this in the logs β look for messages like concurrency reduced from X β Y and concurrency increased from X β Y.
Error Handling (RunConfig)
Control retry behavior and early shutdown for failed generations.
run_config = dd.RunConfig(
max_conversation_restarts=5, # Full conversation restarts (default: 5)
max_conversation_correction_steps=0, # In-conversation corrections (default: 0)
disable_early_shutdown=False, # Enable early shutdown (default)
shutdown_error_rate=0.5, # Shut down if >50% errors
shutdown_error_window=10, # Min tasks before error monitoring
)
designer.set_run_config(run_config)
When to adjust:
- Strict schemas: Increase
max_conversation_restartsto 7, addmax_conversation_correction_steps=2 - Debugging: Set
disable_early_shutdown=Trueto see all errors - Simple text: Reduce
max_conversation_restartsto 3
Common Problems
| Problem | Symptom | Solution |
|---|---|---|
| Low throughput | Low GPU utilization | Increase max_parallel_requests and/or buffer_size. If the throttle has self-reduced due to earlier 429s (check logs for "concurrency reduced" messages), the server may need more capacity or you can wait for AIMD recovery. |
| Frequent 429 β recovery cycles | Logs show repeated concurrency drops and ramp-ups | The max_parallel_requests ceiling is above the server's sustained capacity. This is handled automatically, but you can lower the ceiling to reduce the sawtooth or tune reduce_factor / success_window. |
| Long tail of slow generations | Most records fast, few very slow | Reduce max_conversation_restarts, simplify schemas, improve prompts |
| Multi-model idle periods | One model busy, others idle | Reduce buffer_size for faster cycling, or consolidate models |
| Memory errors | OOM crashes | Reduce buffer_size and max_parallel_requests |
| Too many errors | Generation fails frequently | Check prompts/schemas; adjust shutdown_error_rate or disable early shutdown for debugging |
Tuning Workflow
- Start with defaults for initial development β AIMD handles rate-limit adaptation automatically
- Profile your workload: How many LLM columns? How many records? What models?
- Identify bottleneck: Low GPU util β increase
max_parallel_requests(AIMD will self-correct if you overshoot). Memory issues β decreasebuffer_size. Long tails β tune retry settings. - Check throttle logs: Look for "concurrency reduced" / "concurrency increased" messages to understand whether rate limits are the bottleneck
- Iterate: Make one change at a time, measure impact before next change
Related Documentation
- Deployment Options: Choosing between library and microservice
- Model Configuration: Complete model settings reference
- Inference Parameters: Detailed parameter reference