Skip to main contentSkip to user menuSkip to navigation

Full Request Path

Budgeted end-to-end request breakdowns

Not Started
Loading...

A typical web request travels through 8-10 infrastructure components before reaching your application. Each hop adds latency, but also provides opportunities for optimization. The total journey can range from 50ms for cached content to 500ms+ for dynamic, uncached responses on slow networks.

The biggest wins come from eliminating entire steps: cache at the edge to skip the origin, reuse connections to skip handshakes, and compress responses to reduce transfer time. Every millisecond saved directly improves user experience and conversion rates.

⚡ Quick Decision

Optimize Connection When:

  • • DNS lookups >20ms frequently
  • • TLS handshakes on every request
  • • Connection setup dominates latency

Optimize Infrastructure When:

  • • Load balancer adds >5ms
  • • API Gateway overhead >15ms
  • • Multiple proxy hops exist

Optimize Payload When:

  • • Response sizes >1MB
  • • Mobile users predominant
  • • Transfer time >100ms

💡 For implementation guides and code examples: See our technology deep dives: NGINX, Observability, HTTP/2 & HTTP/3

Request Path Latency Breakdown

Detailed latency breakdown with optimization insights for each component

DNS Resolution20 ms
If cached, this is sub‑millisecond; misses dominate cold paths.
TLS Handshake30 ms
TLS 1.3 is 1‑RTT; connection reuse removes this entirely.
CDN Edge10 ms
Edge POP localizes RTT; prioritize cacheable responses.
Load Balancer + WAF3 ms
Inline controls add small but consistent overhead.
API Gateway10 ms
Auth, routing, and transformations live here; keep rules lean.
Auth Service2 ms
JWT verification is sub‑millisecond with local keys.
Application + DB15 ms
Indexing and hot read paths keep this bounded.
Response (1MB via 4G)200 ms
Dominated by bandwidth and RTT (BDP). Compress aggressively.

Optimization Opportunities

Proven techniques to reduce request latency.

Connection Pooling-30 ms
Eliminate TLS handshake
HTTP/2 Multiplexing-20 ms
Parallel requests
Edge Caching (95% hit)-100 ms
Skip origin entirely
Response Compression-120 ms
60-80% size reduction
DB Connection Pool-10 ms
Reuse connections

💡 Pro Tip

Combining all optimizations can reduce total latency from 290ms to under 50ms for cached content. The biggest wins come from edge caching and response compression.

No quiz questions available
Quiz ID "full-request-path" not found