If you have ever stared at a load balancer’s dashboard showing wildly fluctuating request rates or struggled to visualize traffic distribution across 50 backend servers, the linear scale has failed you. Enter log10 loadshare —a logarithmic lens that compresses exponential disparities into readable, actionable insights.
log10_loadshare = log10( current_loadshare + 1 ) Why add 1? To handle zero values. log10(0) is undefined (negative infinity). By adding 1, an idle server with 0 RPS yields log10(1) = 0 . A server with 9 RPS yields log10(10) = 1 . This creates a clean, zero-bound metric. | Raw Loadshare (RPS) | log10(RPS + 1) | Interpretation | | :--- | :--- | :--- | | 0 | 0.00 | Idle | | 9 | 1.00 | Minimal load | | 99 | 2.00 | Low load | | 999 | 3.00 | Moderate load | | 9,999 | 4.00 | High load | | 99,999 | 5.00 | Extreme load | log10 loadshare
But log10 loadshare scales universally. Both clusters will show values between 1.7 (50 RPS) and 3.7 (5,000 RPS). You can now create a for all clusters. 3. Autoscaling Algorithms Reactive autoscaling (e.g., KEDA, HPA) often uses thresholds like "scale if CPU > 80%". But CPU is a noisy metric. Request-based scaling using raw RPS is better, but it suffers from the "elephant vs. mouse" problem: a 10x spike in RPS on a small service looks identical to a 10% spike on a large service. If you have ever stared at a load
# Extract RPS per backend from HAProxy logs (simplified) awk 'print $NF' /var/log/haproxy.log | sort | uniq -c | \ awk 'print "log10_loadshare=" log($1+1)/log(10) " raw=" $1' Raw loadshare tells you how much traffic a node handles, but not how well it handles it. A powerful composite metric is the Log-Load Latency Ratio (L3R) : To handle zero values
In distributed systems, loadshare represents the proportionate amount of traffic, computational work, or connection handles assigned to a specific node (server, container, or thread) relative to the total system capacity or total incoming requests. | Context | Definition of Loadshare | | :--- | :--- | | Load Balancer | The number of active connections or requests per second (RPS) routed to a single backend server. | | Message Queue | The number of unacknowledged messages a specific consumer is processing. | | Database Shard | The query throughput or data volume stored on a specific shard replica. | | CDN Edge Node | The bandwidth or request count handled by a particular Point of Presence (PoP). |
def imbalance_score(raw_rates): """ Returns a score between 0 (perfect balance) and 1 (severe imbalance). Uses log10 scale to normalize across magnitudes. """ log_vals = log10_loadshare(raw_rates) max_log = max(log_vals) min_log = min(log_vals) # Theoretical maximum delta in log10 space for typical systems is ~5 return (max_log - min_log) / 5.0 backend_rates = [1500, 1200, 300, 1450, 1400] print(f"Log10 values: log10_loadshare(backend_rates)") print(f"Imbalance score: imbalance_score(backend_rates):.2f") Output: Imbalance score: 0.38 (moderate skew) In HAProxy or Nginx Log Analysis If you have raw access logs, you can compute log10 loadshare per backend server using a one-liner in awk :