From d445d1ea26588bc3aa81f8845a32d0bff8142aa2 Mon Sep 17 00:00:00 2001 From: Badrish Chandramouli Date: Sun, 22 Dec 2024 14:26:53 -0800 Subject: [PATCH] We use a batch size of 1 for network BDNs in order to stress the network layer. (#896) --- benchmark/BDN.benchmark/Network/NetworkBase.cs | 11 +++-------- 1 file changed, 3 insertions(+), 8 deletions(-) diff --git a/benchmark/BDN.benchmark/Network/NetworkBase.cs b/benchmark/BDN.benchmark/Network/NetworkBase.cs index a217988640..2e8400ae64 100644 --- a/benchmark/BDN.benchmark/Network/NetworkBase.cs +++ b/benchmark/BDN.benchmark/Network/NetworkBase.cs @@ -28,15 +28,10 @@ public IEnumerable NetworkParamsProvider() } /// - /// Batch size per method invocation - /// With a batchSize of 100, we have a convenient conversion of latency to throughput: - /// 5 us = 20 Mops/sec - /// 10 us = 10 Mops/sec - /// 20 us = 5 Mops/sec - /// 25 us = 4 Mops/sec - /// 100 us = 1 Mops/sec + /// Batch size per method invocation - we use a batch size of 1 for network BDNs + /// in order to stress the network layer. /// - const int batchSize = 100; + const int batchSize = 1; EmbeddedRespServer server; EmbeddedNetworkHandler networkHandler;