From b4df027e4e93d27997749d7c551ecc496dd879f4 Mon Sep 17 00:00:00 2001 From: Jim Harris Date: Tue, 12 Feb 2019 16:09:29 -0700 Subject: [PATCH] test/reactor_perf: remove use of max_delay_us This test doesn't really need to use this - in general we want to test max event throughput with no delays interjected. Signed-off-by: Jim Harris Change-Id: If4aaf90ce815687a5ca725a89dfab5e057f9a5c4 Reviewed-on: https://review.gerrithub.io/c/444306 Tested-by: SPDK CI Jenkins Reviewed-by: Shuhei Matsumoto Reviewed-by: Darek Stojaczyk Reviewed-by: Ben Walker --- test/event/reactor_perf/reactor_perf.c | 7 +------ 1 file changed, 1 insertion(+), 6 deletions(-) diff --git a/test/event/reactor_perf/reactor_perf.c b/test/event/reactor_perf/reactor_perf.c index 8783b48f8..ade0b5804 100644 --- a/test/event/reactor_perf/reactor_perf.c +++ b/test/event/reactor_perf/reactor_perf.c @@ -92,7 +92,6 @@ static void usage(const char *program_name) { printf("%s options\n", program_name); - printf("\t[-d Allowed delay when passing messages between cores in microseconds]\n"); printf("\t[-q Queue depth (default: 1)]\n"); printf("\t[-t time in seconds]\n"); } @@ -107,12 +106,11 @@ main(int argc, char **argv) spdk_app_opts_init(&opts); opts.name = "reactor_perf"; - opts.max_delay_us = 1000; g_time_in_sec = 0; g_queue_depth = 1; - while ((op = getopt(argc, argv, "d:q:t:")) != -1) { + while ((op = getopt(argc, argv, "q:t:")) != -1) { if (op == '?') { usage(argv[0]); exit(1); @@ -123,9 +121,6 @@ main(int argc, char **argv) exit(1); } switch (op) { - case 'd': - opts.max_delay_us = (uint64_t)val; - break; case 'q': g_queue_depth = val; break;