# This file is part of GNUnet # Copyright (C) 2018 GNUnet e.V. # # GNUnet is free software: you can redistribute it and/or modify it # under the terms of the GNU Affero General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # GNUnet is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see . # # SPDX-License-Identfier: AGPL3.0-or-later # Aggregate benchmarking data from multiple threads/processes # generated by util/benchmark.c. # # Can be used as # awk -f collect.awk gnunet-benchmark-{ops,urls}-*.txt # records are of the following forms: # 1:op 2: 3:count 4: 6:time_us 7: # 1:url 2: 3:status 4: 5:count 6: 7:time_us 8: 9:time_us_max 10: # 11:bytes_sent 12: 13:bytes_received 14: function abs(v) { return v < 0 ? -v : v } { if ($1 == "op") { n = $4; t = $6; op[$2]["count"] += n; op[$2]["time_us"] += t; if (n > 0) { op[$2]["time_us_sq"] += n * (t/n) * (t/n); } total_ops += t; } else if ($1 == "url") { n = $6; t = $8; sent = $12 recv = $14 url[$2][$4]["count"] += n; url[$2][$4]["time_us"] += t; if (n > 0) { url[$2][$4]["time_us_sq"] += n * (t/n) * (t/n); } url[$2][$4]["bytes_sent"] += sent; url[$2][$4]["bytes_received"] += recv; max = url[$2][$4]["time_us_max"]; url[$2][$4]["time_us_max"] = (t/n > max ? t/n : max) } else if ($1 == "op_baseline") { # take average time for operations from baseline values with format: # op_baseline time_avg_us op_baseline[$2] = $4; have_baseline = 1; } } function avg(sum, n) { if (n == 0) { return 0; } else { return sum / n; } } function stdev(sum, sum_sq, n) { if (n == 0) { return 0; } else { return sqrt(abs((sum_sq / n) - ((sum / n) * (sum / n)))); } } END { for (x in op) { print "op", x, "count", op[x]["count"], "time_us", op[x]["time_us"], \ "time_avg_us", avg(op[x]["time_us"], op[x]["count"]), \ "stdev", stdev(op[x]["time_us"], op[x]["time_us_sq"], op[x]["count"]); } for (x in url) { for (y in url[x]) { print "url", x, "status", y, \ "count", url[x][y]["count"], "time_us", url[x][y]["time_us"], \ "time_avg_us", avg(url[x][y]["time_us"], url[x][y]["count"]), \ "stdev", stdev(url[x][y]["time_us"], url[x][y]["time_us_sq"], url[x][y]["count"]), \ "time_us_max", url[x][y]["time_us_max"], \ "bytes_sent_avg", avg(url[x][y]["bytes_sent"], url[x][y]["count"]), \ "bytes_received_avg", avg(url[x][y]["bytes_received"], url[x][y]["count"]); } } if (total_ops) { print "total_ops_ms", total_ops; } # Invoke awk with -V baseline_out= to extract baseline average if (baseline_out) { for (x in op) { print "op_baseline", x, "time_avg_us", avg(op[x]["time_us"], op[x]["count"]) > baseline_out } } if (have_baseline) { for (x in op) { total_ops_adjusted += op_baseline[x] * op[x]["count"]; } print "total_ops_adjusted_ms", int(total_ops_adjusted); } }