summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorLars Wirzenius <liw@liw.fi>2015-10-17 19:56:51 +0300
committerLars Wirzenius <liw@liw.fi>2015-10-17 19:56:51 +0300
commitab9a9b5a5767c1951cc5f38bef4c28c10ea10f8f (patch)
tree150af74d0d6beed59b4259c06229293860118bac
parent0e0e8ea5b88a1e7aeeb7acfaa1417a7096823ff1 (diff)
downloadobnam-ab9a9b5a5767c1951cc5f38bef4c28c10ea10f8f.tar.gz
Drop obsolete README.benchmarks
-rw-r--r--README.benchmarks79
1 files changed, 0 insertions, 79 deletions
diff --git a/README.benchmarks b/README.benchmarks
deleted file mode 100644
index 30241aa9..00000000
--- a/README.benchmarks
+++ /dev/null
@@ -1,79 +0,0 @@
-README for Obnam benchmarks
-===========================
-
-I've tried a number of approaches to benchmarks with Obnam over the
-years, but no approach has prevailed. This README describes my current
-approach in the hope that it will evolve into something useful.
-
-Ideally I would optimise Obnam for real-world use, but for now, I will
-be content with the simple synthetic benchmarks described here.
-
-Lars Wirzenius
-
-Overview
---------
-
-I do not want a large number of different benchmarks, at least for
-now. I want a small set that I can and will run systematically, at
-least for each release. Too much data can be just as bad as too little
-data: if it takes too much effort to analyse the data, then that eats
-up from development time. That said, hard numbers are better than
-guesses.
-
-I've decided on the following data sets:
-
-* 10^6 empty files, spread over a 1000 directories with 1000 files
- each. Obnam has (at least with repository format 6) a high overhead
- per file, regardless of the contents of the file, and this is a
- pessimal situation for that.
-
- The interesting numbers here are: number of files backed up per
- second, and size of backup repository.
-
-* A single directory with a single file, 2^12 bytes (1 TiB) long.
- little repetition in the data. This benchmarks the opposite end of
- the spectrum of number of files vs size of data.
-
- The interesting numbers here are number of bytes of actual file data
- backed up per second and size of backup repository.
-
-Later, I may add more data sets. An intriguing idea would be to
-generate data from [Summain] manifests, where everything except the
-actual file data is duplicated from anonymised manifests captured from
-real systems.
-
-[Summain]: http://liw.fi/summain/
-
-For each data set, I will run the following operations:
-
-* An initial backup.
-* A no-op second generation backup.
-* A restore of the second generation, with `obnam restore`.
-* A restore of the second generation, with `obnam mount`.
-
-I will measure the following about each operation:
-
-* Total wall-clock time.
-* Maximum VmRSS memory, as logged by Obnam itself.
-
-I will additionally capture Python profiler output of each operation,
-to allow easier analysis of where time is going.
-
-I will run the benchmarks without compression or encryption, at least
-for now, and in general use the default settings built into Obnam for
-everything, unless there's a need to tweak them to make the benchmark
-work at all.
-
-Benchmark results
------------------
-
-A benchmark run will produce the following:
-
-* A JSON file with the measurements given above.
-* A Python profiling file for each operation for each dataset.
- (Two datasets times four operations gives eight profiles.)
-
-I will run the benchmark for each release of Obnam, starting with
-Obnam 1.6.1. I will not care about Larch versions at this time: I will
-use the installed version. I will store the resulting data sets in a
-separate git repository for reference.