On Thu, Mar 24 2022, Neeraj Singh wrote: > On Thu, Mar 24, 2022 at 9:53 AM Ævar Arnfjörð Bjarmason > <avarab@xxxxxxxxx> wrote: >> >> >> On Thu, Mar 24 2022, Neeraj Singh via GitGitGadget wrote: >> >> > From: Neeraj Singh <neerajsi@xxxxxxxxxxxxx> >> > >> > Add test cases to exercise batch mode for: >> > * 'git add' >> > * 'git stash' >> > * 'git update-index' >> > * 'git unpack-objects' >> > >> > These tests ensure that the added data winds up in the object database. >> > >> > In this change we introduce a new test helper lib-unique-files.sh. The >> > goal of this library is to create a tree of files that have different >> > oids from any other files that may have been created in the current test >> > repo. This helps us avoid missing validation of an object being added >> > due to it already being in the repo. >> > >> > Signed-off-by: Neeraj Singh <neerajsi@xxxxxxxxxxxxx> >> > --- >> > t/lib-unique-files.sh | 32 ++++++++++++++++++++++++++++++++ >> > t/t3700-add.sh | 28 ++++++++++++++++++++++++++++ >> > t/t3903-stash.sh | 20 ++++++++++++++++++++ >> > t/t5300-pack-object.sh | 41 +++++++++++++++++++++++++++-------------- >> > 4 files changed, 107 insertions(+), 14 deletions(-) >> > create mode 100644 t/lib-unique-files.sh >> > >> > diff --git a/t/lib-unique-files.sh b/t/lib-unique-files.sh >> > new file mode 100644 >> > index 00000000000..74efca91dd7 >> > --- /dev/null >> > +++ b/t/lib-unique-files.sh >> > @@ -0,0 +1,32 @@ >> > +# Helper to create files with unique contents >> > + >> > +# Create multiple files with unique contents within this test run. Takes the >> > +# number of directories, the number of files in each directory, and the base >> > +# directory. >> > +# >> > +# test_create_unique_files 2 3 my_dir -- Creates 2 directories with 3 files >> > +# each in my_dir, all with contents >> > +# different from previous invocations >> > +# of this command in this run. >> > + >> > +test_create_unique_files () { >> > + test "$#" -ne 3 && BUG "3 param" >> > + >> > + local dirs="$1" && >> > + local files="$2" && >> > + local basedir="$3" && >> > + local counter=0 && >> > + test_tick && >> > + local basedata=$basedir$test_tick && >> > + rm -rf "$basedir" && >> > + for i in $(test_seq $dirs) >> > + do >> > + local dir=$basedir/dir$i && >> > + mkdir -p "$dir" && >> > + for j in $(test_seq $files) >> > + do >> > + counter=$((counter + 1)) && >> > + echo "$basedata.$counter">"$dir/file$j.txt" >> > + done >> > + done >> > +} >> >> Having written my own perf tests for this series, I still don't get why >> this is needed, at all. >> >> tl;dr: the below: I think this whole workaround is because you missed >> that "test_when_finished" exists, and how it excludes perf timings. >> > > I actually noticed test_when_finished, but I didn't think of your > "setup the next round on cleanup of last" idea. I was debating at the > time adding a "test_perf_setup" helper to do the setup work during > each perf iteration. How about I do that and just create a new repo > in each test_perf_setup step? > >> I.e. I get that if we ran this N times we'd want to wipe our repo >> between tests, as for e.g. "git add" you want it to actually add the >> objects. >> >> It's what I do with the "hyperfine" command in >> https://lore.kernel.org/git/RFC-patch-v2-4.7-61f4f3d7ef4-20220323T140753Z-avarab@xxxxxxxxx/ >> with the "-p" option. >> >> I.e. hyperfine has a way to say "this is setup, but don't measure the >> time", which is 1/2 of what you're working around here and in 10/11. >> >> But as 10/11 shows you're limited to one run with t/perf because you >> want to not include those "setup" numbers, and "test_perf" has no easy >> way to avoid that (but more on that later). >> >> Which b.t.w. I'm really skeptical of as an approach here in any case >> (even if we couldn't exclude it from the numbers). >> >> I.e. yes what "hyperfine" does would be preferrable, but in exchange for >> avoiding that you're comparing samples of 1 runs. >> >> Surely we're better off with N run (even if noisy). Given enough of them >> the difference will shake out, and our estimated +/- will narrow.. >> >> But aside from that, why isn't this just: >> >> for cfg in true false blah >> done >> test_expect_success "setup for $cfg" ' >> git init repo-$cfg && >> for f in $(test_seq 1 100) >> do >> >repo-$cfg/$f >> done >> ' >> >> test_perf "perf test for $cfg" ' >> git -C repo-$cfg >> ' >> done >> >> Which surely is going to be more accurate in the context of our limited >> t/perf environment because creating unique files is not sufficient at >> all to ensure that your tests don't interfere with each other. >> >> That's because in the first iteration we'll create N objects in >> .git/objects/aa/* or whatever, which will *still be there* for your >> second test, which will impact performance. >> >> Whereas if you just make N repos you don't need unique files, and you >> won't be introducing that as a conflating variable. >> >> But anyway, reading perf-lib.sh again I haven't tested, but this whole >> workaround seems truly unnecessary. I.e. in test_run_perf_ we do: >> >> test_run_perf_ () { >> test_cleanup=: >> test_export_="test_cleanup" >> export test_cleanup test_export_ >> "$GTIME" -f "%E %U %S" -o test_time.$i "$TEST_SHELL_PATH" -c ' >> [... code we run and time ...] >> ' >> [... later ...] >> test_eval_ "$test_cleanup" >> } >> >> So can't you just avoid this whole glorious workaround for the low low >> cost of approximately one shellscript string assignment? :) >> >> I.e. if you do: >> >> setup_clean () { >> rm -rf repo >> } >> >> setup_first () { >> git init repo && >> [make a bunch of files or whatever in repo] >> } >> >> setup_next () { >> test_when_finished "setup_clean" && >> setup_first >> } >> >> test_expect_success 'setup initial stuff' ' >> setup_first >> ' >> >> test_perf 'my perf test' ' >> test_when_finished "setup_next" && >> [your perf test here] >> ' >> >> test_expect_success 'cleanup' ' >> # Not really needed, but just for completeness, we are >> # about to nuke the trash dir anyway... >> setup_clean >> ' >> >> I haven't tested (and need to run), but i'm pretty sure that does >> exactly what you want without these workarounds, i.e. you'll get >> "trampoline setup" without that setup being included in the perf >> numbers. >> >> Is it pretty? No, but it's a lot less complex than this unique file >> business & workarounds, and will give you just the numbers you want, and >> most importantly you car run it N times now for better samples. >> >> I.e. "what you want" sans a *tiny* bit of noise that we use to just call >> a function to do: >> >> test_cleanup=setup_next >> >> Which we'll then eval *after* we measure your numbers to setup the next >> test. > > How about I add a new test_perf_setup mechanism to make your idea work > in a straightforward way? Sure, that sounds great. > I still want the test_create_unique_files thing as a way to make > multiple files easily. And for the non-perf tests it makes sense to > have differing contents within a test run. I think running your perf test on some generated data might still make sense, but I think given the above that the *method* really doesn't make any sense. I.e. pretty much the whole structure of t/perf is to write tests that can be run on an arbitrary user-provided repo, some of them do make some content assumptions (or need no repo), but we've tried to have tests there handle arbitrary repos. You ended up with that "generated random files" to get around the X-Y problem of not being able to reset the area without making that part of the metrics, but as demo'd above we can use test_when_finished for that. And once that's resolved it would actually be much more handy to be able to run this on an arbitrary repo, as you can see in my "git hyperfine" one-liner I grabbed the "t" directory, but we could just make our test data all files in the dir (or specify a glob via an env var). I think it still sounds interesting to have a way to make arbitrary test data, but surely that's then better as e.g.: cd t/perf ./make-random-repo /tmp/random-repo && GIT_PERF_REPO=/tmp/random-repo ./run p<your test> I.e. once we've resolved the metrics/play area issue needing to run this on some very specific data is artificial limitation v.s. just being able to point it at a given repo.