Re: [igt-dev] [PATCH i-g-t] i915/gem_ctx_persistence: Race context closure with replace-engines

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On 11/02/2020 19:22, Chris Wilson wrote:
Tvrtko spotted a race condition between replacing a set of hanging
engines and closing the context. So exercise it.

5s is not much time to hit the small window, but a little bit of testing
several times a day is better than nothing.

Signed-off-by: Chris Wilson <chris@xxxxxxxxxxxxxxxxxx>
Cc: Tvrtko Ursulin <tvrtko.ursulin@xxxxxxxxx>
---
  tests/i915/gem_ctx_persistence.c | 93 ++++++++++++++++++++++++++++++++
  1 file changed, 93 insertions(+)

diff --git a/tests/i915/gem_ctx_persistence.c b/tests/i915/gem_ctx_persistence.c
index 22f29d25e..6321dbe67 100644
--- a/tests/i915/gem_ctx_persistence.c
+++ b/tests/i915/gem_ctx_persistence.c
@@ -33,6 +33,7 @@
  #include "i915/gem_engine_topology.h"
  #include "i915/gem_ring.h"
  #include "i915/gem_submission.h"
+#include "igt_aux.h"
  #include "igt_debugfs.h"
  #include "igt_dummyload.h"
  #include "igt_gt.h"
@@ -803,6 +804,95 @@ static void replace_engines(int i915, const struct intel_execution_engine2 *e)
  	gem_quiescent_gpu(i915);
  }
+struct close_race {
+	int pipe[2];
+};
+
+static void race_set_engines(int i915, int fd)
+{
+	I915_DEFINE_CONTEXT_PARAM_ENGINES(engines, 1) = {
+		.engines = {}
+	};
+	struct drm_i915_gem_context_param param = {
+		.param = I915_CONTEXT_PARAM_ENGINES,
+		.value = to_user_pointer(&engines),
+		.size = sizeof(engines),
+	};
+
+	while (read(fd, &param.ctx_id, sizeof(param.ctx_id)) > 0) {
+		if (!param.ctx_id)
+			break;
+		__gem_context_set_param(i915, &param);
+	}
+}
+
+static void close_replace_race(int i915)
+{
+	const int ncpus = sysconf(_SC_NPROCESSORS_ONLN);
+	struct close_race *t;
+	int fence = -1;
+
+	/*
+	 * Tvrtko being the evil genius noticed that if we could successfully
+	 * replace a set of engines after the context had been closed, those
+	 * engines could escape oversight.
+	 */

I think it would read better if you reworded this a bit not to mention names and claimed attributes. :)

+
+	t = malloc(sizeof(*t) * ncpus);
+	igt_assert(t);
+
+	for (int i = 0; i < ncpus; i++)
+		igt_assert(pipe(t[i].pipe) == 0);
+
+	igt_fork(child, ncpus) {
+		close(t[child].pipe[1]);
+		race_set_engines(i915, t[child].pipe[0]);
+	}
+
+	for (int i = 0; i < ncpus; i++)
+		close(t[i].pipe[0]);
+
+	igt_until_timeout(5) {
+		igt_spin_t *spin;
+		uint32_t ctx;
+
+		ctx = gem_context_clone_with_engines(i915, 0);
+		gem_context_set_persistence(i915, ctx, true);
+
+		spin = igt_spin_new(i915, ctx, .flags = IGT_SPIN_FENCE_OUT);
+		for (int i = 0; i < ncpus; i++)
+			write(t[i].pipe[1], &ctx, sizeof(ctx));

It's early so I hope I am not too confused, but drm client in the forked process is a different one. So I think it needs to use threads to be able to share.

Regards,

Tvrtko

+
+		if (fence < 0) {
+			fence = spin->out_fence;
+		} else {
+			int tmp;
+
+			tmp = sync_fence_merge(fence, spin->out_fence);
+			close(fence);
+			close(spin->out_fence);
+
+			fence = tmp;
+		}
+		spin->out_fence = -1;
+
+		gem_context_destroy(i915, ctx);
+	}
+
+	for (int i = 0; i < ncpus; i++) {
+		uint32_t end = 0;
+		write(t[i].pipe[1], &end, sizeof(end));
+		close(t[i].pipe[1]);
+	}
+	igt_waitchildren();
+	free(t);
+
+	igt_assert(sync_fence_wait(fence, MSEC_PER_SEC / 2) == 0);
+	close(fence);
+
+	gem_quiescent_gpu(i915);
+}
+
  static void replace_engines_hostile(int i915,
  				    const struct intel_execution_engine2 *e)
  {
@@ -961,6 +1051,9 @@ igt_main
  					replace_engines_hostile(i915, e);
  			}
  		}
+
+		igt_subtest("close-replace-race")
+			close_replace_race(i915);
  	}
igt_fixture {

_______________________________________________
Intel-gfx mailing list
Intel-gfx@xxxxxxxxxxxxxxxxxxxxx
https://lists.freedesktop.org/mailman/listinfo/intel-gfx



[Index of Archives]     [AMD Graphics]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux