2 * SPDX-License-Identifier: MIT
4 * Copyright © 2014-2016 Intel Corporation
7 #include "gt/intel_engine.h"
9 #include "i915_gem_ioctls.h"
10 #include "i915_gem_object.h"
12 static __always_inline u32
__busy_read_flag(u16 id
)
14 if (id
== (u16
)I915_ENGINE_CLASS_INVALID
)
18 return 0x10000u
<< id
;
21 static __always_inline u32
__busy_write_id(u16 id
)
24 * The uABI guarantees an active writer is also amongst the read
25 * engines. This would be true if we accessed the activity tracking
26 * under the lock, but as we perform the lookup of the object and
27 * its activity locklessly we can not guarantee that the last_write
28 * being active implies that we have set the same engine flag from
29 * last_read - hence we always set both read and write busy for
32 if (id
== (u16
)I915_ENGINE_CLASS_INVALID
)
35 return (id
+ 1) | __busy_read_flag(id
);
38 static __always_inline
unsigned int
39 __busy_set_if_active(const struct dma_fence
*fence
, u32 (*flag
)(u16 id
))
41 const struct i915_request
*rq
;
44 * We have to check the current hw status of the fence as the uABI
45 * guarantees forward progress. We could rely on the idle worker
46 * to eventually flush us, but to minimise latency just ask the
49 * Note we only report on the status of native fences.
51 if (!dma_fence_is_i915(fence
))
54 /* opencode to_request() in order to avoid const warnings */
55 rq
= container_of(fence
, const struct i915_request
, fence
);
56 if (i915_request_completed(rq
))
59 /* Beware type-expansion follies! */
60 BUILD_BUG_ON(!typecheck(u16
, rq
->engine
->uabi_class
));
61 return flag(rq
->engine
->uabi_class
);
64 static __always_inline
unsigned int
65 busy_check_reader(const struct dma_fence
*fence
)
67 return __busy_set_if_active(fence
, __busy_read_flag
);
70 static __always_inline
unsigned int
71 busy_check_writer(const struct dma_fence
*fence
)
76 return __busy_set_if_active(fence
, __busy_write_id
);
80 i915_gem_busy_ioctl(struct drm_device
*dev
, void *data
,
81 struct drm_file
*file
)
83 struct drm_i915_gem_busy
*args
= data
;
84 struct drm_i915_gem_object
*obj
;
85 struct dma_resv_list
*list
;
91 obj
= i915_gem_object_lookup_rcu(file
, args
->handle
);
96 * A discrepancy here is that we do not report the status of
97 * non-i915 fences, i.e. even though we may report the object as idle,
98 * a call to set-domain may still stall waiting for foreign rendering.
99 * This also means that wait-ioctl may report an object as busy,
100 * where busy-ioctl considers it idle.
102 * We trade the ability to warn of foreign fences to report on which
103 * i915 engines are active for the object.
105 * Alternatively, we can trade that extra information on read/write
108 * !dma_resv_test_signaled_rcu(obj->resv, true);
109 * to report the overall busyness. This is what the wait-ioctl does.
113 seq
= raw_read_seqcount(&obj
->base
.resv
->seq
);
115 /* Translate the exclusive fence to the READ *and* WRITE engine */
117 busy_check_writer(rcu_dereference(obj
->base
.resv
->fence_excl
));
119 /* Translate shared fences to READ set of engines */
120 list
= rcu_dereference(obj
->base
.resv
->fence
);
122 unsigned int shared_count
= list
->shared_count
, i
;
124 for (i
= 0; i
< shared_count
; ++i
) {
125 struct dma_fence
*fence
=
126 rcu_dereference(list
->shared
[i
]);
128 args
->busy
|= busy_check_reader(fence
);
132 if (args
->busy
&& read_seqcount_retry(&obj
->base
.resv
->seq
, seq
))