2 * Copyright © 2016 Intel Corporation
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
26 #include "intel_frontbuffer.h"
27 #include "i915_gem_clflush.h"
29 static DEFINE_SPINLOCK(clflush_lock
);
32 struct dma_fence dma
; /* Must be first for dma_fence_free() */
33 struct i915_sw_fence wait
;
34 struct work_struct work
;
35 struct drm_i915_gem_object
*obj
;
38 static const char *i915_clflush_get_driver_name(struct dma_fence
*fence
)
43 static const char *i915_clflush_get_timeline_name(struct dma_fence
*fence
)
48 static bool i915_clflush_enable_signaling(struct dma_fence
*fence
)
53 static void i915_clflush_release(struct dma_fence
*fence
)
55 struct clflush
*clflush
= container_of(fence
, typeof(*clflush
), dma
);
57 i915_sw_fence_fini(&clflush
->wait
);
59 BUILD_BUG_ON(offsetof(typeof(*clflush
), dma
));
60 dma_fence_free(&clflush
->dma
);
63 static const struct dma_fence_ops i915_clflush_ops
= {
64 .get_driver_name
= i915_clflush_get_driver_name
,
65 .get_timeline_name
= i915_clflush_get_timeline_name
,
66 .enable_signaling
= i915_clflush_enable_signaling
,
67 .wait
= dma_fence_default_wait
,
68 .release
= i915_clflush_release
,
71 static void __i915_do_clflush(struct drm_i915_gem_object
*obj
)
73 GEM_BUG_ON(!i915_gem_object_has_pages(obj
));
74 drm_clflush_sg(obj
->mm
.pages
);
75 intel_fb_obj_flush(obj
, ORIGIN_CPU
);
78 static void i915_clflush_work(struct work_struct
*work
)
80 struct clflush
*clflush
= container_of(work
, typeof(*clflush
), work
);
81 struct drm_i915_gem_object
*obj
= clflush
->obj
;
83 if (i915_gem_object_pin_pages(obj
)) {
84 DRM_ERROR("Failed to acquire obj->pages for clflushing\n");
88 __i915_do_clflush(obj
);
90 i915_gem_object_unpin_pages(obj
);
93 i915_gem_object_put(obj
);
95 dma_fence_signal(&clflush
->dma
);
96 dma_fence_put(&clflush
->dma
);
99 static int __i915_sw_fence_call
100 i915_clflush_notify(struct i915_sw_fence
*fence
,
101 enum i915_sw_fence_notify state
)
103 struct clflush
*clflush
= container_of(fence
, typeof(*clflush
), wait
);
107 schedule_work(&clflush
->work
);
111 dma_fence_put(&clflush
->dma
);
118 bool i915_gem_clflush_object(struct drm_i915_gem_object
*obj
,
121 struct clflush
*clflush
;
124 * Stolen memory is always coherent with the GPU as it is explicitly
125 * marked as wc by the system, or the system is cache-coherent.
126 * Similarly, we only access struct pages through the CPU cache, so
127 * anything not backed by physical memory we consider to be always
128 * coherent and not need clflushing.
130 if (!i915_gem_object_has_struct_page(obj
)) {
131 obj
->cache_dirty
= false;
135 /* If the GPU is snooping the contents of the CPU cache,
136 * we do not need to manually clear the CPU cache lines. However,
137 * the caches are only snooped when the render cache is
138 * flushed/invalidated. As we always have to emit invalidations
139 * and flushes when moving into and out of the RENDER domain, correct
140 * snooping behaviour occurs naturally as the result of our domain
143 if (!(flags
& I915_CLFLUSH_FORCE
) &&
144 obj
->cache_coherent
& I915_BO_CACHE_COHERENT_FOR_READ
)
147 trace_i915_gem_object_clflush(obj
);
150 if (!(flags
& I915_CLFLUSH_SYNC
))
151 clflush
= kmalloc(sizeof(*clflush
), GFP_KERNEL
);
153 GEM_BUG_ON(!obj
->cache_dirty
);
155 dma_fence_init(&clflush
->dma
,
158 to_i915(obj
->base
.dev
)->mm
.unordered_timeline
,
160 i915_sw_fence_init(&clflush
->wait
, i915_clflush_notify
);
162 clflush
->obj
= i915_gem_object_get(obj
);
163 INIT_WORK(&clflush
->work
, i915_clflush_work
);
165 dma_fence_get(&clflush
->dma
);
167 i915_sw_fence_await_reservation(&clflush
->wait
,
169 true, I915_FENCE_TIMEOUT
,
172 reservation_object_lock(obj
->resv
, NULL
);
173 reservation_object_add_excl_fence(obj
->resv
, &clflush
->dma
);
174 reservation_object_unlock(obj
->resv
);
176 i915_sw_fence_commit(&clflush
->wait
);
177 } else if (obj
->mm
.pages
) {
178 __i915_do_clflush(obj
);
180 GEM_BUG_ON(obj
->write_domain
!= I915_GEM_DOMAIN_CPU
);
183 obj
->cache_dirty
= false;