2 * Copyright (C) 2013 Red Hat
3 * Author: Rob Clark <robdclark@gmail.com>
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 as published by
7 * the Free Software Foundation.
9 * This program is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
14 * You should have received a copy of the GNU General Public License along with
15 * this program. If not, see <http://www.gnu.org/licenses/>.
21 #include <linux/kref.h>
22 #include <linux/reservation.h>
25 /* Additional internal-use only BO flags: */
26 #define MSM_BO_STOLEN 0x10000000 /* try to use stolen/splash memory */
28 struct msm_gem_address_space
{
30 /* NOTE: mm managed at the page level, size is in # of pages
31 * and position mm_node->start is in # of pages:
34 spinlock_t lock
; /* Protects drm_mm node allocation/removal */
40 struct drm_mm_node node
;
42 struct msm_gem_address_space
*aspace
;
43 struct list_head list
; /* node in msm_gem_object::vmas */
46 struct msm_gem_object
{
47 struct drm_gem_object base
;
52 * Advice: are the backing pages purgeable?
57 * count of active vmap'ing
61 /* And object is either:
62 * inactive - on priv->inactive_list
63 * active - on one one of the gpu's active_list.. well, at
64 * least for now we don't have (I don't think) hw sync between
65 * 2d and 3d one devices which have both, meaning we need to
66 * block on submit if a bo is already on other ring
69 struct list_head mm_list
;
70 struct msm_gpu
*gpu
; /* non-null if active */
72 /* Transiently in the process of submit ioctl, objects associated
73 * with the submit are on submit->bo_list.. this only lasts for
74 * the duration of the ioctl, so one bo can never be on multiple
77 struct list_head submit_entry
;
83 struct list_head vmas
; /* list of msm_gem_vma */
85 /* normally (resv == &_resv) except for imported bo's */
86 struct reservation_object
*resv
;
87 struct reservation_object _resv
;
89 /* For physically contiguous buffers. Used when we don't have
90 * an IOMMU. Also used for stolen/splashscreen buffer.
92 struct drm_mm_node
*vram_node
;
93 struct mutex lock
; /* Protects resources associated with bo */
95 #define to_msm_bo(x) container_of(x, struct msm_gem_object, base)
97 static inline bool is_active(struct msm_gem_object
*msm_obj
)
99 return msm_obj
->gpu
!= NULL
;
102 static inline bool is_purgeable(struct msm_gem_object
*msm_obj
)
104 WARN_ON(!mutex_is_locked(&msm_obj
->base
.dev
->struct_mutex
));
105 return (msm_obj
->madv
== MSM_MADV_DONTNEED
) && msm_obj
->sgt
&&
106 !msm_obj
->base
.dma_buf
&& !msm_obj
->base
.import_attach
;
109 static inline bool is_vunmapable(struct msm_gem_object
*msm_obj
)
111 return (msm_obj
->vmap_count
== 0) && msm_obj
->vaddr
;
114 /* The shrinker can be triggered while we hold objA->lock, and need
115 * to grab objB->lock to purge it. Lockdep just sees these as a single
116 * class of lock, so we use subclasses to teach it the difference.
118 * OBJ_LOCK_NORMAL is implicit (ie. normal mutex_lock() call), and
119 * OBJ_LOCK_SHRINKER is used by shrinker.
121 * It is *essential* that we never go down paths that could trigger the
122 * shrinker for a purgable object. This is ensured by checking that
123 * msm_obj->madv == MSM_MADV_WILLNEED.
130 void msm_gem_purge(struct drm_gem_object
*obj
, enum msm_gem_lock subclass
);
131 void msm_gem_vunmap(struct drm_gem_object
*obj
, enum msm_gem_lock subclass
);
133 /* Created per submit-ioctl, to track bo's and cmdstream bufs, etc,
134 * associated with the cmdstream submission for synchronization (and
135 * make it easier to unwind when things go wrong, etc). This only
136 * lasts for the duration of the submit-ioctl.
138 struct msm_gem_submit
{
139 struct drm_device
*dev
;
141 struct list_head node
; /* node in ring submit list */
142 struct list_head bo_list
;
143 struct ww_acquire_ctx ticket
;
144 uint32_t seqno
; /* Sequence number of the submit on the ring */
145 struct dma_fence
*fence
;
146 struct msm_gpu_submitqueue
*queue
;
147 struct pid
*pid
; /* submitting process */
148 bool valid
; /* true if no cmdstream patching needed */
149 struct msm_ringbuffer
*ring
;
150 unsigned int nr_cmds
;
154 uint32_t size
; /* in dwords */
156 uint32_t idx
; /* cmdstream buffer idx in bos[] */
157 } *cmd
; /* array of size nr_cmds */
160 struct msm_gem_object
*obj
;
165 #endif /* __MSM_GEM_H__ */