Merge tag 'block-5.11-2021-01-10' of git://git.kernel.dk/linux-block
[linux/fpc-iii.git] / drivers / remoteproc / remoteproc_virtio.c
blob0cc617f760682c5de9195c982643162487b0374c
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * Remote processor messaging transport (OMAP platform-specific bits)
5 * Copyright (C) 2011 Texas Instruments, Inc.
6 * Copyright (C) 2011 Google, Inc.
8 * Ohad Ben-Cohen <ohad@wizery.com>
9 * Brian Swetland <swetland@google.com>
12 #include <linux/dma-map-ops.h>
13 #include <linux/export.h>
14 #include <linux/of_reserved_mem.h>
15 #include <linux/remoteproc.h>
16 #include <linux/virtio.h>
17 #include <linux/virtio_config.h>
18 #include <linux/virtio_ids.h>
19 #include <linux/virtio_ring.h>
20 #include <linux/err.h>
21 #include <linux/kref.h>
22 #include <linux/slab.h>
24 #include "remoteproc_internal.h"
26 /* kick the remote processor, and let it know which virtqueue to poke at */
27 static bool rproc_virtio_notify(struct virtqueue *vq)
29 struct rproc_vring *rvring = vq->priv;
30 struct rproc *rproc = rvring->rvdev->rproc;
31 int notifyid = rvring->notifyid;
33 dev_dbg(&rproc->dev, "kicking vq index: %d\n", notifyid);
35 rproc->ops->kick(rproc, notifyid);
36 return true;
39 /**
40 * rproc_vq_interrupt() - tell remoteproc that a virtqueue is interrupted
41 * @rproc: handle to the remote processor
42 * @notifyid: index of the signalled virtqueue (unique per this @rproc)
44 * This function should be called by the platform-specific rproc driver,
45 * when the remote processor signals that a specific virtqueue has pending
46 * messages available.
48 * Returns IRQ_NONE if no message was found in the @notifyid virtqueue,
49 * and otherwise returns IRQ_HANDLED.
51 irqreturn_t rproc_vq_interrupt(struct rproc *rproc, int notifyid)
53 struct rproc_vring *rvring;
55 dev_dbg(&rproc->dev, "vq index %d is interrupted\n", notifyid);
57 rvring = idr_find(&rproc->notifyids, notifyid);
58 if (!rvring || !rvring->vq)
59 return IRQ_NONE;
61 return vring_interrupt(0, rvring->vq);
63 EXPORT_SYMBOL(rproc_vq_interrupt);
65 static struct virtqueue *rp_find_vq(struct virtio_device *vdev,
66 unsigned int id,
67 void (*callback)(struct virtqueue *vq),
68 const char *name, bool ctx)
70 struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
71 struct rproc *rproc = vdev_to_rproc(vdev);
72 struct device *dev = &rproc->dev;
73 struct rproc_mem_entry *mem;
74 struct rproc_vring *rvring;
75 struct fw_rsc_vdev *rsc;
76 struct virtqueue *vq;
77 void *addr;
78 int len, size;
80 /* we're temporarily limited to two virtqueues per rvdev */
81 if (id >= ARRAY_SIZE(rvdev->vring))
82 return ERR_PTR(-EINVAL);
84 if (!name)
85 return NULL;
87 /* Search allocated memory region by name */
88 mem = rproc_find_carveout_by_name(rproc, "vdev%dvring%d", rvdev->index,
89 id);
90 if (!mem || !mem->va)
91 return ERR_PTR(-ENOMEM);
93 rvring = &rvdev->vring[id];
94 addr = mem->va;
95 len = rvring->len;
97 /* zero vring */
98 size = vring_size(len, rvring->align);
99 memset(addr, 0, size);
101 dev_dbg(dev, "vring%d: va %pK qsz %d notifyid %d\n",
102 id, addr, len, rvring->notifyid);
105 * Create the new vq, and tell virtio we're not interested in
106 * the 'weak' smp barriers, since we're talking with a real device.
108 vq = vring_new_virtqueue(id, len, rvring->align, vdev, false, ctx,
109 addr, rproc_virtio_notify, callback, name);
110 if (!vq) {
111 dev_err(dev, "vring_new_virtqueue %s failed\n", name);
112 rproc_free_vring(rvring);
113 return ERR_PTR(-ENOMEM);
116 rvring->vq = vq;
117 vq->priv = rvring;
119 /* Update vring in resource table */
120 rsc = (void *)rproc->table_ptr + rvdev->rsc_offset;
121 rsc->vring[id].da = mem->da;
123 return vq;
126 static void __rproc_virtio_del_vqs(struct virtio_device *vdev)
128 struct virtqueue *vq, *n;
129 struct rproc_vring *rvring;
131 list_for_each_entry_safe(vq, n, &vdev->vqs, list) {
132 rvring = vq->priv;
133 rvring->vq = NULL;
134 vring_del_virtqueue(vq);
138 static void rproc_virtio_del_vqs(struct virtio_device *vdev)
140 __rproc_virtio_del_vqs(vdev);
143 static int rproc_virtio_find_vqs(struct virtio_device *vdev, unsigned int nvqs,
144 struct virtqueue *vqs[],
145 vq_callback_t *callbacks[],
146 const char * const names[],
147 const bool * ctx,
148 struct irq_affinity *desc)
150 int i, ret, queue_idx = 0;
152 for (i = 0; i < nvqs; ++i) {
153 if (!names[i]) {
154 vqs[i] = NULL;
155 continue;
158 vqs[i] = rp_find_vq(vdev, queue_idx++, callbacks[i], names[i],
159 ctx ? ctx[i] : false);
160 if (IS_ERR(vqs[i])) {
161 ret = PTR_ERR(vqs[i]);
162 goto error;
166 return 0;
168 error:
169 __rproc_virtio_del_vqs(vdev);
170 return ret;
173 static u8 rproc_virtio_get_status(struct virtio_device *vdev)
175 struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
176 struct fw_rsc_vdev *rsc;
178 rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;
180 return rsc->status;
183 static void rproc_virtio_set_status(struct virtio_device *vdev, u8 status)
185 struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
186 struct fw_rsc_vdev *rsc;
188 rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;
190 rsc->status = status;
191 dev_dbg(&vdev->dev, "status: %d\n", status);
194 static void rproc_virtio_reset(struct virtio_device *vdev)
196 struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
197 struct fw_rsc_vdev *rsc;
199 rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;
201 rsc->status = 0;
202 dev_dbg(&vdev->dev, "reset !\n");
205 /* provide the vdev features as retrieved from the firmware */
206 static u64 rproc_virtio_get_features(struct virtio_device *vdev)
208 struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
209 struct fw_rsc_vdev *rsc;
211 rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;
213 return rsc->dfeatures;
216 static void rproc_transport_features(struct virtio_device *vdev)
219 * Packed ring isn't enabled on remoteproc for now,
220 * because remoteproc uses vring_new_virtqueue() which
221 * creates virtio rings on preallocated memory.
223 __virtio_clear_bit(vdev, VIRTIO_F_RING_PACKED);
226 static int rproc_virtio_finalize_features(struct virtio_device *vdev)
228 struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
229 struct fw_rsc_vdev *rsc;
231 rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;
233 /* Give virtio_ring a chance to accept features */
234 vring_transport_features(vdev);
236 /* Give virtio_rproc a chance to accept features. */
237 rproc_transport_features(vdev);
239 /* Make sure we don't have any features > 32 bits! */
240 BUG_ON((u32)vdev->features != vdev->features);
243 * Remember the finalized features of our vdev, and provide it
244 * to the remote processor once it is powered on.
246 rsc->gfeatures = vdev->features;
248 return 0;
251 static void rproc_virtio_get(struct virtio_device *vdev, unsigned int offset,
252 void *buf, unsigned int len)
254 struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
255 struct fw_rsc_vdev *rsc;
256 void *cfg;
258 rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;
259 cfg = &rsc->vring[rsc->num_of_vrings];
261 if (offset + len > rsc->config_len || offset + len < len) {
262 dev_err(&vdev->dev, "rproc_virtio_get: access out of bounds\n");
263 return;
266 memcpy(buf, cfg + offset, len);
269 static void rproc_virtio_set(struct virtio_device *vdev, unsigned int offset,
270 const void *buf, unsigned int len)
272 struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
273 struct fw_rsc_vdev *rsc;
274 void *cfg;
276 rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;
277 cfg = &rsc->vring[rsc->num_of_vrings];
279 if (offset + len > rsc->config_len || offset + len < len) {
280 dev_err(&vdev->dev, "rproc_virtio_set: access out of bounds\n");
281 return;
284 memcpy(cfg + offset, buf, len);
287 static const struct virtio_config_ops rproc_virtio_config_ops = {
288 .get_features = rproc_virtio_get_features,
289 .finalize_features = rproc_virtio_finalize_features,
290 .find_vqs = rproc_virtio_find_vqs,
291 .del_vqs = rproc_virtio_del_vqs,
292 .reset = rproc_virtio_reset,
293 .set_status = rproc_virtio_set_status,
294 .get_status = rproc_virtio_get_status,
295 .get = rproc_virtio_get,
296 .set = rproc_virtio_set,
300 * This function is called whenever vdev is released, and is responsible
301 * to decrement the remote processor's refcount which was taken when vdev was
302 * added.
304 * Never call this function directly; it will be called by the driver
305 * core when needed.
307 static void rproc_virtio_dev_release(struct device *dev)
309 struct virtio_device *vdev = dev_to_virtio(dev);
310 struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
311 struct rproc *rproc = vdev_to_rproc(vdev);
313 kfree(vdev);
315 kref_put(&rvdev->refcount, rproc_vdev_release);
317 put_device(&rproc->dev);
321 * rproc_add_virtio_dev() - register an rproc-induced virtio device
322 * @rvdev: the remote vdev
323 * @id: the device type identification (used to match it with a driver).
325 * This function registers a virtio device. This vdev's partent is
326 * the rproc device.
328 * Returns 0 on success or an appropriate error value otherwise.
330 int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
332 struct rproc *rproc = rvdev->rproc;
333 struct device *dev = &rvdev->dev;
334 struct virtio_device *vdev;
335 struct rproc_mem_entry *mem;
336 int ret;
338 if (rproc->ops->kick == NULL) {
339 ret = -EINVAL;
340 dev_err(dev, ".kick method not defined for %s\n", rproc->name);
341 goto out;
344 /* Try to find dedicated vdev buffer carveout */
345 mem = rproc_find_carveout_by_name(rproc, "vdev%dbuffer", rvdev->index);
346 if (mem) {
347 phys_addr_t pa;
349 if (mem->of_resm_idx != -1) {
350 struct device_node *np = rproc->dev.parent->of_node;
352 /* Associate reserved memory to vdev device */
353 ret = of_reserved_mem_device_init_by_idx(dev, np,
354 mem->of_resm_idx);
355 if (ret) {
356 dev_err(dev, "Can't associate reserved memory\n");
357 goto out;
359 } else {
360 if (mem->va) {
361 dev_warn(dev, "vdev %d buffer already mapped\n",
362 rvdev->index);
363 pa = rproc_va_to_pa(mem->va);
364 } else {
365 /* Use dma address as carveout no memmapped yet */
366 pa = (phys_addr_t)mem->dma;
369 /* Associate vdev buffer memory pool to vdev subdev */
370 ret = dma_declare_coherent_memory(dev, pa,
371 mem->da,
372 mem->len);
373 if (ret < 0) {
374 dev_err(dev, "Failed to associate buffer\n");
375 goto out;
378 } else {
379 struct device_node *np = rproc->dev.parent->of_node;
382 * If we don't have dedicated buffer, just attempt to re-assign
383 * the reserved memory from our parent. A default memory-region
384 * at index 0 from the parent's memory-regions is assigned for
385 * the rvdev dev to allocate from. Failure is non-critical and
386 * the allocations will fall back to global pools, so don't
387 * check return value either.
389 of_reserved_mem_device_init_by_idx(dev, np, 0);
392 /* Allocate virtio device */
393 vdev = kzalloc(sizeof(*vdev), GFP_KERNEL);
394 if (!vdev) {
395 ret = -ENOMEM;
396 goto out;
398 vdev->id.device = id,
399 vdev->config = &rproc_virtio_config_ops,
400 vdev->dev.parent = dev;
401 vdev->dev.release = rproc_virtio_dev_release;
404 * We're indirectly making a non-temporary copy of the rproc pointer
405 * here, because drivers probed with this vdev will indirectly
406 * access the wrapping rproc.
408 * Therefore we must increment the rproc refcount here, and decrement
409 * it _only_ when the vdev is released.
411 get_device(&rproc->dev);
413 /* Reference the vdev and vring allocations */
414 kref_get(&rvdev->refcount);
416 ret = register_virtio_device(vdev);
417 if (ret) {
418 put_device(&vdev->dev);
419 dev_err(dev, "failed to register vdev: %d\n", ret);
420 goto out;
423 dev_info(dev, "registered %s (type %d)\n", dev_name(&vdev->dev), id);
425 out:
426 return ret;
430 * rproc_remove_virtio_dev() - remove an rproc-induced virtio device
431 * @dev: the virtio device
432 * @data: must be null
434 * This function unregisters an existing virtio device.
436 int rproc_remove_virtio_dev(struct device *dev, void *data)
438 struct virtio_device *vdev = dev_to_virtio(dev);
440 unregister_virtio_device(vdev);
441 return 0;