mirror of
https://github.com/torvalds/linux.git
synced 2026-03-08 06:04:44 +01:00
This was done entirely with mindless brute force, using
git grep -l '\<k[vmz]*alloc_objs*(.*, GFP_KERNEL)' |
xargs sed -i 's/\(alloc_objs*(.*\), GFP_KERNEL)/\1)/'
to convert the new alloc_obj() users that had a simple GFP_KERNEL
argument to just drop that argument.
Note that due to the extreme simplicity of the scripting, any slightly
more complex cases spread over multiple lines would not be triggered:
they definitely exist, but this covers the vast bulk of the cases, and
the resulting diff is also then easier to check automatically.
For the same reason the 'flex' versions will be done as a separate
conversion.
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
289 lines
6.6 KiB
C
289 lines
6.6 KiB
C
// SPDX-License-Identifier: GPL-2.0-only OR MIT
|
|
/* Copyright (c) 2023 Imagination Technologies Ltd. */
|
|
|
|
#include <uapi/drm/pvr_drm.h>
|
|
|
|
#include <drm/drm_syncobj.h>
|
|
#include <drm/gpu_scheduler.h>
|
|
#include <linux/xarray.h>
|
|
#include <linux/dma-fence-unwrap.h>
|
|
|
|
#include "pvr_device.h"
|
|
#include "pvr_queue.h"
|
|
#include "pvr_sync.h"
|
|
|
|
static int
|
|
pvr_check_sync_op(const struct drm_pvr_sync_op *sync_op)
|
|
{
|
|
u8 handle_type;
|
|
|
|
if (sync_op->flags & ~DRM_PVR_SYNC_OP_FLAGS_MASK)
|
|
return -EINVAL;
|
|
|
|
handle_type = sync_op->flags & DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_MASK;
|
|
if (handle_type != DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ &&
|
|
handle_type != DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_TIMELINE_SYNCOBJ)
|
|
return -EINVAL;
|
|
|
|
if (handle_type == DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ &&
|
|
sync_op->value != 0)
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
pvr_sync_signal_free(struct pvr_sync_signal *sig_sync)
|
|
{
|
|
if (!sig_sync)
|
|
return;
|
|
|
|
drm_syncobj_put(sig_sync->syncobj);
|
|
dma_fence_chain_free(sig_sync->chain);
|
|
dma_fence_put(sig_sync->fence);
|
|
kfree(sig_sync);
|
|
}
|
|
|
|
void
|
|
pvr_sync_signal_array_cleanup(struct xarray *array)
|
|
{
|
|
struct pvr_sync_signal *sig_sync;
|
|
unsigned long i;
|
|
|
|
xa_for_each(array, i, sig_sync)
|
|
pvr_sync_signal_free(sig_sync);
|
|
|
|
xa_destroy(array);
|
|
}
|
|
|
|
static struct pvr_sync_signal *
|
|
pvr_sync_signal_array_add(struct xarray *array, struct drm_file *file, u32 handle, u64 point)
|
|
{
|
|
struct pvr_sync_signal *sig_sync;
|
|
struct dma_fence *cur_fence;
|
|
int err;
|
|
u32 id;
|
|
|
|
sig_sync = kzalloc_obj(*sig_sync);
|
|
if (!sig_sync)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
sig_sync->handle = handle;
|
|
sig_sync->point = point;
|
|
|
|
if (point > 0) {
|
|
sig_sync->chain = dma_fence_chain_alloc();
|
|
if (!sig_sync->chain) {
|
|
err = -ENOMEM;
|
|
goto err_free_sig_sync;
|
|
}
|
|
}
|
|
|
|
sig_sync->syncobj = drm_syncobj_find(file, handle);
|
|
if (!sig_sync->syncobj) {
|
|
err = -EINVAL;
|
|
goto err_free_sig_sync;
|
|
}
|
|
|
|
/* Retrieve the current fence attached to that point. It's
|
|
* perfectly fine to get a NULL fence here, it just means there's
|
|
* no fence attached to that point yet.
|
|
*/
|
|
if (!drm_syncobj_find_fence(file, handle, point, 0, &cur_fence))
|
|
sig_sync->fence = cur_fence;
|
|
|
|
err = xa_alloc(array, &id, sig_sync, xa_limit_32b, GFP_KERNEL);
|
|
if (err)
|
|
goto err_free_sig_sync;
|
|
|
|
return sig_sync;
|
|
|
|
err_free_sig_sync:
|
|
pvr_sync_signal_free(sig_sync);
|
|
return ERR_PTR(err);
|
|
}
|
|
|
|
static struct pvr_sync_signal *
|
|
pvr_sync_signal_array_search(struct xarray *array, u32 handle, u64 point)
|
|
{
|
|
struct pvr_sync_signal *sig_sync;
|
|
unsigned long i;
|
|
|
|
xa_for_each(array, i, sig_sync) {
|
|
if (handle == sig_sync->handle && point == sig_sync->point)
|
|
return sig_sync;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static struct pvr_sync_signal *
|
|
pvr_sync_signal_array_get(struct xarray *array, struct drm_file *file, u32 handle, u64 point)
|
|
{
|
|
struct pvr_sync_signal *sig_sync;
|
|
|
|
sig_sync = pvr_sync_signal_array_search(array, handle, point);
|
|
if (sig_sync)
|
|
return sig_sync;
|
|
|
|
return pvr_sync_signal_array_add(array, file, handle, point);
|
|
}
|
|
|
|
int
|
|
pvr_sync_signal_array_collect_ops(struct xarray *array,
|
|
struct drm_file *file,
|
|
u32 sync_op_count,
|
|
const struct drm_pvr_sync_op *sync_ops)
|
|
{
|
|
for (u32 i = 0; i < sync_op_count; i++) {
|
|
struct pvr_sync_signal *sig_sync;
|
|
int ret;
|
|
|
|
if (!(sync_ops[i].flags & DRM_PVR_SYNC_OP_FLAG_SIGNAL))
|
|
continue;
|
|
|
|
ret = pvr_check_sync_op(&sync_ops[i]);
|
|
if (ret)
|
|
return ret;
|
|
|
|
sig_sync = pvr_sync_signal_array_get(array, file,
|
|
sync_ops[i].handle,
|
|
sync_ops[i].value);
|
|
if (IS_ERR(sig_sync))
|
|
return PTR_ERR(sig_sync);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
pvr_sync_signal_array_update_fences(struct xarray *array,
|
|
u32 sync_op_count,
|
|
const struct drm_pvr_sync_op *sync_ops,
|
|
struct dma_fence *done_fence)
|
|
{
|
|
for (u32 i = 0; i < sync_op_count; i++) {
|
|
struct dma_fence *old_fence;
|
|
struct pvr_sync_signal *sig_sync;
|
|
|
|
if (!(sync_ops[i].flags & DRM_PVR_SYNC_OP_FLAG_SIGNAL))
|
|
continue;
|
|
|
|
sig_sync = pvr_sync_signal_array_search(array, sync_ops[i].handle,
|
|
sync_ops[i].value);
|
|
if (WARN_ON(!sig_sync))
|
|
return -EINVAL;
|
|
|
|
old_fence = sig_sync->fence;
|
|
sig_sync->fence = dma_fence_get(done_fence);
|
|
dma_fence_put(old_fence);
|
|
|
|
if (WARN_ON(!sig_sync->fence))
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
pvr_sync_signal_array_push_fences(struct xarray *array)
|
|
{
|
|
struct pvr_sync_signal *sig_sync;
|
|
unsigned long i;
|
|
|
|
xa_for_each(array, i, sig_sync) {
|
|
if (sig_sync->chain) {
|
|
drm_syncobj_add_point(sig_sync->syncobj, sig_sync->chain,
|
|
sig_sync->fence, sig_sync->point);
|
|
sig_sync->chain = NULL;
|
|
} else {
|
|
drm_syncobj_replace_fence(sig_sync->syncobj, sig_sync->fence);
|
|
}
|
|
}
|
|
}
|
|
|
|
static int
|
|
pvr_sync_add_dep_to_job(struct drm_sched_job *job, struct dma_fence *f)
|
|
{
|
|
struct dma_fence_unwrap iter;
|
|
u32 native_fence_count = 0;
|
|
struct dma_fence *uf;
|
|
int err = 0;
|
|
|
|
dma_fence_unwrap_for_each(uf, &iter, f) {
|
|
if (pvr_queue_fence_is_ufo_backed(uf))
|
|
native_fence_count++;
|
|
}
|
|
|
|
/* No need to unwrap the fence if it's fully non-native. */
|
|
if (!native_fence_count)
|
|
return drm_sched_job_add_dependency(job, f);
|
|
|
|
dma_fence_unwrap_for_each(uf, &iter, f) {
|
|
/* There's no dma_fence_unwrap_stop() helper cleaning up the refs
|
|
* owned by dma_fence_unwrap(), so let's just iterate over all
|
|
* entries without doing anything when something failed.
|
|
*/
|
|
if (err)
|
|
continue;
|
|
|
|
if (pvr_queue_fence_is_ufo_backed(uf)) {
|
|
struct drm_sched_fence *s_fence = to_drm_sched_fence(uf);
|
|
|
|
/* If this is a native dependency, we wait for the scheduled fence,
|
|
* and we will let pvr_queue_run_job() issue FW waits.
|
|
*/
|
|
err = drm_sched_job_add_dependency(job,
|
|
dma_fence_get(&s_fence->scheduled));
|
|
} else {
|
|
err = drm_sched_job_add_dependency(job, dma_fence_get(uf));
|
|
}
|
|
}
|
|
|
|
dma_fence_put(f);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
pvr_sync_add_deps_to_job(struct pvr_file *pvr_file, struct drm_sched_job *job,
|
|
u32 sync_op_count,
|
|
const struct drm_pvr_sync_op *sync_ops,
|
|
struct xarray *signal_array)
|
|
{
|
|
int err = 0;
|
|
|
|
if (!sync_op_count)
|
|
return 0;
|
|
|
|
for (u32 i = 0; i < sync_op_count; i++) {
|
|
struct pvr_sync_signal *sig_sync;
|
|
struct dma_fence *fence;
|
|
|
|
if (sync_ops[i].flags & DRM_PVR_SYNC_OP_FLAG_SIGNAL)
|
|
continue;
|
|
|
|
err = pvr_check_sync_op(&sync_ops[i]);
|
|
if (err)
|
|
return err;
|
|
|
|
sig_sync = pvr_sync_signal_array_search(signal_array, sync_ops[i].handle,
|
|
sync_ops[i].value);
|
|
if (sig_sync) {
|
|
if (WARN_ON(!sig_sync->fence))
|
|
return -EINVAL;
|
|
|
|
fence = dma_fence_get(sig_sync->fence);
|
|
} else {
|
|
err = drm_syncobj_find_fence(from_pvr_file(pvr_file), sync_ops[i].handle,
|
|
sync_ops[i].value, 0, &fence);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
err = pvr_sync_add_dep_to_job(job, fence);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|