| // SPDX-License-Identifier: GPL-2.0-only | 
 | /* | 
 |  * Copyright (C) 2013-2016 Red Hat | 
 |  * Author: Rob Clark <robdclark@gmail.com> | 
 |  */ | 
 |  | 
 | #include <linux/dma-fence.h> | 
 |  | 
 | #include "msm_drv.h" | 
 | #include "msm_fence.h" | 
 | #include "msm_gpu.h" | 
 |  | 
 | static struct msm_gpu *fctx2gpu(struct msm_fence_context *fctx) | 
 | { | 
 | 	struct msm_drm_private *priv = fctx->dev->dev_private; | 
 | 	return priv->gpu; | 
 | } | 
 |  | 
 | static enum hrtimer_restart deadline_timer(struct hrtimer *t) | 
 | { | 
 | 	struct msm_fence_context *fctx = container_of(t, | 
 | 			struct msm_fence_context, deadline_timer); | 
 |  | 
 | 	kthread_queue_work(fctx2gpu(fctx)->worker, &fctx->deadline_work); | 
 |  | 
 | 	return HRTIMER_NORESTART; | 
 | } | 
 |  | 
 | static void deadline_work(struct kthread_work *work) | 
 | { | 
 | 	struct msm_fence_context *fctx = container_of(work, | 
 | 			struct msm_fence_context, deadline_work); | 
 |  | 
 | 	/* If deadline fence has already passed, nothing to do: */ | 
 | 	if (msm_fence_completed(fctx, fctx->next_deadline_fence)) | 
 | 		return; | 
 |  | 
 | 	msm_devfreq_boost(fctx2gpu(fctx), 2); | 
 | } | 
 |  | 
 |  | 
 | struct msm_fence_context * | 
 | msm_fence_context_alloc(struct drm_device *dev, volatile uint32_t *fenceptr, | 
 | 		const char *name) | 
 | { | 
 | 	struct msm_fence_context *fctx; | 
 | 	static int index = 0; | 
 |  | 
 | 	fctx = kzalloc(sizeof(*fctx), GFP_KERNEL); | 
 | 	if (!fctx) | 
 | 		return ERR_PTR(-ENOMEM); | 
 |  | 
 | 	fctx->dev = dev; | 
 | 	strscpy(fctx->name, name, sizeof(fctx->name)); | 
 | 	fctx->context = dma_fence_context_alloc(1); | 
 | 	fctx->index = index++; | 
 | 	fctx->fenceptr = fenceptr; | 
 | 	spin_lock_init(&fctx->spinlock); | 
 |  | 
 | 	/* | 
 | 	 * Start out close to the 32b fence rollover point, so we can | 
 | 	 * catch bugs with fence comparisons. | 
 | 	 */ | 
 | 	fctx->last_fence = 0xffffff00; | 
 | 	fctx->completed_fence = fctx->last_fence; | 
 | 	*fctx->fenceptr = fctx->last_fence; | 
 |  | 
 | 	hrtimer_init(&fctx->deadline_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); | 
 | 	fctx->deadline_timer.function = deadline_timer; | 
 |  | 
 | 	kthread_init_work(&fctx->deadline_work, deadline_work); | 
 |  | 
 | 	fctx->next_deadline = ktime_get(); | 
 |  | 
 | 	return fctx; | 
 | } | 
 |  | 
 | void msm_fence_context_free(struct msm_fence_context *fctx) | 
 | { | 
 | 	kfree(fctx); | 
 | } | 
 |  | 
 | bool msm_fence_completed(struct msm_fence_context *fctx, uint32_t fence) | 
 | { | 
 | 	/* | 
 | 	 * Note: Check completed_fence first, as fenceptr is in a write-combine | 
 | 	 * mapping, so it will be more expensive to read. | 
 | 	 */ | 
 | 	return (int32_t)(fctx->completed_fence - fence) >= 0 || | 
 | 		(int32_t)(*fctx->fenceptr - fence) >= 0; | 
 | } | 
 |  | 
 | /* called from irq handler and workqueue (in recover path) */ | 
 | void msm_update_fence(struct msm_fence_context *fctx, uint32_t fence) | 
 | { | 
 | 	unsigned long flags; | 
 |  | 
 | 	spin_lock_irqsave(&fctx->spinlock, flags); | 
 | 	if (fence_after(fence, fctx->completed_fence)) | 
 | 		fctx->completed_fence = fence; | 
 | 	if (msm_fence_completed(fctx, fctx->next_deadline_fence)) | 
 | 		hrtimer_cancel(&fctx->deadline_timer); | 
 | 	spin_unlock_irqrestore(&fctx->spinlock, flags); | 
 | } | 
 |  | 
 | struct msm_fence { | 
 | 	struct dma_fence base; | 
 | 	struct msm_fence_context *fctx; | 
 | }; | 
 |  | 
 | static inline struct msm_fence *to_msm_fence(struct dma_fence *fence) | 
 | { | 
 | 	return container_of(fence, struct msm_fence, base); | 
 | } | 
 |  | 
 | static const char *msm_fence_get_driver_name(struct dma_fence *fence) | 
 | { | 
 | 	return "msm"; | 
 | } | 
 |  | 
 | static const char *msm_fence_get_timeline_name(struct dma_fence *fence) | 
 | { | 
 | 	struct msm_fence *f = to_msm_fence(fence); | 
 | 	return f->fctx->name; | 
 | } | 
 |  | 
 | static bool msm_fence_signaled(struct dma_fence *fence) | 
 | { | 
 | 	struct msm_fence *f = to_msm_fence(fence); | 
 | 	return msm_fence_completed(f->fctx, f->base.seqno); | 
 | } | 
 |  | 
 | static void msm_fence_set_deadline(struct dma_fence *fence, ktime_t deadline) | 
 | { | 
 | 	struct msm_fence *f = to_msm_fence(fence); | 
 | 	struct msm_fence_context *fctx = f->fctx; | 
 | 	unsigned long flags; | 
 | 	ktime_t now; | 
 |  | 
 | 	spin_lock_irqsave(&fctx->spinlock, flags); | 
 | 	now = ktime_get(); | 
 |  | 
 | 	if (ktime_after(now, fctx->next_deadline) || | 
 | 			ktime_before(deadline, fctx->next_deadline)) { | 
 | 		fctx->next_deadline = deadline; | 
 | 		fctx->next_deadline_fence = | 
 | 			max(fctx->next_deadline_fence, (uint32_t)fence->seqno); | 
 |  | 
 | 		/* | 
 | 		 * Set timer to trigger boost 3ms before deadline, or | 
 | 		 * if we are already less than 3ms before the deadline | 
 | 		 * schedule boost work immediately. | 
 | 		 */ | 
 | 		deadline = ktime_sub(deadline, ms_to_ktime(3)); | 
 |  | 
 | 		if (ktime_after(now, deadline)) { | 
 | 			kthread_queue_work(fctx2gpu(fctx)->worker, | 
 | 					&fctx->deadline_work); | 
 | 		} else { | 
 | 			hrtimer_start(&fctx->deadline_timer, deadline, | 
 | 					HRTIMER_MODE_ABS); | 
 | 		} | 
 | 	} | 
 |  | 
 | 	spin_unlock_irqrestore(&fctx->spinlock, flags); | 
 | } | 
 |  | 
 | static const struct dma_fence_ops msm_fence_ops = { | 
 | 	.get_driver_name = msm_fence_get_driver_name, | 
 | 	.get_timeline_name = msm_fence_get_timeline_name, | 
 | 	.signaled = msm_fence_signaled, | 
 | 	.set_deadline = msm_fence_set_deadline, | 
 | }; | 
 |  | 
 | struct dma_fence * | 
 | msm_fence_alloc(void) | 
 | { | 
 | 	struct msm_fence *f; | 
 |  | 
 | 	f = kzalloc(sizeof(*f), GFP_KERNEL); | 
 | 	if (!f) | 
 | 		return ERR_PTR(-ENOMEM); | 
 |  | 
 | 	return &f->base; | 
 | } | 
 |  | 
 | void | 
 | msm_fence_init(struct dma_fence *fence, struct msm_fence_context *fctx) | 
 | { | 
 | 	struct msm_fence *f = to_msm_fence(fence); | 
 |  | 
 | 	f->fctx = fctx; | 
 |  | 
 | 	/* | 
 | 	 * Until this point, the fence was just some pre-allocated memory, | 
 | 	 * no-one should have taken a reference to it yet. | 
 | 	 */ | 
 | 	WARN_ON(kref_read(&fence->refcount)); | 
 |  | 
 | 	dma_fence_init(&f->base, &msm_fence_ops, &fctx->spinlock, | 
 | 		       fctx->context, ++fctx->last_fence); | 
 | } |