]> git.proxmox.com Git - mirror_ubuntu-kernels.git/commitdiff
drm/amdgpu: always enable EOP interrupt v2
authorChunming Zhou <david1.zhou@amd.com>
Mon, 1 Jun 2015 06:14:32 +0000 (14:14 +0800)
committerAlex Deucher <alexander.deucher@amd.com>
Mon, 17 Aug 2015 20:50:17 +0000 (16:50 -0400)
v2 (chk): always enable EOP interrupt, independent of scheduler,
  remove now unused delayed_irq handling.

Signed-off-by: Chunming Zhou <david1.zhou@amd.com>
Signed-off-by: Christian König <christian.koenig@amd.com>
Reviewed-by: Jammy Zhou <Jammy.Zhou@amd.com>
drivers/gpu/drm/amd/amdgpu/amdgpu.h
drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c

index 9f47b3e013c7e3a01514d68e1646ac3e427227d6..1ec89d2864dd2ed3415f363c0c3a029104686619 100644 (file)
@@ -383,7 +383,6 @@ struct amdgpu_fence_driver {
        uint64_t                        sync_seq[AMDGPU_MAX_RINGS];
        atomic64_t                      last_seq;
        bool                            initialized;
-       bool                            delayed_irq;
        struct amdgpu_irq_src           *irq_src;
        unsigned                        irq_type;
        struct delayed_work             lockup_work;
index e3629c923862cb3130238d179875cb6632cfa0cb..b89dafec9ecf54b14589988c9ebe44efb80e33d6 100644 (file)
@@ -164,8 +164,6 @@ static int amdgpu_fence_check_signaled(wait_queue_t *wait, unsigned mode, int fl
                else
                        FENCE_TRACE(&fence->base, "was already signaled\n");
 
-               amdgpu_irq_put(adev, fence->ring->fence_drv.irq_src,
-                               fence->ring->fence_drv.irq_type);
                __remove_wait_queue(&adev->fence_queue, &fence->fence_wake);
                fence_put(&fence->base);
        } else
@@ -267,12 +265,6 @@ static void amdgpu_fence_check_lockup(struct work_struct *work)
                return;
        }
 
-       if (fence_drv->delayed_irq && ring->adev->ddev->irq_enabled) {
-               fence_drv->delayed_irq = false;
-               amdgpu_irq_update(ring->adev, fence_drv->irq_src,
-                               fence_drv->irq_type);
-       }
-
        if (amdgpu_fence_activity(ring))
                wake_up_all(&ring->adev->fence_queue);
        else if (amdgpu_ring_is_lockup(ring)) {
@@ -420,29 +412,6 @@ static bool amdgpu_fence_enable_signaling(struct fence *f)
        if (atomic64_read(&ring->fence_drv.last_seq) >= fence->seq)
                return false;
 
-       if (down_read_trylock(&adev->exclusive_lock)) {
-               amdgpu_irq_get(adev, ring->fence_drv.irq_src,
-                       ring->fence_drv.irq_type);
-               if (amdgpu_fence_activity(ring))
-                       wake_up_all_locked(&adev->fence_queue);
-
-               /* did fence get signaled after we enabled the sw irq? */
-               if (atomic64_read(&ring->fence_drv.last_seq) >= fence->seq) {
-                       amdgpu_irq_put(adev, ring->fence_drv.irq_src,
-                               ring->fence_drv.irq_type);
-                       up_read(&adev->exclusive_lock);
-                       return false;
-               }
-
-               up_read(&adev->exclusive_lock);
-       } else {
-               /* we're probably in a lockup, lets not fiddle too much */
-               if (amdgpu_irq_get_delayed(adev, ring->fence_drv.irq_src,
-                       ring->fence_drv.irq_type))
-                       ring->fence_drv.delayed_irq = true;
-               amdgpu_fence_schedule_check(ring);
-       }
-
        fence->fence_wake.flags = 0;
        fence->fence_wake.private = NULL;
        fence->fence_wake.func = amdgpu_fence_check_signaled;
@@ -541,8 +510,6 @@ static long amdgpu_fence_wait_seq_timeout(struct amdgpu_device *adev,
 
                        last_seq[i] = atomic64_read(&ring->fence_drv.last_seq);
                        trace_amdgpu_fence_wait_begin(adev->ddev, i, target_seq[i]);
-                       amdgpu_irq_get(adev, ring->fence_drv.irq_src,
-                                      ring->fence_drv.irq_type);
                }
 
                if (intr) {
@@ -561,8 +528,6 @@ static long amdgpu_fence_wait_seq_timeout(struct amdgpu_device *adev,
                        if (!ring || !target_seq[i])
                                continue;
 
-                       amdgpu_irq_put(adev, ring->fence_drv.irq_src,
-                                      ring->fence_drv.irq_type);
                        trace_amdgpu_fence_wait_end(adev->ddev, i, target_seq[i]);
                }
 
@@ -901,9 +866,12 @@ int amdgpu_fence_driver_start_ring(struct amdgpu_ring *ring,
                ring->fence_drv.gpu_addr = adev->uvd.gpu_addr + index;
        }
        amdgpu_fence_write(ring, atomic64_read(&ring->fence_drv.last_seq));
-       ring->fence_drv.initialized = true;
+       amdgpu_irq_get(adev, irq_src, irq_type);
+
        ring->fence_drv.irq_src = irq_src;
        ring->fence_drv.irq_type = irq_type;
+       ring->fence_drv.initialized = true;
+
        dev_info(adev->dev, "fence driver on ring %d use gpu addr 0x%016llx, "
                 "cpu addr 0x%p\n", ring->idx,
                 ring->fence_drv.gpu_addr, ring->fence_drv.cpu_addr);
@@ -980,6 +948,8 @@ void amdgpu_fence_driver_fini(struct amdgpu_device *adev)
                        amdgpu_fence_driver_force_completion(adev);
                }
                wake_up_all(&adev->fence_queue);
+               amdgpu_irq_put(adev, ring->fence_drv.irq_src,
+                              ring->fence_drv.irq_type);
                ring->fence_drv.initialized = false;
        }
        mutex_unlock(&adev->ring_lock);