void i915_gem_retire_requests(struct drm_device *dev);
void i915_gem_reset_lists(struct drm_device *dev);
void i915_gem_clflush_object(struct drm_gem_object *obj);
-void i915_gem_flush_ring(struct drm_device *dev,
- struct drm_file *file_priv,
- struct intel_ring_buffer *ring,
- uint32_t invalidate_domains,
- uint32_t flush_domains);
int i915_gem_object_set_domain(struct drm_gem_object *obj,
uint32_t read_domains,
uint32_t write_domain);
return i915_do_wait_request(dev, seqno, 1, ring);
}
-void
+static void
i915_gem_flush_ring(struct drm_device *dev,
struct drm_file *file_priv,
struct intel_ring_buffer *ring,
struct intel_unpin_work *work;
unsigned long flags, offset;
int pipe = intel_crtc->pipe;
- u32 was_dirty, pf, pipesrc;
+ u32 pf, pipesrc;
int ret;
work = kzalloc(sizeof *work, GFP_KERNEL);
obj = intel_fb->obj;
mutex_lock(&dev->struct_mutex);
- was_dirty = obj->write_domain & I915_GEM_GPU_DOMAINS;
ret = intel_pin_and_fence_fb_obj(dev, obj, true);
if (ret)
goto cleanup_work;
atomic_inc(&obj_priv->pending_flip);
work->pending_flip_obj = obj;
- /* Schedule the pipelined flush */
- if (was_dirty)
- i915_gem_flush_ring(dev, NULL, obj_priv->ring, 0, was_dirty);
-
if (IS_GEN3(dev) || IS_GEN2(dev)) {
u32 flip_mask;