drm/i915/dmabuf: Tighten struct_mutex for unmap_dma_buf
[deliverable/linux.git] / drivers / gpu / drm / i915 / i915_gem_dmabuf.c
index e9c2bfd85b5268425ce9465ee2d23d99ee92c534..616f07830b7e166332676ecedda5069629e6a21c 100644 (file)
@@ -95,14 +95,12 @@ static void i915_gem_unmap_dma_buf(struct dma_buf_attachment *attachment,
 {
        struct drm_i915_gem_object *obj = dma_buf_to_obj(attachment->dmabuf);
 
-       mutex_lock(&obj->base.dev->struct_mutex);
-
        dma_unmap_sg(attachment->dev, sg->sgl, sg->nents, dir);
        sg_free_table(sg);
        kfree(sg);
 
+       mutex_lock(&obj->base.dev->struct_mutex);
        i915_gem_object_unpin_pages(obj);
-
        mutex_unlock(&obj->base.dev->struct_mutex);
 }
 
@@ -193,10 +191,26 @@ static void i915_gem_dmabuf_kunmap(struct dma_buf *dma_buf, unsigned long page_n
 
 static int i915_gem_dmabuf_mmap(struct dma_buf *dma_buf, struct vm_area_struct *vma)
 {
-       return -EINVAL;
+       struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
+       int ret;
+
+       if (obj->base.size < vma->vm_end - vma->vm_start)
+               return -EINVAL;
+
+       if (!obj->base.filp)
+               return -ENODEV;
+
+       ret = obj->base.filp->f_op->mmap(obj->base.filp, vma);
+       if (ret)
+               return ret;
+
+       fput(vma->vm_file);
+       vma->vm_file = get_file(obj->base.filp);
+
+       return 0;
 }
 
-static int i915_gem_begin_cpu_access(struct dma_buf *dma_buf, size_t start, size_t length, enum dma_data_direction direction)
+static int i915_gem_begin_cpu_access(struct dma_buf *dma_buf, enum dma_data_direction direction)
 {
        struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
        struct drm_device *dev = obj->base.dev;
@@ -212,6 +226,27 @@ static int i915_gem_begin_cpu_access(struct dma_buf *dma_buf, size_t start, size
        return ret;
 }
 
+static void i915_gem_end_cpu_access(struct dma_buf *dma_buf, enum dma_data_direction direction)
+{
+       struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
+       struct drm_device *dev = obj->base.dev;
+       struct drm_i915_private *dev_priv = to_i915(dev);
+       bool was_interruptible;
+       int ret;
+
+       mutex_lock(&dev->struct_mutex);
+       was_interruptible = dev_priv->mm.interruptible;
+       dev_priv->mm.interruptible = false;
+
+       ret = i915_gem_object_set_to_gtt_domain(obj, false);
+
+       dev_priv->mm.interruptible = was_interruptible;
+       mutex_unlock(&dev->struct_mutex);
+
+       if (unlikely(ret))
+               DRM_ERROR("unable to flush buffer following CPU access; rendering may be corrupt\n");
+}
+
 static const struct dma_buf_ops i915_dmabuf_ops =  {
        .map_dma_buf = i915_gem_map_dma_buf,
        .unmap_dma_buf = i915_gem_unmap_dma_buf,
@@ -224,6 +259,7 @@ static const struct dma_buf_ops i915_dmabuf_ops =  {
        .vmap = i915_gem_dmabuf_vmap,
        .vunmap = i915_gem_dmabuf_vunmap,
        .begin_cpu_access = i915_gem_begin_cpu_access,
+       .end_cpu_access = i915_gem_end_cpu_access,
 };
 
 struct dma_buf *i915_gem_prime_export(struct drm_device *dev,
This page took 0.026567 seconds and 5 git commands to generate.