2 * Copyright 2011 Red Hat Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
22 * Authors: Dave Airlie
28 #include "nouveau_drv.h"
29 #include <nouveau_drm.h>
31 #include <linux/dma-buf.h>
33 static struct sg_table
*nouveau_gem_map_dma_buf(struct dma_buf_attachment
*attachment
,
34 enum dma_data_direction dir
)
36 struct nouveau_bo
*nvbo
= attachment
->dmabuf
->priv
;
37 struct drm_device
*dev
= nvbo
->gem
->dev
;
38 int npages
= nvbo
->bo
.num_pages
;
42 mutex_lock(&dev
->struct_mutex
);
43 sg
= drm_prime_pages_to_sg(nvbo
->bo
.ttm
->pages
, npages
);
44 nents
= dma_map_sg(attachment
->dev
, sg
->sgl
, sg
->nents
, dir
);
45 mutex_unlock(&dev
->struct_mutex
);
49 static void nouveau_gem_unmap_dma_buf(struct dma_buf_attachment
*attachment
,
50 struct sg_table
*sg
, enum dma_data_direction dir
)
52 dma_unmap_sg(attachment
->dev
, sg
->sgl
, sg
->nents
, dir
);
57 static void nouveau_gem_dmabuf_release(struct dma_buf
*dma_buf
)
59 struct nouveau_bo
*nvbo
= dma_buf
->priv
;
61 if (nvbo
->gem
->export_dma_buf
== dma_buf
) {
62 nvbo
->gem
->export_dma_buf
= NULL
;
63 drm_gem_object_unreference_unlocked(nvbo
->gem
);
67 static void *nouveau_gem_kmap_atomic(struct dma_buf
*dma_buf
, unsigned long page_num
)
72 static void nouveau_gem_kunmap_atomic(struct dma_buf
*dma_buf
, unsigned long page_num
, void *addr
)
76 static void *nouveau_gem_kmap(struct dma_buf
*dma_buf
, unsigned long page_num
)
81 static void nouveau_gem_kunmap(struct dma_buf
*dma_buf
, unsigned long page_num
, void *addr
)
86 static int nouveau_gem_prime_mmap(struct dma_buf
*dma_buf
, struct vm_area_struct
*vma
)
91 static void *nouveau_gem_prime_vmap(struct dma_buf
*dma_buf
)
93 struct nouveau_bo
*nvbo
= dma_buf
->priv
;
94 struct drm_device
*dev
= nvbo
->gem
->dev
;
97 mutex_lock(&dev
->struct_mutex
);
98 if (nvbo
->vmapping_count
) {
99 nvbo
->vmapping_count
++;
103 ret
= ttm_bo_kmap(&nvbo
->bo
, 0, nvbo
->bo
.num_pages
,
104 &nvbo
->dma_buf_vmap
);
106 mutex_unlock(&dev
->struct_mutex
);
109 nvbo
->vmapping_count
= 1;
111 mutex_unlock(&dev
->struct_mutex
);
112 return nvbo
->dma_buf_vmap
.virtual;
115 static void nouveau_gem_prime_vunmap(struct dma_buf
*dma_buf
, void *vaddr
)
117 struct nouveau_bo
*nvbo
= dma_buf
->priv
;
118 struct drm_device
*dev
= nvbo
->gem
->dev
;
120 mutex_lock(&dev
->struct_mutex
);
121 nvbo
->vmapping_count
--;
122 if (nvbo
->vmapping_count
== 0) {
123 ttm_bo_kunmap(&nvbo
->dma_buf_vmap
);
125 mutex_unlock(&dev
->struct_mutex
);
128 static const struct dma_buf_ops nouveau_dmabuf_ops
= {
129 .map_dma_buf
= nouveau_gem_map_dma_buf
,
130 .unmap_dma_buf
= nouveau_gem_unmap_dma_buf
,
131 .release
= nouveau_gem_dmabuf_release
,
132 .kmap
= nouveau_gem_kmap
,
133 .kmap_atomic
= nouveau_gem_kmap_atomic
,
134 .kunmap
= nouveau_gem_kunmap
,
135 .kunmap_atomic
= nouveau_gem_kunmap_atomic
,
136 .mmap
= nouveau_gem_prime_mmap
,
137 .vmap
= nouveau_gem_prime_vmap
,
138 .vunmap
= nouveau_gem_prime_vunmap
,
142 nouveau_prime_new(struct drm_device
*dev
,
145 struct nouveau_bo
**pnvbo
)
147 struct nouveau_bo
*nvbo
;
151 flags
= TTM_PL_FLAG_TT
;
153 ret
= nouveau_bo_new(dev
, size
, 0, flags
, 0, 0,
159 /* we restrict allowed domains on nv50+ to only the types
160 * that were requested at creation time. not possibly on
161 * earlier chips without busting the ABI.
163 nvbo
->valid_domains
= NOUVEAU_GEM_DOMAIN_GART
;
164 nvbo
->gem
= drm_gem_object_alloc(dev
, nvbo
->bo
.mem
.size
);
166 nouveau_bo_ref(NULL
, pnvbo
);
170 nvbo
->gem
->driver_private
= nvbo
;
174 struct dma_buf
*nouveau_gem_prime_export(struct drm_device
*dev
,
175 struct drm_gem_object
*obj
, int flags
)
177 struct nouveau_bo
*nvbo
= nouveau_gem_object(obj
);
180 /* pin buffer into GTT */
181 ret
= nouveau_bo_pin(nvbo
, TTM_PL_FLAG_TT
);
183 return ERR_PTR(-EINVAL
);
185 return dma_buf_export(nvbo
, &nouveau_dmabuf_ops
, obj
->size
, flags
);
188 struct drm_gem_object
*nouveau_gem_prime_import(struct drm_device
*dev
,
189 struct dma_buf
*dma_buf
)
191 struct dma_buf_attachment
*attach
;
193 struct nouveau_bo
*nvbo
;
196 if (dma_buf
->ops
== &nouveau_dmabuf_ops
) {
197 nvbo
= dma_buf
->priv
;
199 if (nvbo
->gem
->dev
== dev
) {
200 drm_gem_object_reference(nvbo
->gem
);
206 attach
= dma_buf_attach(dma_buf
, dev
->dev
);
208 return ERR_PTR(PTR_ERR(attach
));
210 sg
= dma_buf_map_attachment(attach
, DMA_BIDIRECTIONAL
);
216 ret
= nouveau_prime_new(dev
, dma_buf
->size
, sg
, &nvbo
);
220 nvbo
->gem
->import_attach
= attach
;
225 dma_buf_unmap_attachment(attach
, sg
, DMA_BIDIRECTIONAL
);
227 dma_buf_detach(dma_buf
, attach
);