1 /* -*- mode: C; c-file-style: "k&r"; tab-width 4; indent-tabs-mode: t; -*- */ 2 3 /* 4 * Copyright (C) 2011 Texas Instruments, Inc 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a 7 * copy of this software and associated documentation files (the "Software"), 8 * to deal in the Software without restriction, including without limitation 9 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 10 * and/or sell copies of the Software, and to permit persons to whom the 11 * Software is furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice (including the next 14 * paragraph) shall be included in all copies or substantial portions of the 15 * Software. 16 * 17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 18 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 19 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 20 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 21 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 22 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 23 * SOFTWARE. 24 * 25 * Authors: 26 * Rob Clark <rob@ti.com> 27 */ 28 29 #ifdef HAVE_CONFIG_H 30 #include "config.h" 31 #endif 32 33 #include <stdlib.h> 34 #include <linux/stddef.h> 35 #include <linux/types.h> 36 #include <errno.h> 37 #include <sys/mman.h> 38 #include <fcntl.h> 39 #include <unistd.h> 40 #include <pthread.h> 41 42 #include <libdrm.h> 43 #include <xf86drm.h> 44 #include <xf86atomic.h> 45 46 #include "omap_drm.h" 47 #include "omap_drmif.h" 48 49 #define __round_mask(x, y) ((__typeof__(x))((y)-1)) 50 #define round_up(x, y) ((((x)-1) | __round_mask(x, y))+1) 51 #define PAGE_SIZE 4096 52 53 static pthread_mutex_t table_lock = PTHREAD_MUTEX_INITIALIZER; 54 static void * dev_table; 55 56 struct omap_device { 57 int fd; 58 atomic_t refcnt; 59 60 /* The handle_table is used to track GEM bo handles associated w/ 61 * this fd. This is needed, in particular, when importing 62 * dmabuf's because we don't want multiple 'struct omap_bo's 63 * floating around with the same handle. Otherwise, when the 64 * first one is omap_bo_del()'d the handle becomes no longer 65 * valid, and the remaining 'struct omap_bo's are left pointing 66 * to an invalid handle (and possible a GEM bo that is already 67 * free'd). 68 */ 69 void *handle_table; 70 }; 71 72 /* a GEM buffer object allocated from the DRM device */ 73 struct omap_bo { 74 struct omap_device *dev; 75 void *map; /* userspace mmap'ing (if there is one) */ 76 uint32_t size; 77 uint32_t handle; 78 uint32_t name; /* flink global handle (DRI2 name) */ 79 uint64_t offset; /* offset to mmap() */ 80 int fd; /* dmabuf handle */ 81 atomic_t refcnt; 82 }; 83 omap_device_new_impl(int fd)84 static struct omap_device * omap_device_new_impl(int fd) 85 { 86 struct omap_device *dev = calloc(sizeof(*dev), 1); 87 if (!dev) 88 return NULL; 89 dev->fd = fd; 90 atomic_set(&dev->refcnt, 1); 91 dev->handle_table = drmHashCreate(); 92 return dev; 93 } 94 omap_device_new(int fd)95 drm_public struct omap_device * omap_device_new(int fd) 96 { 97 struct omap_device *dev = NULL; 98 99 pthread_mutex_lock(&table_lock); 100 101 if (!dev_table) 102 dev_table = drmHashCreate(); 103 104 if (drmHashLookup(dev_table, fd, (void **)&dev)) { 105 /* not found, create new device */ 106 dev = omap_device_new_impl(fd); 107 drmHashInsert(dev_table, fd, dev); 108 } else { 109 /* found, just incr refcnt */ 110 dev = omap_device_ref(dev); 111 } 112 113 pthread_mutex_unlock(&table_lock); 114 115 return dev; 116 } 117 omap_device_ref(struct omap_device * dev)118 drm_public struct omap_device * omap_device_ref(struct omap_device *dev) 119 { 120 atomic_inc(&dev->refcnt); 121 return dev; 122 } 123 omap_device_del(struct omap_device * dev)124 drm_public void omap_device_del(struct omap_device *dev) 125 { 126 if (!atomic_dec_and_test(&dev->refcnt)) 127 return; 128 pthread_mutex_lock(&table_lock); 129 drmHashDestroy(dev->handle_table); 130 drmHashDelete(dev_table, dev->fd); 131 pthread_mutex_unlock(&table_lock); 132 free(dev); 133 } 134 135 drm_public int omap_get_param(struct omap_device * dev,uint64_t param,uint64_t * value)136 omap_get_param(struct omap_device *dev, uint64_t param, uint64_t *value) 137 { 138 struct drm_omap_param req = { 139 .param = param, 140 }; 141 int ret; 142 143 ret = drmCommandWriteRead(dev->fd, DRM_OMAP_GET_PARAM, &req, sizeof(req)); 144 if (ret) { 145 return ret; 146 } 147 148 *value = req.value; 149 150 return 0; 151 } 152 153 drm_public int omap_set_param(struct omap_device * dev,uint64_t param,uint64_t value)154 omap_set_param(struct omap_device *dev, uint64_t param, uint64_t value) 155 { 156 struct drm_omap_param req = { 157 .param = param, 158 .value = value, 159 }; 160 return drmCommandWrite(dev->fd, DRM_OMAP_SET_PARAM, &req, sizeof(req)); 161 } 162 163 /* lookup a buffer from it's handle, call w/ table_lock held: */ lookup_bo(struct omap_device * dev,uint32_t handle)164 static struct omap_bo * lookup_bo(struct omap_device *dev, 165 uint32_t handle) 166 { 167 struct omap_bo *bo = NULL; 168 if (!drmHashLookup(dev->handle_table, handle, (void **)&bo)) { 169 /* found, incr refcnt and return: */ 170 bo = omap_bo_ref(bo); 171 } 172 return bo; 173 } 174 175 /* allocate a new buffer object, call w/ table_lock held */ bo_from_handle(struct omap_device * dev,uint32_t handle)176 static struct omap_bo * bo_from_handle(struct omap_device *dev, 177 uint32_t handle) 178 { 179 struct omap_bo *bo = calloc(sizeof(*bo), 1); 180 if (!bo) { 181 struct drm_gem_close req = { 182 .handle = handle, 183 }; 184 drmIoctl(dev->fd, DRM_IOCTL_GEM_CLOSE, &req); 185 return NULL; 186 } 187 bo->dev = omap_device_ref(dev); 188 bo->handle = handle; 189 atomic_set(&bo->refcnt, 1); 190 /* add ourselves to the handle table: */ 191 drmHashInsert(dev->handle_table, handle, bo); 192 return bo; 193 } 194 195 /* allocate a new buffer object */ omap_bo_new_impl(struct omap_device * dev,union omap_gem_size size,uint32_t flags)196 static struct omap_bo * omap_bo_new_impl(struct omap_device *dev, 197 union omap_gem_size size, uint32_t flags) 198 { 199 struct omap_bo *bo = NULL; 200 struct drm_omap_gem_new req = { 201 .size = size, 202 .flags = flags, 203 }; 204 205 if (size.bytes == 0) { 206 goto fail; 207 } 208 209 if (drmCommandWriteRead(dev->fd, DRM_OMAP_GEM_NEW, &req, sizeof(req))) { 210 goto fail; 211 } 212 213 pthread_mutex_lock(&table_lock); 214 bo = bo_from_handle(dev, req.handle); 215 pthread_mutex_unlock(&table_lock); 216 217 if (flags & OMAP_BO_TILED) { 218 bo->size = round_up(size.tiled.width, PAGE_SIZE) * size.tiled.height; 219 } else { 220 bo->size = size.bytes; 221 } 222 223 return bo; 224 225 fail: 226 free(bo); 227 return NULL; 228 } 229 230 231 /* allocate a new (un-tiled) buffer object */ 232 drm_public struct omap_bo * omap_bo_new(struct omap_device * dev,uint32_t size,uint32_t flags)233 omap_bo_new(struct omap_device *dev, uint32_t size, uint32_t flags) 234 { 235 union omap_gem_size gsize = { 236 .bytes = size, 237 }; 238 if (flags & OMAP_BO_TILED) { 239 return NULL; 240 } 241 return omap_bo_new_impl(dev, gsize, flags); 242 } 243 244 /* allocate a new buffer object */ 245 drm_public struct omap_bo * omap_bo_new_tiled(struct omap_device * dev,uint32_t width,uint32_t height,uint32_t flags)246 omap_bo_new_tiled(struct omap_device *dev, uint32_t width, 247 uint32_t height, uint32_t flags) 248 { 249 union omap_gem_size gsize = { 250 .tiled = { 251 .width = width, 252 .height = height, 253 }, 254 }; 255 if (!(flags & OMAP_BO_TILED)) { 256 return NULL; 257 } 258 return omap_bo_new_impl(dev, gsize, flags); 259 } 260 omap_bo_ref(struct omap_bo * bo)261 drm_public struct omap_bo *omap_bo_ref(struct omap_bo *bo) 262 { 263 atomic_inc(&bo->refcnt); 264 return bo; 265 } 266 267 /* get buffer info */ get_buffer_info(struct omap_bo * bo)268 static int get_buffer_info(struct omap_bo *bo) 269 { 270 struct drm_omap_gem_info req = { 271 .handle = bo->handle, 272 }; 273 int ret = drmCommandWriteRead(bo->dev->fd, DRM_OMAP_GEM_INFO, 274 &req, sizeof(req)); 275 if (ret) { 276 return ret; 277 } 278 279 /* really all we need for now is mmap offset */ 280 bo->offset = req.offset; 281 bo->size = req.size; 282 283 return 0; 284 } 285 286 /* import a buffer object from DRI2 name */ 287 drm_public struct omap_bo * omap_bo_from_name(struct omap_device * dev,uint32_t name)288 omap_bo_from_name(struct omap_device *dev, uint32_t name) 289 { 290 struct omap_bo *bo = NULL; 291 struct drm_gem_open req = { 292 .name = name, 293 }; 294 295 pthread_mutex_lock(&table_lock); 296 297 if (drmIoctl(dev->fd, DRM_IOCTL_GEM_OPEN, &req)) { 298 goto fail; 299 } 300 301 bo = lookup_bo(dev, req.handle); 302 if (!bo) { 303 bo = bo_from_handle(dev, req.handle); 304 bo->name = name; 305 } 306 307 pthread_mutex_unlock(&table_lock); 308 309 return bo; 310 311 fail: 312 pthread_mutex_unlock(&table_lock); 313 free(bo); 314 return NULL; 315 } 316 317 /* import a buffer from dmabuf fd, does not take ownership of the 318 * fd so caller should close() the fd when it is otherwise done 319 * with it (even if it is still using the 'struct omap_bo *') 320 */ 321 drm_public struct omap_bo * omap_bo_from_dmabuf(struct omap_device * dev,int fd)322 omap_bo_from_dmabuf(struct omap_device *dev, int fd) 323 { 324 struct omap_bo *bo = NULL; 325 struct drm_prime_handle req = { 326 .fd = fd, 327 }; 328 int ret; 329 330 pthread_mutex_lock(&table_lock); 331 332 ret = drmIoctl(dev->fd, DRM_IOCTL_PRIME_FD_TO_HANDLE, &req); 333 if (ret) { 334 goto fail; 335 } 336 337 bo = lookup_bo(dev, req.handle); 338 if (!bo) { 339 bo = bo_from_handle(dev, req.handle); 340 } 341 342 pthread_mutex_unlock(&table_lock); 343 344 return bo; 345 346 fail: 347 pthread_mutex_unlock(&table_lock); 348 free(bo); 349 return NULL; 350 } 351 352 /* destroy a buffer object */ omap_bo_del(struct omap_bo * bo)353 drm_public void omap_bo_del(struct omap_bo *bo) 354 { 355 if (!bo) { 356 return; 357 } 358 359 if (!atomic_dec_and_test(&bo->refcnt)) 360 return; 361 362 if (bo->map) { 363 munmap(bo->map, bo->size); 364 } 365 366 if (bo->fd) { 367 close(bo->fd); 368 } 369 370 if (bo->handle) { 371 struct drm_gem_close req = { 372 .handle = bo->handle, 373 }; 374 pthread_mutex_lock(&table_lock); 375 drmHashDelete(bo->dev->handle_table, bo->handle); 376 drmIoctl(bo->dev->fd, DRM_IOCTL_GEM_CLOSE, &req); 377 pthread_mutex_unlock(&table_lock); 378 } 379 380 omap_device_del(bo->dev); 381 382 free(bo); 383 } 384 385 /* get the global flink/DRI2 buffer name */ omap_bo_get_name(struct omap_bo * bo,uint32_t * name)386 drm_public int omap_bo_get_name(struct omap_bo *bo, uint32_t *name) 387 { 388 if (!bo->name) { 389 struct drm_gem_flink req = { 390 .handle = bo->handle, 391 }; 392 int ret; 393 394 ret = drmIoctl(bo->dev->fd, DRM_IOCTL_GEM_FLINK, &req); 395 if (ret) { 396 return ret; 397 } 398 399 bo->name = req.name; 400 } 401 402 *name = bo->name; 403 404 return 0; 405 } 406 omap_bo_handle(struct omap_bo * bo)407 drm_public uint32_t omap_bo_handle(struct omap_bo *bo) 408 { 409 return bo->handle; 410 } 411 412 /* caller owns the dmabuf fd that is returned and is responsible 413 * to close() it when done 414 */ omap_bo_dmabuf(struct omap_bo * bo)415 drm_public int omap_bo_dmabuf(struct omap_bo *bo) 416 { 417 if (!bo->fd) { 418 struct drm_prime_handle req = { 419 .handle = bo->handle, 420 .flags = DRM_CLOEXEC, 421 }; 422 int ret; 423 424 ret = drmIoctl(bo->dev->fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &req); 425 if (ret) { 426 return ret; 427 } 428 429 bo->fd = req.fd; 430 } 431 return dup(bo->fd); 432 } 433 omap_bo_size(struct omap_bo * bo)434 drm_public uint32_t omap_bo_size(struct omap_bo *bo) 435 { 436 if (!bo->size) { 437 get_buffer_info(bo); 438 } 439 return bo->size; 440 } 441 omap_bo_map(struct omap_bo * bo)442 drm_public void *omap_bo_map(struct omap_bo *bo) 443 { 444 if (!bo->map) { 445 if (!bo->offset) { 446 get_buffer_info(bo); 447 } 448 449 bo->map = mmap(0, bo->size, PROT_READ | PROT_WRITE, 450 MAP_SHARED, bo->dev->fd, bo->offset); 451 if (bo->map == MAP_FAILED) { 452 bo->map = NULL; 453 } 454 } 455 return bo->map; 456 } 457 omap_bo_cpu_prep(struct omap_bo * bo,enum omap_gem_op op)458 drm_public int omap_bo_cpu_prep(struct omap_bo *bo, enum omap_gem_op op) 459 { 460 struct drm_omap_gem_cpu_prep req = { 461 .handle = bo->handle, 462 .op = op, 463 }; 464 return drmCommandWrite(bo->dev->fd, 465 DRM_OMAP_GEM_CPU_PREP, &req, sizeof(req)); 466 } 467 omap_bo_cpu_fini(struct omap_bo * bo,enum omap_gem_op op)468 drm_public int omap_bo_cpu_fini(struct omap_bo *bo, enum omap_gem_op op) 469 { 470 struct drm_omap_gem_cpu_fini req = { 471 .handle = bo->handle, 472 .op = op, 473 .nregions = 0, 474 }; 475 return drmCommandWrite(bo->dev->fd, 476 DRM_OMAP_GEM_CPU_FINI, &req, sizeof(req)); 477 } 478