本文的标题来自Linux Kernel 5.6.0-rc4文档, dma-buf作为一个内核子系统,它的使用场景不局限于drm “PRIME“ multi-GPU支持,它主要由3个组件支撑:

  • dma_buf 代表sg_table, 暴露给用户FD
  • dma_fence 通知机制
  • dma_resv 管理共享的或专有的fences

DMA-BUF

dma-buf是Linux内核中在上下文间,进程间,设备间,子系统间进行 buffer 共享的一种实现。 它十几年前就已经合入内核了。

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
commit 3248877ea1796915419fba7c89315fdbf00cb56a
Author: Dave Airlie <airlied@redhat.com>
Date: Fri Nov 25 15:21:02 2011 +0000

drm: base prime/dma-buf support (v5)

This adds the basic drm dma-buf interface layer, called PRIME. This
commit doesn't add any driver support, it is simply and agreed upon starting
point so we can work towards merging driver support for the next merge window.

Current drivers with work done are nouveau, i915, udl, exynos and omap.

The main APIs exposed to userspace allow translating a 32-bit object handle
to a file descriptor, and a file descriptor to a 32-bit object handle.

The flags value is currently limited to O_CLOEXEC.

Acknowledgements:
Daniel Vetter: lots of review
Rob Clark: cleaned up lots of the internals and did lifetime review.

v2: rename some functions after Chris preferred a green shed
fix IS_ERR_OR_NULL -> IS_ERR
v3: Fix Ville pointed out using buffer + kmalloc
v4: add locking as per ickle review
v5: allow re-exporting the original dma-buf (Daniel)

Reviewed-by: Daniel Vetter <daniel.vetter@ffwll.ch>
Reviewed-by: Rob Clark <rob.clark@linaro.org>
Reviewed-by: Sumit Semwal <sumit.semwal@linaro.org>
Reviewed-by: Inki Dae <inki.dae@samsung.com>
Acked-by: Ben Widawsky <benjamin.widawsky@intel.com>
Signed-off-by: Dave Airlie <airlied@redhat.com>
  • exporter调用
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
/**
* drm_gem_prime_handle_to_fd - PRIME export function for GEM drivers
* @dev: dev to export the buffer from
* @file_priv: drm file-private structure
* @handle: buffer handle to export
* @flags: flags like DRM_CLOEXEC
* @prime_fd: pointer to storage for the fd id of the create dma-buf
*
* This is the PRIME export function which must be used mandatorily by GEM
* drivers to ensure correct lifetime management of the underlying GEM object.
* The actual exporting from GEM object to a dma-buf is done through the
* &drm_gem_object_funcs.export callback.
*/
int drm_gem_prime_handle_to_fd(struct drm_device *dev,
struct drm_file *file_priv, uint32_t handle,
uint32_t flags,
int *prime_fd)
  • importer调用
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
/**
* drm_gem_prime_fd_to_handle - PRIME import function for GEM drivers
* @dev: dev to export the buffer from
* @file_priv: drm file-private structure
* @prime_fd: fd id of the dma-buf which should be imported
* @handle: pointer to storage for the handle of the imported buffer object
*
* This is the PRIME import function which must be used mandatorily by GEM
* drivers to ensure correct lifetime management of the underlying GEM object.
* The actual importing of GEM object from the dma-buf is done through the
* &drm_driver.gem_prime_import driver callback.
*
* Returns 0 on success or a negative error code on failure.
*/
int drm_gem_prime_fd_to_handle(struct drm_device *dev,
struct drm_file *file_priv, int prime_fd,
uint32_t *handle)

从上面两个函数的接口看,它们共同涉及3个数据对象:

  • drm_device
  • dma_buf fd
  • drm_gem_object handle

显然两个函数里的dev一定是不同drm_device, dma_buf fd一定是同一个FD(也即同一个dma-buf, 要不然也不叫共享了),那么handle呢?肯定也是不同handle, 因为handle其实是对device而言的,它是一个设备持有的drm_gem_object的ID. 但这个ID背后的东西(backing storage)可能是同一个东西。

可以打个比方,你去银行要办两种业务,两种业务分别排号,假如你要办的A业务排到7号,B业务也刚好排到7号(注意:号码相同,但是两个号),但是很有可能是同一个业务员为你办理这两种业务,这里的业务就是设备驱动,而那个业务员就是dma-buf(或者它封装的那块显存)。

dma_fence

      
signed long
dma_fence_default_wait(struct dma_fence *fence,
                       bool intr,
                       signed long timeout)
{
    struct default_wait_cb cb;
    unsigned long flags;
    signed long ret = timeout ? timeout : 1;

spin_lock_irqsave(fence->lock, flags);

if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT,
    &fence->flags))
    goto out;

if (intr && signal_pending(current)) {
    ret = -ERESTARTSYS;
    goto out;
}

if (!timeout) {
    ret = 0;
    goto out;
}

cb.base.func = dma_fence_default_wait_cb;
cb.task = current;
list_add(&cb.base.node, &fence->cb_list);

while (!test_bit(DMA_FENCE_FLAG_SIGNALED_BIT,
       &fence->flags) && ret > 0) {
    if (intr)
        __set_current_state(TASK_INTERRUPTIBLE);
    else
        __set_current_state(TASK_UNINTERRUPTIBLE);
    spin_unlock_irqrestore(fence->lock, flags);

    ret = schedule_timeout(ret);

    spin_lock_irqsave(fence->lock, flags);
    if (ret > 0 && intr && signal_pending(current))
        ret = -ERESTARTSYS;
}

if (!list_empty(&cb.base.node))
    list_del(&cb.base.node);
__set_current_state(TASK_RUNNING);

out:
spin_unlock_irqrestore(fence->lock, flags);
return ret;
}




fence->ops->wait();

dma_fence_default_wait 是 dma-fence 默认的 wait 操作。该函数会让当前进程(task) 进入睡眠状态 (可中断睡眠或不可中断睡眠,取决于调用者传入的参数 intr), 直到 dma-fence 被 signaled 或者设置的超时时间到。

1
2
3
cb.base.func = dma_fence_default_wait_cb;
cb.task = current;
list_add(&cb.base.node, &fence->cb_list);

dma_resv

dma_buf, dma_fence, dma_resv, 还有drm_gem_object, 这4者有什么关系?

在这里面dma_fence是用来实现drm_gem_object间的共享和同步的,而dma_resv就是所谓的“胶水”,将这两者联系在一起。