On Wed, Jun 20, 2018 at 08:55:39PM +0200, Mark Kettenis wrote: > This diff adds DRI3/prime support to the OpenBSD kernel. The whole > idea behind this is to allow graphics buffers to be exchanged between > processes (typically X client and X server) by passing file > descriptors. The curremt (DRI2) method of exchanging graphics buffers > uses guessable 32-bit integers. That means that it is fairly easy for > other applications running on the machine to snoop the contents of a > graphics buffer. Apart from the improved security, this also gets us > to execute the code paths in X, Mesa, etc. that Linux uses. Hopefully > that makes us less susceptible to bugs in the abandoned codepaths. > > This diff explicitly does not enable sharing of graphics buffers > between graphics cards. That aspect of DRI3 needs more work. It > would be pretty pointless at this point anyway since this only enables > the feature for inteldrm(4). > > To actually use this, a xenocara patch is needed as well. I'll post > that one in a separate message. > > I've used this quite extensively now on my own laptop for a while. It > works fine with pledged Chrome and pledged Firefox. Now this collides > heavily with the file descriptor unlocking work that mpi@ is doing. > Still, I'd like to get this in soonish such that this can be widely > tested and won't ruin g2k18. > > ok?
So the radeon parts (radeon_prime.c etc) can come later as when DRIVER_PRIME/prime_fd_to_handle/prime_handle_to_fd etc are not set everything will fall back? Though it seems DRIVER_PRIME is already set in radeon... > > > Index: dev/pci/drm/drmP.h > =================================================================== > RCS file: /cvs/src/sys/dev/pci/drm/drmP.h,v > retrieving revision 1.217 > diff -u -p -r1.217 drmP.h > --- dev/pci/drm/drmP.h 19 Feb 2018 08:59:52 -0000 1.217 > +++ dev/pci/drm/drmP.h 20 Jun 2018 18:41:39 -0000 > @@ -352,6 +352,12 @@ struct drm_pending_event { > void (*destroy)(struct drm_pending_event *event); > }; > > +/* initial implementaton using a linked list - todo hashtab */ > +struct drm_prime_file_private { > + struct list_head head; > + struct rwlock lock; > +}; > + > /** File private data */ > struct drm_file { > unsigned always_authenticated :1; > @@ -395,6 +401,8 @@ struct drm_file { > struct list_head event_list; > int event_space; > > + struct drm_prime_file_private prime; > + > struct selinfo rsel; > SPLAY_ENTRY(drm_file) link; > }; > @@ -480,6 +488,34 @@ struct drm_gem_object { > uint32_t pending_read_domains; > uint32_t pending_write_domain; > > + /** > + * dma_buf - dma buf associated with this GEM object > + * > + * Pointer to the dma-buf associated with this gem object (either > + * through importing or exporting). We break the resulting reference > + * loop when the last gem handle for this object is released. > + * > + * Protected by obj->object_name_lock > + */ > + struct dma_buf *dma_buf; > + > + /** > + * import_attach - dma buf attachment backing this object > + * > + * Any foreign dma_buf imported as a gem object has this set to the > + * attachment point for the device. This is invariant over the lifetime > + * of a gem object. > + * > + * The driver's ->gem_free_object callback is responsible for cleaning > + * up the dma_buf attachment and references acquired at import time. > + * > + * Note that the drm gem/prime core does not depend upon drivers setting > + * this field any more. So for drivers where this doesn't make sense > + * (e.g. virtual devices or a displaylink behind an usb bus) they can > + * simply leave it as NULL. > + */ > + struct dma_buf_attachment *import_attach; > + > struct uvm_object uobj; > SPLAY_ENTRY(drm_gem_object) entry; > struct uvm_object *uao; > @@ -646,6 +682,20 @@ struct drm_driver { > int (*gem_fault)(struct drm_gem_object *, struct uvm_faultinfo *, > off_t, vaddr_t, vm_page_t *, int, int, vm_prot_t, int); > > + /* prime: */ > + /* export handle -> fd (see drm_gem_prime_handle_to_fd() helper) */ > + int (*prime_handle_to_fd)(struct drm_device *dev, struct drm_file > *file_priv, > + uint32_t handle, uint32_t flags, int *prime_fd); > + /* import fd -> handle (see drm_gem_prime_fd_to_handle() helper) */ > + int (*prime_fd_to_handle)(struct drm_device *dev, struct drm_file > *file_priv, > + int prime_fd, uint32_t *handle); > + /* export GEM -> dmabuf */ > + struct dma_buf * (*gem_prime_export)(struct drm_device *dev, > + struct drm_gem_object *obj, int flags); > + /* import dmabuf -> GEM */ > + struct drm_gem_object * (*gem_prime_import)(struct drm_device *dev, > + struct dma_buf *dma_buf); > + > int (*dumb_create)(struct drm_file *file_priv, > struct drm_device *dev, struct drm_mode_create_dumb *args); > int (*dumb_map_offset)(struct drm_file *file_priv, > @@ -981,6 +1031,13 @@ static inline wait_queue_head_t *drm_crt > /* Modesetting support */ > extern void drm_vblank_pre_modeset(struct drm_device *dev, unsigned int > pipe); > extern void drm_vblank_post_modeset(struct drm_device *dev, unsigned int > pipe); > + > +extern int drm_gem_prime_handle_to_fd(struct drm_device *dev, > + struct drm_file *file_priv, uint32_t handle, uint32_t flags, > + int *prime_fd); > +extern int drm_gem_prime_fd_to_handle(struct drm_device *dev, > + struct drm_file *file_priv, int prime_fd, uint32_t *handle); > +extern void drm_gem_dmabuf_release(struct dma_buf *dma_buf); > > bool drm_mode_parse_command_line_for_connector(const char *, > struct drm_connector *, struct drm_cmdline_mode *); > Index: dev/pci/drm/drm_drv.c > =================================================================== > RCS file: /cvs/src/sys/dev/pci/drm/drm_drv.c,v > retrieving revision 1.157 > diff -u -p -r1.157 drm_drv.c > --- dev/pci/drm/drm_drv.c 31 Jan 2018 05:04:41 -0000 1.157 > +++ dev/pci/drm/drm_drv.c 20 Jun 2018 18:41:39 -0000 > @@ -207,10 +207,8 @@ static struct drm_ioctl_desc drm_ioctls[ > > DRM_IOCTL_DEF(DRM_IOCTL_MODE_GETRESOURCES, drm_mode_getresources, > DRM_CONTROL_ALLOW|DRM_UNLOCKED), > > -#ifdef notyet > DRM_IOCTL_DEF(DRM_IOCTL_PRIME_HANDLE_TO_FD, > drm_prime_handle_to_fd_ioctl, DRM_AUTH|DRM_UNLOCKED|DRM_RENDER_ALLOW), > DRM_IOCTL_DEF(DRM_IOCTL_PRIME_FD_TO_HANDLE, > drm_prime_fd_to_handle_ioctl, DRM_AUTH|DRM_UNLOCKED|DRM_RENDER_ALLOW), > -#endif > > DRM_IOCTL_DEF(DRM_IOCTL_MODE_GETPLANERESOURCES, drm_mode_getplane_res, > DRM_CONTROL_ALLOW|DRM_UNLOCKED), > DRM_IOCTL_DEF(DRM_IOCTL_MODE_GETCRTC, drm_mode_getcrtc, > DRM_CONTROL_ALLOW|DRM_UNLOCKED), > @@ -736,6 +734,9 @@ drmopen(dev_t kdev, int flags, int fmt, > if (dev->driver->driver_features & DRIVER_GEM) > drm_gem_open(dev, file_priv); > > + if (drm_core_check_feature(dev, DRIVER_PRIME)) > + drm_prime_init_file_private(&file_priv->prime); > + > if (dev->driver->open) { > ret = dev->driver->open(dev, file_priv); > if (ret != 0) { > @@ -829,6 +830,10 @@ drmclose(dev_t kdev, int flags, int fmt, > if (dev->driver->postclose) > dev->driver->postclose(dev, file_priv); > > + > + if (drm_core_check_feature(dev, DRIVER_PRIME)) > + drm_prime_destroy_file_private(&file_priv->prime); > + > SPLAY_REMOVE(drm_file_tree, &dev->files, file_priv); > drm_free(file_priv); > > @@ -1116,12 +1121,10 @@ drm_getcap(struct drm_device *dev, void > case DRM_CAP_DUMB_PREFER_SHADOW: > req->value = dev->mode_config.prefer_shadow; > break; > -#ifdef notyet > case DRM_CAP_PRIME: > req->value |= dev->driver->prime_fd_to_handle ? > DRM_PRIME_CAP_IMPORT : 0; > req->value |= dev->driver->prime_handle_to_fd ? > DRM_PRIME_CAP_EXPORT : 0; > break; > -#endif > case DRM_CAP_TIMESTAMP_MONOTONIC: > req->value = drm_timestamp_monotonic; > break; > Index: dev/pci/drm/drm_gem.c > =================================================================== > RCS file: /cvs/src/sys/dev/pci/drm/drm_gem.c,v > retrieving revision 1.5 > diff -u -p -r1.5 drm_gem.c > --- dev/pci/drm/drm_gem.c 3 Sep 2017 13:28:54 -0000 1.5 > +++ dev/pci/drm/drm_gem.c 20 Jun 2018 18:41:39 -0000 > @@ -28,6 +28,7 @@ > > #include <dev/pci/drm/drmP.h> > #include <dev/pci/drm/drm_vma_manager.h> > +#include "drm_internal.h" > > #include <uvm/uvm.h> > > @@ -318,7 +319,6 @@ EXPORT_SYMBOL(drm_gem_private_object_ini > static void > drm_gem_remove_prime_handles(struct drm_gem_object *obj, struct drm_file > *filp) > { > -#ifdef __linux__ > /* > * Note: obj->dma_buf can't disappear as long as we still hold a > * handle reference in obj->handle_count. > @@ -329,7 +329,6 @@ drm_gem_remove_prime_handles(struct drm_ > obj->dma_buf); > } > mutex_unlock(&filp->prime.lock); > -#endif > } > > /** > @@ -352,13 +351,11 @@ static void drm_gem_object_handle_free(s > > static void drm_gem_object_exported_dma_buf_free(struct drm_gem_object *obj) > { > -#ifdef __linux__ > /* Unbreak the reference cycle if we have an exported dma_buf. */ > if (obj->dma_buf) { > dma_buf_put(obj->dma_buf); > obj->dma_buf = NULL; > } > -#endif > } > > static void > @@ -872,6 +869,8 @@ void > drm_gem_object_release(struct drm_gem_object *obj) > { > struct drm_device *dev = obj->dev; > + > + WARN_ON(obj->dma_buf); > > if (obj->uao) > uao_detach(obj->uao); > Index: dev/pci/drm/drm_internal.h > =================================================================== > RCS file: /cvs/src/sys/dev/pci/drm/drm_internal.h,v > retrieving revision 1.3 > diff -u -p -r1.3 drm_internal.h > --- dev/pci/drm/drm_internal.h 1 Jul 2017 16:12:03 -0000 1.3 > +++ dev/pci/drm/drm_internal.h 20 Jun 2018 18:41:39 -0000 > @@ -45,7 +45,6 @@ void drm_vm_close_locked(struct drm_devi > #endif > > /* drm_prime.c */ > -#ifdef __linux__ > int drm_prime_handle_to_fd_ioctl(struct drm_device *dev, void *data, > struct drm_file *file_priv); > int drm_prime_fd_to_handle_ioctl(struct drm_device *dev, void *data, > @@ -55,7 +54,6 @@ void drm_prime_init_file_private(struct > void drm_prime_destroy_file_private(struct drm_prime_file_private > *prime_fpriv); > void drm_prime_remove_buf_handle_locked(struct drm_prime_file_private > *prime_fpriv, > struct dma_buf *dma_buf); > -#endif > > /* drm_info.c */ > #ifdef __linux__ > Index: dev/pci/drm/drm_linux.c > =================================================================== > RCS file: /cvs/src/sys/dev/pci/drm/drm_linux.c,v > retrieving revision 1.23 > diff -u -p -r1.23 drm_linux.c > --- dev/pci/drm/drm_linux.c 25 Apr 2018 01:27:46 -0000 1.23 > +++ dev/pci/drm/drm_linux.c 20 Jun 2018 18:41:39 -0000 > @@ -19,6 +19,9 @@ > #include <dev/pci/drm/drmP.h> > #include <dev/pci/ppbreg.h> > #include <sys/event.h> > +#include <sys/file.h> > +#include <sys/filedesc.h> > +#include <sys/stat.h> > > struct mutex sch_mtx = MUTEX_INITIALIZER(IPL_SCHED); > void *sch_ident; > @@ -802,4 +805,154 @@ unsigned int > fence_context_alloc(unsigned int num) > { > return __sync_add_and_fetch(&drm_fence_count, num) - num; > +} > + > +int > +dmabuf_read(struct file *fp, off_t *poff, struct uio *uio, struct ucred > *cred) > +{ > + return (ENXIO); > +} > + > +int > +dmabuf_write(struct file *fp, off_t *poff, struct uio *uio, struct ucred > *cred) > +{ > + return (ENXIO); > +} > + > +int > +dmabuf_ioctl(struct file *fp, u_long com, caddr_t data, struct proc *p) > +{ > + return (ENOTTY); > +} > + > +int > +dmabuf_poll(struct file *fp, int events, struct proc *p) > +{ > + return (0); > +} > + > +int > +dmabuf_kqfilter(struct file *fp, struct knote *kn) > +{ > + return (EINVAL); > +} > + > +int > +dmabuf_stat(struct file *fp, struct stat *st, struct proc *p) > +{ > + struct dma_buf *dmabuf = fp->f_data; > + > + memset(st, 0, sizeof(*st)); > + st->st_size = dmabuf->size; > + st->st_mode = S_IFIFO; /* XXX */ > + return (0); > +} > + > +int > +dmabuf_close(struct file *fp, struct proc *p) > +{ > + struct dma_buf *dmabuf = fp->f_data; > + > + fp->f_data = NULL; > + dmabuf->ops->release(dmabuf); > + free(dmabuf, M_DRM, sizeof(struct dma_buf)); > + return (0); > +} > + > +struct fileops dmabufops = { > + .fo_read = dmabuf_read, > + .fo_write = dmabuf_write, > + .fo_ioctl = dmabuf_ioctl, > + .fo_poll = dmabuf_poll, > + .fo_kqfilter = dmabuf_kqfilter, > + .fo_stat = dmabuf_stat, > + .fo_close = dmabuf_close > +}; > + > +struct dma_buf * > +dma_buf_export(const struct dma_buf_export_info *info) > +{ > + struct proc *p = curproc; > + struct filedesc *fdp = p->p_fd; > + struct dma_buf *dmabuf; > + struct file *fp; > + int fd, error; > + > + fdplock(fdp); > + error = falloc(p, &fp, &fd); > + if (error) { > + fdpunlock(fdp); > + return ERR_PTR(-error); > + } > + fdremove(fdp, fd); > + fdpunlock(fdp); > + fp->f_type = DTYPE_DMABUF; > + fp->f_ops = &dmabufops; > + dmabuf = malloc(sizeof(struct dma_buf), M_DRM, M_WAITOK | M_ZERO); > + dmabuf->priv = info->priv; > + dmabuf->ops = info->ops; > + dmabuf->size = info->size; > + dmabuf->file = fp; > + fp->f_data = dmabuf; > + return dmabuf; > +} > + > +struct dma_buf * > +dma_buf_get(int fd) > +{ > + struct proc *p = curproc; > + struct filedesc *fdp = p->p_fd; > + struct file *fp; > + > + if ((fp = fd_getfile(fdp, fd)) == NULL) > + return ERR_PTR(-EBADF); > + > + if (fp->f_type != DTYPE_DMABUF) { > + FRELE(fp, p); > + return ERR_PTR(-EINVAL); > + } > + > + return fp->f_data; > +} > + > +void > +dma_buf_put(struct dma_buf *dmabuf) > +{ > + KASSERT(dmabuf); > + KASSERT(dmabuf->file); > + > + FRELE(dmabuf->file, curproc); > +} > + > +int > +dma_buf_fd(struct dma_buf *dmabuf, int flags) > +{ > + struct proc *p = curproc; > + struct filedesc *fdp = p->p_fd; > + struct file *fp = dmabuf->file; > + int fd, cloexec, error; > + > + cloexec = (flags & O_CLOEXEC) ? UF_EXCLOSE : 0; > + > + fdplock(fdp); > +restart: > + if ((error = fdalloc(p, 0, &fd)) != 0) { > + if (error == ENOSPC) { > + fdexpand(p); > + goto restart; > + } > + fdpunlock(fdp); > + return -error; > + } > + > + fdinsert(fdp, fd, cloexec, fp); > + fdpunlock(fdp); > + > + return fd; > +} > + > +void > +get_dma_buf(struct dma_buf *dmabuf) > +{ > + FREF(dmabuf->file); > } > Index: dev/pci/drm/drm_linux.h > =================================================================== > RCS file: /cvs/src/sys/dev/pci/drm/drm_linux.h,v > retrieving revision 1.88 > diff -u -p -r1.88 drm_linux.h > --- dev/pci/drm/drm_linux.h 25 Apr 2018 01:27:46 -0000 1.88 > +++ dev/pci/drm/drm_linux.h 20 Jun 2018 18:41:39 -0000 > @@ -22,6 +22,7 @@ > #include <sys/param.h> > #include <sys/atomic.h> > #include <sys/errno.h> > +#include <sys/fcntl.h> > #include <sys/kernel.h> > #include <sys/signalvar.h> > #include <sys/stdint.h> > @@ -2713,5 +2714,39 @@ release_firmware(const struct firmware * > } > > void *memchr_inv(const void *, int, size_t); > + > +struct dma_buf_ops; > + > +struct dma_buf { > + const struct dma_buf_ops *ops; > + void *priv; > + size_t size; > + struct file *file; > +}; > + > +struct dma_buf_attachment; > + > +void get_dma_buf(struct dma_buf *); > +struct dma_buf *dma_buf_get(int); > +void dma_buf_put(struct dma_buf *); > +int dma_buf_fd(struct dma_buf *, int); > + > +struct dma_buf_ops { > + void (*release)(struct dma_buf *); > +}; > + > +struct dma_buf_export_info { > + const struct dma_buf_ops *ops; > + size_t size; > + int flags; > + void *priv; > +}; > + > +#define DEFINE_DMA_BUF_EXPORT_INFO(x) struct dma_buf_export_info x > + > +struct dma_buf *dma_buf_export(const struct dma_buf_export_info *); > + > +#define dma_buf_attach(x, y) NULL > +#define dma_buf_detach(x, y) panic("dma_buf_detach") > > #endif > Index: dev/pci/drm/drm_prime.c > =================================================================== > RCS file: dev/pci/drm/drm_prime.c > diff -N dev/pci/drm/drm_prime.c > --- /dev/null 1 Jan 1970 00:00:00 -0000 > +++ dev/pci/drm/drm_prime.c 20 Jun 2018 18:41:39 -0000 > @@ -0,0 +1,807 @@ > +/* > + * Copyright ???? 2012 Red Hat > + * > + * Permission is hereby granted, free of charge, to any person obtaining a > + * copy of this software and associated documentation files (the "Software"), > + * to deal in the Software without restriction, including without limitation > + * the rights to use, copy, modify, merge, publish, distribute, sublicense, > + * and/or sell copies of the Software, and to permit persons to whom the > + * Software is furnished to do so, subject to the following conditions: > + * > + * The above copyright notice and this permission notice (including the next > + * paragraph) shall be included in all copies or substantial portions of the > + * Software. > + * > + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR > + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, > + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL > + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER > + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING > + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER > DEALINGS > + * IN THE SOFTWARE. > + * > + * Authors: > + * Dave Airlie <airl...@redhat.com> > + * Rob Clark <rob.cl...@linaro.org> > + * > + */ > + > +#ifdef __linux__ > +#include <linux/export.h> > +#include <linux/dma-buf.h> > +#endif > +#include <dev/pci/drm/drmP.h> > +#ifdef notyet > +#include <drm/drm_gem.h> > +#endif > + > +#include "drm_internal.h" > + > +/* > + * DMA-BUF/GEM Object references and lifetime overview: > + * > + * On the export the dma_buf holds a reference to the exporting GEM > + * object. It takes this reference in handle_to_fd_ioctl, when it > + * first calls .prime_export and stores the exporting GEM object in > + * the dma_buf priv. This reference is released when the dma_buf > + * object goes away in the driver .release function. > + * > + * On the import the importing GEM object holds a reference to the > + * dma_buf (which in turn holds a ref to the exporting GEM object). > + * It takes that reference in the fd_to_handle ioctl. > + * It calls dma_buf_get, creates an attachment to it and stores the > + * attachment in the GEM object. When this attachment is destroyed > + * when the imported object is destroyed, we remove the attachment > + * and drop the reference to the dma_buf. > + * > + * Thus the chain of references always flows in one direction > + * (avoiding loops): importing_gem -> dmabuf -> exporting_gem > + * > + * Self-importing: if userspace is using PRIME as a replacement for flink > + * then it will get a fd->handle request for a GEM object that it created. > + * Drivers should detect this situation and return back the gem object > + * from the dma-buf private. Prime will do this automatically for drivers > that > + * use the drm_gem_prime_{import,export} helpers. > + */ > + > +struct drm_prime_member { > + struct list_head entry; > + struct dma_buf *dma_buf; > + uint32_t handle; > +}; > + > +#ifdef notyet > +struct drm_prime_attachment { > + struct sg_table *sgt; > + enum dma_data_direction dir; > +}; > +#endif > + > +static int drm_prime_add_buf_handle(struct drm_prime_file_private > *prime_fpriv, > + struct dma_buf *dma_buf, uint32_t handle) > +{ > + struct drm_prime_member *member; > + > + member = kmalloc(sizeof(*member), GFP_KERNEL); > + if (!member) > + return -ENOMEM; > + > + get_dma_buf(dma_buf); > + member->dma_buf = dma_buf; > + member->handle = handle; > + list_add(&member->entry, &prime_fpriv->head); > + return 0; > +} > + > +static struct dma_buf *drm_prime_lookup_buf_by_handle(struct > drm_prime_file_private *prime_fpriv, > + uint32_t handle) > +{ > + struct drm_prime_member *member; > + > + list_for_each_entry(member, &prime_fpriv->head, entry) { > + if (member->handle == handle) > + return member->dma_buf; > + } > + > + return NULL; > +} > + > +static int drm_prime_lookup_buf_handle(struct drm_prime_file_private > *prime_fpriv, > + struct dma_buf *dma_buf, > + uint32_t *handle) > +{ > + struct drm_prime_member *member; > + > + list_for_each_entry(member, &prime_fpriv->head, entry) { > + if (member->dma_buf == dma_buf) { > + *handle = member->handle; > + return 0; > + } > + } > + return -ENOENT; > +} > + > +#ifdef notyet > + > +static int drm_gem_map_attach(struct dma_buf *dma_buf, > + struct device *target_dev, > + struct dma_buf_attachment *attach) > +{ > + struct drm_prime_attachment *prime_attach; > + struct drm_gem_object *obj = dma_buf->priv; > + struct drm_device *dev = obj->dev; > + > + prime_attach = kzalloc(sizeof(*prime_attach), GFP_KERNEL); > + if (!prime_attach) > + return -ENOMEM; > + > + prime_attach->dir = DMA_NONE; > + attach->priv = prime_attach; > + > + if (!dev->driver->gem_prime_pin) > + return 0; > + > + return dev->driver->gem_prime_pin(obj); > +} > + > +static void drm_gem_map_detach(struct dma_buf *dma_buf, > + struct dma_buf_attachment *attach) > +{ > + struct drm_prime_attachment *prime_attach = attach->priv; > + struct drm_gem_object *obj = dma_buf->priv; > + struct drm_device *dev = obj->dev; > + struct sg_table *sgt; > + > + if (dev->driver->gem_prime_unpin) > + dev->driver->gem_prime_unpin(obj); > + > + if (!prime_attach) > + return; > + > + sgt = prime_attach->sgt; > + if (sgt) { > + if (prime_attach->dir != DMA_NONE) > + dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, > + prime_attach->dir); > + sg_free_table(sgt); > + } > + > + kfree(sgt); > + kfree(prime_attach); > + attach->priv = NULL; > +} > + > +#endif > + > +void drm_prime_remove_buf_handle_locked(struct drm_prime_file_private > *prime_fpriv, > + struct dma_buf *dma_buf) > +{ > + struct drm_prime_member *member, *safe; > + > + list_for_each_entry_safe(member, safe, &prime_fpriv->head, entry) { > + if (member->dma_buf == dma_buf) { > + dma_buf_put(dma_buf); > + list_del(&member->entry); > + kfree(member); > + } > + } > +} > + > +#ifdef notyet > + > +static struct sg_table *drm_gem_map_dma_buf(struct dma_buf_attachment > *attach, > + enum dma_data_direction dir) > +{ > + struct drm_prime_attachment *prime_attach = attach->priv; > + struct drm_gem_object *obj = attach->dmabuf->priv; > + struct sg_table *sgt; > + > + if (WARN_ON(dir == DMA_NONE || !prime_attach)) > + return ERR_PTR(-EINVAL); > + > + /* return the cached mapping when possible */ > + if (prime_attach->dir == dir) > + return prime_attach->sgt; > + > + /* > + * two mappings with different directions for the same attachment are > + * not allowed > + */ > + if (WARN_ON(prime_attach->dir != DMA_NONE)) > + return ERR_PTR(-EBUSY); > + > + sgt = obj->dev->driver->gem_prime_get_sg_table(obj); > + > + if (!IS_ERR(sgt)) { > + if (!dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir)) { > + sg_free_table(sgt); > + kfree(sgt); > + sgt = ERR_PTR(-ENOMEM); > + } else { > + prime_attach->sgt = sgt; > + prime_attach->dir = dir; > + } > + } > + > + return sgt; > +} > + > +static void drm_gem_unmap_dma_buf(struct dma_buf_attachment *attach, > + struct sg_table *sgt, > + enum dma_data_direction dir) > +{ > + /* nothing to be done here */ > +} > + > +#endif > + > +/** > + * drm_gem_dmabuf_release - dma_buf release implementation for GEM > + * @dma_buf: buffer to be released > + * > + * Generic release function for dma_bufs exported as PRIME buffers. GEM > drivers > + * must use this in their dma_buf ops structure as the release callback. > + */ > +void drm_gem_dmabuf_release(struct dma_buf *dma_buf) > +{ > + struct drm_gem_object *obj = dma_buf->priv; > + > + /* drop the reference on the export fd holds */ > + drm_gem_object_unreference_unlocked(obj); > +} > +EXPORT_SYMBOL(drm_gem_dmabuf_release); > + > +#ifdef notyet > + > +static void *drm_gem_dmabuf_vmap(struct dma_buf *dma_buf) > +{ > + struct drm_gem_object *obj = dma_buf->priv; > + struct drm_device *dev = obj->dev; > + > + return dev->driver->gem_prime_vmap(obj); > +} > + > +static void drm_gem_dmabuf_vunmap(struct dma_buf *dma_buf, void *vaddr) > +{ > + struct drm_gem_object *obj = dma_buf->priv; > + struct drm_device *dev = obj->dev; > + > + dev->driver->gem_prime_vunmap(obj, vaddr); > +} > + > +static void *drm_gem_dmabuf_kmap_atomic(struct dma_buf *dma_buf, > + unsigned long page_num) > +{ > + return NULL; > +} > + > +static void drm_gem_dmabuf_kunmap_atomic(struct dma_buf *dma_buf, > + unsigned long page_num, void *addr) > +{ > + > +} > +static void *drm_gem_dmabuf_kmap(struct dma_buf *dma_buf, > + unsigned long page_num) > +{ > + return NULL; > +} > + > +static void drm_gem_dmabuf_kunmap(struct dma_buf *dma_buf, > + unsigned long page_num, void *addr) > +{ > + > +} > + > +static int drm_gem_dmabuf_mmap(struct dma_buf *dma_buf, > + struct vm_area_struct *vma) > +{ > + struct drm_gem_object *obj = dma_buf->priv; > + struct drm_device *dev = obj->dev; > + > + if (!dev->driver->gem_prime_mmap) > + return -ENOSYS; > + > + return dev->driver->gem_prime_mmap(obj, vma); > +} > + > +static const struct dma_buf_ops drm_gem_prime_dmabuf_ops = { > + .attach = drm_gem_map_attach, > + .detach = drm_gem_map_detach, > + .map_dma_buf = drm_gem_map_dma_buf, > + .unmap_dma_buf = drm_gem_unmap_dma_buf, > + .release = drm_gem_dmabuf_release, > + .kmap = drm_gem_dmabuf_kmap, > + .kmap_atomic = drm_gem_dmabuf_kmap_atomic, > + .kunmap = drm_gem_dmabuf_kunmap, > + .kunmap_atomic = drm_gem_dmabuf_kunmap_atomic, > + .mmap = drm_gem_dmabuf_mmap, > + .vmap = drm_gem_dmabuf_vmap, > + .vunmap = drm_gem_dmabuf_vunmap, > +}; > + > +/** > + * DOC: PRIME Helpers > + * > + * Drivers can implement @gem_prime_export and @gem_prime_import in terms of > + * simpler APIs by using the helper functions @drm_gem_prime_export and > + * @drm_gem_prime_import. These functions implement dma-buf support in > terms of > + * six lower-level driver callbacks: > + * > + * Export callbacks: > + * > + * - @gem_prime_pin (optional): prepare a GEM object for exporting > + * > + * - @gem_prime_get_sg_table: provide a scatter/gather table of pinned pages > + * > + * - @gem_prime_vmap: vmap a buffer exported by your driver > + * > + * - @gem_prime_vunmap: vunmap a buffer exported by your driver > + * > + * - @gem_prime_mmap (optional): mmap a buffer exported by your driver > + * > + * Import callback: > + * > + * - @gem_prime_import_sg_table (import): produce a GEM object from another > + * driver's scatter/gather table > + */ > + > +/** > + * drm_gem_prime_export - helper library implementation of the export > callback > + * @dev: drm_device to export from > + * @obj: GEM object to export > + * @flags: flags like DRM_CLOEXEC > + * > + * This is the implementation of the gem_prime_export functions for GEM > drivers > + * using the PRIME helpers. > + */ > +struct dma_buf *drm_gem_prime_export(struct drm_device *dev, > + struct drm_gem_object *obj, > + int flags) > +{ > + struct dma_buf_export_info exp_info = { > + .exp_name = KBUILD_MODNAME, /* white lie for debug */ > + .owner = dev->driver->fops->owner, > + .ops = &drm_gem_prime_dmabuf_ops, > + .size = obj->size, > + .flags = flags, > + .priv = obj, > + }; > + > + if (dev->driver->gem_prime_res_obj) > + exp_info.resv = dev->driver->gem_prime_res_obj(obj); > + > + return dma_buf_export(&exp_info); > +} > +EXPORT_SYMBOL(drm_gem_prime_export); > + > +#endif > + > +static struct dma_buf *export_and_register_object(struct drm_device *dev, > + struct drm_gem_object *obj, > + uint32_t flags) > +{ > + struct dma_buf *dmabuf; > + > + /* prevent races with concurrent gem_close. */ > + if (obj->handle_count == 0) { > + dmabuf = ERR_PTR(-ENOENT); > + return dmabuf; > + } > + > + dmabuf = dev->driver->gem_prime_export(dev, obj, flags); > + if (IS_ERR(dmabuf)) { > + /* normally the created dma-buf takes ownership of the ref, > + * but if that fails then drop the ref > + */ > + return dmabuf; > + } > + > + /* > + * Note that callers do not need to clean up the export cache > + * since the check for obj->handle_count guarantees that someone > + * will clean it up. > + */ > + obj->dma_buf = dmabuf; > + get_dma_buf(obj->dma_buf); > + /* Grab a new ref since the callers is now used by the dma-buf */ > + drm_gem_object_reference(obj); > + > + return dmabuf; > +} > + > +/** > + * drm_gem_prime_handle_to_fd - PRIME export function for GEM drivers > + * @dev: dev to export the buffer from > + * @file_priv: drm file-private structure > + * @handle: buffer handle to export > + * @flags: flags like DRM_CLOEXEC > + * @prime_fd: pointer to storage for the fd id of the create dma-buf > + * > + * This is the PRIME export function which must be used mandatorily by GEM > + * drivers to ensure correct lifetime management of the underlying GEM > object. > + * The actual exporting from GEM object to a dma-buf is done through the > + * gem_prime_export driver callback. > + */ > +int drm_gem_prime_handle_to_fd(struct drm_device *dev, > + struct drm_file *file_priv, uint32_t handle, > + uint32_t flags, > + int *prime_fd) > +{ > + struct drm_gem_object *obj; > + int ret = 0; > + struct dma_buf *dmabuf; > + > + mutex_lock(&file_priv->prime.lock); > + obj = drm_gem_object_lookup(dev, file_priv, handle); > + if (!obj) { > + ret = -ENOENT; > + goto out_unlock; > + } > + > + dmabuf = drm_prime_lookup_buf_by_handle(&file_priv->prime, handle); > + if (dmabuf) { > + get_dma_buf(dmabuf); > + goto out_have_handle; > + } > + > + mutex_lock(&dev->object_name_lock); > +#ifdef notyet > + /* re-export the original imported object */ > + if (obj->import_attach) { > + dmabuf = obj->import_attach->dmabuf; > + get_dma_buf(dmabuf); > + goto out_have_obj; > + } > +#endif > + > + if (obj->dma_buf) { > + get_dma_buf(obj->dma_buf); > + dmabuf = obj->dma_buf; > + goto out_have_obj; > + } > + > + dmabuf = export_and_register_object(dev, obj, flags); > + if (IS_ERR(dmabuf)) { > + /* normally the created dma-buf takes ownership of the ref, > + * but if that fails then drop the ref > + */ > + ret = PTR_ERR(dmabuf); > + mutex_unlock(&dev->object_name_lock); > + goto out; > + } > + > +out_have_obj: > + /* > + * If we've exported this buffer then cheat and add it to the import > list > + * so we get the correct handle back. We must do this under the > + * protection of dev->object_name_lock to ensure that a racing gem close > + * ioctl doesn't miss to remove this buffer handle from the cache. > + */ > + ret = drm_prime_add_buf_handle(&file_priv->prime, > + dmabuf, handle); > + mutex_unlock(&dev->object_name_lock); > + if (ret) > + goto fail_put_dmabuf; > + > +out_have_handle: > + ret = dma_buf_fd(dmabuf, flags); > + /* > + * We must _not_ remove the buffer from the handle cache since the newly > + * created dma buf is already linked in the global obj->dma_buf pointer, > + * and that is invariant as long as a userspace gem handle exists. > + * Closing the handle will clean out the cache anyway, so we don't leak. > + */ > + if (ret < 0) { > + goto fail_put_dmabuf; > + } else { > + *prime_fd = ret; > + ret = 0; > + } > + > + goto out; > + > +fail_put_dmabuf: > + dma_buf_put(dmabuf); > +out: > + drm_gem_object_unreference_unlocked(obj); > +out_unlock: > + mutex_unlock(&file_priv->prime.lock); > + > + return ret; > +} > +EXPORT_SYMBOL(drm_gem_prime_handle_to_fd); > + > +#ifdef notyet > + > +/** > + * drm_gem_prime_import - helper library implementation of the import > callback > + * @dev: drm_device to import into > + * @dma_buf: dma-buf object to import > + * > + * This is the implementation of the gem_prime_import functions for GEM > drivers > + * using the PRIME helpers. > + */ > +struct drm_gem_object *drm_gem_prime_import(struct drm_device *dev, > + struct dma_buf *dma_buf) > +{ > + struct dma_buf_attachment *attach; > + struct sg_table *sgt; > + struct drm_gem_object *obj; > + int ret; > + > + if (dma_buf->ops == &drm_gem_prime_dmabuf_ops) { > + obj = dma_buf->priv; > + if (obj->dev == dev) { > + /* > + * Importing dmabuf exported from out own gem increases > + * refcount on gem itself instead of f_count of dmabuf. > + */ > + drm_gem_object_reference(obj); > + return obj; > + } > + } > + > + if (!dev->driver->gem_prime_import_sg_table) > + return ERR_PTR(-EINVAL); > + > + attach = dma_buf_attach(dma_buf, dev->dev); > + if (IS_ERR(attach)) > + return ERR_CAST(attach); > + > + get_dma_buf(dma_buf); > + > + sgt = dma_buf_map_attachment(attach, DMA_BIDIRECTIONAL); > + if (IS_ERR(sgt)) { > + ret = PTR_ERR(sgt); > + goto fail_detach; > + } > + > + obj = dev->driver->gem_prime_import_sg_table(dev, attach, sgt); > + if (IS_ERR(obj)) { > + ret = PTR_ERR(obj); > + goto fail_unmap; > + } > + > + obj->import_attach = attach; > + > + return obj; > + > +fail_unmap: > + dma_buf_unmap_attachment(attach, sgt, DMA_BIDIRECTIONAL); > +fail_detach: > + dma_buf_detach(dma_buf, attach); > + dma_buf_put(dma_buf); > + > + return ERR_PTR(ret); > +} > +EXPORT_SYMBOL(drm_gem_prime_import); > + > +#endif > + > +/** > + * drm_gem_prime_fd_to_handle - PRIME import function for GEM drivers > + * @dev: dev to export the buffer from > + * @file_priv: drm file-private structure > + * @prime_fd: fd id of the dma-buf which should be imported > + * @handle: pointer to storage for the handle of the imported buffer object > + * > + * This is the PRIME import function which must be used mandatorily by GEM > + * drivers to ensure correct lifetime management of the underlying GEM > object. > + * The actual importing of GEM object from the dma-buf is done through the > + * gem_import_export driver callback. > + */ > +int drm_gem_prime_fd_to_handle(struct drm_device *dev, > + struct drm_file *file_priv, int prime_fd, > + uint32_t *handle) > +{ > + struct dma_buf *dma_buf; > + struct drm_gem_object *obj; > + int ret; > + > + dma_buf = dma_buf_get(prime_fd); > + if (IS_ERR(dma_buf)) > + return PTR_ERR(dma_buf); > + > + mutex_lock(&file_priv->prime.lock); > + > + ret = drm_prime_lookup_buf_handle(&file_priv->prime, > + dma_buf, handle); > + if (ret == 0) > + goto out_put; > + > + /* never seen this one, need to import */ > + mutex_lock(&dev->object_name_lock); > + obj = dev->driver->gem_prime_import(dev, dma_buf); > + if (IS_ERR(obj)) { > + ret = PTR_ERR(obj); > + goto out_unlock; > + } > + > + if (obj->dma_buf) { > + WARN_ON(obj->dma_buf != dma_buf); > + } else { > + obj->dma_buf = dma_buf; > + get_dma_buf(dma_buf); > + } > + > + /* drm_gem_handle_create_tail unlocks dev->object_name_lock. */ > + ret = drm_gem_handle_create_tail(file_priv, obj, handle); > + drm_gem_object_unreference_unlocked(obj); > + if (ret) > + goto out_put; > + > + ret = drm_prime_add_buf_handle(&file_priv->prime, > + dma_buf, *handle); > + if (ret) > + goto fail; > + > + mutex_unlock(&file_priv->prime.lock); > + > + dma_buf_put(dma_buf); > + > + return 0; > + > +fail: > + /* hmm, if driver attached, we are relying on the free-object path > + * to detach.. which seems ok.. > + */ > + drm_gem_handle_delete(file_priv, *handle); > +out_unlock: > + mutex_unlock(&dev->object_name_lock); > +out_put: > + dma_buf_put(dma_buf); > + mutex_unlock(&file_priv->prime.lock); > + return ret; > +} > +EXPORT_SYMBOL(drm_gem_prime_fd_to_handle); > + > +int drm_prime_handle_to_fd_ioctl(struct drm_device *dev, void *data, > + struct drm_file *file_priv) > +{ > + struct drm_prime_handle *args = data; > + uint32_t flags; > + > + if (!drm_core_check_feature(dev, DRIVER_PRIME)) > + return -EINVAL; > + > + if (!dev->driver->prime_handle_to_fd) > + return -ENOSYS; > + > + /* check flags are valid */ > + if (args->flags & ~DRM_CLOEXEC) > + return -EINVAL; > + > + /* we only want to pass DRM_CLOEXEC which is == O_CLOEXEC */ > + flags = args->flags & DRM_CLOEXEC; > + > + return dev->driver->prime_handle_to_fd(dev, file_priv, > + args->handle, flags, &args->fd); > +} > + > +int drm_prime_fd_to_handle_ioctl(struct drm_device *dev, void *data, > + struct drm_file *file_priv) > +{ > + struct drm_prime_handle *args = data; > + > + if (!drm_core_check_feature(dev, DRIVER_PRIME)) > + return -EINVAL; > + > + if (!dev->driver->prime_fd_to_handle) > + return -ENOSYS; > + > + return dev->driver->prime_fd_to_handle(dev, file_priv, > + args->fd, &args->handle); > +} > + > +#ifdef notyet > + > +/** > + * drm_prime_pages_to_sg - converts a page array into an sg list > + * @pages: pointer to the array of page pointers to convert > + * @nr_pages: length of the page vector > + * > + * This helper creates an sg table object from a set of pages > + * the driver is responsible for mapping the pages into the > + * importers address space for use with dma_buf itself. > + */ > +struct sg_table *drm_prime_pages_to_sg(struct page **pages, unsigned int > nr_pages) > +{ > + struct sg_table *sg = NULL; > + int ret; > + > + sg = kmalloc(sizeof(struct sg_table), GFP_KERNEL); > + if (!sg) { > + ret = -ENOMEM; > + goto out; > + } > + > + ret = sg_alloc_table_from_pages(sg, pages, nr_pages, 0, > + nr_pages << PAGE_SHIFT, GFP_KERNEL); > + if (ret) > + goto out; > + > + return sg; > +out: > + kfree(sg); > + return ERR_PTR(ret); > +} > +EXPORT_SYMBOL(drm_prime_pages_to_sg); > + > +/** > + * drm_prime_sg_to_page_addr_arrays - convert an sg table into a page array > + * @sgt: scatter-gather table to convert > + * @pages: array of page pointers to store the page array in > + * @addrs: optional array to store the dma bus address of each page > + * @max_pages: size of both the passed-in arrays > + * > + * Exports an sg table into an array of pages and addresses. This is > currently > + * required by the TTM driver in order to do correct fault handling. > + */ > +int drm_prime_sg_to_page_addr_arrays(struct sg_table *sgt, struct page > **pages, > + dma_addr_t *addrs, int max_pages) > +{ > + unsigned count; > + struct scatterlist *sg; > + struct page *page; > + u32 len; > + int pg_index; > + dma_addr_t addr; > + > + pg_index = 0; > + for_each_sg(sgt->sgl, sg, sgt->nents, count) { > + len = sg->length; > + page = sg_page(sg); > + addr = sg_dma_address(sg); > + > + while (len > 0) { > + if (WARN_ON(pg_index >= max_pages)) > + return -1; > + pages[pg_index] = page; > + if (addrs) > + addrs[pg_index] = addr; > + > + page++; > + addr += PAGE_SIZE; > + len -= PAGE_SIZE; > + pg_index++; > + } > + } > + return 0; > +} > +EXPORT_SYMBOL(drm_prime_sg_to_page_addr_arrays); > + > +/** > + * drm_prime_gem_destroy - helper to clean up a PRIME-imported GEM object > + * @obj: GEM object which was created from a dma-buf > + * @sg: the sg-table which was pinned at import time > + * > + * This is the cleanup functions which GEM drivers need to call when they use > + * @drm_gem_prime_import to import dma-bufs. > + */ > +void drm_prime_gem_destroy(struct drm_gem_object *obj, struct sg_table *sg) > +{ > + struct dma_buf_attachment *attach; > + struct dma_buf *dma_buf; > + attach = obj->import_attach; > + if (sg) > + dma_buf_unmap_attachment(attach, sg, DMA_BIDIRECTIONAL); > + dma_buf = attach->dmabuf; > + dma_buf_detach(attach->dmabuf, attach); > + /* remove the reference */ > + dma_buf_put(dma_buf); > +} > +EXPORT_SYMBOL(drm_prime_gem_destroy); > + > +#endif > + > +void drm_prime_init_file_private(struct drm_prime_file_private *prime_fpriv) > +{ > + INIT_LIST_HEAD(&prime_fpriv->head); > + rw_init(&prime_fpriv->lock, "primlk"); > +} > + > +void drm_prime_destroy_file_private(struct drm_prime_file_private > *prime_fpriv) > +{ > + /* by now drm_gem_release should've made sure the list is empty */ > + WARN_ON(!list_empty(&prime_fpriv->head)); > +} > Index: dev/pci/drm/files.drm > =================================================================== > RCS file: /cvs/src/sys/dev/pci/drm/files.drm,v > retrieving revision 1.41 > diff -u -p -r1.41 files.drm > --- dev/pci/drm/files.drm 25 Apr 2018 01:27:46 -0000 1.41 > +++ dev/pci/drm/files.drm 20 Jun 2018 18:41:39 -0000 > @@ -32,6 +32,7 @@ file dev/pci/drm/drm_cache.c drm > file dev/pci/drm/drm_panel.c drm & inteldrm > file dev/pci/drm/drm_dp_mst_topology.c drm > file dev/pci/drm/drm_mipi_dsi.c drm & inteldrm > +file dev/pci/drm/drm_prime.c drm > file dev/pci/drm/linux_hdmi.c drm > file dev/pci/drm/linux_list_sort.c drm > file dev/pci/drm/linux_reservation.c drm > @@ -59,6 +60,7 @@ file dev/pci/drm/i915/i915_drv.c inteld > file dev/pci/drm/i915/i915_gem.c inteldrm > file dev/pci/drm/i915/i915_gem_context.c inteldrm > file dev/pci/drm/i915/i915_gem_batch_pool.c inteldrm > +file dev/pci/drm/i915/i915_gem_dmabuf.c inteldrm > file dev/pci/drm/i915/i915_gem_evict.c inteldrm > file dev/pci/drm/i915/i915_gem_execbuffer.c inteldrm > file dev/pci/drm/i915/i915_gem_fence.c inteldrm > Index: dev/pci/drm/i915/i915_drv.c > =================================================================== > RCS file: /cvs/src/sys/dev/pci/drm/i915/i915_drv.c,v > retrieving revision 1.113 > diff -u -p -r1.113 i915_drv.c > --- dev/pci/drm/i915/i915_drv.c 22 May 2018 20:49:03 -0000 1.113 > +++ dev/pci/drm/i915/i915_drv.c 20 Jun 2018 18:41:39 -0000 > @@ -1793,12 +1793,10 @@ static struct drm_driver driver = { > .gem_fault = i915_gem_fault, > #endif > > -#ifdef notyet > .prime_handle_to_fd = drm_gem_prime_handle_to_fd, > .prime_fd_to_handle = drm_gem_prime_fd_to_handle, > .gem_prime_export = i915_gem_prime_export, > .gem_prime_import = i915_gem_prime_import, > -#endif > > .dumb_create = i915_gem_dumb_create, > .dumb_map_offset = i915_gem_mmap_gtt, > Index: dev/pci/drm/i915/i915_drv.h > =================================================================== > RCS file: /cvs/src/sys/dev/pci/drm/i915/i915_drv.h,v > retrieving revision 1.79 > diff -u -p -r1.79 i915_drv.h > --- dev/pci/drm/i915/i915_drv.h 15 Jan 2018 22:24:17 -0000 1.79 > +++ dev/pci/drm/i915/i915_drv.h 20 Jun 2018 18:41:39 -0000 > @@ -3215,13 +3215,11 @@ i915_gem_get_gtt_alignment(struct drm_de > int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj, > enum i915_cache_level cache_level); > > -#ifdef notyet > struct drm_gem_object *i915_gem_prime_import(struct drm_device *dev, > struct dma_buf *dma_buf); > > struct dma_buf *i915_gem_prime_export(struct drm_device *dev, > struct drm_gem_object *gem_obj, int flags); > -#endif > > u64 i915_gem_obj_ggtt_offset_view(struct drm_i915_gem_object *o, > const struct i915_ggtt_view *view); > Index: dev/pci/drm/i915/i915_gem_dmabuf.c > =================================================================== > RCS file: dev/pci/drm/i915/i915_gem_dmabuf.c > diff -N dev/pci/drm/i915/i915_gem_dmabuf.c > --- /dev/null 1 Jan 1970 00:00:00 -0000 > +++ dev/pci/drm/i915/i915_gem_dmabuf.c 20 Jun 2018 18:41:39 -0000 > @@ -0,0 +1,337 @@ > +/* > + * Copyright 2012 Red Hat Inc > + * > + * Permission is hereby granted, free of charge, to any person obtaining a > + * copy of this software and associated documentation files (the "Software"), > + * to deal in the Software without restriction, including without limitation > + * the rights to use, copy, modify, merge, publish, distribute, sublicense, > + * and/or sell copies of the Software, and to permit persons to whom the > + * Software is furnished to do so, subject to the following conditions: > + * > + * The above copyright notice and this permission notice (including the next > + * paragraph) shall be included in all copies or substantial portions of the > + * Software. > + * > + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR > + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, > + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL > + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER > + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING > + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER > + * DEALINGS IN THE SOFTWARE. > + * > + * Authors: > + * Dave Airlie <airl...@redhat.com> > + */ > +#include <dev/pci/drm/drmP.h> > +#include "i915_drv.h" > +#ifdef __linux__ > +#include <linux/dma-buf.h> > +#endif > + > +static struct drm_i915_gem_object *dma_buf_to_obj(struct dma_buf *buf) > +{ > + return to_intel_bo(buf->priv); > +} > + > +#ifdef notyet > + > +static struct sg_table *i915_gem_map_dma_buf(struct dma_buf_attachment > *attachment, > + enum dma_data_direction dir) > +{ > + struct drm_i915_gem_object *obj = dma_buf_to_obj(attachment->dmabuf); > + struct sg_table *st; > + struct scatterlist *src, *dst; > + int ret, i; > + > + ret = i915_mutex_lock_interruptible(obj->base.dev); > + if (ret) > + goto err; > + > + ret = i915_gem_object_get_pages(obj); > + if (ret) > + goto err_unlock; > + > + i915_gem_object_pin_pages(obj); > + > + /* Copy sg so that we make an independent mapping */ > + st = kmalloc(sizeof(struct sg_table), GFP_KERNEL); > + if (st == NULL) { > + ret = -ENOMEM; > + goto err_unpin; > + } > + > + ret = sg_alloc_table(st, obj->pages->nents, GFP_KERNEL); > + if (ret) > + goto err_free; > + > + src = obj->pages->sgl; > + dst = st->sgl; > + for (i = 0; i < obj->pages->nents; i++) { > + sg_set_page(dst, sg_page(src), src->length, 0); > + dst = sg_next(dst); > + src = sg_next(src); > + } > + > + if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) { > + ret =-ENOMEM; > + goto err_free_sg; > + } > + > + mutex_unlock(&obj->base.dev->struct_mutex); > + return st; > + > +err_free_sg: > + sg_free_table(st); > +err_free: > + kfree(st); > +err_unpin: > + i915_gem_object_unpin_pages(obj); > +err_unlock: > + mutex_unlock(&obj->base.dev->struct_mutex); > +err: > + return ERR_PTR(ret); > +} > + > +static void i915_gem_unmap_dma_buf(struct dma_buf_attachment *attachment, > + struct sg_table *sg, > + enum dma_data_direction dir) > +{ > + struct drm_i915_gem_object *obj = dma_buf_to_obj(attachment->dmabuf); > + > + mutex_lock(&obj->base.dev->struct_mutex); > + > + dma_unmap_sg(attachment->dev, sg->sgl, sg->nents, dir); > + sg_free_table(sg); > + kfree(sg); > + > + i915_gem_object_unpin_pages(obj); > + > + mutex_unlock(&obj->base.dev->struct_mutex); > +} > + > +static void *i915_gem_dmabuf_vmap(struct dma_buf *dma_buf) > +{ > + struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf); > + struct drm_device *dev = obj->base.dev; > + struct sg_page_iter sg_iter; > + struct page **pages; > + int ret, i; > + > + ret = i915_mutex_lock_interruptible(dev); > + if (ret) > + return ERR_PTR(ret); > + > + if (obj->dma_buf_vmapping) { > + obj->vmapping_count++; > + goto out_unlock; > + } > + > + ret = i915_gem_object_get_pages(obj); > + if (ret) > + goto err; > + > + i915_gem_object_pin_pages(obj); > + > + ret = -ENOMEM; > + > + pages = drm_malloc_ab(obj->base.size >> PAGE_SHIFT, sizeof(*pages)); > + if (pages == NULL) > + goto err_unpin; > + > + i = 0; > + for_each_sg_page(obj->pages->sgl, &sg_iter, obj->pages->nents, 0) > + pages[i++] = sg_page_iter_page(&sg_iter); > + > + obj->dma_buf_vmapping = vmap(pages, i, 0, PAGE_KERNEL); > + drm_free_large(pages); > + > + if (!obj->dma_buf_vmapping) > + goto err_unpin; > + > + obj->vmapping_count = 1; > +out_unlock: > + mutex_unlock(&dev->struct_mutex); > + return obj->dma_buf_vmapping; > + > +err_unpin: > + i915_gem_object_unpin_pages(obj); > +err: > + mutex_unlock(&dev->struct_mutex); > + return ERR_PTR(ret); > +} > + > +static void i915_gem_dmabuf_vunmap(struct dma_buf *dma_buf, void *vaddr) > +{ > + struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf); > + struct drm_device *dev = obj->base.dev; > + > + mutex_lock(&dev->struct_mutex); > + if (--obj->vmapping_count == 0) { > + vunmap(obj->dma_buf_vmapping); > + obj->dma_buf_vmapping = NULL; > + > + i915_gem_object_unpin_pages(obj); > + } > + mutex_unlock(&dev->struct_mutex); > +} > + > +static void *i915_gem_dmabuf_kmap_atomic(struct dma_buf *dma_buf, unsigned > long page_num) > +{ > + return NULL; > +} > + > +static void i915_gem_dmabuf_kunmap_atomic(struct dma_buf *dma_buf, unsigned > long page_num, void *addr) > +{ > + > +} > +static void *i915_gem_dmabuf_kmap(struct dma_buf *dma_buf, unsigned long > page_num) > +{ > + return NULL; > +} > + > +static void i915_gem_dmabuf_kunmap(struct dma_buf *dma_buf, unsigned long > page_num, void *addr) > +{ > + > +} > + > +static int i915_gem_dmabuf_mmap(struct dma_buf *dma_buf, struct > vm_area_struct *vma) > +{ > + return -EINVAL; > +} > + > +static int i915_gem_begin_cpu_access(struct dma_buf *dma_buf, size_t start, > size_t length, enum dma_data_direction direction) > +{ > + struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf); > + struct drm_device *dev = obj->base.dev; > + int ret; > + bool write = (direction == DMA_BIDIRECTIONAL || direction == > DMA_TO_DEVICE); > + > + ret = i915_mutex_lock_interruptible(dev); > + if (ret) > + return ret; > + > + ret = i915_gem_object_set_to_cpu_domain(obj, write); > + mutex_unlock(&dev->struct_mutex); > + return ret; > +} > + > +#endif > + > +static const struct dma_buf_ops i915_dmabuf_ops = { > +#ifdef notyet > + .map_dma_buf = i915_gem_map_dma_buf, > + .unmap_dma_buf = i915_gem_unmap_dma_buf, > +#endif > + .release = drm_gem_dmabuf_release, > +#ifdef notyet > + .kmap = i915_gem_dmabuf_kmap, > + .kmap_atomic = i915_gem_dmabuf_kmap_atomic, > + .kunmap = i915_gem_dmabuf_kunmap, > + .kunmap_atomic = i915_gem_dmabuf_kunmap_atomic, > + .mmap = i915_gem_dmabuf_mmap, > + .vmap = i915_gem_dmabuf_vmap, > + .vunmap = i915_gem_dmabuf_vunmap, > + .begin_cpu_access = i915_gem_begin_cpu_access, > +#endif > +}; > + > +struct dma_buf *i915_gem_prime_export(struct drm_device *dev, > + struct drm_gem_object *gem_obj, int flags) > +{ > + struct drm_i915_gem_object *obj = to_intel_bo(gem_obj); > + DEFINE_DMA_BUF_EXPORT_INFO(exp_info); > + > + exp_info.ops = &i915_dmabuf_ops; > + exp_info.size = gem_obj->size; > + exp_info.flags = flags; > + exp_info.priv = gem_obj; > + > + > + if (obj->ops->dmabuf_export) { > + int ret = obj->ops->dmabuf_export(obj); > + if (ret) > + return ERR_PTR(ret); > + } > + > + return dma_buf_export(&exp_info); > +} > + > +#ifdef notyet > + > +static int i915_gem_object_get_pages_dmabuf(struct drm_i915_gem_object *obj) > +{ > + struct sg_table *sg; > + > + sg = dma_buf_map_attachment(obj->base.import_attach, DMA_BIDIRECTIONAL); > + if (IS_ERR(sg)) > + return PTR_ERR(sg); > + > + obj->pages = sg; > + return 0; > +} > + > +static void i915_gem_object_put_pages_dmabuf(struct drm_i915_gem_object *obj) > +{ > + dma_buf_unmap_attachment(obj->base.import_attach, > + obj->pages, DMA_BIDIRECTIONAL); > +} > + > +static const struct drm_i915_gem_object_ops i915_gem_object_dmabuf_ops = { > + .get_pages = i915_gem_object_get_pages_dmabuf, > + .put_pages = i915_gem_object_put_pages_dmabuf, > +}; > + > +#endif > + > +struct drm_gem_object *i915_gem_prime_import(struct drm_device *dev, > + struct dma_buf *dma_buf) > +{ > + struct dma_buf_attachment *attach; > + struct drm_i915_gem_object *obj; > + int ret; > + > + /* is this one of own objects? */ > + if (dma_buf->ops == &i915_dmabuf_ops) { > + obj = dma_buf_to_obj(dma_buf); > + /* is it from our device? */ > + if (obj->base.dev == dev) { > + /* > + * Importing dmabuf exported from out own gem increases > + * refcount on gem itself instead of f_count of dmabuf. > + */ > + drm_gem_object_reference(&obj->base); > + return &obj->base; > + } > + } > + > + /* need to attach */ > + attach = dma_buf_attach(dma_buf, dev->dev); > + if (IS_ERR(attach)) > + return ERR_CAST(attach); > + > +#ifdef notyet > + get_dma_buf(dma_buf); > + > + obj = i915_gem_object_alloc(dev); > + if (obj == NULL) { > + ret = -ENOMEM; > + goto fail_detach; > + } > + > + drm_gem_private_object_init(dev, &obj->base, dma_buf->size); > + i915_gem_object_init(obj, &i915_gem_object_dmabuf_ops); > + obj->base.import_attach = attach; > + > + return &obj->base; > + > +fail_detach: > + dma_buf_detach(dma_buf, attach); > + dma_buf_put(dma_buf); > + > + return ERR_PTR(ret); > +#else > + ret = 0; > + panic(__func__); > +#endif > +} > Index: kern/kern_descrip.c > =================================================================== > RCS file: /cvs/src/sys/kern/kern_descrip.c,v > retrieving revision 1.167 > diff -u -p -r1.167 kern_descrip.c > --- kern/kern_descrip.c 20 Jun 2018 10:52:49 -0000 1.167 > +++ kern/kern_descrip.c 20 Jun 2018 18:41:39 -0000 > @@ -686,10 +686,12 @@ fdinsert(struct filedesc *fdp, int fd, i > fdpassertlocked(fdp); > > mtx_enter(&fhdlk); > - if ((fq = fdp->fd_ofiles[0]) != NULL) { > - LIST_INSERT_AFTER(fq, fp, f_list); > - } else { > - LIST_INSERT_HEAD(&filehead, fp, f_list); > + if ((fp->f_iflags & FIF_INSERTED) == 0) { > + if ((fq = fdp->fd_ofiles[0]) != NULL) { > + LIST_INSERT_AFTER(fq, fp, f_list); > + } else { > + LIST_INSERT_HEAD(&filehead, fp, f_list); > + } > } > KASSERT(fdp->fd_ofiles[fd] == NULL); > fdp->fd_ofiles[fd] = fp; > @@ -712,6 +714,7 @@ fdrelease(struct proc *p, int fd) > { > struct filedesc *fdp = p->p_fd; > struct file *fp; > + int error; > > fdpassertlocked(fdp); > > @@ -720,7 +723,10 @@ fdrelease(struct proc *p, int fd) > return (EBADF); > fdremove(fdp, fd); > knote_fdclose(p, fd); > - return (closef(fp, p)); > + fdpunlock(fdp); > + error = closef(fp, p); > + fdplock(fdp); > + return error; > } > > /* > Index: kern/kern_pledge.c > =================================================================== > RCS file: /cvs/src/sys/kern/kern_pledge.c,v > retrieving revision 1.233 > diff -u -p -r1.233 kern_pledge.c > --- kern/kern_pledge.c 20 Jun 2018 10:48:55 -0000 1.233 > +++ kern/kern_pledge.c 20 Jun 2018 18:41:39 -0000 > @@ -709,6 +709,7 @@ pledge_recvfd(struct proc *p, struct fil > switch (fp->f_type) { > case DTYPE_SOCKET: > case DTYPE_PIPE: > + case DTYPE_DMABUF: > return (0); > case DTYPE_VNODE: > vp = fp->f_data; > @@ -736,6 +737,7 @@ pledge_sendfd(struct proc *p, struct fil > switch (fp->f_type) { > case DTYPE_SOCKET: > case DTYPE_PIPE: > + case DTYPE_DMABUF: > return (0); > case DTYPE_VNODE: > vp = fp->f_data; > Index: sys/file.h > =================================================================== > RCS file: /cvs/src/sys/sys/file.h,v > retrieving revision 1.49 > diff -u -p -r1.49 file.h > --- sys/file.h 20 Jun 2018 10:52:49 -0000 1.49 > +++ sys/file.h 20 Jun 2018 18:41:40 -0000 > @@ -77,6 +77,7 @@ struct file { > #define DTYPE_SOCKET 2 /* communications endpoint */ > #define DTYPE_PIPE 3 /* pipe */ > #define DTYPE_KQUEUE 4 /* event queue */ > +#define DTYPE_DMABUF 5 /* DMA buffer (for DRM) */ > short f_type; /* [I] descriptor type */ > long f_count; /* [F] reference count */ > struct ucred *f_cred; /* [I] credentials associated with descriptor */ >