Merge tag 'v4.6-rc3' into drm-intel-next-queued
authorDaniel Vetter <daniel.vetter@ffwll.ch>
Mon, 11 Apr 2016 17:21:06 +0000 (19:21 +0200)
committerDaniel Vetter <daniel.vetter@ffwll.ch>
Mon, 11 Apr 2016 17:25:13 +0000 (19:25 +0200)
Linux 4.6-rc3

Backmerge requested by Chris Wilson to make his patches apply cleanly.
Tiny conflict in vmalloc.c with the (properly acked and all) patch in
drm-intel-next:

commit 4da56b99d99e5a7df2b7f11e87bfea935f909732
Author: Chris Wilson <chris@chris-wilson.co.uk>
Date:   Mon Apr 4 14:46:42 2016 +0100

    mm/vmap: Add a notifier for when we run out of vmap address space

and Linus' tree.

Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>
1  2 
drivers/gpu/drm/i915/i915_gem.c
drivers/gpu/drm/i915/i915_gem_dmabuf.c
drivers/gpu/drm/i915/i915_gem_userptr.c
mm/vmalloc.c

@@@ -130,9 -130,9 +130,9 @@@ in
  i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data,
                            struct drm_file *file)
  {
 -      struct drm_i915_private *dev_priv = dev->dev_private;
 +      struct drm_i915_private *dev_priv = to_i915(dev);
 +      struct i915_ggtt *ggtt = &dev_priv->ggtt;
        struct drm_i915_gem_get_aperture *args = data;
 -      struct i915_gtt *ggtt = &dev_priv->gtt;
        struct i915_vma *vma;
        size_t pinned;
  
                        pinned += vma->node.size;
        mutex_unlock(&dev->struct_mutex);
  
 -      args->aper_size = dev_priv->gtt.base.total;
 +      args->aper_size = ggtt->base.total;
        args->aper_available_size = args->aper_size - pinned;
  
        return 0;
@@@ -177,7 -177,7 +177,7 @@@ i915_gem_object_get_pages_phys(struct d
                drm_clflush_virt_range(vaddr, PAGE_SIZE);
                kunmap_atomic(src);
  
-               page_cache_release(page);
+               put_page(page);
                vaddr += PAGE_SIZE;
        }
  
@@@ -243,7 -243,7 +243,7 @@@ i915_gem_object_put_pages_phys(struct d
                        set_page_dirty(page);
                        if (obj->madv == I915_MADV_WILLNEED)
                                mark_page_accessed(page);
-                       page_cache_release(page);
+                       put_page(page);
                        vaddr += PAGE_SIZE;
                }
                obj->dirty = 0;
@@@ -765,8 -765,7 +765,8 @@@ i915_gem_gtt_pwrite_fast(struct drm_dev
                         struct drm_i915_gem_pwrite *args,
                         struct drm_file *file)
  {
 -      struct drm_i915_private *dev_priv = dev->dev_private;
 +      struct drm_i915_private *dev_priv = to_i915(dev);
 +      struct i915_ggtt *ggtt = &dev_priv->ggtt;
        ssize_t remain;
        loff_t offset, page_base;
        char __user *user_data;
                 * source page isn't available.  Return the error and we'll
                 * retry in the slow path.
                 */
 -              if (fast_user_write(dev_priv->gtt.mappable, page_base,
 +              if (fast_user_write(ggtt->mappable, page_base,
                                    page_offset, user_data, page_length)) {
                        ret = -EFAULT;
                        goto out_flush;
@@@ -1142,9 -1141,9 +1142,9 @@@ static void fake_irq(unsigned long data
  }
  
  static bool missed_irq(struct drm_i915_private *dev_priv,
 -                     struct intel_engine_cs *ring)
 +                     struct intel_engine_cs *engine)
  {
 -      return test_bit(ring->id, &dev_priv->gpu_error.missed_irq_rings);
 +      return test_bit(engine->id, &dev_priv->gpu_error.missed_irq_rings);
  }
  
  static unsigned long local_clock_us(unsigned *cpu)
@@@ -1194,7 -1193,7 +1194,7 @@@ static int __i915_spin_request(struct d
         * takes to sleep on a request, on the order of a microsecond.
         */
  
 -      if (req->ring->irq_refcount)
 +      if (req->engine->irq_refcount)
                return -EBUSY;
  
        /* Only spin if we know the GPU is processing this request */
@@@ -1244,11 -1243,11 +1244,11 @@@ int __i915_wait_request(struct drm_i915
                        s64 *timeout,
                        struct intel_rps_client *rps)
  {
 -      struct intel_engine_cs *ring = i915_gem_request_get_ring(req);
 -      struct drm_device *dev = ring->dev;
 +      struct intel_engine_cs *engine = i915_gem_request_get_engine(req);
 +      struct drm_device *dev = engine->dev;
        struct drm_i915_private *dev_priv = dev->dev_private;
        const bool irq_test_in_progress =
 -              ACCESS_ONCE(dev_priv->gpu_error.test_irq_rings) & intel_ring_flag(ring);
 +              ACCESS_ONCE(dev_priv->gpu_error.test_irq_rings) & intel_engine_flag(engine);
        int state = interruptible ? TASK_INTERRUPTIBLE : TASK_UNINTERRUPTIBLE;
        DEFINE_WAIT(wait);
        unsigned long timeout_expire;
        if (ret == 0)
                goto out;
  
 -      if (!irq_test_in_progress && WARN_ON(!ring->irq_get(ring))) {
 +      if (!irq_test_in_progress && WARN_ON(!engine->irq_get(engine))) {
                ret = -ENODEV;
                goto out;
        }
        for (;;) {
                struct timer_list timer;
  
 -              prepare_to_wait(&ring->irq_queue, &wait, state);
 +              prepare_to_wait(&engine->irq_queue, &wait, state);
  
                /* We need to check whether any gpu reset happened in between
                 * the caller grabbing the seqno and now ... */
                }
  
                timer.function = NULL;
 -              if (timeout || missed_irq(dev_priv, ring)) {
 +              if (timeout || missed_irq(dev_priv, engine)) {
                        unsigned long expire;
  
                        setup_timer_on_stack(&timer, fake_irq, (unsigned long)current);
 -                      expire = missed_irq(dev_priv, ring) ? jiffies + 1 : timeout_expire;
 +                      expire = missed_irq(dev_priv, engine) ? jiffies + 1 : timeout_expire;
                        mod_timer(&timer, expire);
                }
  
                }
        }
        if (!irq_test_in_progress)
 -              ring->irq_put(ring);
 +              engine->irq_put(engine);
  
 -      finish_wait(&ring->irq_queue, &wait);
 +      finish_wait(&engine->irq_queue, &wait);
  
  out:
        trace_i915_gem_request_wait_end(req);
  int i915_gem_request_add_to_client(struct drm_i915_gem_request *req,
                                   struct drm_file *file)
  {
 -      struct drm_i915_private *dev_private;
        struct drm_i915_file_private *file_priv;
  
        WARN_ON(!req || !file || req->file_priv);
        if (req->file_priv)
                return -EINVAL;
  
 -      dev_private = req->ring->dev->dev_private;
        file_priv = file->driver_priv;
  
        spin_lock(&file_priv->mm.lock);
@@@ -1433,7 -1434,7 +1433,7 @@@ static void i915_gem_request_retire(str
  static void
  __i915_gem_request_retire__upto(struct drm_i915_gem_request *req)
  {
 -      struct intel_engine_cs *engine = req->ring;
 +      struct intel_engine_cs *engine = req->engine;
        struct drm_i915_gem_request *tmp;
  
        lockdep_assert_held(&engine->dev->struct_mutex);
@@@ -1465,7 -1466,7 +1465,7 @@@ i915_wait_request(struct drm_i915_gem_r
  
        BUG_ON(req == NULL);
  
 -      dev = req->ring->dev;
 +      dev = req->engine->dev;
        dev_priv = dev->dev_private;
        interruptible = dev_priv->mm.interruptible;
  
@@@ -1504,14 -1505,14 +1504,14 @@@ i915_gem_object_wait_rendering(struct d
                        if (ret)
                                return ret;
  
 -                      i = obj->last_write_req->ring->id;
 +                      i = obj->last_write_req->engine->id;
                        if (obj->last_read_req[i] == obj->last_write_req)
                                i915_gem_object_retire__read(obj, i);
                        else
                                i915_gem_object_retire__write(obj);
                }
        } else {
 -              for (i = 0; i < I915_NUM_RINGS; i++) {
 +              for (i = 0; i < I915_NUM_ENGINES; i++) {
                        if (obj->last_read_req[i] == NULL)
                                continue;
  
@@@ -1531,7 -1532,7 +1531,7 @@@ static voi
  i915_gem_object_retire_request(struct drm_i915_gem_object *obj,
                               struct drm_i915_gem_request *req)
  {
 -      int ring = req->ring->id;
 +      int ring = req->engine->id;
  
        if (obj->last_read_req[ring] == req)
                i915_gem_object_retire__read(obj, ring);
@@@ -1551,7 -1552,7 +1551,7 @@@ i915_gem_object_wait_rendering__nonbloc
  {
        struct drm_device *dev = obj->base.dev;
        struct drm_i915_private *dev_priv = dev->dev_private;
 -      struct drm_i915_gem_request *requests[I915_NUM_RINGS];
 +      struct drm_i915_gem_request *requests[I915_NUM_ENGINES];
        unsigned reset_counter;
        int ret, i, n = 0;
  
  
                requests[n++] = i915_gem_request_reference(req);
        } else {
 -              for (i = 0; i < I915_NUM_RINGS; i++) {
 +              for (i = 0; i < I915_NUM_ENGINES; i++) {
                        struct drm_i915_gem_request *req;
  
                        req = obj->last_read_req[i];
@@@ -1791,8 -1792,7 +1791,8 @@@ int i915_gem_fault(struct vm_area_struc
  {
        struct drm_i915_gem_object *obj = to_intel_bo(vma->vm_private_data);
        struct drm_device *dev = obj->base.dev;
 -      struct drm_i915_private *dev_priv = dev->dev_private;
 +      struct drm_i915_private *dev_priv = to_i915(dev);
 +      struct i915_ggtt *ggtt = &dev_priv->ggtt;
        struct i915_ggtt_view view = i915_ggtt_view_normal;
        pgoff_t page_offset;
        unsigned long pfn;
        }
  
        /* Use a partial view if the object is bigger than the aperture. */
 -      if (obj->base.size >= dev_priv->gtt.mappable_end &&
 +      if (obj->base.size >= ggtt->mappable_end &&
            obj->tiling_mode == I915_TILING_NONE) {
                static const unsigned int chunk_size = 256; // 1 MiB
  
                goto unpin;
  
        /* Finally, remap it using the new GTT offset */
 -      pfn = dev_priv->gtt.mappable_base +
 +      pfn = ggtt->mappable_base +
                i915_gem_obj_ggtt_offset_view(obj, &view);
        pfn >>= PAGE_SHIFT;
  
@@@ -2206,7 -2206,7 +2206,7 @@@ i915_gem_object_put_pages_gtt(struct dr
                if (obj->madv == I915_MADV_WILLNEED)
                        mark_page_accessed(page);
  
-               page_cache_release(page);
+               put_page(page);
        }
        obj->dirty = 0;
  
@@@ -2232,14 -2232,6 +2232,14 @@@ i915_gem_object_put_pages(struct drm_i9
         * lists early. */
        list_del(&obj->global_list);
  
 +      if (obj->mapping) {
 +              if (is_vmalloc_addr(obj->mapping))
 +                      vunmap(obj->mapping);
 +              else
 +                      kunmap(kmap_to_page(obj->mapping));
 +              obj->mapping = NULL;
 +      }
 +
        ops->put_pages(obj);
        obj->pages = NULL;
  
@@@ -2354,7 -2346,7 +2354,7 @@@ i915_gem_object_get_pages_gtt(struct dr
  err_pages:
        sg_mark_end(sg);
        for_each_sg_page(st->sgl, &sg_iter, st->nents, 0)
-               page_cache_release(sg_page_iter_page(&sg_iter));
+               put_page(sg_page_iter_page(&sg_iter));
        sg_free_table(st);
        kfree(st);
  
@@@ -2408,64 -2400,21 +2408,64 @@@ i915_gem_object_get_pages(struct drm_i9
        return 0;
  }
  
 +void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj)
 +{
 +      int ret;
 +
 +      lockdep_assert_held(&obj->base.dev->struct_mutex);
 +
 +      ret = i915_gem_object_get_pages(obj);
 +      if (ret)
 +              return ERR_PTR(ret);
 +
 +      i915_gem_object_pin_pages(obj);
 +
 +      if (obj->mapping == NULL) {
 +              struct page **pages;
 +
 +              pages = NULL;
 +              if (obj->base.size == PAGE_SIZE)
 +                      obj->mapping = kmap(sg_page(obj->pages->sgl));
 +              else
 +                      pages = drm_malloc_gfp(obj->base.size >> PAGE_SHIFT,
 +                                             sizeof(*pages),
 +                                             GFP_TEMPORARY);
 +              if (pages != NULL) {
 +                      struct sg_page_iter sg_iter;
 +                      int n;
 +
 +                      n = 0;
 +                      for_each_sg_page(obj->pages->sgl, &sg_iter,
 +                                       obj->pages->nents, 0)
 +                              pages[n++] = sg_page_iter_page(&sg_iter);
 +
 +                      obj->mapping = vmap(pages, n, 0, PAGE_KERNEL);
 +                      drm_free_large(pages);
 +              }
 +              if (obj->mapping == NULL) {
 +                      i915_gem_object_unpin_pages(obj);
 +                      return ERR_PTR(-ENOMEM);
 +              }
 +      }
 +
 +      return obj->mapping;
 +}
 +
  void i915_vma_move_to_active(struct i915_vma *vma,
                             struct drm_i915_gem_request *req)
  {
        struct drm_i915_gem_object *obj = vma->obj;
 -      struct intel_engine_cs *ring;
 +      struct intel_engine_cs *engine;
  
 -      ring = i915_gem_request_get_ring(req);
 +      engine = i915_gem_request_get_engine(req);
  
        /* Add a reference if we're newly entering the active list. */
        if (obj->active == 0)
                drm_gem_object_reference(&obj->base);
 -      obj->active |= intel_ring_flag(ring);
 +      obj->active |= intel_engine_flag(engine);
  
 -      list_move_tail(&obj->ring_list[ring->id], &ring->active_list);
 -      i915_gem_request_assign(&obj->last_read_req[ring->id], req);
 +      list_move_tail(&obj->engine_list[engine->id], &engine->active_list);
 +      i915_gem_request_assign(&obj->last_read_req[engine->id], req);
  
        list_move_tail(&vma->vm_link, &vma->vm->active_list);
  }
@@@ -2474,7 -2423,7 +2474,7 @@@ static voi
  i915_gem_object_retire__write(struct drm_i915_gem_object *obj)
  {
        RQ_BUG_ON(obj->last_write_req == NULL);
 -      RQ_BUG_ON(!(obj->active & intel_ring_flag(obj->last_write_req->ring)));
 +      RQ_BUG_ON(!(obj->active & intel_engine_flag(obj->last_write_req->engine)));
  
        i915_gem_request_assign(&obj->last_write_req, NULL);
        intel_fb_obj_flush(obj, true, ORIGIN_CS);
@@@ -2488,10 -2437,10 +2488,10 @@@ i915_gem_object_retire__read(struct drm
        RQ_BUG_ON(obj->last_read_req[ring] == NULL);
        RQ_BUG_ON(!(obj->active & (1 << ring)));
  
 -      list_del_init(&obj->ring_list[ring]);
 +      list_del_init(&obj->engine_list[ring]);
        i915_gem_request_assign(&obj->last_read_req[ring], NULL);
  
 -      if (obj->last_write_req && obj->last_write_req->ring->id == ring)
 +      if (obj->last_write_req && obj->last_write_req->engine->id == ring)
                i915_gem_object_retire__write(obj);
  
        obj->active &= ~(1 << ring);
@@@ -2518,20 -2467,24 +2518,20 @@@ static in
  i915_gem_init_seqno(struct drm_device *dev, u32 seqno)
  {
        struct drm_i915_private *dev_priv = dev->dev_private;
 -      struct intel_engine_cs *ring;
 -      int ret, i, j;
 +      struct intel_engine_cs *engine;
 +      int ret;
  
        /* Carefully retire all requests without writing to the rings */
 -      for_each_ring(ring, dev_priv, i) {
 -              ret = intel_ring_idle(ring);
 +      for_each_engine(engine, dev_priv) {
 +              ret = intel_engine_idle(engine);
                if (ret)
                        return ret;
        }
        i915_gem_retire_requests(dev);
  
        /* Finally reset hw state */
 -      for_each_ring(ring, dev_priv, i) {
 -              intel_ring_init_seqno(ring, seqno);
 -
 -              for (j = 0; j < ARRAY_SIZE(ring->semaphore.sync_seqno); j++)
 -                      ring->semaphore.sync_seqno[j] = 0;
 -      }
 +      for_each_engine(engine, dev_priv)
 +              intel_ring_init_seqno(engine, seqno);
  
        return 0;
  }
@@@ -2589,7 -2542,7 +2589,7 @@@ void __i915_add_request(struct drm_i915
                        struct drm_i915_gem_object *obj,
                        bool flush_caches)
  {
 -      struct intel_engine_cs *ring;
 +      struct intel_engine_cs *engine;
        struct drm_i915_private *dev_priv;
        struct intel_ringbuffer *ringbuf;
        u32 request_start;
        if (WARN_ON(request == NULL))
                return;
  
 -      ring = request->ring;
 -      dev_priv = ring->dev->dev_private;
 +      engine = request->engine;
 +      dev_priv = request->i915;
        ringbuf = request->ringbuf;
  
        /*
                WARN(ret, "*_ring_flush_all_caches failed: %d!\n", ret);
        }
  
 +      trace_i915_gem_request_add(request);
 +
 +      request->head = request_start;
 +
 +      /* Whilst this request exists, batch_obj will be on the
 +       * active_list, and so will hold the active reference. Only when this
 +       * request is retired will the the batch_obj be moved onto the
 +       * inactive_list and lose its active reference. Hence we do not need
 +       * to explicitly hold another reference here.
 +       */
 +      request->batch_obj = obj;
 +
 +      /* Seal the request and mark it as pending execution. Note that
 +       * we may inspect this state, without holding any locks, during
 +       * hangcheck. Hence we apply the barrier to ensure that we do not
 +       * see a more recent value in the hws than we are tracking.
 +       */
 +      request->emitted_jiffies = jiffies;
 +      request->previous_seqno = engine->last_submitted_seqno;
 +      smp_store_mb(engine->last_submitted_seqno, request->seqno);
 +      list_add_tail(&request->list, &engine->request_list);
 +
        /* Record the position of the start of the request so that
         * should we detect the updated seqno part-way through the
         * GPU processing the request, we never over-estimate the
        request->postfix = intel_ring_get_tail(ringbuf);
  
        if (i915.enable_execlists)
 -              ret = ring->emit_request(request);
 +              ret = engine->emit_request(request);
        else {
 -              ret = ring->add_request(request);
 +              ret = engine->add_request(request);
  
                request->tail = intel_ring_get_tail(ringbuf);
        }
        /* Not allowed to fail! */
        WARN(ret, "emit|add_request failed: %d!\n", ret);
  
 -      request->head = request_start;
 -
 -      /* Whilst this request exists, batch_obj will be on the
 -       * active_list, and so will hold the active reference. Only when this
 -       * request is retired will the the batch_obj be moved onto the
 -       * inactive_list and lose its active reference. Hence we do not need
 -       * to explicitly hold another reference here.
 -       */
 -      request->batch_obj = obj;
 -
 -      request->emitted_jiffies = jiffies;
 -      request->previous_seqno = ring->last_submitted_seqno;
 -      ring->last_submitted_seqno = request->seqno;
 -      list_add_tail(&request->list, &ring->request_list);
 -
 -      trace_i915_gem_request_add(request);
 -
 -      i915_queue_hangcheck(ring->dev);
 +      i915_queue_hangcheck(engine->dev);
  
        queue_delayed_work(dev_priv->wq,
                           &dev_priv->mm.retire_work,
@@@ -2732,7 -2680,7 +2732,7 @@@ void i915_gem_request_free(struct kref 
  
        if (ctx) {
                if (i915.enable_execlists && ctx != req->i915->kernel_context)
 -                      intel_lr_context_unpin(ctx, req->ring);
 +                      intel_lr_context_unpin(ctx, req->engine);
  
                i915_gem_context_unreference(ctx);
        }
  }
  
  static inline int
 -__i915_gem_request_alloc(struct intel_engine_cs *ring,
 +__i915_gem_request_alloc(struct intel_engine_cs *engine,
                         struct intel_context *ctx,
                         struct drm_i915_gem_request **req_out)
  {
 -      struct drm_i915_private *dev_priv = to_i915(ring->dev);
 +      struct drm_i915_private *dev_priv = to_i915(engine->dev);
        struct drm_i915_gem_request *req;
        int ret;
  
        if (req == NULL)
                return -ENOMEM;
  
 -      ret = i915_gem_get_seqno(ring->dev, &req->seqno);
 +      ret = i915_gem_get_seqno(engine->dev, &req->seqno);
        if (ret)
                goto err;
  
        kref_init(&req->ref);
        req->i915 = dev_priv;
 -      req->ring = ring;
 +      req->engine = engine;
        req->ctx  = ctx;
        i915_gem_context_reference(req->ctx);
  
@@@ -2839,11 -2787,11 +2839,11 @@@ void i915_gem_request_cancel(struct drm
  }
  
  struct drm_i915_gem_request *
 -i915_gem_find_active_request(struct intel_engine_cs *ring)
 +i915_gem_find_active_request(struct intel_engine_cs *engine)
  {
        struct drm_i915_gem_request *request;
  
 -      list_for_each_entry(request, &ring->request_list, list) {
 +      list_for_each_entry(request, &engine->request_list, list) {
                if (i915_gem_request_completed(request, false))
                        continue;
  
        return NULL;
  }
  
 -static void i915_gem_reset_ring_status(struct drm_i915_private *dev_priv,
 -                                     struct intel_engine_cs *ring)
 +static void i915_gem_reset_engine_status(struct drm_i915_private *dev_priv,
 +                                     struct intel_engine_cs *engine)
  {
        struct drm_i915_gem_request *request;
        bool ring_hung;
  
 -      request = i915_gem_find_active_request(ring);
 +      request = i915_gem_find_active_request(engine);
  
        if (request == NULL)
                return;
  
 -      ring_hung = ring->hangcheck.score >= HANGCHECK_SCORE_RING_HUNG;
 +      ring_hung = engine->hangcheck.score >= HANGCHECK_SCORE_RING_HUNG;
  
        i915_set_reset_status(dev_priv, request->ctx, ring_hung);
  
 -      list_for_each_entry_continue(request, &ring->request_list, list)
 +      list_for_each_entry_continue(request, &engine->request_list, list)
                i915_set_reset_status(dev_priv, request->ctx, false);
  }
  
 -static void i915_gem_reset_ring_cleanup(struct drm_i915_private *dev_priv,
 -                                      struct intel_engine_cs *ring)
 +static void i915_gem_reset_engine_cleanup(struct drm_i915_private *dev_priv,
 +                                      struct intel_engine_cs *engine)
  {
        struct intel_ringbuffer *buffer;
  
 -      while (!list_empty(&ring->active_list)) {
 +      while (!list_empty(&engine->active_list)) {
                struct drm_i915_gem_object *obj;
  
 -              obj = list_first_entry(&ring->active_list,
 +              obj = list_first_entry(&engine->active_list,
                                       struct drm_i915_gem_object,
 -                                     ring_list[ring->id]);
 +                                     engine_list[engine->id]);
  
 -              i915_gem_object_retire__read(obj, ring->id);
 +              i915_gem_object_retire__read(obj, engine->id);
        }
  
        /*
         */
  
        if (i915.enable_execlists) {
 -              spin_lock_irq(&ring->execlist_lock);
 +              /* Ensure irq handler finishes or is cancelled. */
 +              tasklet_kill(&engine->irq_tasklet);
  
 +              spin_lock_bh(&engine->execlist_lock);
                /* list_splice_tail_init checks for empty lists */
 -              list_splice_tail_init(&ring->execlist_queue,
 -                                    &ring->execlist_retired_req_list);
 +              list_splice_tail_init(&engine->execlist_queue,
 +                                    &engine->execlist_retired_req_list);
 +              spin_unlock_bh(&engine->execlist_lock);
  
 -              spin_unlock_irq(&ring->execlist_lock);
 -              intel_execlists_retire_requests(ring);
 +              intel_execlists_retire_requests(engine);
        }
  
        /*
         * implicit references on things like e.g. ppgtt address spaces through
         * the request.
         */
 -      while (!list_empty(&ring->request_list)) {
 +      while (!list_empty(&engine->request_list)) {
                struct drm_i915_gem_request *request;
  
 -              request = list_first_entry(&ring->request_list,
 +              request = list_first_entry(&engine->request_list,
                                           struct drm_i915_gem_request,
                                           list);
  
         * upon reset is less than when we start. Do one more pass over
         * all the ringbuffers to reset last_retired_head.
         */
 -      list_for_each_entry(buffer, &ring->buffers, link) {
 +      list_for_each_entry(buffer, &engine->buffers, link) {
                buffer->last_retired_head = buffer->tail;
                intel_ring_update_space(buffer);
        }
 +
 +      intel_ring_init_seqno(engine, engine->last_submitted_seqno);
  }
  
  void i915_gem_reset(struct drm_device *dev)
  {
        struct drm_i915_private *dev_priv = dev->dev_private;
 -      struct intel_engine_cs *ring;
 -      int i;
 +      struct intel_engine_cs *engine;
  
        /*
         * Before we free the objects from the requests, we need to inspect
         * them for finding the guilty party. As the requests only borrow
         * their reference to the objects, the inspection must be done first.
         */
 -      for_each_ring(ring, dev_priv, i)
 -              i915_gem_reset_ring_status(dev_priv, ring);
 +      for_each_engine(engine, dev_priv)
 +              i915_gem_reset_engine_status(dev_priv, engine);
  
 -      for_each_ring(ring, dev_priv, i)
 -              i915_gem_reset_ring_cleanup(dev_priv, ring);
 +      for_each_engine(engine, dev_priv)
 +              i915_gem_reset_engine_cleanup(dev_priv, engine);
  
        i915_gem_context_reset(dev);
  
   * This function clears the request list as sequence numbers are passed.
   */
  void
 -i915_gem_retire_requests_ring(struct intel_engine_cs *ring)
 +i915_gem_retire_requests_ring(struct intel_engine_cs *engine)
  {
 -      WARN_ON(i915_verify_lists(ring->dev));
 +      WARN_ON(i915_verify_lists(engine->dev));
  
        /* Retire requests first as we use it above for the early return.
         * If we retire requests last, we may use a later seqno and so clear
         * the requests lists without clearing the active list, leading to
         * confusion.
         */
 -      while (!list_empty(&ring->request_list)) {
 +      while (!list_empty(&engine->request_list)) {
                struct drm_i915_gem_request *request;
  
 -              request = list_first_entry(&ring->request_list,
 +              request = list_first_entry(&engine->request_list,
                                           struct drm_i915_gem_request,
                                           list);
  
         * by the ringbuffer to the flushing/inactive lists as appropriate,
         * before we free the context associated with the requests.
         */
 -      while (!list_empty(&ring->active_list)) {
 +      while (!list_empty(&engine->active_list)) {
                struct drm_i915_gem_object *obj;
  
 -              obj = list_first_entry(&ring->active_list,
 -                                    struct drm_i915_gem_object,
 -                                    ring_list[ring->id]);
 +              obj = list_first_entry(&engine->active_list,
 +                                     struct drm_i915_gem_object,
 +                                     engine_list[engine->id]);
  
 -              if (!list_empty(&obj->last_read_req[ring->id]->list))
 +              if (!list_empty(&obj->last_read_req[engine->id]->list))
                        break;
  
 -              i915_gem_object_retire__read(obj, ring->id);
 +              i915_gem_object_retire__read(obj, engine->id);
        }
  
 -      if (unlikely(ring->trace_irq_req &&
 -                   i915_gem_request_completed(ring->trace_irq_req, true))) {
 -              ring->irq_put(ring);
 -              i915_gem_request_assign(&ring->trace_irq_req, NULL);
 +      if (unlikely(engine->trace_irq_req &&
 +                   i915_gem_request_completed(engine->trace_irq_req, true))) {
 +              engine->irq_put(engine);
 +              i915_gem_request_assign(&engine->trace_irq_req, NULL);
        }
  
 -      WARN_ON(i915_verify_lists(ring->dev));
 +      WARN_ON(i915_verify_lists(engine->dev));
  }
  
  bool
  i915_gem_retire_requests(struct drm_device *dev)
  {
        struct drm_i915_private *dev_priv = dev->dev_private;
 -      struct intel_engine_cs *ring;
 +      struct intel_engine_cs *engine;
        bool idle = true;
 -      int i;
  
 -      for_each_ring(ring, dev_priv, i) {
 -              i915_gem_retire_requests_ring(ring);
 -              idle &= list_empty(&ring->request_list);
 +      for_each_engine(engine, dev_priv) {
 +              i915_gem_retire_requests_ring(engine);
 +              idle &= list_empty(&engine->request_list);
                if (i915.enable_execlists) {
 -                      spin_lock_irq(&ring->execlist_lock);
 -                      idle &= list_empty(&ring->execlist_queue);
 -                      spin_unlock_irq(&ring->execlist_lock);
 +                      spin_lock_bh(&engine->execlist_lock);
 +                      idle &= list_empty(&engine->execlist_queue);
 +                      spin_unlock_bh(&engine->execlist_lock);
  
 -                      intel_execlists_retire_requests(ring);
 +                      intel_execlists_retire_requests(engine);
                }
        }
  
@@@ -3065,21 -3011,25 +3065,21 @@@ i915_gem_idle_work_handler(struct work_
        struct drm_i915_private *dev_priv =
                container_of(work, typeof(*dev_priv), mm.idle_work.work);
        struct drm_device *dev = dev_priv->dev;
 -      struct intel_engine_cs *ring;
 -      int i;
 +      struct intel_engine_cs *engine;
  
 -      for_each_ring(ring, dev_priv, i)
 -              if (!list_empty(&ring->request_list))
 +      for_each_engine(engine, dev_priv)
 +              if (!list_empty(&engine->request_list))
                        return;
  
        /* we probably should sync with hangcheck here, using cancel_work_sync.
 -       * Also locking seems to be fubar here, ring->request_list is protected
 +       * Also locking seems to be fubar here, engine->request_list is protected
         * by dev->struct_mutex. */
  
        intel_mark_idle(dev);
  
        if (mutex_trylock(&dev->struct_mutex)) {
 -              struct intel_engine_cs *ring;
 -              int i;
 -
 -              for_each_ring(ring, dev_priv, i)
 -                      i915_gem_batch_pool_fini(&ring->batch_pool);
 +              for_each_engine(engine, dev_priv)
 +                      i915_gem_batch_pool_fini(&engine->batch_pool);
  
                mutex_unlock(&dev->struct_mutex);
        }
@@@ -3098,7 -3048,7 +3098,7 @@@ i915_gem_object_flush_active(struct drm
        if (!obj->active)
                return 0;
  
 -      for (i = 0; i < I915_NUM_RINGS; i++) {
 +      for (i = 0; i < I915_NUM_ENGINES; i++) {
                struct drm_i915_gem_request *req;
  
                req = obj->last_read_req[i];
@@@ -3146,7 -3096,7 +3146,7 @@@ i915_gem_wait_ioctl(struct drm_device *
        struct drm_i915_private *dev_priv = dev->dev_private;
        struct drm_i915_gem_wait *args = data;
        struct drm_i915_gem_object *obj;
 -      struct drm_i915_gem_request *req[I915_NUM_RINGS];
 +      struct drm_i915_gem_request *req[I915_NUM_ENGINES];
        unsigned reset_counter;
        int i, n = 0;
        int ret;
        drm_gem_object_unreference(&obj->base);
        reset_counter = atomic_read(&dev_priv->gpu_error.reset_counter);
  
 -      for (i = 0; i < I915_NUM_RINGS; i++) {
 +      for (i = 0; i < I915_NUM_ENGINES; i++) {
                if (obj->last_read_req[i] == NULL)
                        continue;
  
@@@ -3216,7 -3166,7 +3216,7 @@@ __i915_gem_object_sync(struct drm_i915_
        struct intel_engine_cs *from;
        int ret;
  
 -      from = i915_gem_request_get_ring(from_req);
 +      from = i915_gem_request_get_engine(from_req);
        if (to == from)
                return 0;
  
@@@ -3310,7 -3260,7 +3310,7 @@@ i915_gem_object_sync(struct drm_i915_ge
                     struct drm_i915_gem_request **to_req)
  {
        const bool readonly = obj->base.pending_write_domain == 0;
 -      struct drm_i915_gem_request *req[I915_NUM_RINGS];
 +      struct drm_i915_gem_request *req[I915_NUM_ENGINES];
        int ret, i, n;
  
        if (!obj->active)
                if (obj->last_write_req)
                        req[n++] = obj->last_write_req;
        } else {
 -              for (i = 0; i < I915_NUM_RINGS; i++)
 +              for (i = 0; i < I915_NUM_ENGINES; i++)
                        if (obj->last_read_req[i])
                                req[n++] = obj->last_read_req[i];
        }
@@@ -3441,15 -3391,15 +3441,15 @@@ int __i915_vma_unbind_no_wait(struct i9
  int i915_gpu_idle(struct drm_device *dev)
  {
        struct drm_i915_private *dev_priv = dev->dev_private;
 -      struct intel_engine_cs *ring;
 -      int ret, i;
 +      struct intel_engine_cs *engine;
 +      int ret;
  
        /* Flush everything onto the inactive list. */
 -      for_each_ring(ring, dev_priv, i) {
 +      for_each_engine(engine, dev_priv) {
                if (!i915.enable_execlists) {
                        struct drm_i915_gem_request *req;
  
 -                      req = i915_gem_request_alloc(ring, NULL);
 +                      req = i915_gem_request_alloc(engine, NULL);
                        if (IS_ERR(req))
                                return PTR_ERR(req);
  
                        i915_add_request_no_flush(req);
                }
  
 -              ret = intel_ring_idle(ring);
 +              ret = intel_engine_idle(engine);
                if (ret)
                        return ret;
        }
@@@ -3516,8 -3466,7 +3516,8 @@@ i915_gem_object_bind_to_vm(struct drm_i
                           uint64_t flags)
  {
        struct drm_device *dev = obj->base.dev;
 -      struct drm_i915_private *dev_priv = dev->dev_private;
 +      struct drm_i915_private *dev_priv = to_i915(dev);
 +      struct i915_ggtt *ggtt = &dev_priv->ggtt;
        u32 fence_alignment, unfenced_alignment;
        u32 search_flag, alloc_flag;
        u64 start, end;
        start = flags & PIN_OFFSET_BIAS ? flags & PIN_OFFSET_MASK : 0;
        end = vm->total;
        if (flags & PIN_MAPPABLE)
 -              end = min_t(u64, end, dev_priv->gtt.mappable_end);
 +              end = min_t(u64, end, ggtt->mappable_end);
        if (flags & PIN_ZONE_4G)
                end = min_t(u64, end, (1ULL << 32) - PAGE_SIZE);
  
@@@ -3771,9 -3720,6 +3771,9 @@@ i915_gem_object_flush_cpu_write_domain(
  int
  i915_gem_object_set_to_gtt_domain(struct drm_i915_gem_object *obj, bool write)
  {
 +      struct drm_device *dev = obj->base.dev;
 +      struct drm_i915_private *dev_priv = to_i915(dev);
 +      struct i915_ggtt *ggtt = &dev_priv->ggtt;
        uint32_t old_write_domain, old_read_domains;
        struct i915_vma *vma;
        int ret;
        vma = i915_gem_obj_to_ggtt(obj);
        if (vma && drm_mm_node_allocated(&vma->node) && !obj->active)
                list_move_tail(&vma->vm_link,
 -                             &to_i915(obj->base.dev)->gtt.base.inactive_list);
 +                             &ggtt->base.inactive_list);
  
        return 0;
  }
@@@ -4003,7 -3949,7 +4003,7 @@@ int i915_gem_set_caching_ioctl(struct d
                 * cacheline, whereas normally such cachelines would get
                 * invalidated.
                 */
 -              if (IS_BXT_REVID(dev, 0, BXT_REVID_A1))
 +              if (!HAS_LLC(dev) && !HAS_SNOOP(dev))
                        return -ENODEV;
  
                level = I915_CACHE_LLC;
@@@ -4265,7 -4211,7 +4265,7 @@@ void __i915_vma_set_map_and_fenceable(s
                     (vma->node.start & (fence_alignment - 1)) == 0);
  
        mappable = (vma->node.start + fence_size <=
 -                  to_i915(obj->base.dev)->gtt.mappable_end);
 +                  to_i915(obj->base.dev)->ggtt.mappable_end);
  
        obj->map_and_fenceable = mappable && fenceable;
  }
@@@ -4297,6 -4243,9 +4297,6 @@@ i915_gem_object_do_pin(struct drm_i915_
        vma = ggtt_view ? i915_gem_obj_to_ggtt_view(obj, ggtt_view) :
                          i915_gem_obj_to_vma(obj, vm);
  
 -      if (IS_ERR(vma))
 -              return PTR_ERR(vma);
 -
        if (vma) {
                if (WARN_ON(vma->pin_count == DRM_I915_GEM_OBJECT_MAX_PIN_COUNT))
                        return -EBUSY;
@@@ -4359,13 -4308,10 +4359,13 @@@ i915_gem_object_ggtt_pin(struct drm_i91
                         uint32_t alignment,
                         uint64_t flags)
  {
 -      if (WARN_ONCE(!view, "no view specified"))
 -              return -EINVAL;
 +      struct drm_device *dev = obj->base.dev;
 +      struct drm_i915_private *dev_priv = to_i915(dev);
 +      struct i915_ggtt *ggtt = &dev_priv->ggtt;
 +
 +      BUG_ON(!view);
  
 -      return i915_gem_object_do_pin(obj, i915_obj_to_ggtt(obj), view,
 +      return i915_gem_object_do_pin(obj, &ggtt->base, view,
                                      alignment, flags | PIN_GLOBAL);
  }
  
@@@ -4413,15 -4359,15 +4413,15 @@@ i915_gem_busy_ioctl(struct drm_device *
        if (obj->active) {
                int i;
  
 -              for (i = 0; i < I915_NUM_RINGS; i++) {
 +              for (i = 0; i < I915_NUM_ENGINES; i++) {
                        struct drm_i915_gem_request *req;
  
                        req = obj->last_read_req[i];
                        if (req)
 -                              args->busy |= 1 << (16 + req->ring->exec_id);
 +                              args->busy |= 1 << (16 + req->engine->exec_id);
                }
                if (obj->last_write_req)
 -                      args->busy |= obj->last_write_req->ring->exec_id;
 +                      args->busy |= obj->last_write_req->engine->exec_id;
        }
  
  unref:
@@@ -4501,8 -4447,8 +4501,8 @@@ void i915_gem_object_init(struct drm_i9
        int i;
  
        INIT_LIST_HEAD(&obj->global_list);
 -      for (i = 0; i < I915_NUM_RINGS; i++)
 -              INIT_LIST_HEAD(&obj->ring_list[i]);
 +      for (i = 0; i < I915_NUM_ENGINES; i++)
 +              INIT_LIST_HEAD(&obj->engine_list[i]);
        INIT_LIST_HEAD(&obj->obj_exec_link);
        INIT_LIST_HEAD(&obj->vma_list);
        INIT_LIST_HEAD(&obj->batch_pool_link);
@@@ -4677,15 -4623,14 +4677,15 @@@ struct i915_vma *i915_gem_obj_to_vma(st
  struct i915_vma *i915_gem_obj_to_ggtt_view(struct drm_i915_gem_object *obj,
                                           const struct i915_ggtt_view *view)
  {
 -      struct i915_address_space *ggtt = i915_obj_to_ggtt(obj);
 +      struct drm_device *dev = obj->base.dev;
 +      struct drm_i915_private *dev_priv = to_i915(dev);
 +      struct i915_ggtt *ggtt = &dev_priv->ggtt;
        struct i915_vma *vma;
  
 -      if (WARN_ONCE(!view, "no view specified"))
 -              return ERR_PTR(-EINVAL);
 +      BUG_ON(!view);
  
        list_for_each_entry(vma, &obj->vma_list, obj_link)
 -              if (vma->vm == ggtt &&
 +              if (vma->vm == &ggtt->base &&
                    i915_ggtt_view_equal(&vma->ggtt_view, view))
                        return vma;
        return NULL;
@@@ -4708,13 -4653,14 +4708,13 @@@ void i915_gem_vma_destroy(struct i915_v
  }
  
  static void
 -i915_gem_stop_ringbuffers(struct drm_device *dev)
 +i915_gem_stop_engines(struct drm_device *dev)
  {
        struct drm_i915_private *dev_priv = dev->dev_private;
 -      struct intel_engine_cs *ring;
 -      int i;
 +      struct intel_engine_cs *engine;
  
 -      for_each_ring(ring, dev_priv, i)
 -              dev_priv->gt.stop_ring(ring);
 +      for_each_engine(engine, dev_priv)
 +              dev_priv->gt.stop_engine(engine);
  }
  
  int
@@@ -4730,7 -4676,7 +4730,7 @@@ i915_gem_suspend(struct drm_device *dev
  
        i915_gem_retire_requests(dev);
  
 -      i915_gem_stop_ringbuffers(dev);
 +      i915_gem_stop_engines(dev);
        mutex_unlock(&dev->struct_mutex);
  
        cancel_delayed_work_sync(&dev_priv->gpu_error.hangcheck_work);
@@@ -4751,8 -4697,8 +4751,8 @@@ err
  
  int i915_gem_l3_remap(struct drm_i915_gem_request *req, int slice)
  {
 -      struct intel_engine_cs *ring = req->ring;
 -      struct drm_device *dev = ring->dev;
 +      struct intel_engine_cs *engine = req->engine;
 +      struct drm_device *dev = engine->dev;
        struct drm_i915_private *dev_priv = dev->dev_private;
        u32 *remap_info = dev_priv->l3_parity.remap_info[slice];
        int i, ret;
         * at initialization time.
         */
        for (i = 0; i < GEN7_L3LOG_SIZE / 4; i++) {
 -              intel_ring_emit(ring, MI_LOAD_REGISTER_IMM(1));
 -              intel_ring_emit_reg(ring, GEN7_L3LOG(slice, i));
 -              intel_ring_emit(ring, remap_info[i]);
 +              intel_ring_emit(engine, MI_LOAD_REGISTER_IMM(1));
 +              intel_ring_emit_reg(engine, GEN7_L3LOG(slice, i));
 +              intel_ring_emit(engine, remap_info[i]);
        }
  
 -      intel_ring_advance(ring);
 +      intel_ring_advance(engine);
  
        return ret;
  }
@@@ -4832,7 -4778,7 +4832,7 @@@ static void init_unused_rings(struct dr
        }
  }
  
 -int i915_gem_init_rings(struct drm_device *dev)
 +int i915_gem_init_engines(struct drm_device *dev)
  {
        struct drm_i915_private *dev_priv = dev->dev_private;
        int ret;
        return 0;
  
  cleanup_vebox_ring:
 -      intel_cleanup_ring_buffer(&dev_priv->ring[VECS]);
 +      intel_cleanup_engine(&dev_priv->engine[VECS]);
  cleanup_blt_ring:
 -      intel_cleanup_ring_buffer(&dev_priv->ring[BCS]);
 +      intel_cleanup_engine(&dev_priv->engine[BCS]);
  cleanup_bsd_ring:
 -      intel_cleanup_ring_buffer(&dev_priv->ring[VCS]);
 +      intel_cleanup_engine(&dev_priv->engine[VCS]);
  cleanup_render_ring:
 -      intel_cleanup_ring_buffer(&dev_priv->ring[RCS]);
 +      intel_cleanup_engine(&dev_priv->engine[RCS]);
  
        return ret;
  }
@@@ -4883,8 -4829,8 +4883,8 @@@ in
  i915_gem_init_hw(struct drm_device *dev)
  {
        struct drm_i915_private *dev_priv = dev->dev_private;
 -      struct intel_engine_cs *ring;
 -      int ret, i, j;
 +      struct intel_engine_cs *engine;
 +      int ret, j;
  
        if (INTEL_INFO(dev)->gen < 6 && !intel_enable_gtt())
                return -EIO;
        }
  
        /* Need to do basic initialisation of all rings first: */
 -      for_each_ring(ring, dev_priv, i) {
 -              ret = ring->init_hw(ring);
 +      for_each_engine(engine, dev_priv) {
 +              ret = engine->init_hw(engine);
                if (ret)
                        goto out;
        }
                goto out;
  
        /* Now it is safe to go back round and do everything else: */
 -      for_each_ring(ring, dev_priv, i) {
 +      for_each_engine(engine, dev_priv) {
                struct drm_i915_gem_request *req;
  
 -              req = i915_gem_request_alloc(ring, NULL);
 +              req = i915_gem_request_alloc(engine, NULL);
                if (IS_ERR(req)) {
                        ret = PTR_ERR(req);
 -                      i915_gem_cleanup_ringbuffer(dev);
 +                      i915_gem_cleanup_engines(dev);
                        goto out;
                }
  
 -              if (ring->id == RCS) {
 +              if (engine->id == RCS) {
                        for (j = 0; j < NUM_L3_SLICES(dev); j++)
                                i915_gem_l3_remap(req, j);
                }
  
                ret = i915_ppgtt_init_ring(req);
                if (ret && ret != -EIO) {
 -                      DRM_ERROR("PPGTT enable ring #%d failed %d\n", i, ret);
 +                      DRM_ERROR("PPGTT enable %s failed %d\n",
 +                                engine->name, ret);
                        i915_gem_request_cancel(req);
 -                      i915_gem_cleanup_ringbuffer(dev);
 +                      i915_gem_cleanup_engines(dev);
                        goto out;
                }
  
                ret = i915_gem_context_enable(req);
                if (ret && ret != -EIO) {
 -                      DRM_ERROR("Context enable ring #%d failed %d\n", i, ret);
 +                      DRM_ERROR("Context enable %s failed %d\n",
 +                                engine->name, ret);
                        i915_gem_request_cancel(req);
 -                      i915_gem_cleanup_ringbuffer(dev);
 +                      i915_gem_cleanup_engines(dev);
                        goto out;
                }
  
@@@ -5008,14 -4952,14 +5008,14 @@@ int i915_gem_init(struct drm_device *de
  
        if (!i915.enable_execlists) {
                dev_priv->gt.execbuf_submit = i915_gem_ringbuffer_submission;
 -              dev_priv->gt.init_rings = i915_gem_init_rings;
 -              dev_priv->gt.cleanup_ring = intel_cleanup_ring_buffer;
 -              dev_priv->gt.stop_ring = intel_stop_ring_buffer;
 +              dev_priv->gt.init_engines = i915_gem_init_engines;
 +              dev_priv->gt.cleanup_engine = intel_cleanup_engine;
 +              dev_priv->gt.stop_engine = intel_stop_engine;
        } else {
                dev_priv->gt.execbuf_submit = intel_execlists_submission;
 -              dev_priv->gt.init_rings = intel_logical_rings_init;
 -              dev_priv->gt.cleanup_ring = intel_logical_ring_cleanup;
 -              dev_priv->gt.stop_ring = intel_logical_ring_stop;
 +              dev_priv->gt.init_engines = intel_logical_rings_init;
 +              dev_priv->gt.cleanup_engine = intel_logical_ring_cleanup;
 +              dev_priv->gt.stop_engine = intel_logical_ring_stop;
        }
  
        /* This is just a security blanket to placate dragons.
        if (ret)
                goto out_unlock;
  
 -      i915_gem_init_global_gtt(dev);
 +      i915_gem_init_ggtt(dev);
  
        ret = i915_gem_context_init(dev);
        if (ret)
                goto out_unlock;
  
 -      ret = dev_priv->gt.init_rings(dev);
 +      ret = dev_priv->gt.init_engines(dev);
        if (ret)
                goto out_unlock;
  
@@@ -5059,52 -5003,29 +5059,52 @@@ out_unlock
  }
  
  void
 -i915_gem_cleanup_ringbuffer(struct drm_device *dev)
 +i915_gem_cleanup_engines(struct drm_device *dev)
  {
        struct drm_i915_private *dev_priv = dev->dev_private;
 -      struct intel_engine_cs *ring;
 -      int i;
 +      struct intel_engine_cs *engine;
  
 -      for_each_ring(ring, dev_priv, i)
 -              dev_priv->gt.cleanup_ring(ring);
 +      for_each_engine(engine, dev_priv)
 +              dev_priv->gt.cleanup_engine(engine);
  
 -    if (i915.enable_execlists)
 -            /*
 -             * Neither the BIOS, ourselves or any other kernel
 -             * expects the system to be in execlists mode on startup,
 -             * so we need to reset the GPU back to legacy mode.
 -             */
 -            intel_gpu_reset(dev);
 +      if (i915.enable_execlists)
 +              /*
 +               * Neither the BIOS, ourselves or any other kernel
 +               * expects the system to be in execlists mode on startup,
 +               * so we need to reset the GPU back to legacy mode.
 +               */
 +              intel_gpu_reset(dev, ALL_ENGINES);
  }
  
  static void
 -init_ring_lists(struct intel_engine_cs *ring)
 +init_engine_lists(struct intel_engine_cs *engine)
 +{
 +      INIT_LIST_HEAD(&engine->active_list);
 +      INIT_LIST_HEAD(&engine->request_list);
 +}
 +
 +void
 +i915_gem_load_init_fences(struct drm_i915_private *dev_priv)
  {
 -      INIT_LIST_HEAD(&ring->active_list);
 -      INIT_LIST_HEAD(&ring->request_list);
 +      struct drm_device *dev = dev_priv->dev;
 +
 +      if (INTEL_INFO(dev_priv)->gen >= 7 && !IS_VALLEYVIEW(dev_priv) &&
 +          !IS_CHERRYVIEW(dev_priv))
 +              dev_priv->num_fence_regs = 32;
 +      else if (INTEL_INFO(dev_priv)->gen >= 4 || IS_I945G(dev_priv) ||
 +               IS_I945GM(dev_priv) || IS_G33(dev_priv))
 +              dev_priv->num_fence_regs = 16;
 +      else
 +              dev_priv->num_fence_regs = 8;
 +
 +      if (intel_vgpu_active(dev))
 +              dev_priv->num_fence_regs =
 +                              I915_READ(vgtif_reg(avail_rs.fence_num));
 +
 +      /* Initialize fence registers to zero */
 +      i915_gem_restore_fences(dev);
 +
 +      i915_gem_detect_bit_6_swizzle(dev);
  }
  
  void
@@@ -5134,8 -5055,8 +5134,8 @@@ i915_gem_load_init(struct drm_device *d
        INIT_LIST_HEAD(&dev_priv->mm.unbound_list);
        INIT_LIST_HEAD(&dev_priv->mm.bound_list);
        INIT_LIST_HEAD(&dev_priv->mm.fence_list);
 -      for (i = 0; i < I915_NUM_RINGS; i++)
 -              init_ring_lists(&dev_priv->ring[i]);
 +      for (i = 0; i < I915_NUM_ENGINES; i++)
 +              init_engine_lists(&dev_priv->engine[i]);
        for (i = 0; i < I915_MAX_NUM_FENCES; i++)
                INIT_LIST_HEAD(&dev_priv->fence_regs[i].lru_list);
        INIT_DELAYED_WORK(&dev_priv->mm.retire_work,
  
        dev_priv->relative_constants_mode = I915_EXEC_CONSTANTS_REL_GENERAL;
  
 -      if (INTEL_INFO(dev)->gen >= 7 && !IS_VALLEYVIEW(dev) && !IS_CHERRYVIEW(dev))
 -              dev_priv->num_fence_regs = 32;
 -      else if (INTEL_INFO(dev)->gen >= 4 || IS_I945G(dev) || IS_I945GM(dev) || IS_G33(dev))
 -              dev_priv->num_fence_regs = 16;
 -      else
 -              dev_priv->num_fence_regs = 8;
 -
 -      if (intel_vgpu_active(dev))
 -              dev_priv->num_fence_regs =
 -                              I915_READ(vgtif_reg(avail_rs.fence_num));
 -
        /*
         * Set initial sequence number for requests.
         * Using this number allows the wraparound to happen early,
        dev_priv->next_seqno = ((u32)~0 - 0x1100);
        dev_priv->last_seqno = ((u32)~0 - 0x1101);
  
 -      /* Initialize fence registers to zero */
        INIT_LIST_HEAD(&dev_priv->mm.fence_list);
 -      i915_gem_restore_fences(dev);
  
 -      i915_gem_detect_bit_6_swizzle(dev);
        init_waitqueue_head(&dev_priv->pending_flip_queue);
  
        dev_priv->mm.interruptible = true;
@@@ -5278,12 -5213,11 +5278,12 @@@ u64 i915_gem_obj_offset(struct drm_i915
  u64 i915_gem_obj_ggtt_offset_view(struct drm_i915_gem_object *o,
                                  const struct i915_ggtt_view *view)
  {
 -      struct i915_address_space *ggtt = i915_obj_to_ggtt(o);
 +      struct drm_i915_private *dev_priv = to_i915(o->base.dev);
 +      struct i915_ggtt *ggtt = &dev_priv->ggtt;
        struct i915_vma *vma;
  
        list_for_each_entry(vma, &o->vma_list, obj_link)
 -              if (vma->vm == ggtt &&
 +              if (vma->vm == &ggtt->base &&
                    i915_ggtt_view_equal(&vma->ggtt_view, view))
                        return vma->node.start;
  
@@@ -5310,12 -5244,11 +5310,12 @@@ bool i915_gem_obj_bound(struct drm_i915
  bool i915_gem_obj_ggtt_bound_view(struct drm_i915_gem_object *o,
                                  const struct i915_ggtt_view *view)
  {
 -      struct i915_address_space *ggtt = i915_obj_to_ggtt(o);
 +      struct drm_i915_private *dev_priv = to_i915(o->base.dev);
 +      struct i915_ggtt *ggtt = &dev_priv->ggtt;
        struct i915_vma *vma;
  
        list_for_each_entry(vma, &o->vma_list, obj_link)
 -              if (vma->vm == ggtt &&
 +              if (vma->vm == &ggtt->base &&
                    i915_ggtt_view_equal(&vma->ggtt_view, view) &&
                    drm_mm_node_allocated(&vma->node))
                        return true;
@@@ -95,12 -95,14 +95,12 @@@ static void i915_gem_unmap_dma_buf(stru
  {
        struct drm_i915_gem_object *obj = dma_buf_to_obj(attachment->dmabuf);
  
 -      mutex_lock(&obj->base.dev->struct_mutex);
 -
        dma_unmap_sg(attachment->dev, sg->sgl, sg->nents, dir);
        sg_free_table(sg);
        kfree(sg);
  
 +      mutex_lock(&obj->base.dev->struct_mutex);
        i915_gem_object_unpin_pages(obj);
 -
        mutex_unlock(&obj->base.dev->struct_mutex);
  }
  
@@@ -108,17 -110,51 +108,17 @@@ static void *i915_gem_dmabuf_vmap(struc
  {
        struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
        struct drm_device *dev = obj->base.dev;
 -      struct sg_page_iter sg_iter;
 -      struct page **pages;
 -      int ret, i;
 +      void *addr;
 +      int ret;
  
        ret = i915_mutex_lock_interruptible(dev);
        if (ret)
                return ERR_PTR(ret);
  
 -      if (obj->dma_buf_vmapping) {
 -              obj->vmapping_count++;
 -              goto out_unlock;
 -      }
 -
 -      ret = i915_gem_object_get_pages(obj);
 -      if (ret)
 -              goto err;
 -
 -      i915_gem_object_pin_pages(obj);
 -
 -      ret = -ENOMEM;
 -
 -      pages = drm_malloc_ab(obj->base.size >> PAGE_SHIFT, sizeof(*pages));
 -      if (pages == NULL)
 -              goto err_unpin;
 -
 -      i = 0;
 -      for_each_sg_page(obj->pages->sgl, &sg_iter, obj->pages->nents, 0)
 -              pages[i++] = sg_page_iter_page(&sg_iter);
 -
 -      obj->dma_buf_vmapping = vmap(pages, i, 0, PAGE_KERNEL);
 -      drm_free_large(pages);
 -
 -      if (!obj->dma_buf_vmapping)
 -              goto err_unpin;
 -
 -      obj->vmapping_count = 1;
 -out_unlock:
 +      addr = i915_gem_object_pin_map(obj);
        mutex_unlock(&dev->struct_mutex);
 -      return obj->dma_buf_vmapping;
  
 -err_unpin:
 -      i915_gem_object_unpin_pages(obj);
 -err:
 -      mutex_unlock(&dev->struct_mutex);
 -      return ERR_PTR(ret);
 +      return addr;
  }
  
  static void i915_gem_dmabuf_vunmap(struct dma_buf *dma_buf, void *vaddr)
        struct drm_device *dev = obj->base.dev;
  
        mutex_lock(&dev->struct_mutex);
 -      if (--obj->vmapping_count == 0) {
 -              vunmap(obj->dma_buf_vmapping);
 -              obj->dma_buf_vmapping = NULL;
 -
 -              i915_gem_object_unpin_pages(obj);
 -      }
 +      i915_gem_object_unpin_map(obj);
        mutex_unlock(&dev->struct_mutex);
  }
  
@@@ -187,25 -228,20 +187,20 @@@ static int i915_gem_begin_cpu_access(st
        return ret;
  }
  
- static void i915_gem_end_cpu_access(struct dma_buf *dma_buf, enum dma_data_direction direction)
+ static int i915_gem_end_cpu_access(struct dma_buf *dma_buf, enum dma_data_direction direction)
  {
        struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf);
        struct drm_device *dev = obj->base.dev;
-       struct drm_i915_private *dev_priv = to_i915(dev);
-       bool was_interruptible;
        int ret;
  
-       mutex_lock(&dev->struct_mutex);
-       was_interruptible = dev_priv->mm.interruptible;
-       dev_priv->mm.interruptible = false;
+       ret = i915_mutex_lock_interruptible(dev);
+       if (ret)
+               return ret;
  
        ret = i915_gem_object_set_to_gtt_domain(obj, false);
-       dev_priv->mm.interruptible = was_interruptible;
        mutex_unlock(&dev->struct_mutex);
  
-       if (unlikely(ret))
-               DRM_ERROR("unable to flush buffer following CPU access; rendering may be corrupt\n");
+       return ret;
  }
  
  static const struct dma_buf_ops i915_dmabuf_ops =  {
@@@ -494,17 -494,20 +494,17 @@@ __i915_gem_userptr_get_pages_worker(str
        ret = -ENOMEM;
        pinned = 0;
  
 -      pvec = kmalloc(npages*sizeof(struct page *),
 -                     GFP_TEMPORARY | __GFP_NOWARN | __GFP_NORETRY);
 -      if (pvec == NULL)
 -              pvec = drm_malloc_ab(npages, sizeof(struct page *));
 +      pvec = drm_malloc_gfp(npages, sizeof(struct page *), GFP_TEMPORARY);
        if (pvec != NULL) {
                struct mm_struct *mm = obj->userptr.mm->mm;
  
                down_read(&mm->mmap_sem);
                while (pinned < npages) {
-                       ret = get_user_pages(work->task, mm,
-                                            obj->userptr.ptr + pinned * PAGE_SIZE,
-                                            npages - pinned,
-                                            !obj->userptr.read_only, 0,
-                                            pvec + pinned, NULL);
+                       ret = get_user_pages_remote(work->task, mm,
+                                       obj->userptr.ptr + pinned * PAGE_SIZE,
+                                       npages - pinned,
+                                       !obj->userptr.read_only, 0,
+                                       pvec + pinned, NULL);
                        if (ret < 0)
                                break;
  
@@@ -631,11 -634,14 +631,11 @@@ i915_gem_userptr_get_pages(struct drm_i
        pvec = NULL;
        pinned = 0;
        if (obj->userptr.mm->mm == current->mm) {
 -              pvec = kmalloc(num_pages*sizeof(struct page *),
 -                             GFP_TEMPORARY | __GFP_NOWARN | __GFP_NORETRY);
 +              pvec = drm_malloc_gfp(num_pages, sizeof(struct page *),
 +                                    GFP_TEMPORARY);
                if (pvec == NULL) {
 -                      pvec = drm_malloc_ab(num_pages, sizeof(struct page *));
 -                      if (pvec == NULL) {
 -                              __i915_gem_userptr_set_active(obj, false);
 -                              return -ENOMEM;
 -                      }
 +                      __i915_gem_userptr_set_active(obj, false);
 +                      return -ENOMEM;
                }
  
                pinned = __get_user_pages_fast(obj->userptr.ptr, num_pages,
@@@ -677,7 -683,7 +677,7 @@@ i915_gem_userptr_put_pages(struct drm_i
                        set_page_dirty(page);
  
                mark_page_accessed(page);
-               page_cache_release(page);
+               put_page(page);
        }
        obj->dirty = 0;
  
@@@ -752,13 -758,6 +752,13 @@@ i915_gem_userptr_ioctl(struct drm_devic
        int ret;
        u32 handle;
  
 +      if (!HAS_LLC(dev) && !HAS_SNOOP(dev)) {
 +              /* We cannot support coherent userptr objects on hw without
 +               * LLC and broken snooping.
 +               */
 +              return -ENODEV;
 +      }
 +
        if (args->flags & ~(I915_USERPTR_READ_ONLY |
                            I915_USERPTR_UNSYNCHRONIZED))
                return -EINVAL;
diff --combined mm/vmalloc.c
@@@ -21,7 -21,6 +21,7 @@@
  #include <linux/debugobjects.h>
  #include <linux/kallsyms.h>
  #include <linux/list.h>
 +#include <linux/notifier.h>
  #include <linux/rbtree.h>
  #include <linux/radix-tree.h>
  #include <linux/rcupdate.h>
@@@ -345,8 -344,6 +345,8 @@@ static void __insert_vmap_area(struct v
  
  static void purge_vmap_area_lazy(void);
  
 +static BLOCKING_NOTIFIER_HEAD(vmap_notify_list);
 +
  /*
   * Allocate a region of KVA of the specified size and alignment, within the
   * vstart and vend.
@@@ -366,8 -363,6 +366,8 @@@ static struct vmap_area *alloc_vmap_are
        BUG_ON(offset_in_page(size));
        BUG_ON(!is_power_of_2(align));
  
 +      might_sleep_if(gfpflags_allow_blocking(gfp_mask));
 +
        va = kmalloc_node(sizeof(struct vmap_area),
                        gfp_mask & GFP_RECLAIM_MASK, node);
        if (unlikely(!va))
@@@ -473,35 -468,13 +473,35 @@@ overflow
                purged = 1;
                goto retry;
        }
 +
 +      if (gfpflags_allow_blocking(gfp_mask)) {
 +              unsigned long freed = 0;
 +              blocking_notifier_call_chain(&vmap_notify_list, 0, &freed);
 +              if (freed > 0) {
 +                      purged = 0;
 +                      goto retry;
 +              }
 +      }
 +
        if (printk_ratelimit())
-               pr_warn("vmap allocation for size %lu failed: "
-                       "use vmalloc=<size> to increase size.\n", size);
+               pr_warn("vmap allocation for size %lu failed: use vmalloc=<size> to increase size\n",
+                       size);
        kfree(va);
        return ERR_PTR(-EBUSY);
  }
  
 +int register_vmap_purge_notifier(struct notifier_block *nb)
 +{
 +      return blocking_notifier_chain_register(&vmap_notify_list, nb);
 +}
 +EXPORT_SYMBOL_GPL(register_vmap_purge_notifier);
 +
 +int unregister_vmap_purge_notifier(struct notifier_block *nb)
 +{
 +      return blocking_notifier_chain_unregister(&vmap_notify_list, nb);
 +}
 +EXPORT_SYMBOL_GPL(unregister_vmap_purge_notifier);
 +
  static void __free_vmap_area(struct vmap_area *va)
  {
        BUG_ON(RB_EMPTY_NODE(&va->rb_node));
@@@ -558,22 -531,21 +558,21 @@@ static void unmap_vmap_area(struct vmap
  static void vmap_debug_free_range(unsigned long start, unsigned long end)
  {
        /*
-        * Unmap page tables and force a TLB flush immediately if
-        * CONFIG_DEBUG_PAGEALLOC is set. This catches use after free
-        * bugs similarly to those in linear kernel virtual address
-        * space after a page has been freed.
+        * Unmap page tables and force a TLB flush immediately if pagealloc
+        * debugging is enabled.  This catches use after free bugs similarly to
+        * those in linear kernel virtual address space after a page has been
+        * freed.
         *
-        * All the lazy freeing logic is still retained, in order to
-        * minimise intrusiveness of this debugging feature.
+        * All the lazy freeing logic is still retained, in order to minimise
+        * intrusiveness of this debugging feature.
         *
-        * This is going to be *slow* (linear kernel virtual address
-        * debugging doesn't do a broadcast TLB flush so it is a lot
-        * faster).
+        * This is going to be *slow* (linear kernel virtual address debugging
+        * doesn't do a broadcast TLB flush so it is a lot faster).
         */
- #ifdef CONFIG_DEBUG_PAGEALLOC
-       vunmap_page_range(start, end);
-       flush_tlb_kernel_range(start, end);
- #endif
+       if (debug_pagealloc_enabled()) {
+               vunmap_page_range(start, end);
+               flush_tlb_kernel_range(start, end);
+       }
  }
  
  /*
@@@ -1113,7 -1085,7 +1112,7 @@@ void vm_unmap_ram(const void *mem, unsi
        BUG_ON(!addr);
        BUG_ON(addr < VMALLOC_START);
        BUG_ON(addr > VMALLOC_END);
-       BUG_ON(!IS_ALIGNED(addr, PAGE_SIZE));
+       BUG_ON(!PAGE_ALIGNED(addr));
  
        debug_check_no_locks_freed(mem, size);
        vmap_debug_free_range(addr, addr+size);