blob: 6728ea5c71d4c2916a37daa20ec767252a96d0ff [file] [log] [blame]
Eric Anholt673a3942008-07-30 12:06:12 -07001/*
Daniel Vetterbe6a0372015-03-18 10:46:04 +01002 * Copyright © 2008-2015 Intel Corporation
Eric Anholt673a3942008-07-30 12:06:12 -07003 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21 * IN THE SOFTWARE.
22 *
23 * Authors:
24 * Eric Anholt <eric@anholt.net>
25 *
26 */
27
David Herrmann0de23972013-07-24 21:07:52 +020028#include <drm/drm_vma_manager.h>
Daniel Vetterfcd70cd2019-01-17 22:03:34 +010029#include <drm/drm_pci.h>
David Howells760285e2012-10-02 18:01:07 +010030#include <drm/i915_drm.h>
Chris Wilson6b5e90f52016-11-14 20:41:05 +000031#include <linux/dma-fence-array.h>
Chris Wilsonfe3288b2017-02-12 17:20:01 +000032#include <linux/kthread.h>
Chris Wilsonc13d87e2016-07-20 09:21:15 +010033#include <linux/reservation.h>
Hugh Dickins5949eac2011-06-27 16:18:18 -070034#include <linux/shmem_fs.h>
Tejun Heo5a0e3ad2010-03-24 17:04:11 +090035#include <linux/slab.h>
Chris Wilson20e49332016-11-22 14:41:21 +000036#include <linux/stop_machine.h>
Eric Anholt673a3942008-07-30 12:06:12 -070037#include <linux/swap.h>
Jesse Barnes79e53942008-11-07 14:24:08 -080038#include <linux/pci.h>
Daniel Vetter1286ff72012-05-10 15:25:09 +020039#include <linux/dma-buf.h>
Daniel Vetterfcd70cd2019-01-17 22:03:34 +010040#include <linux/mman.h>
Eric Anholt673a3942008-07-30 12:06:12 -070041
Chris Wilson9f588922019-01-16 15:33:04 +000042#include "i915_drv.h"
43#include "i915_gem_clflush.h"
44#include "i915_gemfs.h"
45#include "i915_reset.h"
46#include "i915_trace.h"
47#include "i915_vgpu.h"
48
49#include "intel_drv.h"
50#include "intel_frontbuffer.h"
51#include "intel_mocs.h"
52#include "intel_workarounds.h"
53
Chris Wilsonfbbd37b2016-10-28 13:58:42 +010054static void i915_gem_flush_free_objects(struct drm_i915_private *i915);
Chris Wilson61050802012-04-17 15:31:31 +010055
Chris Wilson2c225692013-08-09 12:26:45 +010056static bool cpu_write_needs_clflush(struct drm_i915_gem_object *obj)
57{
Chris Wilsone27ab732017-06-15 13:38:49 +010058 if (obj->cache_dirty)
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +053059 return false;
60
Chris Wilsonb8f55be2017-08-11 12:11:16 +010061 if (!(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE))
Chris Wilson2c225692013-08-09 12:26:45 +010062 return true;
63
Chris Wilsonbd3d2252017-10-13 21:26:14 +010064 return obj->pin_global; /* currently in use by HW, keep flushed */
Chris Wilson2c225692013-08-09 12:26:45 +010065}
66
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +053067static int
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +010068insert_mappable_node(struct i915_ggtt *ggtt,
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +053069 struct drm_mm_node *node, u32 size)
70{
71 memset(node, 0, sizeof(*node));
Chris Wilson82ad6442018-06-05 16:37:58 +010072 return drm_mm_insert_node_in_range(&ggtt->vm.mm, node,
Chris Wilson4e64e552017-02-02 21:04:38 +000073 size, 0, I915_COLOR_UNEVICTABLE,
74 0, ggtt->mappable_end,
75 DRM_MM_INSERT_LOW);
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +053076}
77
78static void
79remove_mappable_node(struct drm_mm_node *node)
80{
81 drm_mm_remove_node(node);
82}
83
Chris Wilson73aa8082010-09-30 11:46:12 +010084/* some bookkeeping */
85static void i915_gem_info_add_obj(struct drm_i915_private *dev_priv,
Chris Wilson3ef7f222016-10-18 13:02:48 +010086 u64 size)
Chris Wilson73aa8082010-09-30 11:46:12 +010087{
Daniel Vetterc20e8352013-07-24 22:40:23 +020088 spin_lock(&dev_priv->mm.object_stat_lock);
Chris Wilson73aa8082010-09-30 11:46:12 +010089 dev_priv->mm.object_count++;
90 dev_priv->mm.object_memory += size;
Daniel Vetterc20e8352013-07-24 22:40:23 +020091 spin_unlock(&dev_priv->mm.object_stat_lock);
Chris Wilson73aa8082010-09-30 11:46:12 +010092}
93
94static void i915_gem_info_remove_obj(struct drm_i915_private *dev_priv,
Chris Wilson3ef7f222016-10-18 13:02:48 +010095 u64 size)
Chris Wilson73aa8082010-09-30 11:46:12 +010096{
Daniel Vetterc20e8352013-07-24 22:40:23 +020097 spin_lock(&dev_priv->mm.object_stat_lock);
Chris Wilson73aa8082010-09-30 11:46:12 +010098 dev_priv->mm.object_count--;
99 dev_priv->mm.object_memory -= size;
Daniel Vetterc20e8352013-07-24 22:40:23 +0200100 spin_unlock(&dev_priv->mm.object_stat_lock);
Chris Wilson73aa8082010-09-30 11:46:12 +0100101}
102
Chris Wilson21dd3732011-01-26 15:55:56 +0000103static int
Daniel Vetter33196de2012-11-14 17:14:05 +0100104i915_gem_wait_for_error(struct i915_gpu_error *error)
Chris Wilson30dbf0c2010-09-25 10:19:17 +0100105{
Chris Wilson30dbf0c2010-09-25 10:19:17 +0100106 int ret;
107
Chris Wilson4c7d62c2016-10-28 13:58:32 +0100108 might_sleep();
109
Daniel Vetter0a6759c2012-07-04 22:18:41 +0200110 /*
111 * Only wait 10 seconds for the gpu reset to complete to avoid hanging
112 * userspace. If it takes that long something really bad is going on and
113 * we should simply try to bail out and fail as gracefully as possible.
114 */
Daniel Vetter1f83fee2012-11-15 17:17:22 +0100115 ret = wait_event_interruptible_timeout(error->reset_queue,
Chris Wilson8c185ec2017-03-16 17:13:02 +0000116 !i915_reset_backoff(error),
Chris Wilsonb52992c2016-10-28 13:58:24 +0100117 I915_RESET_TIMEOUT);
Daniel Vetter0a6759c2012-07-04 22:18:41 +0200118 if (ret == 0) {
119 DRM_ERROR("Timed out waiting for the gpu reset to complete\n");
120 return -EIO;
121 } else if (ret < 0) {
Chris Wilson30dbf0c2010-09-25 10:19:17 +0100122 return ret;
Chris Wilsond98c52cf2016-04-13 17:35:05 +0100123 } else {
124 return 0;
Daniel Vetter0a6759c2012-07-04 22:18:41 +0200125 }
Chris Wilson30dbf0c2010-09-25 10:19:17 +0100126}
127
Chris Wilson54cf91d2010-11-25 18:00:26 +0000128int i915_mutex_lock_interruptible(struct drm_device *dev)
Chris Wilson76c1dec2010-09-25 11:22:51 +0100129{
Chris Wilsonfac5e232016-07-04 11:34:36 +0100130 struct drm_i915_private *dev_priv = to_i915(dev);
Chris Wilson76c1dec2010-09-25 11:22:51 +0100131 int ret;
132
Daniel Vetter33196de2012-11-14 17:14:05 +0100133 ret = i915_gem_wait_for_error(&dev_priv->gpu_error);
Chris Wilson76c1dec2010-09-25 11:22:51 +0100134 if (ret)
135 return ret;
136
137 ret = mutex_lock_interruptible(&dev->struct_mutex);
138 if (ret)
139 return ret;
140
Chris Wilson76c1dec2010-09-25 11:22:51 +0100141 return 0;
142}
Chris Wilson30dbf0c2010-09-25 10:19:17 +0100143
Chris Wilsone4d20062018-04-06 16:51:44 +0100144static u32 __i915_gem_park(struct drm_i915_private *i915)
145{
Chris Wilson506d1f62019-01-14 14:21:11 +0000146 intel_wakeref_t wakeref;
147
Chris Wilson4dfacb02018-05-31 09:22:43 +0100148 GEM_TRACE("\n");
149
Chris Wilsone4d20062018-04-06 16:51:44 +0100150 lockdep_assert_held(&i915->drm.struct_mutex);
151 GEM_BUG_ON(i915->gt.active_requests);
Chris Wilson643b4502018-04-30 14:15:03 +0100152 GEM_BUG_ON(!list_empty(&i915->gt.active_rings));
Chris Wilsone4d20062018-04-06 16:51:44 +0100153
154 if (!i915->gt.awake)
155 return I915_EPOCH_INVALID;
156
157 GEM_BUG_ON(i915->gt.epoch == I915_EPOCH_INVALID);
158
159 /*
160 * Be paranoid and flush a concurrent interrupt to make sure
161 * we don't reactivate any irq tasklets after parking.
162 *
163 * FIXME: Note that even though we have waited for execlists to be idle,
164 * there may still be an in-flight interrupt even though the CSB
165 * is now empty. synchronize_irq() makes sure that a residual interrupt
166 * is completed before we continue, but it doesn't prevent the HW from
167 * raising a spurious interrupt later. To complete the shield we should
168 * coordinate disabling the CS irq with flushing the interrupts.
169 */
170 synchronize_irq(i915->drm.irq);
171
172 intel_engines_park(i915);
Chris Wilsona89d1f92018-05-02 17:38:39 +0100173 i915_timelines_park(i915);
Chris Wilsone4d20062018-04-06 16:51:44 +0100174
175 i915_pmu_gt_parked(i915);
Chris Wilson3365e222018-05-03 20:51:14 +0100176 i915_vma_parked(i915);
Chris Wilsone4d20062018-04-06 16:51:44 +0100177
Chris Wilson506d1f62019-01-14 14:21:11 +0000178 wakeref = fetch_and_zero(&i915->gt.awake);
179 GEM_BUG_ON(!wakeref);
Chris Wilsone4d20062018-04-06 16:51:44 +0100180
181 if (INTEL_GEN(i915) >= 6)
182 gen6_rps_idle(i915);
183
Chris Wilson8d761e72019-01-14 14:21:28 +0000184 intel_display_power_put(i915, POWER_DOMAIN_GT_IRQ, wakeref);
Chris Wilsone4d20062018-04-06 16:51:44 +0100185
186 return i915->gt.epoch;
187}
188
189void i915_gem_park(struct drm_i915_private *i915)
190{
Chris Wilson4dfacb02018-05-31 09:22:43 +0100191 GEM_TRACE("\n");
192
Chris Wilsone4d20062018-04-06 16:51:44 +0100193 lockdep_assert_held(&i915->drm.struct_mutex);
194 GEM_BUG_ON(i915->gt.active_requests);
195
196 if (!i915->gt.awake)
197 return;
198
199 /* Defer the actual call to __i915_gem_park() to prevent ping-pongs */
200 mod_delayed_work(i915->wq, &i915->gt.idle_work, msecs_to_jiffies(100));
201}
202
203void i915_gem_unpark(struct drm_i915_private *i915)
204{
Chris Wilson4dfacb02018-05-31 09:22:43 +0100205 GEM_TRACE("\n");
206
Chris Wilsone4d20062018-04-06 16:51:44 +0100207 lockdep_assert_held(&i915->drm.struct_mutex);
208 GEM_BUG_ON(!i915->gt.active_requests);
Chris Wilson8d761e72019-01-14 14:21:28 +0000209 assert_rpm_wakelock_held(i915);
Chris Wilsone4d20062018-04-06 16:51:44 +0100210
211 if (i915->gt.awake)
212 return;
213
Chris Wilsone4d20062018-04-06 16:51:44 +0100214 /*
215 * It seems that the DMC likes to transition between the DC states a lot
216 * when there are no connected displays (no active power domains) during
217 * command submission.
218 *
219 * This activity has negative impact on the performance of the chip with
220 * huge latencies observed in the interrupt handler and elsewhere.
221 *
222 * Work around it by grabbing a GT IRQ power domain whilst there is any
223 * GT activity, preventing any DC state transitions.
224 */
Chris Wilson8d761e72019-01-14 14:21:28 +0000225 i915->gt.awake = intel_display_power_get(i915, POWER_DOMAIN_GT_IRQ);
226 GEM_BUG_ON(!i915->gt.awake);
Chris Wilsone4d20062018-04-06 16:51:44 +0100227
Chris Wilsone4d20062018-04-06 16:51:44 +0100228 if (unlikely(++i915->gt.epoch == 0)) /* keep 0 as invalid */
229 i915->gt.epoch = 1;
230
231 intel_enable_gt_powersave(i915);
232 i915_update_gfx_val(i915);
233 if (INTEL_GEN(i915) >= 6)
234 gen6_rps_busy(i915);
235 i915_pmu_gt_unparked(i915);
236
237 intel_engines_unpark(i915);
238
239 i915_queue_hangcheck(i915);
240
241 queue_delayed_work(i915->wq,
242 &i915->gt.retire_work,
243 round_jiffies_up_relative(HZ));
244}
245
Eric Anholt673a3942008-07-30 12:06:12 -0700246int
Eric Anholt5a125c32008-10-22 21:40:13 -0700247i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data,
Chris Wilson05394f32010-11-08 19:18:58 +0000248 struct drm_file *file)
Eric Anholt5a125c32008-10-22 21:40:13 -0700249{
Chris Wilson09d7e462019-01-28 10:23:53 +0000250 struct i915_ggtt *ggtt = &to_i915(dev)->ggtt;
Joonas Lahtinen72e96d642016-03-30 16:57:10 +0300251 struct drm_i915_gem_get_aperture *args = data;
Tvrtko Ursulinca1543b2015-07-01 11:51:10 +0100252 struct i915_vma *vma;
Weinan Liff8f7972017-05-31 10:35:52 +0800253 u64 pinned;
Eric Anholt5a125c32008-10-22 21:40:13 -0700254
Chris Wilson09d7e462019-01-28 10:23:53 +0000255 mutex_lock(&ggtt->vm.mutex);
256
Chris Wilson82ad6442018-06-05 16:37:58 +0100257 pinned = ggtt->vm.reserved;
Chris Wilson499197d2019-01-28 10:23:52 +0000258 list_for_each_entry(vma, &ggtt->vm.bound_list, vm_link)
Chris Wilson20dfbde2016-08-04 16:32:30 +0100259 if (i915_vma_is_pinned(vma))
Tvrtko Ursulinca1543b2015-07-01 11:51:10 +0100260 pinned += vma->node.size;
Chris Wilson09d7e462019-01-28 10:23:53 +0000261
262 mutex_unlock(&ggtt->vm.mutex);
Eric Anholt5a125c32008-10-22 21:40:13 -0700263
Chris Wilson82ad6442018-06-05 16:37:58 +0100264 args->aper_size = ggtt->vm.total;
Akshay Joshi0206e352011-08-16 15:34:10 -0400265 args->aper_available_size = args->aper_size - pinned;
Chris Wilson6299f992010-11-24 12:23:44 +0000266
Eric Anholt5a125c32008-10-22 21:40:13 -0700267 return 0;
268}
269
Matthew Auldb91b09e2017-10-06 23:18:17 +0100270static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj)
Chris Wilson00731152014-05-21 12:42:56 +0100271{
Al Viro93c76a32015-12-04 23:45:44 -0500272 struct address_space *mapping = obj->base.filp->f_mapping;
Chris Wilsondbb43512016-12-07 13:34:11 +0000273 drm_dma_handle_t *phys;
Chris Wilson6a2c4232014-11-04 04:51:40 -0800274 struct sg_table *st;
275 struct scatterlist *sg;
Chris Wilsondbb43512016-12-07 13:34:11 +0000276 char *vaddr;
Chris Wilson6a2c4232014-11-04 04:51:40 -0800277 int i;
Matthew Auldb91b09e2017-10-06 23:18:17 +0100278 int err;
Chris Wilson00731152014-05-21 12:42:56 +0100279
Chris Wilson6a2c4232014-11-04 04:51:40 -0800280 if (WARN_ON(i915_gem_object_needs_bit17_swizzle(obj)))
Matthew Auldb91b09e2017-10-06 23:18:17 +0100281 return -EINVAL;
Chris Wilson00731152014-05-21 12:42:56 +0100282
Chris Wilsondbb43512016-12-07 13:34:11 +0000283 /* Always aligning to the object size, allows a single allocation
284 * to handle all possible callers, and given typical object sizes,
285 * the alignment of the buddy allocation will naturally match.
286 */
287 phys = drm_pci_alloc(obj->base.dev,
Ville Syrjälä750fae22017-09-07 17:32:03 +0300288 roundup_pow_of_two(obj->base.size),
Chris Wilsondbb43512016-12-07 13:34:11 +0000289 roundup_pow_of_two(obj->base.size));
290 if (!phys)
Matthew Auldb91b09e2017-10-06 23:18:17 +0100291 return -ENOMEM;
Chris Wilsondbb43512016-12-07 13:34:11 +0000292
293 vaddr = phys->vaddr;
Chris Wilson6a2c4232014-11-04 04:51:40 -0800294 for (i = 0; i < obj->base.size / PAGE_SIZE; i++) {
295 struct page *page;
296 char *src;
297
298 page = shmem_read_mapping_page(mapping, i);
Chris Wilsondbb43512016-12-07 13:34:11 +0000299 if (IS_ERR(page)) {
Matthew Auldb91b09e2017-10-06 23:18:17 +0100300 err = PTR_ERR(page);
Chris Wilsondbb43512016-12-07 13:34:11 +0000301 goto err_phys;
302 }
Chris Wilson6a2c4232014-11-04 04:51:40 -0800303
304 src = kmap_atomic(page);
305 memcpy(vaddr, src, PAGE_SIZE);
306 drm_clflush_virt_range(vaddr, PAGE_SIZE);
307 kunmap_atomic(src);
308
Kirill A. Shutemov09cbfea2016-04-01 15:29:47 +0300309 put_page(page);
Chris Wilson6a2c4232014-11-04 04:51:40 -0800310 vaddr += PAGE_SIZE;
311 }
312
Chris Wilsonc0336662016-05-06 15:40:21 +0100313 i915_gem_chipset_flush(to_i915(obj->base.dev));
Chris Wilson6a2c4232014-11-04 04:51:40 -0800314
315 st = kmalloc(sizeof(*st), GFP_KERNEL);
Chris Wilsondbb43512016-12-07 13:34:11 +0000316 if (!st) {
Matthew Auldb91b09e2017-10-06 23:18:17 +0100317 err = -ENOMEM;
Chris Wilsondbb43512016-12-07 13:34:11 +0000318 goto err_phys;
319 }
Chris Wilson6a2c4232014-11-04 04:51:40 -0800320
321 if (sg_alloc_table(st, 1, GFP_KERNEL)) {
322 kfree(st);
Matthew Auldb91b09e2017-10-06 23:18:17 +0100323 err = -ENOMEM;
Chris Wilsondbb43512016-12-07 13:34:11 +0000324 goto err_phys;
Chris Wilson6a2c4232014-11-04 04:51:40 -0800325 }
326
327 sg = st->sgl;
328 sg->offset = 0;
329 sg->length = obj->base.size;
330
Chris Wilsondbb43512016-12-07 13:34:11 +0000331 sg_dma_address(sg) = phys->busaddr;
Chris Wilson6a2c4232014-11-04 04:51:40 -0800332 sg_dma_len(sg) = obj->base.size;
333
Chris Wilsondbb43512016-12-07 13:34:11 +0000334 obj->phys_handle = phys;
Matthew Auldb91b09e2017-10-06 23:18:17 +0100335
Matthew Aulda5c081662017-10-06 23:18:18 +0100336 __i915_gem_object_set_pages(obj, st, sg->length);
Matthew Auldb91b09e2017-10-06 23:18:17 +0100337
338 return 0;
Chris Wilsondbb43512016-12-07 13:34:11 +0000339
340err_phys:
341 drm_pci_free(obj->base.dev, phys);
Matthew Auldb91b09e2017-10-06 23:18:17 +0100342
343 return err;
Chris Wilson6a2c4232014-11-04 04:51:40 -0800344}
345
Chris Wilsone27ab732017-06-15 13:38:49 +0100346static void __start_cpu_write(struct drm_i915_gem_object *obj)
347{
Christian Königc0a51fd2018-02-16 13:43:38 +0100348 obj->read_domains = I915_GEM_DOMAIN_CPU;
349 obj->write_domain = I915_GEM_DOMAIN_CPU;
Chris Wilsone27ab732017-06-15 13:38:49 +0100350 if (cpu_write_needs_clflush(obj))
351 obj->cache_dirty = true;
352}
353
Chris Wilson6a2c4232014-11-04 04:51:40 -0800354static void
Chris Wilson2b3c8312016-11-11 14:58:09 +0000355__i915_gem_object_release_shmem(struct drm_i915_gem_object *obj,
Chris Wilsone5facdf2016-12-23 14:57:57 +0000356 struct sg_table *pages,
357 bool needs_clflush)
Chris Wilson6a2c4232014-11-04 04:51:40 -0800358{
Chris Wilsona4f5ea62016-10-28 13:58:35 +0100359 GEM_BUG_ON(obj->mm.madv == __I915_MADV_PURGED);
Chris Wilson6a2c4232014-11-04 04:51:40 -0800360
Chris Wilsona4f5ea62016-10-28 13:58:35 +0100361 if (obj->mm.madv == I915_MADV_DONTNEED)
362 obj->mm.dirty = false;
Chris Wilson6a2c4232014-11-04 04:51:40 -0800363
Chris Wilsone5facdf2016-12-23 14:57:57 +0000364 if (needs_clflush &&
Christian Königc0a51fd2018-02-16 13:43:38 +0100365 (obj->read_domains & I915_GEM_DOMAIN_CPU) == 0 &&
Chris Wilsonb8f55be2017-08-11 12:11:16 +0100366 !(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_READ))
Chris Wilson2b3c8312016-11-11 14:58:09 +0000367 drm_clflush_sg(pages);
Chris Wilson03ac84f2016-10-28 13:58:36 +0100368
Chris Wilsone27ab732017-06-15 13:38:49 +0100369 __start_cpu_write(obj);
Chris Wilson03ac84f2016-10-28 13:58:36 +0100370}
371
372static void
373i915_gem_object_put_pages_phys(struct drm_i915_gem_object *obj,
374 struct sg_table *pages)
375{
Chris Wilsone5facdf2016-12-23 14:57:57 +0000376 __i915_gem_object_release_shmem(obj, pages, false);
Chris Wilson03ac84f2016-10-28 13:58:36 +0100377
Chris Wilsona4f5ea62016-10-28 13:58:35 +0100378 if (obj->mm.dirty) {
Al Viro93c76a32015-12-04 23:45:44 -0500379 struct address_space *mapping = obj->base.filp->f_mapping;
Chris Wilson6a2c4232014-11-04 04:51:40 -0800380 char *vaddr = obj->phys_handle->vaddr;
Chris Wilson00731152014-05-21 12:42:56 +0100381 int i;
382
383 for (i = 0; i < obj->base.size / PAGE_SIZE; i++) {
Chris Wilson6a2c4232014-11-04 04:51:40 -0800384 struct page *page;
385 char *dst;
Chris Wilson00731152014-05-21 12:42:56 +0100386
Chris Wilson6a2c4232014-11-04 04:51:40 -0800387 page = shmem_read_mapping_page(mapping, i);
388 if (IS_ERR(page))
389 continue;
390
391 dst = kmap_atomic(page);
392 drm_clflush_virt_range(vaddr, PAGE_SIZE);
393 memcpy(dst, vaddr, PAGE_SIZE);
394 kunmap_atomic(dst);
395
396 set_page_dirty(page);
Chris Wilsona4f5ea62016-10-28 13:58:35 +0100397 if (obj->mm.madv == I915_MADV_WILLNEED)
Chris Wilson00731152014-05-21 12:42:56 +0100398 mark_page_accessed(page);
Kirill A. Shutemov09cbfea2016-04-01 15:29:47 +0300399 put_page(page);
Chris Wilson00731152014-05-21 12:42:56 +0100400 vaddr += PAGE_SIZE;
401 }
Chris Wilsona4f5ea62016-10-28 13:58:35 +0100402 obj->mm.dirty = false;
Chris Wilson00731152014-05-21 12:42:56 +0100403 }
404
Chris Wilson03ac84f2016-10-28 13:58:36 +0100405 sg_free_table(pages);
406 kfree(pages);
Chris Wilsondbb43512016-12-07 13:34:11 +0000407
408 drm_pci_free(obj->base.dev, obj->phys_handle);
Chris Wilson6a2c4232014-11-04 04:51:40 -0800409}
410
411static void
412i915_gem_object_release_phys(struct drm_i915_gem_object *obj)
413{
Chris Wilsona4f5ea62016-10-28 13:58:35 +0100414 i915_gem_object_unpin_pages(obj);
Chris Wilson6a2c4232014-11-04 04:51:40 -0800415}
416
417static const struct drm_i915_gem_object_ops i915_gem_phys_ops = {
418 .get_pages = i915_gem_object_get_pages_phys,
419 .put_pages = i915_gem_object_put_pages_phys,
420 .release = i915_gem_object_release_phys,
421};
422
Chris Wilson581ab1f2017-02-15 16:39:00 +0000423static const struct drm_i915_gem_object_ops i915_gem_object_ops;
424
Chris Wilson35a96112016-08-14 18:44:40 +0100425int i915_gem_object_unbind(struct drm_i915_gem_object *obj)
Chris Wilsonaa653a62016-08-04 07:52:27 +0100426{
427 struct i915_vma *vma;
428 LIST_HEAD(still_in_list);
Chris Wilson02bef8f2016-08-14 18:44:41 +0100429 int ret;
Chris Wilsonaa653a62016-08-04 07:52:27 +0100430
Chris Wilson02bef8f2016-08-14 18:44:41 +0100431 lockdep_assert_held(&obj->base.dev->struct_mutex);
432
433 /* Closed vma are removed from the obj->vma_list - but they may
434 * still have an active binding on the object. To remove those we
435 * must wait for all rendering to complete to the object (as unbinding
436 * must anyway), and retire the requests.
Chris Wilsonaa653a62016-08-04 07:52:27 +0100437 */
Chris Wilson5888fc92017-12-04 13:25:13 +0000438 ret = i915_gem_object_set_to_cpu_domain(obj, false);
Chris Wilson02bef8f2016-08-14 18:44:41 +0100439 if (ret)
440 return ret;
441
Chris Wilson528cbd12019-01-28 10:23:54 +0000442 spin_lock(&obj->vma.lock);
443 while (!ret && (vma = list_first_entry_or_null(&obj->vma.list,
444 struct i915_vma,
445 obj_link))) {
Chris Wilsonaa653a62016-08-04 07:52:27 +0100446 list_move_tail(&vma->obj_link, &still_in_list);
Chris Wilson528cbd12019-01-28 10:23:54 +0000447 spin_unlock(&obj->vma.lock);
448
Chris Wilsonaa653a62016-08-04 07:52:27 +0100449 ret = i915_vma_unbind(vma);
Chris Wilson528cbd12019-01-28 10:23:54 +0000450
451 spin_lock(&obj->vma.lock);
Chris Wilsonaa653a62016-08-04 07:52:27 +0100452 }
Chris Wilson528cbd12019-01-28 10:23:54 +0000453 list_splice(&still_in_list, &obj->vma.list);
454 spin_unlock(&obj->vma.lock);
Chris Wilsonaa653a62016-08-04 07:52:27 +0100455
456 return ret;
457}
458
Chris Wilsone95433c72016-10-28 13:58:27 +0100459static long
460i915_gem_object_wait_fence(struct dma_fence *fence,
461 unsigned int flags,
462 long timeout,
Sagar Arun Kamble562d9ba2017-10-10 22:30:06 +0100463 struct intel_rps_client *rps_client)
Chris Wilsone95433c72016-10-28 13:58:27 +0100464{
Chris Wilsone61e0f52018-02-21 09:56:36 +0000465 struct i915_request *rq;
Chris Wilsone95433c72016-10-28 13:58:27 +0100466
467 BUILD_BUG_ON(I915_WAIT_INTERRUPTIBLE != 0x1);
468
469 if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
470 return timeout;
471
472 if (!dma_fence_is_i915(fence))
473 return dma_fence_wait_timeout(fence,
474 flags & I915_WAIT_INTERRUPTIBLE,
475 timeout);
476
477 rq = to_request(fence);
Chris Wilsone61e0f52018-02-21 09:56:36 +0000478 if (i915_request_completed(rq))
Chris Wilsone95433c72016-10-28 13:58:27 +0100479 goto out;
480
Chris Wilsone9af4ea2018-01-18 13:16:09 +0000481 /*
482 * This client is about to stall waiting for the GPU. In many cases
Chris Wilsone95433c72016-10-28 13:58:27 +0100483 * this is undesirable and limits the throughput of the system, as
484 * many clients cannot continue processing user input/output whilst
485 * blocked. RPS autotuning may take tens of milliseconds to respond
486 * to the GPU load and thus incurs additional latency for the client.
487 * We can circumvent that by promoting the GPU frequency to maximum
488 * before we wait. This makes the GPU throttle up much more quickly
489 * (good for benchmarks and user experience, e.g. window animations),
490 * but at a cost of spending more power processing the workload
491 * (bad for battery). Not all clients even want their results
492 * immediately and for them we should just let the GPU select its own
493 * frequency to maximise efficiency. To prevent a single client from
494 * forcing the clocks too high for the whole system, we only allow
495 * each client to waitboost once in a busy period.
496 */
Chris Wilsone61e0f52018-02-21 09:56:36 +0000497 if (rps_client && !i915_request_started(rq)) {
Chris Wilsone95433c72016-10-28 13:58:27 +0100498 if (INTEL_GEN(rq->i915) >= 6)
Sagar Arun Kamble562d9ba2017-10-10 22:30:06 +0100499 gen6_rps_boost(rq, rps_client);
Chris Wilsone95433c72016-10-28 13:58:27 +0100500 }
501
Chris Wilsone61e0f52018-02-21 09:56:36 +0000502 timeout = i915_request_wait(rq, flags, timeout);
Chris Wilsone95433c72016-10-28 13:58:27 +0100503
504out:
Chris Wilsone61e0f52018-02-21 09:56:36 +0000505 if (flags & I915_WAIT_LOCKED && i915_request_completed(rq))
506 i915_request_retire_upto(rq);
Chris Wilsone95433c72016-10-28 13:58:27 +0100507
Chris Wilsone95433c72016-10-28 13:58:27 +0100508 return timeout;
509}
510
511static long
512i915_gem_object_wait_reservation(struct reservation_object *resv,
513 unsigned int flags,
514 long timeout,
Sagar Arun Kamble562d9ba2017-10-10 22:30:06 +0100515 struct intel_rps_client *rps_client)
Chris Wilsone95433c72016-10-28 13:58:27 +0100516{
Chris Wilsone54ca972017-02-17 15:13:04 +0000517 unsigned int seq = __read_seqcount_begin(&resv->seq);
Chris Wilsone95433c72016-10-28 13:58:27 +0100518 struct dma_fence *excl;
Chris Wilsone54ca972017-02-17 15:13:04 +0000519 bool prune_fences = false;
Chris Wilsone95433c72016-10-28 13:58:27 +0100520
521 if (flags & I915_WAIT_ALL) {
522 struct dma_fence **shared;
523 unsigned int count, i;
524 int ret;
525
526 ret = reservation_object_get_fences_rcu(resv,
527 &excl, &count, &shared);
528 if (ret)
529 return ret;
530
531 for (i = 0; i < count; i++) {
532 timeout = i915_gem_object_wait_fence(shared[i],
533 flags, timeout,
Sagar Arun Kamble562d9ba2017-10-10 22:30:06 +0100534 rps_client);
Chris Wilsond892e932017-02-12 21:53:43 +0000535 if (timeout < 0)
Chris Wilsone95433c72016-10-28 13:58:27 +0100536 break;
537
538 dma_fence_put(shared[i]);
539 }
540
541 for (; i < count; i++)
542 dma_fence_put(shared[i]);
543 kfree(shared);
Chris Wilsone54ca972017-02-17 15:13:04 +0000544
Chris Wilsonfa730552018-03-07 17:13:03 +0000545 /*
546 * If both shared fences and an exclusive fence exist,
547 * then by construction the shared fences must be later
548 * than the exclusive fence. If we successfully wait for
549 * all the shared fences, we know that the exclusive fence
550 * must all be signaled. If all the shared fences are
551 * signaled, we can prune the array and recover the
552 * floating references on the fences/requests.
553 */
Chris Wilsone54ca972017-02-17 15:13:04 +0000554 prune_fences = count && timeout >= 0;
Chris Wilsone95433c72016-10-28 13:58:27 +0100555 } else {
556 excl = reservation_object_get_excl_rcu(resv);
557 }
558
Chris Wilsonfa730552018-03-07 17:13:03 +0000559 if (excl && timeout >= 0)
Sagar Arun Kamble562d9ba2017-10-10 22:30:06 +0100560 timeout = i915_gem_object_wait_fence(excl, flags, timeout,
561 rps_client);
Chris Wilsone95433c72016-10-28 13:58:27 +0100562
563 dma_fence_put(excl);
564
Chris Wilsonfa730552018-03-07 17:13:03 +0000565 /*
566 * Opportunistically prune the fences iff we know they have *all* been
Chris Wilson03d1cac2017-03-08 13:26:28 +0000567 * signaled and that the reservation object has not been changed (i.e.
568 * no new fences have been added).
569 */
Chris Wilsone54ca972017-02-17 15:13:04 +0000570 if (prune_fences && !__read_seqcount_retry(&resv->seq, seq)) {
Chris Wilson03d1cac2017-03-08 13:26:28 +0000571 if (reservation_object_trylock(resv)) {
572 if (!__read_seqcount_retry(&resv->seq, seq))
573 reservation_object_add_excl_fence(resv, NULL);
574 reservation_object_unlock(resv);
575 }
Chris Wilsone54ca972017-02-17 15:13:04 +0000576 }
577
Chris Wilsone95433c72016-10-28 13:58:27 +0100578 return timeout;
579}
580
Chris Wilsonb7268c52018-04-18 19:40:52 +0100581static void __fence_set_priority(struct dma_fence *fence,
582 const struct i915_sched_attr *attr)
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000583{
Chris Wilsone61e0f52018-02-21 09:56:36 +0000584 struct i915_request *rq;
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000585 struct intel_engine_cs *engine;
586
Chris Wilsonc218ee02018-01-06 10:56:18 +0000587 if (dma_fence_is_signaled(fence) || !dma_fence_is_i915(fence))
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000588 return;
589
590 rq = to_request(fence);
591 engine = rq->engine;
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000592
Chris Wilson4f6d8fc2018-05-07 14:57:25 +0100593 local_bh_disable();
594 rcu_read_lock(); /* RCU serialisation for set-wedged protection */
Chris Wilson47650db2018-03-07 13:42:25 +0000595 if (engine->schedule)
Chris Wilsonb7268c52018-04-18 19:40:52 +0100596 engine->schedule(rq, attr);
Chris Wilson47650db2018-03-07 13:42:25 +0000597 rcu_read_unlock();
Chris Wilson4f6d8fc2018-05-07 14:57:25 +0100598 local_bh_enable(); /* kick the tasklets if queues were reprioritised */
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000599}
600
Chris Wilsonb7268c52018-04-18 19:40:52 +0100601static void fence_set_priority(struct dma_fence *fence,
602 const struct i915_sched_attr *attr)
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000603{
604 /* Recurse once into a fence-array */
605 if (dma_fence_is_array(fence)) {
606 struct dma_fence_array *array = to_dma_fence_array(fence);
607 int i;
608
609 for (i = 0; i < array->num_fences; i++)
Chris Wilsonb7268c52018-04-18 19:40:52 +0100610 __fence_set_priority(array->fences[i], attr);
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000611 } else {
Chris Wilsonb7268c52018-04-18 19:40:52 +0100612 __fence_set_priority(fence, attr);
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000613 }
614}
615
616int
617i915_gem_object_wait_priority(struct drm_i915_gem_object *obj,
618 unsigned int flags,
Chris Wilsonb7268c52018-04-18 19:40:52 +0100619 const struct i915_sched_attr *attr)
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000620{
621 struct dma_fence *excl;
622
623 if (flags & I915_WAIT_ALL) {
624 struct dma_fence **shared;
625 unsigned int count, i;
626 int ret;
627
628 ret = reservation_object_get_fences_rcu(obj->resv,
629 &excl, &count, &shared);
630 if (ret)
631 return ret;
632
633 for (i = 0; i < count; i++) {
Chris Wilsonb7268c52018-04-18 19:40:52 +0100634 fence_set_priority(shared[i], attr);
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000635 dma_fence_put(shared[i]);
636 }
637
638 kfree(shared);
639 } else {
640 excl = reservation_object_get_excl_rcu(obj->resv);
641 }
642
643 if (excl) {
Chris Wilsonb7268c52018-04-18 19:40:52 +0100644 fence_set_priority(excl, attr);
Chris Wilson6b5e90f52016-11-14 20:41:05 +0000645 dma_fence_put(excl);
646 }
647 return 0;
648}
649
Chris Wilson00e60f22016-08-04 16:32:40 +0100650/**
Chris Wilsone95433c72016-10-28 13:58:27 +0100651 * Waits for rendering to the object to be completed
Chris Wilson00e60f22016-08-04 16:32:40 +0100652 * @obj: i915 gem object
Chris Wilsone95433c72016-10-28 13:58:27 +0100653 * @flags: how to wait (under a lock, for all rendering or just for writes etc)
654 * @timeout: how long to wait
Chris Wilsona0a8b1c2017-11-09 14:06:44 +0000655 * @rps_client: client (user process) to charge for any waitboosting
Chris Wilson00e60f22016-08-04 16:32:40 +0100656 */
657int
Chris Wilsone95433c72016-10-28 13:58:27 +0100658i915_gem_object_wait(struct drm_i915_gem_object *obj,
659 unsigned int flags,
660 long timeout,
Sagar Arun Kamble562d9ba2017-10-10 22:30:06 +0100661 struct intel_rps_client *rps_client)
Chris Wilson00e60f22016-08-04 16:32:40 +0100662{
Chris Wilsone95433c72016-10-28 13:58:27 +0100663 might_sleep();
Chris Wilsone95433c72016-10-28 13:58:27 +0100664 GEM_BUG_ON(timeout < 0);
Chris Wilson00e60f22016-08-04 16:32:40 +0100665
Chris Wilsond07f0e52016-10-28 13:58:44 +0100666 timeout = i915_gem_object_wait_reservation(obj->resv,
667 flags, timeout,
Sagar Arun Kamble562d9ba2017-10-10 22:30:06 +0100668 rps_client);
Chris Wilsone95433c72016-10-28 13:58:27 +0100669 return timeout < 0 ? timeout : 0;
Chris Wilson00e60f22016-08-04 16:32:40 +0100670}
671
672static struct intel_rps_client *to_rps_client(struct drm_file *file)
673{
674 struct drm_i915_file_private *fpriv = file->driver_priv;
675
Sagar Arun Kamble562d9ba2017-10-10 22:30:06 +0100676 return &fpriv->rps_client;
Chris Wilson00e60f22016-08-04 16:32:40 +0100677}
678
Chris Wilson00731152014-05-21 12:42:56 +0100679static int
680i915_gem_phys_pwrite(struct drm_i915_gem_object *obj,
681 struct drm_i915_gem_pwrite *args,
Chris Wilson03ac84f2016-10-28 13:58:36 +0100682 struct drm_file *file)
Chris Wilson00731152014-05-21 12:42:56 +0100683{
Chris Wilson00731152014-05-21 12:42:56 +0100684 void *vaddr = obj->phys_handle->vaddr + args->offset;
Gustavo Padovan3ed605b2016-04-26 12:32:27 -0300685 char __user *user_data = u64_to_user_ptr(args->data_ptr);
Chris Wilson6a2c4232014-11-04 04:51:40 -0800686
687 /* We manually control the domain here and pretend that it
688 * remains coherent i.e. in the GTT domain, like shmem_pwrite.
689 */
Rodrigo Vivi77a0d1c2015-06-18 11:43:24 -0700690 intel_fb_obj_invalidate(obj, ORIGIN_CPU);
Chris Wilson10466d22017-01-06 15:22:38 +0000691 if (copy_from_user(vaddr, user_data, args->size))
692 return -EFAULT;
Chris Wilson00731152014-05-21 12:42:56 +0100693
Chris Wilson6a2c4232014-11-04 04:51:40 -0800694 drm_clflush_virt_range(vaddr, args->size);
Chris Wilson10466d22017-01-06 15:22:38 +0000695 i915_gem_chipset_flush(to_i915(obj->base.dev));
Paulo Zanoni063e4e62015-02-13 17:23:45 -0200696
Chris Wilsond59b21e2017-02-22 11:40:49 +0000697 intel_fb_obj_flush(obj, ORIGIN_CPU);
Chris Wilson10466d22017-01-06 15:22:38 +0000698 return 0;
Chris Wilson00731152014-05-21 12:42:56 +0100699}
700
Tvrtko Ursulin187685c2016-12-01 14:16:36 +0000701void *i915_gem_object_alloc(struct drm_i915_private *dev_priv)
Chris Wilson42dcedd2012-11-15 11:32:30 +0000702{
Chris Wilsonefab6d82015-04-07 16:20:57 +0100703 return kmem_cache_zalloc(dev_priv->objects, GFP_KERNEL);
Chris Wilson42dcedd2012-11-15 11:32:30 +0000704}
705
706void i915_gem_object_free(struct drm_i915_gem_object *obj)
707{
Chris Wilsonfac5e232016-07-04 11:34:36 +0100708 struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
Chris Wilsonefab6d82015-04-07 16:20:57 +0100709 kmem_cache_free(dev_priv->objects, obj);
Chris Wilson42dcedd2012-11-15 11:32:30 +0000710}
711
Dave Airlieff72145b2011-02-07 12:16:14 +1000712static int
713i915_gem_create(struct drm_file *file,
Tvrtko Ursulin12d79d72016-12-01 14:16:37 +0000714 struct drm_i915_private *dev_priv,
Jani Nikula739f3ab2019-01-16 11:15:19 +0200715 u64 size,
716 u32 *handle_p)
Eric Anholt673a3942008-07-30 12:06:12 -0700717{
Chris Wilson05394f32010-11-08 19:18:58 +0000718 struct drm_i915_gem_object *obj;
Pekka Paalanena1a2d1d2009-08-23 12:40:55 +0300719 int ret;
720 u32 handle;
Eric Anholt673a3942008-07-30 12:06:12 -0700721
Dave Airlieff72145b2011-02-07 12:16:14 +1000722 size = roundup(size, PAGE_SIZE);
Chris Wilson8ffc0242011-09-14 14:14:28 +0200723 if (size == 0)
724 return -EINVAL;
Eric Anholt673a3942008-07-30 12:06:12 -0700725
726 /* Allocate the new object */
Tvrtko Ursulin12d79d72016-12-01 14:16:37 +0000727 obj = i915_gem_object_create(dev_priv, size);
Chris Wilsonfe3db792016-04-25 13:32:13 +0100728 if (IS_ERR(obj))
729 return PTR_ERR(obj);
Eric Anholt673a3942008-07-30 12:06:12 -0700730
Chris Wilson05394f32010-11-08 19:18:58 +0000731 ret = drm_gem_handle_create(file, &obj->base, &handle);
Chris Wilson202f2fe2010-10-14 13:20:40 +0100732 /* drop reference from allocate - handle holds it now */
Chris Wilsonf0cd5182016-10-28 13:58:43 +0100733 i915_gem_object_put(obj);
Daniel Vetterd861e332013-07-24 23:25:03 +0200734 if (ret)
735 return ret;
Chris Wilson202f2fe2010-10-14 13:20:40 +0100736
Dave Airlieff72145b2011-02-07 12:16:14 +1000737 *handle_p = handle;
Eric Anholt673a3942008-07-30 12:06:12 -0700738 return 0;
739}
740
Dave Airlieff72145b2011-02-07 12:16:14 +1000741int
742i915_gem_dumb_create(struct drm_file *file,
743 struct drm_device *dev,
744 struct drm_mode_create_dumb *args)
745{
746 /* have to work out size/pitch and return them */
Paulo Zanonide45eaf2013-10-18 18:48:24 -0300747 args->pitch = ALIGN(args->width * DIV_ROUND_UP(args->bpp, 8), 64);
Dave Airlieff72145b2011-02-07 12:16:14 +1000748 args->size = args->pitch * args->height;
Tvrtko Ursulin12d79d72016-12-01 14:16:37 +0000749 return i915_gem_create(file, to_i915(dev),
Dave Airlieda6b51d2014-12-24 13:11:17 +1000750 args->size, &args->handle);
Dave Airlieff72145b2011-02-07 12:16:14 +1000751}
752
Chris Wilsone27ab732017-06-15 13:38:49 +0100753static bool gpu_write_needs_clflush(struct drm_i915_gem_object *obj)
754{
755 return !(obj->cache_level == I915_CACHE_NONE ||
756 obj->cache_level == I915_CACHE_WT);
757}
758
Dave Airlieff72145b2011-02-07 12:16:14 +1000759/**
760 * Creates a new mm object and returns a handle to it.
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +0100761 * @dev: drm device pointer
762 * @data: ioctl data blob
763 * @file: drm file pointer
Dave Airlieff72145b2011-02-07 12:16:14 +1000764 */
765int
766i915_gem_create_ioctl(struct drm_device *dev, void *data,
767 struct drm_file *file)
768{
Tvrtko Ursulin12d79d72016-12-01 14:16:37 +0000769 struct drm_i915_private *dev_priv = to_i915(dev);
Dave Airlieff72145b2011-02-07 12:16:14 +1000770 struct drm_i915_gem_create *args = data;
Daniel Vetter63ed2cb2012-04-23 16:50:50 +0200771
Tvrtko Ursulin12d79d72016-12-01 14:16:37 +0000772 i915_gem_flush_free_objects(dev_priv);
Chris Wilsonfbbd37b2016-10-28 13:58:42 +0100773
Tvrtko Ursulin12d79d72016-12-01 14:16:37 +0000774 return i915_gem_create(file, dev_priv,
Dave Airlieda6b51d2014-12-24 13:11:17 +1000775 args->size, &args->handle);
Dave Airlieff72145b2011-02-07 12:16:14 +1000776}
777
Chris Wilsonef749212017-04-12 12:01:10 +0100778static inline enum fb_op_origin
779fb_write_origin(struct drm_i915_gem_object *obj, unsigned int domain)
780{
781 return (domain == I915_GEM_DOMAIN_GTT ?
782 obj->frontbuffer_ggtt_origin : ORIGIN_CPU);
783}
784
Chris Wilson7125397b2017-12-06 12:49:14 +0000785void i915_gem_flush_ggtt_writes(struct drm_i915_private *dev_priv)
Chris Wilsonef749212017-04-12 12:01:10 +0100786{
Chris Wilson538ef962019-01-14 14:21:18 +0000787 intel_wakeref_t wakeref;
788
Chris Wilson7125397b2017-12-06 12:49:14 +0000789 /*
790 * No actual flushing is required for the GTT write domain for reads
791 * from the GTT domain. Writes to it "immediately" go to main memory
792 * as far as we know, so there's no chipset flush. It also doesn't
793 * land in the GPU render cache.
Chris Wilsonef749212017-04-12 12:01:10 +0100794 *
795 * However, we do have to enforce the order so that all writes through
796 * the GTT land before any writes to the device, such as updates to
797 * the GATT itself.
798 *
799 * We also have to wait a bit for the writes to land from the GTT.
800 * An uncached read (i.e. mmio) seems to be ideal for the round-trip
801 * timing. This issue has only been observed when switching quickly
802 * between GTT writes and CPU reads from inside the kernel on recent hw,
803 * and it appears to only affect discrete GTT blocks (i.e. on LLC
Chris Wilson7125397b2017-12-06 12:49:14 +0000804 * system agents we cannot reproduce this behaviour, until Cannonlake
805 * that was!).
Chris Wilsonef749212017-04-12 12:01:10 +0100806 */
Chris Wilson7125397b2017-12-06 12:49:14 +0000807
Chris Wilson900ccf32018-07-20 11:19:10 +0100808 wmb();
809
810 if (INTEL_INFO(dev_priv)->has_coherent_ggtt)
811 return;
812
Chris Wilsona8bd3b82018-07-17 10:26:55 +0100813 i915_gem_chipset_flush(dev_priv);
Chris Wilsonef749212017-04-12 12:01:10 +0100814
Chris Wilsond4225a52019-01-14 14:21:23 +0000815 with_intel_runtime_pm(dev_priv, wakeref) {
816 spin_lock_irq(&dev_priv->uncore.lock);
Chris Wilson7125397b2017-12-06 12:49:14 +0000817
Chris Wilsond4225a52019-01-14 14:21:23 +0000818 POSTING_READ_FW(RING_HEAD(RENDER_RING_BASE));
Chris Wilson7125397b2017-12-06 12:49:14 +0000819
Chris Wilsond4225a52019-01-14 14:21:23 +0000820 spin_unlock_irq(&dev_priv->uncore.lock);
821 }
Chris Wilson7125397b2017-12-06 12:49:14 +0000822}
823
824static void
825flush_write_domain(struct drm_i915_gem_object *obj, unsigned int flush_domains)
826{
827 struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
828 struct i915_vma *vma;
829
Christian Königc0a51fd2018-02-16 13:43:38 +0100830 if (!(obj->write_domain & flush_domains))
Chris Wilson7125397b2017-12-06 12:49:14 +0000831 return;
832
Christian Königc0a51fd2018-02-16 13:43:38 +0100833 switch (obj->write_domain) {
Chris Wilsonef749212017-04-12 12:01:10 +0100834 case I915_GEM_DOMAIN_GTT:
Chris Wilson7125397b2017-12-06 12:49:14 +0000835 i915_gem_flush_ggtt_writes(dev_priv);
Chris Wilsonef749212017-04-12 12:01:10 +0100836
837 intel_fb_obj_flush(obj,
838 fb_write_origin(obj, I915_GEM_DOMAIN_GTT));
Chris Wilson7125397b2017-12-06 12:49:14 +0000839
Chris Wilsone2189dd2017-12-07 21:14:07 +0000840 for_each_ggtt_vma(vma, obj) {
Chris Wilson7125397b2017-12-06 12:49:14 +0000841 if (vma->iomap)
842 continue;
843
844 i915_vma_unset_ggtt_write(vma);
845 }
Chris Wilsonef749212017-04-12 12:01:10 +0100846 break;
847
Chris Wilsonadd00e62018-07-06 12:54:02 +0100848 case I915_GEM_DOMAIN_WC:
849 wmb();
850 break;
851
Chris Wilsonef749212017-04-12 12:01:10 +0100852 case I915_GEM_DOMAIN_CPU:
853 i915_gem_clflush_object(obj, I915_CLFLUSH_SYNC);
854 break;
Chris Wilsone27ab732017-06-15 13:38:49 +0100855
856 case I915_GEM_DOMAIN_RENDER:
857 if (gpu_write_needs_clflush(obj))
858 obj->cache_dirty = true;
859 break;
Chris Wilsonef749212017-04-12 12:01:10 +0100860 }
861
Christian Königc0a51fd2018-02-16 13:43:38 +0100862 obj->write_domain = 0;
Chris Wilsonef749212017-04-12 12:01:10 +0100863}
864
Brad Volkin4c914c02014-02-18 10:15:45 -0800865/*
866 * Pins the specified object's pages and synchronizes the object with
867 * GPU accesses. Sets needs_clflush to non-zero if the caller should
868 * flush the object from the CPU cache.
869 */
870int i915_gem_obj_prepare_shmem_read(struct drm_i915_gem_object *obj,
Chris Wilson43394c72016-08-18 17:16:47 +0100871 unsigned int *needs_clflush)
Brad Volkin4c914c02014-02-18 10:15:45 -0800872{
873 int ret;
874
Chris Wilsone95433c72016-10-28 13:58:27 +0100875 lockdep_assert_held(&obj->base.dev->struct_mutex);
Brad Volkin4c914c02014-02-18 10:15:45 -0800876
Chris Wilsone95433c72016-10-28 13:58:27 +0100877 *needs_clflush = 0;
Chris Wilson43394c72016-08-18 17:16:47 +0100878 if (!i915_gem_object_has_struct_page(obj))
879 return -ENODEV;
Brad Volkin4c914c02014-02-18 10:15:45 -0800880
Chris Wilsone95433c72016-10-28 13:58:27 +0100881 ret = i915_gem_object_wait(obj,
882 I915_WAIT_INTERRUPTIBLE |
883 I915_WAIT_LOCKED,
884 MAX_SCHEDULE_TIMEOUT,
885 NULL);
Chris Wilsonc13d87e2016-07-20 09:21:15 +0100886 if (ret)
887 return ret;
888
Chris Wilsona4f5ea62016-10-28 13:58:35 +0100889 ret = i915_gem_object_pin_pages(obj);
Chris Wilson97649512016-08-18 17:16:50 +0100890 if (ret)
891 return ret;
892
Chris Wilsonb8f55be2017-08-11 12:11:16 +0100893 if (obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_READ ||
894 !static_cpu_has(X86_FEATURE_CLFLUSH)) {
Chris Wilson7f5f95d2017-03-10 00:09:42 +0000895 ret = i915_gem_object_set_to_cpu_domain(obj, false);
896 if (ret)
897 goto err_unpin;
898 else
899 goto out;
900 }
901
Chris Wilsonef749212017-04-12 12:01:10 +0100902 flush_write_domain(obj, ~I915_GEM_DOMAIN_CPU);
Chris Wilsona314d5c2016-08-18 17:16:48 +0100903
Chris Wilson43394c72016-08-18 17:16:47 +0100904 /* If we're not in the cpu read domain, set ourself into the gtt
905 * read domain and manually flush cachelines (if required). This
906 * optimizes for the case when the gpu will dirty the data
907 * anyway again before the next pread happens.
908 */
Chris Wilsone27ab732017-06-15 13:38:49 +0100909 if (!obj->cache_dirty &&
Christian Königc0a51fd2018-02-16 13:43:38 +0100910 !(obj->read_domains & I915_GEM_DOMAIN_CPU))
Chris Wilson7f5f95d2017-03-10 00:09:42 +0000911 *needs_clflush = CLFLUSH_BEFORE;
Brad Volkin4c914c02014-02-18 10:15:45 -0800912
Chris Wilson7f5f95d2017-03-10 00:09:42 +0000913out:
Chris Wilson97649512016-08-18 17:16:50 +0100914 /* return with the pages pinned */
Chris Wilson43394c72016-08-18 17:16:47 +0100915 return 0;
Chris Wilson97649512016-08-18 17:16:50 +0100916
917err_unpin:
918 i915_gem_object_unpin_pages(obj);
919 return ret;
Chris Wilson43394c72016-08-18 17:16:47 +0100920}
921
922int i915_gem_obj_prepare_shmem_write(struct drm_i915_gem_object *obj,
923 unsigned int *needs_clflush)
924{
925 int ret;
926
Chris Wilsone95433c72016-10-28 13:58:27 +0100927 lockdep_assert_held(&obj->base.dev->struct_mutex);
928
Chris Wilson43394c72016-08-18 17:16:47 +0100929 *needs_clflush = 0;
930 if (!i915_gem_object_has_struct_page(obj))
931 return -ENODEV;
932
Chris Wilsone95433c72016-10-28 13:58:27 +0100933 ret = i915_gem_object_wait(obj,
934 I915_WAIT_INTERRUPTIBLE |
935 I915_WAIT_LOCKED |
936 I915_WAIT_ALL,
937 MAX_SCHEDULE_TIMEOUT,
938 NULL);
Chris Wilson43394c72016-08-18 17:16:47 +0100939 if (ret)
940 return ret;
941
Chris Wilsona4f5ea62016-10-28 13:58:35 +0100942 ret = i915_gem_object_pin_pages(obj);
Chris Wilson97649512016-08-18 17:16:50 +0100943 if (ret)
944 return ret;
945
Chris Wilsonb8f55be2017-08-11 12:11:16 +0100946 if (obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE ||
947 !static_cpu_has(X86_FEATURE_CLFLUSH)) {
Chris Wilson7f5f95d2017-03-10 00:09:42 +0000948 ret = i915_gem_object_set_to_cpu_domain(obj, true);
949 if (ret)
950 goto err_unpin;
951 else
952 goto out;
953 }
954
Chris Wilsonef749212017-04-12 12:01:10 +0100955 flush_write_domain(obj, ~I915_GEM_DOMAIN_CPU);
Chris Wilsona314d5c2016-08-18 17:16:48 +0100956
Chris Wilson43394c72016-08-18 17:16:47 +0100957 /* If we're not in the cpu write domain, set ourself into the
958 * gtt write domain and manually flush cachelines (as required).
959 * This optimizes for the case when the gpu will use the data
960 * right away and we therefore have to clflush anyway.
961 */
Chris Wilsone27ab732017-06-15 13:38:49 +0100962 if (!obj->cache_dirty) {
Chris Wilson7f5f95d2017-03-10 00:09:42 +0000963 *needs_clflush |= CLFLUSH_AFTER;
Chris Wilson43394c72016-08-18 17:16:47 +0100964
Chris Wilsone27ab732017-06-15 13:38:49 +0100965 /*
966 * Same trick applies to invalidate partially written
967 * cachelines read before writing.
968 */
Christian Königc0a51fd2018-02-16 13:43:38 +0100969 if (!(obj->read_domains & I915_GEM_DOMAIN_CPU))
Chris Wilsone27ab732017-06-15 13:38:49 +0100970 *needs_clflush |= CLFLUSH_BEFORE;
971 }
Chris Wilson43394c72016-08-18 17:16:47 +0100972
Chris Wilson7f5f95d2017-03-10 00:09:42 +0000973out:
Chris Wilson43394c72016-08-18 17:16:47 +0100974 intel_fb_obj_invalidate(obj, ORIGIN_CPU);
Chris Wilsona4f5ea62016-10-28 13:58:35 +0100975 obj->mm.dirty = true;
Chris Wilson97649512016-08-18 17:16:50 +0100976 /* return with the pages pinned */
Chris Wilson43394c72016-08-18 17:16:47 +0100977 return 0;
Chris Wilson97649512016-08-18 17:16:50 +0100978
979err_unpin:
980 i915_gem_object_unpin_pages(obj);
981 return ret;
Brad Volkin4c914c02014-02-18 10:15:45 -0800982}
983
Daniel Vetterd174bd62012-03-25 19:47:40 +0200984static int
Chris Wilsonb9d126e2019-01-05 12:07:58 +0000985shmem_pread(struct page *page, int offset, int len, char __user *user_data,
986 bool needs_clflush)
Daniel Vetterd174bd62012-03-25 19:47:40 +0200987{
988 char *vaddr;
989 int ret;
990
991 vaddr = kmap(page);
Daniel Vetterd174bd62012-03-25 19:47:40 +0200992
Chris Wilsonb9d126e2019-01-05 12:07:58 +0000993 if (needs_clflush)
994 drm_clflush_virt_range(vaddr + offset, len);
995
996 ret = __copy_to_user(user_data, vaddr + offset, len);
997
Daniel Vetterd174bd62012-03-25 19:47:40 +0200998 kunmap(page);
999
Chris Wilsonb9d126e2019-01-05 12:07:58 +00001000 return ret ? -EFAULT : 0;
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001001}
1002
1003static int
1004i915_gem_shmem_pread(struct drm_i915_gem_object *obj,
1005 struct drm_i915_gem_pread *args)
1006{
1007 char __user *user_data;
1008 u64 remain;
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001009 unsigned int needs_clflush;
1010 unsigned int idx, offset;
1011 int ret;
1012
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001013 ret = mutex_lock_interruptible(&obj->base.dev->struct_mutex);
1014 if (ret)
1015 return ret;
1016
1017 ret = i915_gem_obj_prepare_shmem_read(obj, &needs_clflush);
1018 mutex_unlock(&obj->base.dev->struct_mutex);
1019 if (ret)
1020 return ret;
1021
1022 remain = args->size;
1023 user_data = u64_to_user_ptr(args->data_ptr);
1024 offset = offset_in_page(args->offset);
1025 for (idx = args->offset >> PAGE_SHIFT; remain; idx++) {
1026 struct page *page = i915_gem_object_get_page(obj, idx);
Chris Wilsona5e856a52018-10-12 15:02:28 +01001027 unsigned int length = min_t(u64, remain, PAGE_SIZE - offset);
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001028
1029 ret = shmem_pread(page, offset, length, user_data,
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001030 needs_clflush);
1031 if (ret)
1032 break;
1033
1034 remain -= length;
1035 user_data += length;
1036 offset = 0;
1037 }
1038
1039 i915_gem_obj_finish_shmem_access(obj);
1040 return ret;
1041}
1042
1043static inline bool
1044gtt_user_read(struct io_mapping *mapping,
1045 loff_t base, int offset,
1046 char __user *user_data, int length)
1047{
Ville Syrjäläafe722b2017-09-01 20:12:52 +03001048 void __iomem *vaddr;
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001049 unsigned long unwritten;
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301050
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301051 /* We can use the cpu mem copy function because this is X86. */
Ville Syrjäläafe722b2017-09-01 20:12:52 +03001052 vaddr = io_mapping_map_atomic_wc(mapping, base);
1053 unwritten = __copy_to_user_inatomic(user_data,
1054 (void __force *)vaddr + offset,
1055 length);
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001056 io_mapping_unmap_atomic(vaddr);
1057 if (unwritten) {
Ville Syrjäläafe722b2017-09-01 20:12:52 +03001058 vaddr = io_mapping_map_wc(mapping, base, PAGE_SIZE);
1059 unwritten = copy_to_user(user_data,
1060 (void __force *)vaddr + offset,
1061 length);
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001062 io_mapping_unmap(vaddr);
1063 }
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301064 return unwritten;
1065}
1066
1067static int
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001068i915_gem_gtt_pread(struct drm_i915_gem_object *obj,
1069 const struct drm_i915_gem_pread *args)
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301070{
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001071 struct drm_i915_private *i915 = to_i915(obj->base.dev);
1072 struct i915_ggtt *ggtt = &i915->ggtt;
Chris Wilson538ef962019-01-14 14:21:18 +00001073 intel_wakeref_t wakeref;
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301074 struct drm_mm_node node;
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001075 struct i915_vma *vma;
1076 void __user *user_data;
1077 u64 remain, offset;
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301078 int ret;
1079
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001080 ret = mutex_lock_interruptible(&i915->drm.struct_mutex);
1081 if (ret)
1082 return ret;
1083
Chris Wilson538ef962019-01-14 14:21:18 +00001084 wakeref = intel_runtime_pm_get(i915);
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001085 vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0,
Chris Wilsona3259ca2017-10-09 09:44:00 +01001086 PIN_MAPPABLE |
1087 PIN_NONFAULT |
1088 PIN_NONBLOCK);
Chris Wilson18034582016-08-18 17:16:45 +01001089 if (!IS_ERR(vma)) {
1090 node.start = i915_ggtt_offset(vma);
1091 node.allocated = false;
Chris Wilson49ef5292016-08-18 17:17:00 +01001092 ret = i915_vma_put_fence(vma);
Chris Wilson18034582016-08-18 17:16:45 +01001093 if (ret) {
1094 i915_vma_unpin(vma);
1095 vma = ERR_PTR(ret);
1096 }
1097 }
Chris Wilson058d88c2016-08-15 10:49:06 +01001098 if (IS_ERR(vma)) {
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001099 ret = insert_mappable_node(ggtt, &node, PAGE_SIZE);
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301100 if (ret)
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001101 goto out_unlock;
1102 GEM_BUG_ON(!node.allocated);
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301103 }
1104
1105 ret = i915_gem_object_set_to_gtt_domain(obj, false);
1106 if (ret)
1107 goto out_unpin;
1108
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001109 mutex_unlock(&i915->drm.struct_mutex);
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301110
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001111 user_data = u64_to_user_ptr(args->data_ptr);
1112 remain = args->size;
1113 offset = args->offset;
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301114
1115 while (remain > 0) {
1116 /* Operation in this page
1117 *
1118 * page_base = page offset within aperture
1119 * page_offset = offset within page
1120 * page_length = bytes to copy for this page
1121 */
1122 u32 page_base = node.start;
1123 unsigned page_offset = offset_in_page(offset);
1124 unsigned page_length = PAGE_SIZE - page_offset;
1125 page_length = remain < page_length ? remain : page_length;
1126 if (node.allocated) {
1127 wmb();
Chris Wilson82ad6442018-06-05 16:37:58 +01001128 ggtt->vm.insert_page(&ggtt->vm,
1129 i915_gem_object_get_dma_address(obj, offset >> PAGE_SHIFT),
1130 node.start, I915_CACHE_NONE, 0);
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301131 wmb();
1132 } else {
1133 page_base += offset & PAGE_MASK;
1134 }
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001135
Matthew Auld73ebd502017-12-11 15:18:20 +00001136 if (gtt_user_read(&ggtt->iomap, page_base, page_offset,
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001137 user_data, page_length)) {
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301138 ret = -EFAULT;
1139 break;
1140 }
1141
1142 remain -= page_length;
1143 user_data += page_length;
1144 offset += page_length;
1145 }
1146
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001147 mutex_lock(&i915->drm.struct_mutex);
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301148out_unpin:
1149 if (node.allocated) {
1150 wmb();
Chris Wilson82ad6442018-06-05 16:37:58 +01001151 ggtt->vm.clear_range(&ggtt->vm, node.start, node.size);
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301152 remove_mappable_node(&node);
1153 } else {
Chris Wilson058d88c2016-08-15 10:49:06 +01001154 i915_vma_unpin(vma);
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301155 }
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001156out_unlock:
Chris Wilson538ef962019-01-14 14:21:18 +00001157 intel_runtime_pm_put(i915, wakeref);
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001158 mutex_unlock(&i915->drm.struct_mutex);
Chris Wilsonf60d7f02012-09-04 21:02:56 +01001159
Eric Anholteb014592009-03-10 11:44:52 -07001160 return ret;
1161}
1162
Eric Anholt673a3942008-07-30 12:06:12 -07001163/**
1164 * Reads data from the object referenced by handle.
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +01001165 * @dev: drm device pointer
1166 * @data: ioctl data blob
1167 * @file: drm file pointer
Eric Anholt673a3942008-07-30 12:06:12 -07001168 *
1169 * On error, the contents of *data are undefined.
1170 */
1171int
1172i915_gem_pread_ioctl(struct drm_device *dev, void *data,
Chris Wilson05394f32010-11-08 19:18:58 +00001173 struct drm_file *file)
Eric Anholt673a3942008-07-30 12:06:12 -07001174{
1175 struct drm_i915_gem_pread *args = data;
Chris Wilson05394f32010-11-08 19:18:58 +00001176 struct drm_i915_gem_object *obj;
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001177 int ret;
Eric Anholt673a3942008-07-30 12:06:12 -07001178
Chris Wilson51311d02010-11-17 09:10:42 +00001179 if (args->size == 0)
1180 return 0;
1181
Linus Torvalds96d4f262019-01-03 18:57:57 -08001182 if (!access_ok(u64_to_user_ptr(args->data_ptr),
Chris Wilson51311d02010-11-17 09:10:42 +00001183 args->size))
1184 return -EFAULT;
1185
Chris Wilson03ac0642016-07-20 13:31:51 +01001186 obj = i915_gem_object_lookup(file, args->handle);
Chris Wilson258a5ed2016-08-05 10:14:16 +01001187 if (!obj)
1188 return -ENOENT;
Eric Anholt673a3942008-07-30 12:06:12 -07001189
Chris Wilson7dcd2492010-09-26 20:21:44 +01001190 /* Bounds check source. */
Matthew Auld966d5bf2016-12-13 20:32:22 +00001191 if (range_overflows_t(u64, args->offset, args->size, obj->base.size)) {
Chris Wilsonce9d4192010-09-26 20:50:05 +01001192 ret = -EINVAL;
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001193 goto out;
Chris Wilsonce9d4192010-09-26 20:50:05 +01001194 }
1195
Chris Wilsondb53a302011-02-03 11:57:46 +00001196 trace_i915_gem_object_pread(obj, args->offset, args->size);
1197
Chris Wilsone95433c72016-10-28 13:58:27 +01001198 ret = i915_gem_object_wait(obj,
1199 I915_WAIT_INTERRUPTIBLE,
1200 MAX_SCHEDULE_TIMEOUT,
1201 to_rps_client(file));
Chris Wilson258a5ed2016-08-05 10:14:16 +01001202 if (ret)
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001203 goto out;
Chris Wilson258a5ed2016-08-05 10:14:16 +01001204
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001205 ret = i915_gem_object_pin_pages(obj);
Chris Wilson258a5ed2016-08-05 10:14:16 +01001206 if (ret)
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001207 goto out;
Chris Wilson258a5ed2016-08-05 10:14:16 +01001208
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001209 ret = i915_gem_shmem_pread(obj, args);
Chris Wilson9c870d02016-10-24 13:42:15 +01001210 if (ret == -EFAULT || ret == -ENODEV)
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001211 ret = i915_gem_gtt_pread(obj, args);
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301212
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001213 i915_gem_object_unpin_pages(obj);
1214out:
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001215 i915_gem_object_put(obj);
Eric Anholteb014592009-03-10 11:44:52 -07001216 return ret;
Eric Anholt673a3942008-07-30 12:06:12 -07001217}
1218
Keith Packard0839ccb2008-10-30 19:38:48 -07001219/* This is the fast write path which cannot handle
1220 * page faults in the source data
Linus Torvalds9b7530cc2008-10-20 14:16:43 -07001221 */
Linus Torvalds9b7530cc2008-10-20 14:16:43 -07001222
Chris Wilsonfe115622016-10-28 13:58:40 +01001223static inline bool
1224ggtt_write(struct io_mapping *mapping,
1225 loff_t base, int offset,
1226 char __user *user_data, int length)
Keith Packard0839ccb2008-10-30 19:38:48 -07001227{
Ville Syrjäläafe722b2017-09-01 20:12:52 +03001228 void __iomem *vaddr;
Keith Packard0839ccb2008-10-30 19:38:48 -07001229 unsigned long unwritten;
1230
Ben Widawsky4f0c7cf2012-04-16 14:07:47 -07001231 /* We can use the cpu mem copy function because this is X86. */
Ville Syrjäläafe722b2017-09-01 20:12:52 +03001232 vaddr = io_mapping_map_atomic_wc(mapping, base);
1233 unwritten = __copy_from_user_inatomic_nocache((void __force *)vaddr + offset,
Keith Packard0839ccb2008-10-30 19:38:48 -07001234 user_data, length);
Chris Wilsonfe115622016-10-28 13:58:40 +01001235 io_mapping_unmap_atomic(vaddr);
1236 if (unwritten) {
Ville Syrjäläafe722b2017-09-01 20:12:52 +03001237 vaddr = io_mapping_map_wc(mapping, base, PAGE_SIZE);
1238 unwritten = copy_from_user((void __force *)vaddr + offset,
1239 user_data, length);
Chris Wilsonfe115622016-10-28 13:58:40 +01001240 io_mapping_unmap(vaddr);
1241 }
Keith Packard0839ccb2008-10-30 19:38:48 -07001242
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001243 return unwritten;
1244}
1245
Eric Anholt3de09aa2009-03-09 09:42:23 -07001246/**
1247 * This is the fast pwrite path, where we copy the data directly from the
1248 * user into the GTT, uncached.
Chris Wilsonfe115622016-10-28 13:58:40 +01001249 * @obj: i915 GEM object
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +01001250 * @args: pwrite arguments structure
Eric Anholt3de09aa2009-03-09 09:42:23 -07001251 */
Eric Anholt673a3942008-07-30 12:06:12 -07001252static int
Chris Wilsonfe115622016-10-28 13:58:40 +01001253i915_gem_gtt_pwrite_fast(struct drm_i915_gem_object *obj,
1254 const struct drm_i915_gem_pwrite *args)
Eric Anholt673a3942008-07-30 12:06:12 -07001255{
Chris Wilsonfe115622016-10-28 13:58:40 +01001256 struct drm_i915_private *i915 = to_i915(obj->base.dev);
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301257 struct i915_ggtt *ggtt = &i915->ggtt;
Chris Wilson538ef962019-01-14 14:21:18 +00001258 intel_wakeref_t wakeref;
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301259 struct drm_mm_node node;
Chris Wilsonfe115622016-10-28 13:58:40 +01001260 struct i915_vma *vma;
1261 u64 remain, offset;
1262 void __user *user_data;
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301263 int ret;
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301264
Chris Wilsonfe115622016-10-28 13:58:40 +01001265 ret = mutex_lock_interruptible(&i915->drm.struct_mutex);
1266 if (ret)
1267 return ret;
Daniel Vetter935aaa62012-03-25 19:47:35 +02001268
Chris Wilson8bd818152017-10-19 07:37:33 +01001269 if (i915_gem_object_has_struct_page(obj)) {
1270 /*
1271 * Avoid waking the device up if we can fallback, as
1272 * waking/resuming is very slow (worst-case 10-100 ms
1273 * depending on PCI sleeps and our own resume time).
1274 * This easily dwarfs any performance advantage from
1275 * using the cache bypass of indirect GGTT access.
1276 */
Chris Wilson538ef962019-01-14 14:21:18 +00001277 wakeref = intel_runtime_pm_get_if_in_use(i915);
1278 if (!wakeref) {
Chris Wilson8bd818152017-10-19 07:37:33 +01001279 ret = -EFAULT;
1280 goto out_unlock;
1281 }
1282 } else {
1283 /* No backing pages, no fallback, we must force GGTT access */
Chris Wilson538ef962019-01-14 14:21:18 +00001284 wakeref = intel_runtime_pm_get(i915);
Chris Wilson8bd818152017-10-19 07:37:33 +01001285 }
1286
Chris Wilson058d88c2016-08-15 10:49:06 +01001287 vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0,
Chris Wilsona3259ca2017-10-09 09:44:00 +01001288 PIN_MAPPABLE |
1289 PIN_NONFAULT |
1290 PIN_NONBLOCK);
Chris Wilson18034582016-08-18 17:16:45 +01001291 if (!IS_ERR(vma)) {
1292 node.start = i915_ggtt_offset(vma);
1293 node.allocated = false;
Chris Wilson49ef5292016-08-18 17:17:00 +01001294 ret = i915_vma_put_fence(vma);
Chris Wilson18034582016-08-18 17:16:45 +01001295 if (ret) {
1296 i915_vma_unpin(vma);
1297 vma = ERR_PTR(ret);
1298 }
1299 }
Chris Wilson058d88c2016-08-15 10:49:06 +01001300 if (IS_ERR(vma)) {
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001301 ret = insert_mappable_node(ggtt, &node, PAGE_SIZE);
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301302 if (ret)
Chris Wilson8bd818152017-10-19 07:37:33 +01001303 goto out_rpm;
Chris Wilsonfe115622016-10-28 13:58:40 +01001304 GEM_BUG_ON(!node.allocated);
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301305 }
Daniel Vetter935aaa62012-03-25 19:47:35 +02001306
1307 ret = i915_gem_object_set_to_gtt_domain(obj, true);
1308 if (ret)
1309 goto out_unpin;
1310
Chris Wilsonfe115622016-10-28 13:58:40 +01001311 mutex_unlock(&i915->drm.struct_mutex);
1312
Chris Wilsonb19482d2016-08-18 17:16:43 +01001313 intel_fb_obj_invalidate(obj, ORIGIN_CPU);
Paulo Zanoni063e4e62015-02-13 17:23:45 -02001314
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301315 user_data = u64_to_user_ptr(args->data_ptr);
1316 offset = args->offset;
1317 remain = args->size;
1318 while (remain) {
Eric Anholt673a3942008-07-30 12:06:12 -07001319 /* Operation in this page
1320 *
Keith Packard0839ccb2008-10-30 19:38:48 -07001321 * page_base = page offset within aperture
1322 * page_offset = offset within page
1323 * page_length = bytes to copy for this page
Eric Anholt673a3942008-07-30 12:06:12 -07001324 */
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301325 u32 page_base = node.start;
Chris Wilsonbb6dc8d2016-10-28 13:58:39 +01001326 unsigned int page_offset = offset_in_page(offset);
1327 unsigned int page_length = PAGE_SIZE - page_offset;
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301328 page_length = remain < page_length ? remain : page_length;
1329 if (node.allocated) {
1330 wmb(); /* flush the write before we modify the GGTT */
Chris Wilson82ad6442018-06-05 16:37:58 +01001331 ggtt->vm.insert_page(&ggtt->vm,
1332 i915_gem_object_get_dma_address(obj, offset >> PAGE_SHIFT),
1333 node.start, I915_CACHE_NONE, 0);
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301334 wmb(); /* flush modifications to the GGTT (insert_page) */
1335 } else {
1336 page_base += offset & PAGE_MASK;
1337 }
Keith Packard0839ccb2008-10-30 19:38:48 -07001338 /* If we get a fault while copying data, then (presumably) our
Eric Anholt3de09aa2009-03-09 09:42:23 -07001339 * source page isn't available. Return the error and we'll
1340 * retry in the slow path.
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301341 * If the object is non-shmem backed, we retry again with the
1342 * path that handles page fault.
Keith Packard0839ccb2008-10-30 19:38:48 -07001343 */
Matthew Auld73ebd502017-12-11 15:18:20 +00001344 if (ggtt_write(&ggtt->iomap, page_base, page_offset,
Chris Wilsonfe115622016-10-28 13:58:40 +01001345 user_data, page_length)) {
1346 ret = -EFAULT;
1347 break;
Daniel Vetter935aaa62012-03-25 19:47:35 +02001348 }
Eric Anholt673a3942008-07-30 12:06:12 -07001349
Keith Packard0839ccb2008-10-30 19:38:48 -07001350 remain -= page_length;
1351 user_data += page_length;
1352 offset += page_length;
Eric Anholt673a3942008-07-30 12:06:12 -07001353 }
Chris Wilsond59b21e2017-02-22 11:40:49 +00001354 intel_fb_obj_flush(obj, ORIGIN_CPU);
Chris Wilsonfe115622016-10-28 13:58:40 +01001355
1356 mutex_lock(&i915->drm.struct_mutex);
Daniel Vetter935aaa62012-03-25 19:47:35 +02001357out_unpin:
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301358 if (node.allocated) {
1359 wmb();
Chris Wilson82ad6442018-06-05 16:37:58 +01001360 ggtt->vm.clear_range(&ggtt->vm, node.start, node.size);
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301361 remove_mappable_node(&node);
1362 } else {
Chris Wilson058d88c2016-08-15 10:49:06 +01001363 i915_vma_unpin(vma);
Ankitprasad Sharma4f1959e2016-06-10 14:23:01 +05301364 }
Chris Wilson8bd818152017-10-19 07:37:33 +01001365out_rpm:
Chris Wilson538ef962019-01-14 14:21:18 +00001366 intel_runtime_pm_put(i915, wakeref);
Chris Wilson8bd818152017-10-19 07:37:33 +01001367out_unlock:
Chris Wilsonfe115622016-10-28 13:58:40 +01001368 mutex_unlock(&i915->drm.struct_mutex);
Eric Anholt3de09aa2009-03-09 09:42:23 -07001369 return ret;
Eric Anholt673a3942008-07-30 12:06:12 -07001370}
1371
Chris Wilsonfe115622016-10-28 13:58:40 +01001372/* Per-page copy function for the shmem pwrite fastpath.
1373 * Flushes invalid cachelines before writing to the target if
1374 * needs_clflush_before is set and flushes out any written cachelines after
1375 * writing if needs_clflush is set.
1376 */
Eric Anholt40123c12009-03-09 13:42:30 -07001377static int
Chris Wilsonfe115622016-10-28 13:58:40 +01001378shmem_pwrite(struct page *page, int offset, int len, char __user *user_data,
Chris Wilsonfe115622016-10-28 13:58:40 +01001379 bool needs_clflush_before,
1380 bool needs_clflush_after)
Eric Anholt40123c12009-03-09 13:42:30 -07001381{
Chris Wilsonb9d126e2019-01-05 12:07:58 +00001382 char *vaddr;
Chris Wilsonfe115622016-10-28 13:58:40 +01001383 int ret;
Eric Anholt40123c12009-03-09 13:42:30 -07001384
Chris Wilsonb9d126e2019-01-05 12:07:58 +00001385 vaddr = kmap(page);
Chris Wilsonfe115622016-10-28 13:58:40 +01001386
Chris Wilsonb9d126e2019-01-05 12:07:58 +00001387 if (needs_clflush_before)
1388 drm_clflush_virt_range(vaddr + offset, len);
Chris Wilsonfe115622016-10-28 13:58:40 +01001389
Chris Wilsonb9d126e2019-01-05 12:07:58 +00001390 ret = __copy_from_user(vaddr + offset, user_data, len);
1391 if (!ret && needs_clflush_after)
1392 drm_clflush_virt_range(vaddr + offset, len);
Chris Wilsonfe115622016-10-28 13:58:40 +01001393
Chris Wilsonb9d126e2019-01-05 12:07:58 +00001394 kunmap(page);
1395
1396 return ret ? -EFAULT : 0;
Chris Wilsonfe115622016-10-28 13:58:40 +01001397}
1398
1399static int
1400i915_gem_shmem_pwrite(struct drm_i915_gem_object *obj,
1401 const struct drm_i915_gem_pwrite *args)
1402{
1403 struct drm_i915_private *i915 = to_i915(obj->base.dev);
1404 void __user *user_data;
1405 u64 remain;
Chris Wilsonfe115622016-10-28 13:58:40 +01001406 unsigned int partial_cacheline_write;
1407 unsigned int needs_clflush;
1408 unsigned int offset, idx;
1409 int ret;
1410
1411 ret = mutex_lock_interruptible(&i915->drm.struct_mutex);
Chris Wilson43394c72016-08-18 17:16:47 +01001412 if (ret)
1413 return ret;
Eric Anholt40123c12009-03-09 13:42:30 -07001414
Chris Wilsonfe115622016-10-28 13:58:40 +01001415 ret = i915_gem_obj_prepare_shmem_write(obj, &needs_clflush);
1416 mutex_unlock(&i915->drm.struct_mutex);
1417 if (ret)
1418 return ret;
1419
Chris Wilsonfe115622016-10-28 13:58:40 +01001420 /* If we don't overwrite a cacheline completely we need to be
1421 * careful to have up-to-date data by first clflushing. Don't
1422 * overcomplicate things and flush the entire patch.
1423 */
1424 partial_cacheline_write = 0;
1425 if (needs_clflush & CLFLUSH_BEFORE)
1426 partial_cacheline_write = boot_cpu_data.x86_clflush_size - 1;
1427
Chris Wilson43394c72016-08-18 17:16:47 +01001428 user_data = u64_to_user_ptr(args->data_ptr);
Chris Wilson43394c72016-08-18 17:16:47 +01001429 remain = args->size;
Chris Wilsonfe115622016-10-28 13:58:40 +01001430 offset = offset_in_page(args->offset);
1431 for (idx = args->offset >> PAGE_SHIFT; remain; idx++) {
1432 struct page *page = i915_gem_object_get_page(obj, idx);
Chris Wilsona5e856a52018-10-12 15:02:28 +01001433 unsigned int length = min_t(u64, remain, PAGE_SIZE - offset);
Chris Wilsone5281cc2010-10-28 13:45:36 +01001434
Chris Wilsonfe115622016-10-28 13:58:40 +01001435 ret = shmem_pwrite(page, offset, length, user_data,
Chris Wilsonfe115622016-10-28 13:58:40 +01001436 (offset | length) & partial_cacheline_write,
1437 needs_clflush & CLFLUSH_AFTER);
1438 if (ret)
Chris Wilson9da3da62012-06-01 15:20:22 +01001439 break;
1440
Chris Wilsonfe115622016-10-28 13:58:40 +01001441 remain -= length;
1442 user_data += length;
1443 offset = 0;
Eric Anholt40123c12009-03-09 13:42:30 -07001444 }
1445
Chris Wilsond59b21e2017-02-22 11:40:49 +00001446 intel_fb_obj_flush(obj, ORIGIN_CPU);
Chris Wilsonfe115622016-10-28 13:58:40 +01001447 i915_gem_obj_finish_shmem_access(obj);
Eric Anholt40123c12009-03-09 13:42:30 -07001448 return ret;
Eric Anholt673a3942008-07-30 12:06:12 -07001449}
1450
1451/**
1452 * Writes data to the object referenced by handle.
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +01001453 * @dev: drm device
1454 * @data: ioctl data blob
1455 * @file: drm file
Eric Anholt673a3942008-07-30 12:06:12 -07001456 *
1457 * On error, the contents of the buffer that were to be modified are undefined.
1458 */
1459int
1460i915_gem_pwrite_ioctl(struct drm_device *dev, void *data,
Chris Wilsonfbd5a262010-10-14 15:03:58 +01001461 struct drm_file *file)
Eric Anholt673a3942008-07-30 12:06:12 -07001462{
1463 struct drm_i915_gem_pwrite *args = data;
Chris Wilson05394f32010-11-08 19:18:58 +00001464 struct drm_i915_gem_object *obj;
Chris Wilson51311d02010-11-17 09:10:42 +00001465 int ret;
1466
1467 if (args->size == 0)
1468 return 0;
1469
Linus Torvalds96d4f262019-01-03 18:57:57 -08001470 if (!access_ok(u64_to_user_ptr(args->data_ptr), args->size))
Chris Wilson51311d02010-11-17 09:10:42 +00001471 return -EFAULT;
1472
Chris Wilson03ac0642016-07-20 13:31:51 +01001473 obj = i915_gem_object_lookup(file, args->handle);
Chris Wilson258a5ed2016-08-05 10:14:16 +01001474 if (!obj)
1475 return -ENOENT;
Eric Anholt673a3942008-07-30 12:06:12 -07001476
Chris Wilson7dcd2492010-09-26 20:21:44 +01001477 /* Bounds check destination. */
Matthew Auld966d5bf2016-12-13 20:32:22 +00001478 if (range_overflows_t(u64, args->offset, args->size, obj->base.size)) {
Chris Wilsonce9d4192010-09-26 20:50:05 +01001479 ret = -EINVAL;
Chris Wilson258a5ed2016-08-05 10:14:16 +01001480 goto err;
Chris Wilsonce9d4192010-09-26 20:50:05 +01001481 }
1482
Chris Wilsonf8c1cce2018-07-12 19:53:14 +01001483 /* Writes not allowed into this read-only object */
1484 if (i915_gem_object_is_readonly(obj)) {
1485 ret = -EINVAL;
1486 goto err;
1487 }
1488
Chris Wilsondb53a302011-02-03 11:57:46 +00001489 trace_i915_gem_object_pwrite(obj, args->offset, args->size);
1490
Chris Wilson7c55e2c2017-03-07 12:03:38 +00001491 ret = -ENODEV;
1492 if (obj->ops->pwrite)
1493 ret = obj->ops->pwrite(obj, args);
1494 if (ret != -ENODEV)
1495 goto err;
1496
Chris Wilsone95433c72016-10-28 13:58:27 +01001497 ret = i915_gem_object_wait(obj,
1498 I915_WAIT_INTERRUPTIBLE |
1499 I915_WAIT_ALL,
1500 MAX_SCHEDULE_TIMEOUT,
1501 to_rps_client(file));
Chris Wilson258a5ed2016-08-05 10:14:16 +01001502 if (ret)
1503 goto err;
1504
Chris Wilsonfe115622016-10-28 13:58:40 +01001505 ret = i915_gem_object_pin_pages(obj);
Chris Wilson258a5ed2016-08-05 10:14:16 +01001506 if (ret)
Chris Wilsonfe115622016-10-28 13:58:40 +01001507 goto err;
Chris Wilson258a5ed2016-08-05 10:14:16 +01001508
Daniel Vetter935aaa62012-03-25 19:47:35 +02001509 ret = -EFAULT;
Eric Anholt673a3942008-07-30 12:06:12 -07001510 /* We can only do the GTT pwrite on untiled buffers, as otherwise
1511 * it would end up going through the fenced access, and we'll get
1512 * different detiling behavior between reading and writing.
1513 * pread/pwrite currently are reading and writing from the CPU
1514 * perspective, requiring manual detiling by the client.
1515 */
Chris Wilson6eae0052016-06-20 15:05:52 +01001516 if (!i915_gem_object_has_struct_page(obj) ||
Chris Wilson9c870d02016-10-24 13:42:15 +01001517 cpu_write_needs_clflush(obj))
Daniel Vetter935aaa62012-03-25 19:47:35 +02001518 /* Note that the gtt paths might fail with non-page-backed user
1519 * pointers (e.g. gtt mappings when moving data between
Chris Wilson9c870d02016-10-24 13:42:15 +01001520 * textures). Fallback to the shmem path in that case.
1521 */
Chris Wilsonfe115622016-10-28 13:58:40 +01001522 ret = i915_gem_gtt_pwrite_fast(obj, args);
Eric Anholt673a3942008-07-30 12:06:12 -07001523
Chris Wilsond1054ee2016-07-16 18:42:36 +01001524 if (ret == -EFAULT || ret == -ENOSPC) {
Chris Wilson6a2c4232014-11-04 04:51:40 -08001525 if (obj->phys_handle)
1526 ret = i915_gem_phys_pwrite(obj, args, file);
Ankitprasad Sharmab50a5372016-06-10 14:23:03 +05301527 else
Chris Wilsonfe115622016-10-28 13:58:40 +01001528 ret = i915_gem_shmem_pwrite(obj, args);
Chris Wilson6a2c4232014-11-04 04:51:40 -08001529 }
Daniel Vetter5c0480f2011-12-14 13:57:30 +01001530
Chris Wilsonfe115622016-10-28 13:58:40 +01001531 i915_gem_object_unpin_pages(obj);
Chris Wilson258a5ed2016-08-05 10:14:16 +01001532err:
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001533 i915_gem_object_put(obj);
Chris Wilson258a5ed2016-08-05 10:14:16 +01001534 return ret;
Eric Anholt673a3942008-07-30 12:06:12 -07001535}
1536
Chris Wilson40e62d52016-10-28 13:58:41 +01001537static void i915_gem_object_bump_inactive_ggtt(struct drm_i915_gem_object *obj)
1538{
Chris Wilson09d7e462019-01-28 10:23:53 +00001539 struct drm_i915_private *i915 = to_i915(obj->base.dev);
Chris Wilson40e62d52016-10-28 13:58:41 +01001540 struct list_head *list;
1541 struct i915_vma *vma;
1542
Chris Wilsonf2123812017-10-16 12:40:37 +01001543 GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
1544
Chris Wilson09d7e462019-01-28 10:23:53 +00001545 mutex_lock(&i915->ggtt.vm.mutex);
Chris Wilsone2189dd2017-12-07 21:14:07 +00001546 for_each_ggtt_vma(vma, obj) {
Chris Wilson40e62d52016-10-28 13:58:41 +01001547 if (!drm_mm_node_allocated(&vma->node))
1548 continue;
1549
Chris Wilson499197d2019-01-28 10:23:52 +00001550 list_move_tail(&vma->vm_link, &vma->vm->bound_list);
Chris Wilson40e62d52016-10-28 13:58:41 +01001551 }
Chris Wilson09d7e462019-01-28 10:23:53 +00001552 mutex_unlock(&i915->ggtt.vm.mutex);
Chris Wilson40e62d52016-10-28 13:58:41 +01001553
Chris Wilsonf2123812017-10-16 12:40:37 +01001554 spin_lock(&i915->mm.obj_lock);
Chris Wilson40e62d52016-10-28 13:58:41 +01001555 list = obj->bind_count ? &i915->mm.bound_list : &i915->mm.unbound_list;
Chris Wilsonf2123812017-10-16 12:40:37 +01001556 list_move_tail(&obj->mm.link, list);
1557 spin_unlock(&i915->mm.obj_lock);
Chris Wilson40e62d52016-10-28 13:58:41 +01001558}
1559
Eric Anholt673a3942008-07-30 12:06:12 -07001560/**
Eric Anholt2ef7eea2008-11-10 10:53:25 -08001561 * Called when user space prepares to use an object with the CPU, either
1562 * through the mmap ioctl's mapping or a GTT mapping.
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +01001563 * @dev: drm device
1564 * @data: ioctl data blob
1565 * @file: drm file
Eric Anholt673a3942008-07-30 12:06:12 -07001566 */
1567int
1568i915_gem_set_domain_ioctl(struct drm_device *dev, void *data,
Chris Wilson05394f32010-11-08 19:18:58 +00001569 struct drm_file *file)
Eric Anholt673a3942008-07-30 12:06:12 -07001570{
1571 struct drm_i915_gem_set_domain *args = data;
Chris Wilson05394f32010-11-08 19:18:58 +00001572 struct drm_i915_gem_object *obj;
Jani Nikula739f3ab2019-01-16 11:15:19 +02001573 u32 read_domains = args->read_domains;
1574 u32 write_domain = args->write_domain;
Chris Wilson40e62d52016-10-28 13:58:41 +01001575 int err;
Eric Anholt673a3942008-07-30 12:06:12 -07001576
Eric Anholt2ef7eea2008-11-10 10:53:25 -08001577 /* Only handle setting domains to types used by the CPU. */
Chris Wilsonb8f90962016-08-05 10:14:07 +01001578 if ((write_domain | read_domains) & I915_GEM_GPU_DOMAINS)
Eric Anholt2ef7eea2008-11-10 10:53:25 -08001579 return -EINVAL;
1580
1581 /* Having something in the write domain implies it's in the read
1582 * domain, and only that read domain. Enforce that in the request.
1583 */
1584 if (write_domain != 0 && read_domains != write_domain)
1585 return -EINVAL;
1586
Chris Wilson03ac0642016-07-20 13:31:51 +01001587 obj = i915_gem_object_lookup(file, args->handle);
Chris Wilsonb8f90962016-08-05 10:14:07 +01001588 if (!obj)
1589 return -ENOENT;
Jesse Barnes652c3932009-08-17 13:31:43 -07001590
Chris Wilson3236f572012-08-24 09:35:09 +01001591 /* Try to flush the object off the GPU without holding the lock.
1592 * We will repeat the flush holding the lock in the normal manner
1593 * to catch cases where we are gazumped.
1594 */
Chris Wilson40e62d52016-10-28 13:58:41 +01001595 err = i915_gem_object_wait(obj,
Chris Wilsone95433c72016-10-28 13:58:27 +01001596 I915_WAIT_INTERRUPTIBLE |
Chris Wilsone9eaf822018-10-01 15:47:55 +01001597 I915_WAIT_PRIORITY |
Chris Wilsone95433c72016-10-28 13:58:27 +01001598 (write_domain ? I915_WAIT_ALL : 0),
1599 MAX_SCHEDULE_TIMEOUT,
1600 to_rps_client(file));
Chris Wilson40e62d52016-10-28 13:58:41 +01001601 if (err)
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001602 goto out;
Chris Wilsonb8f90962016-08-05 10:14:07 +01001603
Tina Zhanga03f3952017-11-14 10:25:13 +00001604 /*
1605 * Proxy objects do not control access to the backing storage, ergo
1606 * they cannot be used as a means to manipulate the cache domain
1607 * tracking for that backing storage. The proxy object is always
1608 * considered to be outside of any cache domain.
1609 */
1610 if (i915_gem_object_is_proxy(obj)) {
1611 err = -ENXIO;
1612 goto out;
1613 }
1614
1615 /*
1616 * Flush and acquire obj->pages so that we are coherent through
Chris Wilson40e62d52016-10-28 13:58:41 +01001617 * direct access in memory with previous cached writes through
1618 * shmemfs and that our cache domain tracking remains valid.
1619 * For example, if the obj->filp was moved to swap without us
1620 * being notified and releasing the pages, we would mistakenly
1621 * continue to assume that the obj remained out of the CPU cached
1622 * domain.
1623 */
1624 err = i915_gem_object_pin_pages(obj);
1625 if (err)
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001626 goto out;
Chris Wilson40e62d52016-10-28 13:58:41 +01001627
1628 err = i915_mutex_lock_interruptible(dev);
1629 if (err)
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001630 goto out_unpin;
Chris Wilson3236f572012-08-24 09:35:09 +01001631
Chris Wilsone22d8e32017-04-12 12:01:11 +01001632 if (read_domains & I915_GEM_DOMAIN_WC)
1633 err = i915_gem_object_set_to_wc_domain(obj, write_domain);
1634 else if (read_domains & I915_GEM_DOMAIN_GTT)
1635 err = i915_gem_object_set_to_gtt_domain(obj, write_domain);
Chris Wilson43566de2015-01-02 16:29:29 +05301636 else
Chris Wilsone22d8e32017-04-12 12:01:11 +01001637 err = i915_gem_object_set_to_cpu_domain(obj, write_domain);
Chris Wilson40e62d52016-10-28 13:58:41 +01001638
1639 /* And bump the LRU for this access */
1640 i915_gem_object_bump_inactive_ggtt(obj);
1641
1642 mutex_unlock(&dev->struct_mutex);
Eric Anholt2ef7eea2008-11-10 10:53:25 -08001643
Daniel Vetter031b6982015-06-26 19:35:16 +02001644 if (write_domain != 0)
Chris Wilsonef749212017-04-12 12:01:10 +01001645 intel_fb_obj_invalidate(obj,
1646 fb_write_origin(obj, write_domain));
Daniel Vetter031b6982015-06-26 19:35:16 +02001647
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001648out_unpin:
Chris Wilson40e62d52016-10-28 13:58:41 +01001649 i915_gem_object_unpin_pages(obj);
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001650out:
1651 i915_gem_object_put(obj);
Chris Wilson40e62d52016-10-28 13:58:41 +01001652 return err;
Eric Anholt673a3942008-07-30 12:06:12 -07001653}
1654
1655/**
1656 * Called when user space has done writes to this buffer
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +01001657 * @dev: drm device
1658 * @data: ioctl data blob
1659 * @file: drm file
Eric Anholt673a3942008-07-30 12:06:12 -07001660 */
1661int
1662i915_gem_sw_finish_ioctl(struct drm_device *dev, void *data,
Chris Wilson05394f32010-11-08 19:18:58 +00001663 struct drm_file *file)
Eric Anholt673a3942008-07-30 12:06:12 -07001664{
1665 struct drm_i915_gem_sw_finish *args = data;
Chris Wilson05394f32010-11-08 19:18:58 +00001666 struct drm_i915_gem_object *obj;
Chris Wilson1d7cfea2010-10-17 09:45:41 +01001667
Chris Wilson03ac0642016-07-20 13:31:51 +01001668 obj = i915_gem_object_lookup(file, args->handle);
Chris Wilsonc21724c2016-08-05 10:14:19 +01001669 if (!obj)
1670 return -ENOENT;
Eric Anholt673a3942008-07-30 12:06:12 -07001671
Tina Zhanga03f3952017-11-14 10:25:13 +00001672 /*
1673 * Proxy objects are barred from CPU access, so there is no
1674 * need to ban sw_finish as it is a nop.
1675 */
1676
Eric Anholt673a3942008-07-30 12:06:12 -07001677 /* Pinned buffers may be scanout, so flush the cache */
Chris Wilson5a97bcc2017-02-22 11:40:46 +00001678 i915_gem_object_flush_if_display(obj);
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001679 i915_gem_object_put(obj);
Chris Wilson5a97bcc2017-02-22 11:40:46 +00001680
1681 return 0;
Eric Anholt673a3942008-07-30 12:06:12 -07001682}
1683
Joonas Lahtinen5c4604e2019-02-07 10:54:53 +02001684static inline bool
1685__vma_matches(struct vm_area_struct *vma, struct file *filp,
1686 unsigned long addr, unsigned long size)
1687{
1688 if (vma->vm_file != filp)
1689 return false;
1690
1691 return vma->vm_start == addr && (vma->vm_end - vma->vm_start) == size;
1692}
1693
Eric Anholt673a3942008-07-30 12:06:12 -07001694/**
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +01001695 * i915_gem_mmap_ioctl - Maps the contents of an object, returning the address
1696 * it is mapped to.
1697 * @dev: drm device
1698 * @data: ioctl data blob
1699 * @file: drm file
Eric Anholt673a3942008-07-30 12:06:12 -07001700 *
1701 * While the mapping holds a reference on the contents of the object, it doesn't
1702 * imply a ref on the object itself.
Daniel Vetter34367382014-10-16 12:28:18 +02001703 *
1704 * IMPORTANT:
1705 *
1706 * DRM driver writers who look a this function as an example for how to do GEM
1707 * mmap support, please don't implement mmap support like here. The modern way
1708 * to implement DRM mmap support is with an mmap offset ioctl (like
1709 * i915_gem_mmap_gtt) and then using the mmap syscall on the DRM fd directly.
1710 * That way debug tooling like valgrind will understand what's going on, hiding
1711 * the mmap call in a driver private ioctl will break that. The i915 driver only
1712 * does cpu mmaps this way because we didn't know better.
Eric Anholt673a3942008-07-30 12:06:12 -07001713 */
1714int
1715i915_gem_mmap_ioctl(struct drm_device *dev, void *data,
Chris Wilson05394f32010-11-08 19:18:58 +00001716 struct drm_file *file)
Eric Anholt673a3942008-07-30 12:06:12 -07001717{
1718 struct drm_i915_gem_mmap *args = data;
Chris Wilson03ac0642016-07-20 13:31:51 +01001719 struct drm_i915_gem_object *obj;
Eric Anholt673a3942008-07-30 12:06:12 -07001720 unsigned long addr;
1721
Akash Goel1816f922015-01-02 16:29:30 +05301722 if (args->flags & ~(I915_MMAP_WC))
1723 return -EINVAL;
1724
Borislav Petkov568a58e2016-03-29 17:42:01 +02001725 if (args->flags & I915_MMAP_WC && !boot_cpu_has(X86_FEATURE_PAT))
Akash Goel1816f922015-01-02 16:29:30 +05301726 return -ENODEV;
1727
Chris Wilson03ac0642016-07-20 13:31:51 +01001728 obj = i915_gem_object_lookup(file, args->handle);
1729 if (!obj)
Chris Wilsonbf79cb92010-08-04 14:19:46 +01001730 return -ENOENT;
Eric Anholt673a3942008-07-30 12:06:12 -07001731
Daniel Vetter1286ff72012-05-10 15:25:09 +02001732 /* prime objects have no backing filp to GEM mmap
1733 * pages from.
1734 */
Chris Wilson03ac0642016-07-20 13:31:51 +01001735 if (!obj->base.filp) {
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001736 i915_gem_object_put(obj);
Tina Zhang274b2462017-11-14 10:25:12 +00001737 return -ENXIO;
Daniel Vetter1286ff72012-05-10 15:25:09 +02001738 }
1739
Chris Wilson03ac0642016-07-20 13:31:51 +01001740 addr = vm_mmap(obj->base.filp, 0, args->size,
Eric Anholt673a3942008-07-30 12:06:12 -07001741 PROT_READ | PROT_WRITE, MAP_SHARED,
1742 args->offset);
Joonas Lahtinenebfb6972019-02-07 10:54:54 +02001743 if (IS_ERR_VALUE(addr))
1744 goto err;
1745
Akash Goel1816f922015-01-02 16:29:30 +05301746 if (args->flags & I915_MMAP_WC) {
1747 struct mm_struct *mm = current->mm;
1748 struct vm_area_struct *vma;
1749
Michal Hocko80a89a52016-05-23 16:26:11 -07001750 if (down_write_killable(&mm->mmap_sem)) {
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001751 i915_gem_object_put(obj);
Michal Hocko80a89a52016-05-23 16:26:11 -07001752 return -EINTR;
1753 }
Akash Goel1816f922015-01-02 16:29:30 +05301754 vma = find_vma(mm, addr);
Joonas Lahtinen5c4604e2019-02-07 10:54:53 +02001755 if (vma && __vma_matches(vma, obj->base.filp, addr, args->size))
Akash Goel1816f922015-01-02 16:29:30 +05301756 vma->vm_page_prot =
1757 pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
1758 else
1759 addr = -ENOMEM;
1760 up_write(&mm->mmap_sem);
Joonas Lahtinenebfb6972019-02-07 10:54:54 +02001761 if (IS_ERR_VALUE(addr))
1762 goto err;
Chris Wilsonaeecc962016-06-17 14:46:39 -03001763
1764 /* This may race, but that's ok, it only gets set */
Chris Wilson50349242016-08-18 17:17:04 +01001765 WRITE_ONCE(obj->frontbuffer_ggtt_origin, ORIGIN_CPU);
Akash Goel1816f922015-01-02 16:29:30 +05301766 }
Chris Wilsonf0cd5182016-10-28 13:58:43 +01001767 i915_gem_object_put(obj);
Eric Anholt673a3942008-07-30 12:06:12 -07001768
Jani Nikula739f3ab2019-01-16 11:15:19 +02001769 args->addr_ptr = (u64)addr;
Eric Anholt673a3942008-07-30 12:06:12 -07001770
1771 return 0;
Joonas Lahtinenebfb6972019-02-07 10:54:54 +02001772
1773err:
1774 i915_gem_object_put(obj);
1775
1776 return addr;
Eric Anholt673a3942008-07-30 12:06:12 -07001777}
1778
Chris Wilsond899ace2018-07-25 16:54:47 +01001779static unsigned int tile_row_pages(const struct drm_i915_gem_object *obj)
Chris Wilson03af84f2016-08-18 17:17:01 +01001780{
Chris Wilson6649a0b62017-01-09 16:16:08 +00001781 return i915_gem_object_get_tile_row_size(obj) >> PAGE_SHIFT;
Chris Wilson03af84f2016-08-18 17:17:01 +01001782}
1783
Jesse Barnesde151cf2008-11-12 10:03:55 -08001784/**
Chris Wilson4cc69072016-08-25 19:05:19 +01001785 * i915_gem_mmap_gtt_version - report the current feature set for GTT mmaps
1786 *
1787 * A history of the GTT mmap interface:
1788 *
1789 * 0 - Everything had to fit into the GTT. Both parties of a memcpy had to
1790 * aligned and suitable for fencing, and still fit into the available
1791 * mappable space left by the pinned display objects. A classic problem
1792 * we called the page-fault-of-doom where we would ping-pong between
1793 * two objects that could not fit inside the GTT and so the memcpy
1794 * would page one object in at the expense of the other between every
1795 * single byte.
1796 *
1797 * 1 - Objects can be any size, and have any compatible fencing (X Y, or none
1798 * as set via i915_gem_set_tiling() [DRM_I915_GEM_SET_TILING]). If the
1799 * object is too large for the available space (or simply too large
1800 * for the mappable aperture!), a view is created instead and faulted
1801 * into userspace. (This view is aligned and sized appropriately for
1802 * fenced access.)
1803 *
Chris Wilsone22d8e32017-04-12 12:01:11 +01001804 * 2 - Recognise WC as a separate cache domain so that we can flush the
1805 * delayed writes via GTT before performing direct access via WC.
1806 *
Chris Wilson4cc69072016-08-25 19:05:19 +01001807 * Restrictions:
1808 *
1809 * * snoopable objects cannot be accessed via the GTT. It can cause machine
1810 * hangs on some architectures, corruption on others. An attempt to service
1811 * a GTT page fault from a snoopable object will generate a SIGBUS.
1812 *
1813 * * the object must be able to fit into RAM (physical memory, though no
1814 * limited to the mappable aperture).
1815 *
1816 *
1817 * Caveats:
1818 *
1819 * * a new GTT page fault will synchronize rendering from the GPU and flush
1820 * all data to system memory. Subsequent access will not be synchronized.
1821 *
1822 * * all mappings are revoked on runtime device suspend.
1823 *
1824 * * there are only 8, 16 or 32 fence registers to share between all users
1825 * (older machines require fence register for display and blitter access
1826 * as well). Contention of the fence registers will cause the previous users
1827 * to be unmapped and any new access will generate new page faults.
1828 *
1829 * * running out of memory while servicing a fault may generate a SIGBUS,
1830 * rather than the expected SIGSEGV.
1831 */
1832int i915_gem_mmap_gtt_version(void)
1833{
Chris Wilsone22d8e32017-04-12 12:01:11 +01001834 return 2;
Chris Wilson4cc69072016-08-25 19:05:19 +01001835}
1836
Chris Wilson2d4281b2017-01-10 09:56:32 +00001837static inline struct i915_ggtt_view
Chris Wilsond899ace2018-07-25 16:54:47 +01001838compute_partial_view(const struct drm_i915_gem_object *obj,
Chris Wilson2d4281b2017-01-10 09:56:32 +00001839 pgoff_t page_offset,
1840 unsigned int chunk)
1841{
1842 struct i915_ggtt_view view;
1843
1844 if (i915_gem_object_is_tiled(obj))
1845 chunk = roundup(chunk, tile_row_pages(obj));
1846
Chris Wilson2d4281b2017-01-10 09:56:32 +00001847 view.type = I915_GGTT_VIEW_PARTIAL;
Chris Wilson8bab11932017-01-14 00:28:25 +00001848 view.partial.offset = rounddown(page_offset, chunk);
1849 view.partial.size =
Chris Wilson2d4281b2017-01-10 09:56:32 +00001850 min_t(unsigned int, chunk,
Chris Wilson8bab11932017-01-14 00:28:25 +00001851 (obj->base.size >> PAGE_SHIFT) - view.partial.offset);
Chris Wilson2d4281b2017-01-10 09:56:32 +00001852
1853 /* If the partial covers the entire object, just create a normal VMA. */
1854 if (chunk >= obj->base.size >> PAGE_SHIFT)
1855 view.type = I915_GGTT_VIEW_NORMAL;
1856
1857 return view;
1858}
1859
Chris Wilson4cc69072016-08-25 19:05:19 +01001860/**
Jesse Barnesde151cf2008-11-12 10:03:55 -08001861 * i915_gem_fault - fault a page into the GTT
Geliang Tangd9072a32015-09-15 05:58:44 -07001862 * @vmf: fault info
Jesse Barnesde151cf2008-11-12 10:03:55 -08001863 *
1864 * The fault handler is set up by drm_gem_mmap() when a object is GTT mapped
1865 * from userspace. The fault handler takes care of binding the object to
1866 * the GTT (if needed), allocating and programming a fence register (again,
1867 * only if needed based on whether the old reg is still valid or the object
1868 * is tiled) and inserting a new PTE into the faulting process.
1869 *
1870 * Note that the faulting process may involve evicting existing objects
1871 * from the GTT and/or fence registers to make room. So performance may
1872 * suffer if the GTT working set is large or there are few fence registers
1873 * left.
Chris Wilson4cc69072016-08-25 19:05:19 +01001874 *
1875 * The current feature set supported by i915_gem_fault() and thus GTT mmaps
1876 * is exposed via I915_PARAM_MMAP_GTT_VERSION (see i915_gem_mmap_gtt_version).
Jesse Barnesde151cf2008-11-12 10:03:55 -08001877 */
Chris Wilson52137012018-06-06 22:45:20 +01001878vm_fault_t i915_gem_fault(struct vm_fault *vmf)
Jesse Barnesde151cf2008-11-12 10:03:55 -08001879{
Chris Wilson420980c2018-06-05 14:57:46 +01001880#define MIN_CHUNK_PAGES (SZ_1M >> PAGE_SHIFT)
Dave Jiang11bac802017-02-24 14:56:41 -08001881 struct vm_area_struct *area = vmf->vma;
Chris Wilson058d88c2016-08-15 10:49:06 +01001882 struct drm_i915_gem_object *obj = to_intel_bo(area->vm_private_data);
Chris Wilson05394f32010-11-08 19:18:58 +00001883 struct drm_device *dev = obj->base.dev;
Joonas Lahtinen72e96d642016-03-30 16:57:10 +03001884 struct drm_i915_private *dev_priv = to_i915(dev);
1885 struct i915_ggtt *ggtt = &dev_priv->ggtt;
Chris Wilsonaae7c062018-09-03 09:33:34 +01001886 bool write = area->vm_flags & VM_WRITE;
Chris Wilson538ef962019-01-14 14:21:18 +00001887 intel_wakeref_t wakeref;
Chris Wilson058d88c2016-08-15 10:49:06 +01001888 struct i915_vma *vma;
Jesse Barnesde151cf2008-11-12 10:03:55 -08001889 pgoff_t page_offset;
Chris Wilsonb8f90962016-08-05 10:14:07 +01001890 int ret;
Paulo Zanonif65c9162013-11-27 18:20:34 -02001891
Chris Wilson3e977ac2018-07-12 19:53:13 +01001892 /* Sanity check that we allow writing into this object */
1893 if (i915_gem_object_is_readonly(obj) && write)
1894 return VM_FAULT_SIGBUS;
1895
Jesse Barnesde151cf2008-11-12 10:03:55 -08001896 /* We don't use vmf->pgoff since that has the fake offset */
Jan Kara1a29d852016-12-14 15:07:01 -08001897 page_offset = (vmf->address - area->vm_start) >> PAGE_SHIFT;
Jesse Barnesde151cf2008-11-12 10:03:55 -08001898
Chris Wilsondb53a302011-02-03 11:57:46 +00001899 trace_i915_gem_object_fault(obj, page_offset, true, write);
1900
Chris Wilson6e4930f2014-02-07 18:37:06 -02001901 /* Try to flush the object off the GPU first without holding the lock.
Chris Wilsonb8f90962016-08-05 10:14:07 +01001902 * Upon acquiring the lock, we will perform our sanity checks and then
Chris Wilson6e4930f2014-02-07 18:37:06 -02001903 * repeat the flush holding the lock in the normal manner to catch cases
1904 * where we are gazumped.
1905 */
Chris Wilsone95433c72016-10-28 13:58:27 +01001906 ret = i915_gem_object_wait(obj,
1907 I915_WAIT_INTERRUPTIBLE,
1908 MAX_SCHEDULE_TIMEOUT,
1909 NULL);
Chris Wilson6e4930f2014-02-07 18:37:06 -02001910 if (ret)
Chris Wilsonb8f90962016-08-05 10:14:07 +01001911 goto err;
1912
Chris Wilson40e62d52016-10-28 13:58:41 +01001913 ret = i915_gem_object_pin_pages(obj);
1914 if (ret)
1915 goto err;
1916
Chris Wilson538ef962019-01-14 14:21:18 +00001917 wakeref = intel_runtime_pm_get(dev_priv);
Chris Wilsonb8f90962016-08-05 10:14:07 +01001918
1919 ret = i915_mutex_lock_interruptible(dev);
1920 if (ret)
1921 goto err_rpm;
Chris Wilson6e4930f2014-02-07 18:37:06 -02001922
Chris Wilsoneb119bd2012-12-16 12:43:36 +00001923 /* Access to snoopable pages through the GTT is incoherent. */
Tvrtko Ursulin0031fb92016-11-04 14:42:44 +00001924 if (obj->cache_level != I915_CACHE_NONE && !HAS_LLC(dev_priv)) {
Chris Wilsonddeff6e2014-05-28 16:16:41 +01001925 ret = -EFAULT;
Chris Wilsonb8f90962016-08-05 10:14:07 +01001926 goto err_unlock;
Chris Wilsoneb119bd2012-12-16 12:43:36 +00001927 }
1928
Chris Wilson82118872016-08-18 17:17:05 +01001929
Chris Wilsona61007a82016-08-18 17:17:02 +01001930 /* Now pin it into the GTT as needed */
Chris Wilson7e7367d2018-06-30 10:05:09 +01001931 vma = i915_gem_object_ggtt_pin(obj, NULL, 0, 0,
1932 PIN_MAPPABLE |
1933 PIN_NONBLOCK |
1934 PIN_NONFAULT);
Chris Wilsona61007a82016-08-18 17:17:02 +01001935 if (IS_ERR(vma)) {
Chris Wilsona61007a82016-08-18 17:17:02 +01001936 /* Use a partial view if it is bigger than available space */
Chris Wilson2d4281b2017-01-10 09:56:32 +00001937 struct i915_ggtt_view view =
Chris Wilson8201c1f2017-01-10 09:56:33 +00001938 compute_partial_view(obj, page_offset, MIN_CHUNK_PAGES);
Chris Wilson7e7367d2018-06-30 10:05:09 +01001939 unsigned int flags;
Chris Wilsonaa136d92016-08-18 17:17:03 +01001940
Chris Wilson7e7367d2018-06-30 10:05:09 +01001941 flags = PIN_MAPPABLE;
1942 if (view.type == I915_GGTT_VIEW_NORMAL)
1943 flags |= PIN_NONBLOCK; /* avoid warnings for pinned */
1944
1945 /*
1946 * Userspace is now writing through an untracked VMA, abandon
Chris Wilson50349242016-08-18 17:17:04 +01001947 * all hope that the hardware is able to track future writes.
1948 */
1949 obj->frontbuffer_ggtt_origin = ORIGIN_CPU;
1950
Chris Wilson7e7367d2018-06-30 10:05:09 +01001951 vma = i915_gem_object_ggtt_pin(obj, &view, 0, 0, flags);
1952 if (IS_ERR(vma) && !view.type) {
1953 flags = PIN_MAPPABLE;
1954 view.type = I915_GGTT_VIEW_PARTIAL;
1955 vma = i915_gem_object_ggtt_pin(obj, &view, 0, 0, flags);
1956 }
Chris Wilsona61007a82016-08-18 17:17:02 +01001957 }
Chris Wilson058d88c2016-08-15 10:49:06 +01001958 if (IS_ERR(vma)) {
1959 ret = PTR_ERR(vma);
Chris Wilsonb8f90962016-08-05 10:14:07 +01001960 goto err_unlock;
Chris Wilson058d88c2016-08-15 10:49:06 +01001961 }
Jesse Barnesde151cf2008-11-12 10:03:55 -08001962
Chris Wilsonc9839302012-11-20 10:45:17 +00001963 ret = i915_gem_object_set_to_gtt_domain(obj, write);
1964 if (ret)
Chris Wilsonb8f90962016-08-05 10:14:07 +01001965 goto err_unpin;
Chris Wilsonc9839302012-11-20 10:45:17 +00001966
Chris Wilson3bd40732017-10-09 09:43:56 +01001967 ret = i915_vma_pin_fence(vma);
Chris Wilsonc9839302012-11-20 10:45:17 +00001968 if (ret)
Chris Wilsonb8f90962016-08-05 10:14:07 +01001969 goto err_unpin;
Chris Wilson7d1c4802010-08-07 21:45:03 +01001970
Chris Wilsonb90b91d2014-06-10 12:14:40 +01001971 /* Finally, remap it using the new GTT offset */
Chris Wilsonc58305a2016-08-19 16:54:28 +01001972 ret = remap_io_mapping(area,
Chris Wilson8bab11932017-01-14 00:28:25 +00001973 area->vm_start + (vma->ggtt_view.partial.offset << PAGE_SHIFT),
Matthew Auld73ebd502017-12-11 15:18:20 +00001974 (ggtt->gmadr.start + vma->node.start) >> PAGE_SHIFT,
Chris Wilsonc58305a2016-08-19 16:54:28 +01001975 min_t(u64, vma->size, area->vm_end - area->vm_start),
Matthew Auld73ebd502017-12-11 15:18:20 +00001976 &ggtt->iomap);
Chris Wilsona65adaf2017-10-09 09:43:57 +01001977 if (ret)
1978 goto err_fence;
Chris Wilsona61007a82016-08-18 17:17:02 +01001979
Chris Wilsona65adaf2017-10-09 09:43:57 +01001980 /* Mark as being mmapped into userspace for later revocation */
1981 assert_rpm_wakelock_held(dev_priv);
1982 if (!i915_vma_set_userfault(vma) && !obj->userfault_count++)
1983 list_add(&obj->userfault_link, &dev_priv->mm.userfault_list);
1984 GEM_BUG_ON(!obj->userfault_count);
1985
Chris Wilson7125397b2017-12-06 12:49:14 +00001986 i915_vma_set_ggtt_write(vma);
1987
Chris Wilsona65adaf2017-10-09 09:43:57 +01001988err_fence:
Chris Wilson3bd40732017-10-09 09:43:56 +01001989 i915_vma_unpin_fence(vma);
Chris Wilsonb8f90962016-08-05 10:14:07 +01001990err_unpin:
Chris Wilson058d88c2016-08-15 10:49:06 +01001991 __i915_vma_unpin(vma);
Chris Wilsonb8f90962016-08-05 10:14:07 +01001992err_unlock:
Jesse Barnesde151cf2008-11-12 10:03:55 -08001993 mutex_unlock(&dev->struct_mutex);
Chris Wilsonb8f90962016-08-05 10:14:07 +01001994err_rpm:
Chris Wilson538ef962019-01-14 14:21:18 +00001995 intel_runtime_pm_put(dev_priv, wakeref);
Chris Wilson40e62d52016-10-28 13:58:41 +01001996 i915_gem_object_unpin_pages(obj);
Chris Wilsonb8f90962016-08-05 10:14:07 +01001997err:
Jesse Barnesde151cf2008-11-12 10:03:55 -08001998 switch (ret) {
Chris Wilsond9bc7e92011-02-07 13:09:31 +00001999 case -EIO:
Daniel Vetter2232f032014-09-04 09:36:18 +02002000 /*
2001 * We eat errors when the gpu is terminally wedged to avoid
2002 * userspace unduly crashing (gl has no provisions for mmaps to
2003 * fail). But any other -EIO isn't ours (e.g. swap in failure)
2004 * and so needs to be reported.
2005 */
Chris Wilson52137012018-06-06 22:45:20 +01002006 if (!i915_terminally_wedged(&dev_priv->gpu_error))
2007 return VM_FAULT_SIGBUS;
Gustavo A. R. Silvaf0d759f2018-06-28 17:35:41 -05002008 /* else: fall through */
Chris Wilson045e7692010-11-07 09:18:22 +00002009 case -EAGAIN:
Daniel Vetter571c6082013-09-12 17:57:28 +02002010 /*
2011 * EAGAIN means the gpu is hung and we'll wait for the error
2012 * handler to reset everything when re-faulting in
2013 * i915_mutex_lock_interruptible.
Chris Wilsond9bc7e92011-02-07 13:09:31 +00002014 */
Chris Wilsonc7150892009-09-23 00:43:56 +01002015 case 0:
2016 case -ERESTARTSYS:
Chris Wilsonbed636a2011-02-11 20:31:19 +00002017 case -EINTR:
Dmitry Rogozhkine79e0fe2012-10-03 17:15:26 +03002018 case -EBUSY:
2019 /*
2020 * EBUSY is ok: this just means that another thread
2021 * already did the job.
2022 */
Chris Wilson52137012018-06-06 22:45:20 +01002023 return VM_FAULT_NOPAGE;
Jesse Barnesde151cf2008-11-12 10:03:55 -08002024 case -ENOMEM:
Chris Wilson52137012018-06-06 22:45:20 +01002025 return VM_FAULT_OOM;
Daniel Vettera7c2e1a2012-10-17 11:17:16 +02002026 case -ENOSPC:
Chris Wilson45d67812014-01-31 11:34:57 +00002027 case -EFAULT:
Chris Wilson52137012018-06-06 22:45:20 +01002028 return VM_FAULT_SIGBUS;
Jesse Barnesde151cf2008-11-12 10:03:55 -08002029 default:
Daniel Vettera7c2e1a2012-10-17 11:17:16 +02002030 WARN_ONCE(ret, "unhandled error in i915_gem_fault: %i\n", ret);
Chris Wilson52137012018-06-06 22:45:20 +01002031 return VM_FAULT_SIGBUS;
Jesse Barnesde151cf2008-11-12 10:03:55 -08002032 }
2033}
2034
Chris Wilsona65adaf2017-10-09 09:43:57 +01002035static void __i915_gem_object_release_mmap(struct drm_i915_gem_object *obj)
2036{
2037 struct i915_vma *vma;
2038
2039 GEM_BUG_ON(!obj->userfault_count);
2040
2041 obj->userfault_count = 0;
2042 list_del(&obj->userfault_link);
2043 drm_vma_node_unmap(&obj->base.vma_node,
2044 obj->base.dev->anon_inode->i_mapping);
2045
Chris Wilsone2189dd2017-12-07 21:14:07 +00002046 for_each_ggtt_vma(vma, obj)
Chris Wilsona65adaf2017-10-09 09:43:57 +01002047 i915_vma_unset_userfault(vma);
Chris Wilsona65adaf2017-10-09 09:43:57 +01002048}
2049
Jesse Barnesde151cf2008-11-12 10:03:55 -08002050/**
Chris Wilson901782b2009-07-10 08:18:50 +01002051 * i915_gem_release_mmap - remove physical page mappings
2052 * @obj: obj in question
2053 *
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02002054 * Preserve the reservation of the mmapping with the DRM core code, but
Chris Wilson901782b2009-07-10 08:18:50 +01002055 * relinquish ownership of the pages back to the system.
2056 *
2057 * It is vital that we remove the page mapping if we have mapped a tiled
2058 * object through the GTT and then lose the fence register due to
2059 * resource pressure. Similarly if the object has been moved out of the
2060 * aperture, than pages mapped into userspace must be revoked. Removing the
2061 * mapping will then trigger a page fault on the next user access, allowing
2062 * fixup by i915_gem_fault().
2063 */
Eric Anholtd05ca302009-07-10 13:02:26 -07002064void
Chris Wilson05394f32010-11-08 19:18:58 +00002065i915_gem_release_mmap(struct drm_i915_gem_object *obj)
Chris Wilson901782b2009-07-10 08:18:50 +01002066{
Chris Wilson275f0392016-10-24 13:42:14 +01002067 struct drm_i915_private *i915 = to_i915(obj->base.dev);
Chris Wilson538ef962019-01-14 14:21:18 +00002068 intel_wakeref_t wakeref;
Chris Wilson275f0392016-10-24 13:42:14 +01002069
Chris Wilson349f2cc2016-04-13 17:35:12 +01002070 /* Serialisation between user GTT access and our code depends upon
2071 * revoking the CPU's PTE whilst the mutex is held. The next user
2072 * pagefault then has to wait until we release the mutex.
Chris Wilson9c870d02016-10-24 13:42:15 +01002073 *
2074 * Note that RPM complicates somewhat by adding an additional
2075 * requirement that operations to the GGTT be made holding the RPM
2076 * wakeref.
Chris Wilson349f2cc2016-04-13 17:35:12 +01002077 */
Chris Wilson275f0392016-10-24 13:42:14 +01002078 lockdep_assert_held(&i915->drm.struct_mutex);
Chris Wilson538ef962019-01-14 14:21:18 +00002079 wakeref = intel_runtime_pm_get(i915);
Chris Wilson349f2cc2016-04-13 17:35:12 +01002080
Chris Wilsona65adaf2017-10-09 09:43:57 +01002081 if (!obj->userfault_count)
Chris Wilson9c870d02016-10-24 13:42:15 +01002082 goto out;
Chris Wilson901782b2009-07-10 08:18:50 +01002083
Chris Wilsona65adaf2017-10-09 09:43:57 +01002084 __i915_gem_object_release_mmap(obj);
Chris Wilson349f2cc2016-04-13 17:35:12 +01002085
2086 /* Ensure that the CPU's PTE are revoked and there are not outstanding
2087 * memory transactions from userspace before we return. The TLB
2088 * flushing implied above by changing the PTE above *should* be
2089 * sufficient, an extra barrier here just provides us with a bit
2090 * of paranoid documentation about our requirement to serialise
2091 * memory writes before touching registers / GSM.
2092 */
2093 wmb();
Chris Wilson9c870d02016-10-24 13:42:15 +01002094
2095out:
Chris Wilson538ef962019-01-14 14:21:18 +00002096 intel_runtime_pm_put(i915, wakeref);
Chris Wilson901782b2009-07-10 08:18:50 +01002097}
2098
Chris Wilson7c108fd2016-10-24 13:42:18 +01002099void i915_gem_runtime_suspend(struct drm_i915_private *dev_priv)
Chris Wilsoneedd10f2014-06-16 08:57:44 +01002100{
Chris Wilson3594a3e2016-10-24 13:42:16 +01002101 struct drm_i915_gem_object *obj, *on;
Chris Wilson7c108fd2016-10-24 13:42:18 +01002102 int i;
Chris Wilsoneedd10f2014-06-16 08:57:44 +01002103
Chris Wilson3594a3e2016-10-24 13:42:16 +01002104 /*
2105 * Only called during RPM suspend. All users of the userfault_list
2106 * must be holding an RPM wakeref to ensure that this can not
2107 * run concurrently with themselves (and use the struct_mutex for
2108 * protection between themselves).
2109 */
2110
2111 list_for_each_entry_safe(obj, on,
Chris Wilsona65adaf2017-10-09 09:43:57 +01002112 &dev_priv->mm.userfault_list, userfault_link)
2113 __i915_gem_object_release_mmap(obj);
Chris Wilson7c108fd2016-10-24 13:42:18 +01002114
2115 /* The fence will be lost when the device powers down. If any were
2116 * in use by hardware (i.e. they are pinned), we should not be powering
2117 * down! All other fences will be reacquired by the user upon waking.
2118 */
2119 for (i = 0; i < dev_priv->num_fence_regs; i++) {
2120 struct drm_i915_fence_reg *reg = &dev_priv->fence_regs[i];
2121
Chris Wilsone0ec3ec2017-02-03 12:57:17 +00002122 /* Ideally we want to assert that the fence register is not
2123 * live at this point (i.e. that no piece of code will be
2124 * trying to write through fence + GTT, as that both violates
2125 * our tracking of activity and associated locking/barriers,
2126 * but also is illegal given that the hw is powered down).
2127 *
2128 * Previously we used reg->pin_count as a "liveness" indicator.
2129 * That is not sufficient, and we need a more fine-grained
2130 * tool if we want to have a sanity check here.
2131 */
Chris Wilson7c108fd2016-10-24 13:42:18 +01002132
2133 if (!reg->vma)
2134 continue;
2135
Chris Wilsona65adaf2017-10-09 09:43:57 +01002136 GEM_BUG_ON(i915_vma_has_userfault(reg->vma));
Chris Wilson7c108fd2016-10-24 13:42:18 +01002137 reg->dirty = true;
2138 }
Chris Wilsoneedd10f2014-06-16 08:57:44 +01002139}
2140
Chris Wilsond8cb5082012-08-11 15:41:03 +01002141static int i915_gem_object_create_mmap_offset(struct drm_i915_gem_object *obj)
2142{
Chris Wilsonfac5e232016-07-04 11:34:36 +01002143 struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
Chris Wilsonf3f61842016-08-05 10:14:14 +01002144 int err;
Chris Wilsond8cb5082012-08-11 15:41:03 +01002145
Chris Wilsonf3f61842016-08-05 10:14:14 +01002146 err = drm_gem_create_mmap_offset(&obj->base);
Chris Wilsonb42a13d2017-01-06 15:22:40 +00002147 if (likely(!err))
Chris Wilsonf3f61842016-08-05 10:14:14 +01002148 return 0;
Daniel Vetterda494d72012-12-20 15:11:16 +01002149
Chris Wilsonb42a13d2017-01-06 15:22:40 +00002150 /* Attempt to reap some mmap space from dead objects */
2151 do {
Chris Wilsonec625fb2018-07-09 13:20:42 +01002152 err = i915_gem_wait_for_idle(dev_priv,
2153 I915_WAIT_INTERRUPTIBLE,
2154 MAX_SCHEDULE_TIMEOUT);
Chris Wilsonb42a13d2017-01-06 15:22:40 +00002155 if (err)
2156 break;
Chris Wilsond8cb5082012-08-11 15:41:03 +01002157
Chris Wilsonb42a13d2017-01-06 15:22:40 +00002158 i915_gem_drain_freed_objects(dev_priv);
Chris Wilsonf3f61842016-08-05 10:14:14 +01002159 err = drm_gem_create_mmap_offset(&obj->base);
Chris Wilsonb42a13d2017-01-06 15:22:40 +00002160 if (!err)
2161 break;
2162
2163 } while (flush_delayed_work(&dev_priv->gt.retire_work));
Daniel Vetterda494d72012-12-20 15:11:16 +01002164
Chris Wilsonf3f61842016-08-05 10:14:14 +01002165 return err;
Chris Wilsond8cb5082012-08-11 15:41:03 +01002166}
2167
2168static void i915_gem_object_free_mmap_offset(struct drm_i915_gem_object *obj)
2169{
Chris Wilsond8cb5082012-08-11 15:41:03 +01002170 drm_gem_free_mmap_offset(&obj->base);
2171}
2172
Dave Airlieda6b51d2014-12-24 13:11:17 +10002173int
Dave Airlieff72145b2011-02-07 12:16:14 +10002174i915_gem_mmap_gtt(struct drm_file *file,
2175 struct drm_device *dev,
Jani Nikula739f3ab2019-01-16 11:15:19 +02002176 u32 handle,
2177 u64 *offset)
Jesse Barnesde151cf2008-11-12 10:03:55 -08002178{
Chris Wilson05394f32010-11-08 19:18:58 +00002179 struct drm_i915_gem_object *obj;
Jesse Barnesde151cf2008-11-12 10:03:55 -08002180 int ret;
2181
Chris Wilson03ac0642016-07-20 13:31:51 +01002182 obj = i915_gem_object_lookup(file, handle);
Chris Wilsonf3f61842016-08-05 10:14:14 +01002183 if (!obj)
2184 return -ENOENT;
Chris Wilsonab182822009-09-22 18:46:17 +01002185
Chris Wilsond8cb5082012-08-11 15:41:03 +01002186 ret = i915_gem_object_create_mmap_offset(obj);
Chris Wilsonf3f61842016-08-05 10:14:14 +01002187 if (ret == 0)
2188 *offset = drm_vma_node_offset_addr(&obj->base.vma_node);
Jesse Barnesde151cf2008-11-12 10:03:55 -08002189
Chris Wilsonf0cd5182016-10-28 13:58:43 +01002190 i915_gem_object_put(obj);
Chris Wilson1d7cfea2010-10-17 09:45:41 +01002191 return ret;
Jesse Barnesde151cf2008-11-12 10:03:55 -08002192}
2193
Dave Airlieff72145b2011-02-07 12:16:14 +10002194/**
2195 * i915_gem_mmap_gtt_ioctl - prepare an object for GTT mmap'ing
2196 * @dev: DRM device
2197 * @data: GTT mapping ioctl data
2198 * @file: GEM object info
2199 *
2200 * Simply returns the fake offset to userspace so it can mmap it.
2201 * The mmap call will end up in drm_gem_mmap(), which will set things
2202 * up so we can get faults in the handler above.
2203 *
2204 * The fault handler will take care of binding the object into the GTT
2205 * (since it may have been evicted to make room for something), allocating
2206 * a fence register, and mapping the appropriate aperture address into
2207 * userspace.
2208 */
2209int
2210i915_gem_mmap_gtt_ioctl(struct drm_device *dev, void *data,
2211 struct drm_file *file)
2212{
2213 struct drm_i915_gem_mmap_gtt *args = data;
2214
Dave Airlieda6b51d2014-12-24 13:11:17 +10002215 return i915_gem_mmap_gtt(file, dev, args->handle, &args->offset);
Dave Airlieff72145b2011-02-07 12:16:14 +10002216}
2217
Daniel Vetter225067e2012-08-20 10:23:20 +02002218/* Immediately discard the backing storage */
2219static void
2220i915_gem_object_truncate(struct drm_i915_gem_object *obj)
Chris Wilsone5281cc2010-10-28 13:45:36 +01002221{
Chris Wilson4d6294bf2012-08-11 15:41:05 +01002222 i915_gem_object_free_mmap_offset(obj);
Daniel Vetter1286ff72012-05-10 15:25:09 +02002223
Chris Wilson4d6294bf2012-08-11 15:41:05 +01002224 if (obj->base.filp == NULL)
2225 return;
2226
Daniel Vetter225067e2012-08-20 10:23:20 +02002227 /* Our goal here is to return as much of the memory as
2228 * is possible back to the system as we are called from OOM.
2229 * To do this we must instruct the shmfs to drop all of its
2230 * backing pages, *now*.
Chris Wilsone5281cc2010-10-28 13:45:36 +01002231 */
Chris Wilson55372522014-03-25 13:23:06 +00002232 shmem_truncate_range(file_inode(obj->base.filp), 0, (loff_t)-1);
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002233 obj->mm.madv = __I915_MADV_PURGED;
Chris Wilson4e5462e2017-03-07 13:20:31 +00002234 obj->mm.pages = ERR_PTR(-EFAULT);
Chris Wilsone5281cc2010-10-28 13:45:36 +01002235}
Chris Wilsone5281cc2010-10-28 13:45:36 +01002236
Chris Wilson55372522014-03-25 13:23:06 +00002237/* Try to discard unwanted pages */
Chris Wilson03ac84f2016-10-28 13:58:36 +01002238void __i915_gem_object_invalidate(struct drm_i915_gem_object *obj)
Daniel Vetter225067e2012-08-20 10:23:20 +02002239{
Chris Wilson55372522014-03-25 13:23:06 +00002240 struct address_space *mapping;
2241
Chris Wilson1233e2d2016-10-28 13:58:37 +01002242 lockdep_assert_held(&obj->mm.lock);
Chris Wilsonf1fa4f42017-10-13 21:26:13 +01002243 GEM_BUG_ON(i915_gem_object_has_pages(obj));
Chris Wilson1233e2d2016-10-28 13:58:37 +01002244
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002245 switch (obj->mm.madv) {
Chris Wilson55372522014-03-25 13:23:06 +00002246 case I915_MADV_DONTNEED:
2247 i915_gem_object_truncate(obj);
2248 case __I915_MADV_PURGED:
2249 return;
2250 }
2251
2252 if (obj->base.filp == NULL)
2253 return;
2254
Al Viro93c76a32015-12-04 23:45:44 -05002255 mapping = obj->base.filp->f_mapping,
Chris Wilson55372522014-03-25 13:23:06 +00002256 invalidate_mapping_pages(mapping, 0, (loff_t)-1);
Chris Wilsone5281cc2010-10-28 13:45:36 +01002257}
2258
Kuo-Hsin Yang64e3d122018-11-06 13:23:24 +00002259/*
2260 * Move pages to appropriate lru and release the pagevec, decrementing the
2261 * ref count of those pages.
2262 */
2263static void check_release_pagevec(struct pagevec *pvec)
2264{
2265 check_move_unevictable_pages(pvec);
2266 __pagevec_release(pvec);
2267 cond_resched();
2268}
2269
Chris Wilson5cdf5882010-09-27 15:51:07 +01002270static void
Chris Wilson03ac84f2016-10-28 13:58:36 +01002271i915_gem_object_put_pages_gtt(struct drm_i915_gem_object *obj,
2272 struct sg_table *pages)
Eric Anholt673a3942008-07-30 12:06:12 -07002273{
Dave Gordon85d12252016-05-20 11:54:06 +01002274 struct sgt_iter sgt_iter;
Kuo-Hsin Yang64e3d122018-11-06 13:23:24 +00002275 struct pagevec pvec;
Dave Gordon85d12252016-05-20 11:54:06 +01002276 struct page *page;
Daniel Vetter1286ff72012-05-10 15:25:09 +02002277
Chris Wilsone5facdf2016-12-23 14:57:57 +00002278 __i915_gem_object_release_shmem(obj, pages, true);
Eric Anholt856fa192009-03-19 14:10:50 -07002279
Chris Wilson03ac84f2016-10-28 13:58:36 +01002280 i915_gem_gtt_finish_pages(obj, pages);
Imre Deake2273302015-07-09 12:59:05 +03002281
Daniel Vetter6dacfd22011-09-12 21:30:02 +02002282 if (i915_gem_object_needs_bit17_swizzle(obj))
Chris Wilson03ac84f2016-10-28 13:58:36 +01002283 i915_gem_object_save_bit_17_swizzle(obj, pages);
Eric Anholt280b7132009-03-12 16:56:27 -07002284
Kuo-Hsin Yang64e3d122018-11-06 13:23:24 +00002285 mapping_clear_unevictable(file_inode(obj->base.filp)->i_mapping);
2286
2287 pagevec_init(&pvec);
Chris Wilson03ac84f2016-10-28 13:58:36 +01002288 for_each_sgt_page(page, sgt_iter, pages) {
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002289 if (obj->mm.dirty)
Chris Wilson9da3da62012-06-01 15:20:22 +01002290 set_page_dirty(page);
Chris Wilson3ef94da2009-09-14 16:50:29 +01002291
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002292 if (obj->mm.madv == I915_MADV_WILLNEED)
Chris Wilson9da3da62012-06-01 15:20:22 +01002293 mark_page_accessed(page);
Chris Wilson3ef94da2009-09-14 16:50:29 +01002294
Kuo-Hsin Yang64e3d122018-11-06 13:23:24 +00002295 if (!pagevec_add(&pvec, page))
2296 check_release_pagevec(&pvec);
Chris Wilson3ef94da2009-09-14 16:50:29 +01002297 }
Kuo-Hsin Yang64e3d122018-11-06 13:23:24 +00002298 if (pagevec_count(&pvec))
2299 check_release_pagevec(&pvec);
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002300 obj->mm.dirty = false;
Eric Anholt673a3942008-07-30 12:06:12 -07002301
Chris Wilson03ac84f2016-10-28 13:58:36 +01002302 sg_free_table(pages);
2303 kfree(pages);
Chris Wilson37e680a2012-06-07 15:38:42 +01002304}
2305
Chris Wilson96d77632016-10-28 13:58:33 +01002306static void __i915_gem_object_reset_page_iter(struct drm_i915_gem_object *obj)
2307{
2308 struct radix_tree_iter iter;
Ville Syrjäläc23aa712017-09-01 20:12:51 +03002309 void __rcu **slot;
Chris Wilson96d77632016-10-28 13:58:33 +01002310
Chris Wilsonbea6e982017-10-26 14:00:31 +01002311 rcu_read_lock();
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002312 radix_tree_for_each_slot(slot, &obj->mm.get_page.radix, &iter, 0)
2313 radix_tree_delete(&obj->mm.get_page.radix, iter.index);
Chris Wilsonbea6e982017-10-26 14:00:31 +01002314 rcu_read_unlock();
Chris Wilson96d77632016-10-28 13:58:33 +01002315}
2316
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01002317static struct sg_table *
2318__i915_gem_object_unset_pages(struct drm_i915_gem_object *obj)
Chris Wilson37e680a2012-06-07 15:38:42 +01002319{
Chris Wilsonf2123812017-10-16 12:40:37 +01002320 struct drm_i915_private *i915 = to_i915(obj->base.dev);
Chris Wilson03ac84f2016-10-28 13:58:36 +01002321 struct sg_table *pages;
Chris Wilson37e680a2012-06-07 15:38:42 +01002322
Chris Wilson03ac84f2016-10-28 13:58:36 +01002323 pages = fetch_and_zero(&obj->mm.pages);
Chris Wilson484d9a82019-01-15 12:44:42 +00002324 if (IS_ERR_OR_NULL(pages))
2325 return pages;
Chris Wilsona2165e32012-12-03 11:49:00 +00002326
Chris Wilsonf2123812017-10-16 12:40:37 +01002327 spin_lock(&i915->mm.obj_lock);
2328 list_del(&obj->mm.link);
2329 spin_unlock(&i915->mm.obj_lock);
2330
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002331 if (obj->mm.mapping) {
Chris Wilson4b30cb22016-08-18 17:16:42 +01002332 void *ptr;
2333
Chris Wilson0ce81782017-05-17 13:09:59 +01002334 ptr = page_mask_bits(obj->mm.mapping);
Chris Wilson4b30cb22016-08-18 17:16:42 +01002335 if (is_vmalloc_addr(ptr))
2336 vunmap(ptr);
Chris Wilsonfb8621d2016-04-08 12:11:14 +01002337 else
Chris Wilson4b30cb22016-08-18 17:16:42 +01002338 kunmap(kmap_to_page(ptr));
2339
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002340 obj->mm.mapping = NULL;
Chris Wilson0a798eb2016-04-08 12:11:11 +01002341 }
2342
Chris Wilson96d77632016-10-28 13:58:33 +01002343 __i915_gem_object_reset_page_iter(obj);
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01002344 obj->mm.page_sizes.phys = obj->mm.page_sizes.sg = 0;
Chris Wilson96d77632016-10-28 13:58:33 +01002345
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01002346 return pages;
2347}
2348
Chris Wilson484d9a82019-01-15 12:44:42 +00002349int __i915_gem_object_put_pages(struct drm_i915_gem_object *obj,
2350 enum i915_mm_subclass subclass)
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01002351{
2352 struct sg_table *pages;
Chris Wilson484d9a82019-01-15 12:44:42 +00002353 int ret;
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01002354
2355 if (i915_gem_object_has_pinned_pages(obj))
Chris Wilson484d9a82019-01-15 12:44:42 +00002356 return -EBUSY;
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01002357
2358 GEM_BUG_ON(obj->bind_count);
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01002359
2360 /* May be called by shrinker from within get_pages() (on another bo) */
2361 mutex_lock_nested(&obj->mm.lock, subclass);
Chris Wilson484d9a82019-01-15 12:44:42 +00002362 if (unlikely(atomic_read(&obj->mm.pages_pin_count))) {
2363 ret = -EBUSY;
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01002364 goto unlock;
Chris Wilson484d9a82019-01-15 12:44:42 +00002365 }
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01002366
2367 /*
2368 * ->put_pages might need to allocate memory for the bit17 swizzle
2369 * array, hence protect them from being reaped by removing them from gtt
2370 * lists early.
2371 */
2372 pages = __i915_gem_object_unset_pages(obj);
Chris Wilson484d9a82019-01-15 12:44:42 +00002373
2374 /*
2375 * XXX Temporary hijinx to avoid updating all backends to handle
2376 * NULL pages. In the future, when we have more asynchronous
2377 * get_pages backends we should be better able to handle the
2378 * cancellation of the async task in a more uniform manner.
2379 */
2380 if (!pages && !i915_gem_object_needs_async_cancel(obj))
2381 pages = ERR_PTR(-EINVAL);
2382
Chris Wilson4e5462e2017-03-07 13:20:31 +00002383 if (!IS_ERR(pages))
2384 obj->ops->put_pages(obj, pages);
2385
Chris Wilson484d9a82019-01-15 12:44:42 +00002386 ret = 0;
Chris Wilson1233e2d2016-10-28 13:58:37 +01002387unlock:
2388 mutex_unlock(&obj->mm.lock);
Chris Wilson484d9a82019-01-15 12:44:42 +00002389
2390 return ret;
Chris Wilson6c085a72012-08-20 11:40:46 +02002391}
2392
Tvrtko Ursulinf8e57862018-09-26 09:03:53 +01002393bool i915_sg_trim(struct sg_table *orig_st)
Tvrtko Ursulin0c40ce12016-11-09 15:13:43 +00002394{
2395 struct sg_table new_st;
2396 struct scatterlist *sg, *new_sg;
2397 unsigned int i;
2398
2399 if (orig_st->nents == orig_st->orig_nents)
Chris Wilson935a2f72017-02-13 17:15:13 +00002400 return false;
Tvrtko Ursulin0c40ce12016-11-09 15:13:43 +00002401
Chris Wilson8bfc478f2016-12-23 14:57:58 +00002402 if (sg_alloc_table(&new_st, orig_st->nents, GFP_KERNEL | __GFP_NOWARN))
Chris Wilson935a2f72017-02-13 17:15:13 +00002403 return false;
Tvrtko Ursulin0c40ce12016-11-09 15:13:43 +00002404
2405 new_sg = new_st.sgl;
2406 for_each_sg(orig_st->sgl, sg, orig_st->nents, i) {
2407 sg_set_page(new_sg, sg_page(sg), sg->length, 0);
Matthew Auldc6d22ab2018-09-20 15:27:06 +01002408 sg_dma_address(new_sg) = sg_dma_address(sg);
2409 sg_dma_len(new_sg) = sg_dma_len(sg);
2410
Tvrtko Ursulin0c40ce12016-11-09 15:13:43 +00002411 new_sg = sg_next(new_sg);
2412 }
Chris Wilsonc2dc6cc2016-12-19 12:43:46 +00002413 GEM_BUG_ON(new_sg); /* Should walk exactly nents and hit the end */
Tvrtko Ursulin0c40ce12016-11-09 15:13:43 +00002414
2415 sg_free_table(orig_st);
2416
2417 *orig_st = new_st;
Chris Wilson935a2f72017-02-13 17:15:13 +00002418 return true;
Tvrtko Ursulin0c40ce12016-11-09 15:13:43 +00002419}
2420
Matthew Auldb91b09e2017-10-06 23:18:17 +01002421static int i915_gem_object_get_pages_gtt(struct drm_i915_gem_object *obj)
Eric Anholt673a3942008-07-30 12:06:12 -07002422{
Chris Wilsonfac5e232016-07-04 11:34:36 +01002423 struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
Chris Wilsond766ef52016-12-19 12:43:45 +00002424 const unsigned long page_count = obj->base.size / PAGE_SIZE;
2425 unsigned long i;
Eric Anholt673a3942008-07-30 12:06:12 -07002426 struct address_space *mapping;
Chris Wilson9da3da62012-06-01 15:20:22 +01002427 struct sg_table *st;
2428 struct scatterlist *sg;
Dave Gordon85d12252016-05-20 11:54:06 +01002429 struct sgt_iter sgt_iter;
Eric Anholt673a3942008-07-30 12:06:12 -07002430 struct page *page;
Imre Deak90797e62013-02-18 19:28:03 +02002431 unsigned long last_pfn = 0; /* suppress gcc warning */
Tvrtko Ursulin56024522017-08-03 10:14:17 +01002432 unsigned int max_segment = i915_sg_segment_size();
Matthew Auld84e89782017-10-09 12:00:24 +01002433 unsigned int sg_page_sizes;
Kuo-Hsin Yang64e3d122018-11-06 13:23:24 +00002434 struct pagevec pvec;
Chris Wilson4846bf02017-06-09 12:03:46 +01002435 gfp_t noreclaim;
Imre Deake2273302015-07-09 12:59:05 +03002436 int ret;
Eric Anholt673a3942008-07-30 12:06:12 -07002437
Chris Wilsone0ff7a72018-09-03 09:33:36 +01002438 /*
2439 * Assert that the object is not currently in any GPU domain. As it
Chris Wilson6c085a72012-08-20 11:40:46 +02002440 * wasn't in the GTT, there shouldn't be any way it could have been in
2441 * a GPU cache
2442 */
Christian Königc0a51fd2018-02-16 13:43:38 +01002443 GEM_BUG_ON(obj->read_domains & I915_GEM_GPU_DOMAINS);
2444 GEM_BUG_ON(obj->write_domain & I915_GEM_GPU_DOMAINS);
Chris Wilson6c085a72012-08-20 11:40:46 +02002445
Chris Wilsone0ff7a72018-09-03 09:33:36 +01002446 /*
2447 * If there's no chance of allocating enough pages for the whole
2448 * object, bail early.
2449 */
Arun KSca79b0c2018-12-28 00:34:29 -08002450 if (page_count > totalram_pages())
Chris Wilsone0ff7a72018-09-03 09:33:36 +01002451 return -ENOMEM;
2452
Chris Wilson9da3da62012-06-01 15:20:22 +01002453 st = kmalloc(sizeof(*st), GFP_KERNEL);
2454 if (st == NULL)
Matthew Auldb91b09e2017-10-06 23:18:17 +01002455 return -ENOMEM;
Eric Anholt673a3942008-07-30 12:06:12 -07002456
Chris Wilsond766ef52016-12-19 12:43:45 +00002457rebuild_st:
Chris Wilson9da3da62012-06-01 15:20:22 +01002458 if (sg_alloc_table(st, page_count, GFP_KERNEL)) {
Chris Wilson9da3da62012-06-01 15:20:22 +01002459 kfree(st);
Matthew Auldb91b09e2017-10-06 23:18:17 +01002460 return -ENOMEM;
Chris Wilson9da3da62012-06-01 15:20:22 +01002461 }
2462
Chris Wilsone0ff7a72018-09-03 09:33:36 +01002463 /*
2464 * Get the list of pages out of our struct file. They'll be pinned
Chris Wilson9da3da62012-06-01 15:20:22 +01002465 * at this point until we release them.
2466 *
2467 * Fail silently without starting the shrinker
2468 */
Al Viro93c76a32015-12-04 23:45:44 -05002469 mapping = obj->base.filp->f_mapping;
Kuo-Hsin Yang64e3d122018-11-06 13:23:24 +00002470 mapping_set_unevictable(mapping);
Chris Wilson0f6ab552017-06-09 12:03:48 +01002471 noreclaim = mapping_gfp_constraint(mapping, ~__GFP_RECLAIM);
Chris Wilson4846bf02017-06-09 12:03:46 +01002472 noreclaim |= __GFP_NORETRY | __GFP_NOWARN;
2473
Imre Deak90797e62013-02-18 19:28:03 +02002474 sg = st->sgl;
2475 st->nents = 0;
Matthew Auld84e89782017-10-09 12:00:24 +01002476 sg_page_sizes = 0;
Imre Deak90797e62013-02-18 19:28:03 +02002477 for (i = 0; i < page_count; i++) {
Chris Wilson4846bf02017-06-09 12:03:46 +01002478 const unsigned int shrink[] = {
2479 I915_SHRINK_BOUND | I915_SHRINK_UNBOUND | I915_SHRINK_PURGEABLE,
2480 0,
2481 }, *s = shrink;
2482 gfp_t gfp = noreclaim;
2483
2484 do {
Chris Wilsone6db7f42018-11-05 17:06:40 +00002485 cond_resched();
Chris Wilson6c085a72012-08-20 11:40:46 +02002486 page = shmem_read_mapping_page_gfp(mapping, i, gfp);
Chris Wilson4846bf02017-06-09 12:03:46 +01002487 if (likely(!IS_ERR(page)))
2488 break;
2489
2490 if (!*s) {
2491 ret = PTR_ERR(page);
2492 goto err_sg;
2493 }
2494
Chris Wilson912d5722017-09-06 16:19:30 -07002495 i915_gem_shrink(dev_priv, 2 * page_count, NULL, *s++);
Chris Wilson24f8e002017-03-22 11:05:21 +00002496
Chris Wilsone0ff7a72018-09-03 09:33:36 +01002497 /*
2498 * We've tried hard to allocate the memory by reaping
Chris Wilson6c085a72012-08-20 11:40:46 +02002499 * our own buffer, now let the real VM do its job and
2500 * go down in flames if truly OOM.
Chris Wilson24f8e002017-03-22 11:05:21 +00002501 *
2502 * However, since graphics tend to be disposable,
2503 * defer the oom here by reporting the ENOMEM back
2504 * to userspace.
Chris Wilson6c085a72012-08-20 11:40:46 +02002505 */
Chris Wilson4846bf02017-06-09 12:03:46 +01002506 if (!*s) {
2507 /* reclaim and warn, but no oom */
2508 gfp = mapping_gfp_mask(mapping);
Chris Wilsoneaf41802017-06-09 12:03:47 +01002509
Chris Wilsone0ff7a72018-09-03 09:33:36 +01002510 /*
2511 * Our bo are always dirty and so we require
Chris Wilsoneaf41802017-06-09 12:03:47 +01002512 * kswapd to reclaim our pages (direct reclaim
2513 * does not effectively begin pageout of our
2514 * buffers on its own). However, direct reclaim
2515 * only waits for kswapd when under allocation
2516 * congestion. So as a result __GFP_RECLAIM is
2517 * unreliable and fails to actually reclaim our
2518 * dirty pages -- unless you try over and over
2519 * again with !__GFP_NORETRY. However, we still
2520 * want to fail this allocation rather than
2521 * trigger the out-of-memory killer and for
Michal Hockodbb32952017-07-12 14:36:55 -07002522 * this we want __GFP_RETRY_MAYFAIL.
Chris Wilsoneaf41802017-06-09 12:03:47 +01002523 */
Michal Hockodbb32952017-07-12 14:36:55 -07002524 gfp |= __GFP_RETRY_MAYFAIL;
Imre Deake2273302015-07-09 12:59:05 +03002525 }
Chris Wilson4846bf02017-06-09 12:03:46 +01002526 } while (1);
2527
Chris Wilson871dfbd2016-10-11 09:20:21 +01002528 if (!i ||
2529 sg->length >= max_segment ||
2530 page_to_pfn(page) != last_pfn + 1) {
Matthew Aulda5c081662017-10-06 23:18:18 +01002531 if (i) {
Matthew Auld84e89782017-10-09 12:00:24 +01002532 sg_page_sizes |= sg->length;
Imre Deak90797e62013-02-18 19:28:03 +02002533 sg = sg_next(sg);
Matthew Aulda5c081662017-10-06 23:18:18 +01002534 }
Imre Deak90797e62013-02-18 19:28:03 +02002535 st->nents++;
2536 sg_set_page(sg, page, PAGE_SIZE, 0);
2537 } else {
2538 sg->length += PAGE_SIZE;
2539 }
2540 last_pfn = page_to_pfn(page);
Daniel Vetter3bbbe702013-10-07 17:15:45 -03002541
2542 /* Check that the i965g/gm workaround works. */
2543 WARN_ON((gfp & __GFP_DMA32) && (last_pfn >= 0x00100000UL));
Eric Anholt673a3942008-07-30 12:06:12 -07002544 }
Matthew Aulda5c081662017-10-06 23:18:18 +01002545 if (sg) { /* loop terminated early; short sg table */
Matthew Auld84e89782017-10-09 12:00:24 +01002546 sg_page_sizes |= sg->length;
Konrad Rzeszutek Wilk426729d2013-06-24 11:47:48 -04002547 sg_mark_end(sg);
Matthew Aulda5c081662017-10-06 23:18:18 +01002548 }
Chris Wilson74ce6b62012-10-19 15:51:06 +01002549
Tvrtko Ursulin0c40ce12016-11-09 15:13:43 +00002550 /* Trim unused sg entries to avoid wasting memory. */
2551 i915_sg_trim(st);
2552
Chris Wilson03ac84f2016-10-28 13:58:36 +01002553 ret = i915_gem_gtt_prepare_pages(obj, st);
Chris Wilsond766ef52016-12-19 12:43:45 +00002554 if (ret) {
Chris Wilsone0ff7a72018-09-03 09:33:36 +01002555 /*
2556 * DMA remapping failed? One possible cause is that
Chris Wilsond766ef52016-12-19 12:43:45 +00002557 * it could not reserve enough large entries, asking
2558 * for PAGE_SIZE chunks instead may be helpful.
2559 */
2560 if (max_segment > PAGE_SIZE) {
2561 for_each_sgt_page(page, sgt_iter, st)
2562 put_page(page);
2563 sg_free_table(st);
2564
2565 max_segment = PAGE_SIZE;
2566 goto rebuild_st;
2567 } else {
2568 dev_warn(&dev_priv->drm.pdev->dev,
2569 "Failed to DMA remap %lu pages\n",
2570 page_count);
2571 goto err_pages;
2572 }
2573 }
Imre Deake2273302015-07-09 12:59:05 +03002574
Eric Anholt673a3942008-07-30 12:06:12 -07002575 if (i915_gem_object_needs_bit17_swizzle(obj))
Chris Wilson03ac84f2016-10-28 13:58:36 +01002576 i915_gem_object_do_bit_17_swizzle(obj, st);
Eric Anholt673a3942008-07-30 12:06:12 -07002577
Matthew Auld84e89782017-10-09 12:00:24 +01002578 __i915_gem_object_set_pages(obj, st, sg_page_sizes);
Matthew Auldb91b09e2017-10-06 23:18:17 +01002579
2580 return 0;
Eric Anholt673a3942008-07-30 12:06:12 -07002581
Chris Wilsonb17993b2016-11-14 11:29:30 +00002582err_sg:
Imre Deak90797e62013-02-18 19:28:03 +02002583 sg_mark_end(sg);
Chris Wilsonb17993b2016-11-14 11:29:30 +00002584err_pages:
Kuo-Hsin Yang64e3d122018-11-06 13:23:24 +00002585 mapping_clear_unevictable(mapping);
2586 pagevec_init(&pvec);
2587 for_each_sgt_page(page, sgt_iter, st) {
2588 if (!pagevec_add(&pvec, page))
2589 check_release_pagevec(&pvec);
2590 }
2591 if (pagevec_count(&pvec))
2592 check_release_pagevec(&pvec);
Chris Wilson9da3da62012-06-01 15:20:22 +01002593 sg_free_table(st);
2594 kfree(st);
Chris Wilson0820baf2014-03-25 13:23:03 +00002595
Chris Wilsone0ff7a72018-09-03 09:33:36 +01002596 /*
2597 * shmemfs first checks if there is enough memory to allocate the page
Chris Wilson0820baf2014-03-25 13:23:03 +00002598 * and reports ENOSPC should there be insufficient, along with the usual
2599 * ENOMEM for a genuine allocation failure.
2600 *
2601 * We use ENOSPC in our driver to mean that we have run out of aperture
2602 * space and so want to translate the error from shmemfs back to our
2603 * usual understanding of ENOMEM.
2604 */
Imre Deake2273302015-07-09 12:59:05 +03002605 if (ret == -ENOSPC)
2606 ret = -ENOMEM;
2607
Matthew Auldb91b09e2017-10-06 23:18:17 +01002608 return ret;
Chris Wilson03ac84f2016-10-28 13:58:36 +01002609}
2610
2611void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj,
Matthew Aulda5c081662017-10-06 23:18:18 +01002612 struct sg_table *pages,
Matthew Auld84e89782017-10-09 12:00:24 +01002613 unsigned int sg_page_sizes)
Chris Wilson03ac84f2016-10-28 13:58:36 +01002614{
Matthew Aulda5c081662017-10-06 23:18:18 +01002615 struct drm_i915_private *i915 = to_i915(obj->base.dev);
2616 unsigned long supported = INTEL_INFO(i915)->page_sizes;
2617 int i;
2618
Chris Wilson1233e2d2016-10-28 13:58:37 +01002619 lockdep_assert_held(&obj->mm.lock);
Chris Wilson03ac84f2016-10-28 13:58:36 +01002620
2621 obj->mm.get_page.sg_pos = pages->sgl;
2622 obj->mm.get_page.sg_idx = 0;
2623
2624 obj->mm.pages = pages;
Chris Wilson2c3a3f42016-11-04 10:30:01 +00002625
2626 if (i915_gem_object_is_tiled(obj) &&
Chris Wilsonf2123812017-10-16 12:40:37 +01002627 i915->quirks & QUIRK_PIN_SWIZZLED_PAGES) {
Chris Wilson2c3a3f42016-11-04 10:30:01 +00002628 GEM_BUG_ON(obj->mm.quirked);
2629 __i915_gem_object_pin_pages(obj);
2630 obj->mm.quirked = true;
2631 }
Matthew Aulda5c081662017-10-06 23:18:18 +01002632
Matthew Auld84e89782017-10-09 12:00:24 +01002633 GEM_BUG_ON(!sg_page_sizes);
2634 obj->mm.page_sizes.phys = sg_page_sizes;
Matthew Aulda5c081662017-10-06 23:18:18 +01002635
2636 /*
Matthew Auld84e89782017-10-09 12:00:24 +01002637 * Calculate the supported page-sizes which fit into the given
2638 * sg_page_sizes. This will give us the page-sizes which we may be able
2639 * to use opportunistically when later inserting into the GTT. For
2640 * example if phys=2G, then in theory we should be able to use 1G, 2M,
2641 * 64K or 4K pages, although in practice this will depend on a number of
2642 * other factors.
Matthew Aulda5c081662017-10-06 23:18:18 +01002643 */
2644 obj->mm.page_sizes.sg = 0;
2645 for_each_set_bit(i, &supported, ilog2(I915_GTT_MAX_PAGE_SIZE) + 1) {
2646 if (obj->mm.page_sizes.phys & ~0u << i)
2647 obj->mm.page_sizes.sg |= BIT(i);
2648 }
Matthew Aulda5c081662017-10-06 23:18:18 +01002649 GEM_BUG_ON(!HAS_PAGE_SIZES(i915, obj->mm.page_sizes.sg));
Chris Wilsonf2123812017-10-16 12:40:37 +01002650
2651 spin_lock(&i915->mm.obj_lock);
2652 list_add(&obj->mm.link, &i915->mm.unbound_list);
2653 spin_unlock(&i915->mm.obj_lock);
Chris Wilson03ac84f2016-10-28 13:58:36 +01002654}
2655
2656static int ____i915_gem_object_get_pages(struct drm_i915_gem_object *obj)
2657{
Matthew Auldb91b09e2017-10-06 23:18:17 +01002658 int err;
Chris Wilson03ac84f2016-10-28 13:58:36 +01002659
2660 if (unlikely(obj->mm.madv != I915_MADV_WILLNEED)) {
2661 DRM_DEBUG("Attempting to obtain a purgeable object\n");
2662 return -EFAULT;
2663 }
2664
Matthew Auldb91b09e2017-10-06 23:18:17 +01002665 err = obj->ops->get_pages(obj);
Matthew Auldb65a9b92017-12-18 10:38:55 +00002666 GEM_BUG_ON(!err && !i915_gem_object_has_pages(obj));
Chris Wilson03ac84f2016-10-28 13:58:36 +01002667
Matthew Auldb91b09e2017-10-06 23:18:17 +01002668 return err;
Eric Anholt673a3942008-07-30 12:06:12 -07002669}
2670
Chris Wilson37e680a2012-06-07 15:38:42 +01002671/* Ensure that the associated pages are gathered from the backing storage
Chris Wilson1233e2d2016-10-28 13:58:37 +01002672 * and pinned into our object. i915_gem_object_pin_pages() may be called
Chris Wilson37e680a2012-06-07 15:38:42 +01002673 * multiple times before they are released by a single call to
Chris Wilson1233e2d2016-10-28 13:58:37 +01002674 * i915_gem_object_unpin_pages() - once the pages are no longer referenced
Chris Wilson37e680a2012-06-07 15:38:42 +01002675 * either as a result of memory pressure (reaping pages under the shrinker)
2676 * or as the object is itself released.
2677 */
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002678int __i915_gem_object_get_pages(struct drm_i915_gem_object *obj)
Chris Wilson37e680a2012-06-07 15:38:42 +01002679{
Chris Wilson03ac84f2016-10-28 13:58:36 +01002680 int err;
Chris Wilson37e680a2012-06-07 15:38:42 +01002681
Chris Wilson1233e2d2016-10-28 13:58:37 +01002682 err = mutex_lock_interruptible(&obj->mm.lock);
2683 if (err)
2684 return err;
Chris Wilson4c7d62c2016-10-28 13:58:32 +01002685
Chris Wilsonf1fa4f42017-10-13 21:26:13 +01002686 if (unlikely(!i915_gem_object_has_pages(obj))) {
Chris Wilson88c880b2017-09-06 14:52:20 +01002687 GEM_BUG_ON(i915_gem_object_has_pinned_pages(obj));
2688
Chris Wilson2c3a3f42016-11-04 10:30:01 +00002689 err = ____i915_gem_object_get_pages(obj);
2690 if (err)
2691 goto unlock;
2692
2693 smp_mb__before_atomic();
Chris Wilson1233e2d2016-10-28 13:58:37 +01002694 }
Chris Wilson2c3a3f42016-11-04 10:30:01 +00002695 atomic_inc(&obj->mm.pages_pin_count);
Chris Wilson43e28f02013-01-08 10:53:09 +00002696
Chris Wilson1233e2d2016-10-28 13:58:37 +01002697unlock:
2698 mutex_unlock(&obj->mm.lock);
Chris Wilson03ac84f2016-10-28 13:58:36 +01002699 return err;
Eric Anholt673a3942008-07-30 12:06:12 -07002700}
2701
Dave Gordondd6034c2016-05-20 11:54:04 +01002702/* The 'mapping' part of i915_gem_object_pin_map() below */
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002703static void *i915_gem_object_map(const struct drm_i915_gem_object *obj,
2704 enum i915_map_type type)
Dave Gordondd6034c2016-05-20 11:54:04 +01002705{
2706 unsigned long n_pages = obj->base.size >> PAGE_SHIFT;
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002707 struct sg_table *sgt = obj->mm.pages;
Dave Gordon85d12252016-05-20 11:54:06 +01002708 struct sgt_iter sgt_iter;
2709 struct page *page;
Dave Gordonb338fa42016-05-20 11:54:05 +01002710 struct page *stack_pages[32];
2711 struct page **pages = stack_pages;
Dave Gordondd6034c2016-05-20 11:54:04 +01002712 unsigned long i = 0;
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002713 pgprot_t pgprot;
Dave Gordondd6034c2016-05-20 11:54:04 +01002714 void *addr;
2715
2716 /* A single page can always be kmapped */
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002717 if (n_pages == 1 && type == I915_MAP_WB)
Dave Gordondd6034c2016-05-20 11:54:04 +01002718 return kmap(sg_page(sgt->sgl));
2719
Dave Gordonb338fa42016-05-20 11:54:05 +01002720 if (n_pages > ARRAY_SIZE(stack_pages)) {
2721 /* Too big for stack -- allocate temporary array instead */
Michal Hocko0ee931c2017-09-13 16:28:29 -07002722 pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL);
Dave Gordonb338fa42016-05-20 11:54:05 +01002723 if (!pages)
2724 return NULL;
2725 }
Dave Gordondd6034c2016-05-20 11:54:04 +01002726
Dave Gordon85d12252016-05-20 11:54:06 +01002727 for_each_sgt_page(page, sgt_iter, sgt)
2728 pages[i++] = page;
Dave Gordondd6034c2016-05-20 11:54:04 +01002729
2730 /* Check that we have the expected number of pages */
2731 GEM_BUG_ON(i != n_pages);
2732
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002733 switch (type) {
Chris Wilsona575c672017-08-28 11:46:31 +01002734 default:
2735 MISSING_CASE(type);
2736 /* fallthrough to use PAGE_KERNEL anyway */
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002737 case I915_MAP_WB:
2738 pgprot = PAGE_KERNEL;
2739 break;
2740 case I915_MAP_WC:
2741 pgprot = pgprot_writecombine(PAGE_KERNEL_IO);
2742 break;
2743 }
2744 addr = vmap(pages, n_pages, 0, pgprot);
Dave Gordondd6034c2016-05-20 11:54:04 +01002745
Dave Gordonb338fa42016-05-20 11:54:05 +01002746 if (pages != stack_pages)
Michal Hocko20981052017-05-17 14:23:12 +02002747 kvfree(pages);
Dave Gordondd6034c2016-05-20 11:54:04 +01002748
2749 return addr;
2750}
2751
2752/* get, pin, and map the pages of the object into kernel space */
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002753void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
2754 enum i915_map_type type)
Chris Wilson0a798eb2016-04-08 12:11:11 +01002755{
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002756 enum i915_map_type has_type;
2757 bool pinned;
2758 void *ptr;
Chris Wilson0a798eb2016-04-08 12:11:11 +01002759 int ret;
2760
Tina Zhanga03f3952017-11-14 10:25:13 +00002761 if (unlikely(!i915_gem_object_has_struct_page(obj)))
2762 return ERR_PTR(-ENXIO);
Chris Wilson0a798eb2016-04-08 12:11:11 +01002763
Chris Wilson1233e2d2016-10-28 13:58:37 +01002764 ret = mutex_lock_interruptible(&obj->mm.lock);
Chris Wilson0a798eb2016-04-08 12:11:11 +01002765 if (ret)
2766 return ERR_PTR(ret);
2767
Chris Wilsona575c672017-08-28 11:46:31 +01002768 pinned = !(type & I915_MAP_OVERRIDE);
2769 type &= ~I915_MAP_OVERRIDE;
2770
Chris Wilson1233e2d2016-10-28 13:58:37 +01002771 if (!atomic_inc_not_zero(&obj->mm.pages_pin_count)) {
Chris Wilsonf1fa4f42017-10-13 21:26:13 +01002772 if (unlikely(!i915_gem_object_has_pages(obj))) {
Chris Wilson88c880b2017-09-06 14:52:20 +01002773 GEM_BUG_ON(i915_gem_object_has_pinned_pages(obj));
2774
Chris Wilson2c3a3f42016-11-04 10:30:01 +00002775 ret = ____i915_gem_object_get_pages(obj);
2776 if (ret)
2777 goto err_unlock;
Chris Wilson1233e2d2016-10-28 13:58:37 +01002778
Chris Wilson2c3a3f42016-11-04 10:30:01 +00002779 smp_mb__before_atomic();
2780 }
2781 atomic_inc(&obj->mm.pages_pin_count);
Chris Wilson1233e2d2016-10-28 13:58:37 +01002782 pinned = false;
2783 }
Chris Wilsonf1fa4f42017-10-13 21:26:13 +01002784 GEM_BUG_ON(!i915_gem_object_has_pages(obj));
Chris Wilson0a798eb2016-04-08 12:11:11 +01002785
Chris Wilson0ce81782017-05-17 13:09:59 +01002786 ptr = page_unpack_bits(obj->mm.mapping, &has_type);
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002787 if (ptr && has_type != type) {
2788 if (pinned) {
2789 ret = -EBUSY;
Chris Wilson1233e2d2016-10-28 13:58:37 +01002790 goto err_unpin;
Chris Wilson0a798eb2016-04-08 12:11:11 +01002791 }
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002792
2793 if (is_vmalloc_addr(ptr))
2794 vunmap(ptr);
2795 else
2796 kunmap(kmap_to_page(ptr));
2797
Chris Wilsona4f5ea62016-10-28 13:58:35 +01002798 ptr = obj->mm.mapping = NULL;
Chris Wilson0a798eb2016-04-08 12:11:11 +01002799 }
2800
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002801 if (!ptr) {
2802 ptr = i915_gem_object_map(obj, type);
2803 if (!ptr) {
2804 ret = -ENOMEM;
Chris Wilson1233e2d2016-10-28 13:58:37 +01002805 goto err_unpin;
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002806 }
2807
Chris Wilson0ce81782017-05-17 13:09:59 +01002808 obj->mm.mapping = page_pack_bits(ptr, type);
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002809 }
2810
Chris Wilson1233e2d2016-10-28 13:58:37 +01002811out_unlock:
2812 mutex_unlock(&obj->mm.lock);
Chris Wilsond31d7cb2016-08-12 12:39:58 +01002813 return ptr;
2814
Chris Wilson1233e2d2016-10-28 13:58:37 +01002815err_unpin:
2816 atomic_dec(&obj->mm.pages_pin_count);
2817err_unlock:
2818 ptr = ERR_PTR(ret);
2819 goto out_unlock;
Chris Wilson0a798eb2016-04-08 12:11:11 +01002820}
2821
Chris Wilson7c55e2c2017-03-07 12:03:38 +00002822static int
2823i915_gem_object_pwrite_gtt(struct drm_i915_gem_object *obj,
2824 const struct drm_i915_gem_pwrite *arg)
2825{
2826 struct address_space *mapping = obj->base.filp->f_mapping;
2827 char __user *user_data = u64_to_user_ptr(arg->data_ptr);
2828 u64 remain, offset;
2829 unsigned int pg;
2830
2831 /* Before we instantiate/pin the backing store for our use, we
2832 * can prepopulate the shmemfs filp efficiently using a write into
2833 * the pagecache. We avoid the penalty of instantiating all the
2834 * pages, important if the user is just writing to a few and never
2835 * uses the object on the GPU, and using a direct write into shmemfs
2836 * allows it to avoid the cost of retrieving a page (either swapin
2837 * or clearing-before-use) before it is overwritten.
2838 */
Chris Wilsonf1fa4f42017-10-13 21:26:13 +01002839 if (i915_gem_object_has_pages(obj))
Chris Wilson7c55e2c2017-03-07 12:03:38 +00002840 return -ENODEV;
2841
Chris Wilsona6d65e42017-10-16 21:27:32 +01002842 if (obj->mm.madv != I915_MADV_WILLNEED)
2843 return -EFAULT;
2844
Chris Wilson7c55e2c2017-03-07 12:03:38 +00002845 /* Before the pages are instantiated the object is treated as being
2846 * in the CPU domain. The pages will be clflushed as required before
2847 * use, and we can freely write into the pages directly. If userspace
2848 * races pwrite with any other operation; corruption will ensue -
2849 * that is userspace's prerogative!
2850 */
2851
2852 remain = arg->size;
2853 offset = arg->offset;
2854 pg = offset_in_page(offset);
2855
2856 do {
2857 unsigned int len, unwritten;
2858 struct page *page;
2859 void *data, *vaddr;
2860 int err;
2861
2862 len = PAGE_SIZE - pg;
2863 if (len > remain)
2864 len = remain;
2865
2866 err = pagecache_write_begin(obj->base.filp, mapping,
2867 offset, len, 0,
2868 &page, &data);
2869 if (err < 0)
2870 return err;
2871
2872 vaddr = kmap(page);
2873 unwritten = copy_from_user(vaddr + pg, user_data, len);
2874 kunmap(page);
2875
2876 err = pagecache_write_end(obj->base.filp, mapping,
2877 offset, len, len - unwritten,
2878 page, data);
2879 if (err < 0)
2880 return err;
2881
2882 if (unwritten)
2883 return -EFAULT;
2884
2885 remain -= len;
2886 user_data += len;
2887 offset += len;
2888 pg = 0;
2889 } while (remain);
2890
2891 return 0;
2892}
2893
Chris Wilson85474442019-01-29 18:54:50 +00002894static bool match_ring(struct i915_request *rq)
2895{
2896 struct drm_i915_private *dev_priv = rq->i915;
2897 u32 ring = I915_READ(RING_START(rq->engine->mmio_base));
2898
2899 return ring == i915_ggtt_offset(rq->ring->vma);
2900}
2901
Chris Wilsone61e0f52018-02-21 09:56:36 +00002902struct i915_request *
Tvrtko Ursulin0bc40be2016-03-16 11:00:37 +00002903i915_gem_find_active_request(struct intel_engine_cs *engine)
Chris Wilson9375e442010-09-19 12:21:28 +01002904{
Chris Wilsone61e0f52018-02-21 09:56:36 +00002905 struct i915_request *request, *active = NULL;
Chris Wilson754c9fd2017-02-23 07:44:14 +00002906 unsigned long flags;
Mika Kuoppalaaa60c662013-06-12 15:13:20 +03002907
Chris Wilsoncc7cc532018-05-29 14:29:18 +01002908 /*
2909 * We are called by the error capture, reset and to dump engine
2910 * state at random points in time. In particular, note that neither is
2911 * crucially ordered with an interrupt. After a hang, the GPU is dead
2912 * and we assume that no more writes can happen (we waited long enough
2913 * for all writes that were in transaction to be flushed) - adding an
Chris Wilsonf69a02c2016-07-01 17:23:16 +01002914 * extra delay for a recent interrupt is pointless. Hence, we do
2915 * not need an engine->irq_seqno_barrier() before the seqno reads.
Chris Wilsoncc7cc532018-05-29 14:29:18 +01002916 * At all other times, we must assume the GPU is still running, but
2917 * we only care about the snapshot of this moment.
Chris Wilsonf69a02c2016-07-01 17:23:16 +01002918 */
Chris Wilsona89d1f92018-05-02 17:38:39 +01002919 spin_lock_irqsave(&engine->timeline.lock, flags);
2920 list_for_each_entry(request, &engine->timeline.requests, link) {
Chris Wilson5013eb82019-01-28 18:18:11 +00002921 if (i915_request_completed(request))
Chris Wilson4db080f2013-12-04 11:37:09 +00002922 continue;
Mika Kuoppalaaa60c662013-06-12 15:13:20 +03002923
Chris Wilson85474442019-01-29 18:54:50 +00002924 if (!i915_request_started(request))
2925 break;
2926
2927 /* More than one preemptible request may match! */
2928 if (!match_ring(request))
2929 break;
2930
Chris Wilson754c9fd2017-02-23 07:44:14 +00002931 active = request;
2932 break;
2933 }
Chris Wilsona89d1f92018-05-02 17:38:39 +01002934 spin_unlock_irqrestore(&engine->timeline.lock, flags);
Chris Wilson754c9fd2017-02-23 07:44:14 +00002935
2936 return active;
Mika Kuoppalab6b0fac2014-01-30 19:04:43 +02002937}
2938
Daniel Vetter75ef9da2010-08-21 00:25:16 +02002939static void
Eric Anholt673a3942008-07-30 12:06:12 -07002940i915_gem_retire_work_handler(struct work_struct *work)
2941{
Chris Wilsonb29c19b2013-09-25 17:34:56 +01002942 struct drm_i915_private *dev_priv =
Chris Wilson67d97da2016-07-04 08:08:31 +01002943 container_of(work, typeof(*dev_priv), gt.retire_work.work);
Chris Wilson91c8a322016-07-05 10:40:23 +01002944 struct drm_device *dev = &dev_priv->drm;
Eric Anholt673a3942008-07-30 12:06:12 -07002945
Chris Wilson891b48c2010-09-29 12:26:37 +01002946 /* Come back later if the device is busy... */
Chris Wilsonb29c19b2013-09-25 17:34:56 +01002947 if (mutex_trylock(&dev->struct_mutex)) {
Chris Wilsone61e0f52018-02-21 09:56:36 +00002948 i915_retire_requests(dev_priv);
Chris Wilsonb29c19b2013-09-25 17:34:56 +01002949 mutex_unlock(&dev->struct_mutex);
2950 }
Chris Wilson67d97da2016-07-04 08:08:31 +01002951
Chris Wilson88923042018-01-29 14:41:04 +00002952 /*
2953 * Keep the retire handler running until we are finally idle.
Chris Wilson67d97da2016-07-04 08:08:31 +01002954 * We do not need to do this test under locking as in the worst-case
2955 * we queue the retire worker once too often.
2956 */
Chris Wilson88923042018-01-29 14:41:04 +00002957 if (READ_ONCE(dev_priv->gt.awake))
Chris Wilson67d97da2016-07-04 08:08:31 +01002958 queue_delayed_work(dev_priv->wq,
2959 &dev_priv->gt.retire_work,
Chris Wilsonbcb45082012-10-05 17:02:57 +01002960 round_jiffies_up_relative(HZ));
Chris Wilsonb29c19b2013-09-25 17:34:56 +01002961}
Chris Wilson891b48c2010-09-29 12:26:37 +01002962
Chris Wilson84a10742018-01-24 11:36:08 +00002963static void shrink_caches(struct drm_i915_private *i915)
2964{
2965 /*
2966 * kmem_cache_shrink() discards empty slabs and reorders partially
2967 * filled slabs to prioritise allocating from the mostly full slabs,
2968 * with the aim of reducing fragmentation.
2969 */
2970 kmem_cache_shrink(i915->priorities);
2971 kmem_cache_shrink(i915->dependencies);
2972 kmem_cache_shrink(i915->requests);
2973 kmem_cache_shrink(i915->luts);
2974 kmem_cache_shrink(i915->vmas);
2975 kmem_cache_shrink(i915->objects);
2976}
2977
2978struct sleep_rcu_work {
2979 union {
2980 struct rcu_head rcu;
2981 struct work_struct work;
2982 };
2983 struct drm_i915_private *i915;
2984 unsigned int epoch;
2985};
2986
2987static inline bool
2988same_epoch(struct drm_i915_private *i915, unsigned int epoch)
2989{
2990 /*
2991 * There is a small chance that the epoch wrapped since we started
2992 * sleeping. If we assume that epoch is at least a u32, then it will
2993 * take at least 2^32 * 100ms for it to wrap, or about 326 years.
2994 */
2995 return epoch == READ_ONCE(i915->gt.epoch);
2996}
2997
2998static void __sleep_work(struct work_struct *work)
2999{
3000 struct sleep_rcu_work *s = container_of(work, typeof(*s), work);
3001 struct drm_i915_private *i915 = s->i915;
3002 unsigned int epoch = s->epoch;
3003
3004 kfree(s);
3005 if (same_epoch(i915, epoch))
3006 shrink_caches(i915);
3007}
3008
3009static void __sleep_rcu(struct rcu_head *rcu)
3010{
3011 struct sleep_rcu_work *s = container_of(rcu, typeof(*s), rcu);
3012 struct drm_i915_private *i915 = s->i915;
3013
Chris Wilsona1db9c52018-11-08 09:21:01 +00003014 destroy_rcu_head(&s->rcu);
3015
Chris Wilson84a10742018-01-24 11:36:08 +00003016 if (same_epoch(i915, s->epoch)) {
3017 INIT_WORK(&s->work, __sleep_work);
3018 queue_work(i915->wq, &s->work);
3019 } else {
3020 kfree(s);
3021 }
3022}
3023
Chris Wilson5427f202017-10-23 22:32:34 +01003024static inline bool
3025new_requests_since_last_retire(const struct drm_i915_private *i915)
3026{
3027 return (READ_ONCE(i915->gt.active_requests) ||
3028 work_pending(&i915->gt.idle_work.work));
3029}
3030
Chris Wilson1934f5de2018-05-31 23:40:57 +01003031static void assert_kernel_context_is_current(struct drm_i915_private *i915)
3032{
3033 struct intel_engine_cs *engine;
3034 enum intel_engine_id id;
3035
3036 if (i915_terminally_wedged(&i915->gpu_error))
3037 return;
3038
3039 GEM_BUG_ON(i915->gt.active_requests);
3040 for_each_engine(engine, i915, id) {
Chris Wilson21950ee2019-02-05 13:00:05 +00003041 GEM_BUG_ON(__i915_active_request_peek(&engine->timeline.last_request));
Chris Wilson1934f5de2018-05-31 23:40:57 +01003042 GEM_BUG_ON(engine->last_retired_context !=
3043 to_intel_context(i915->kernel_context, engine));
3044 }
3045}
3046
Chris Wilsonb29c19b2013-09-25 17:34:56 +01003047static void
3048i915_gem_idle_work_handler(struct work_struct *work)
3049{
3050 struct drm_i915_private *dev_priv =
Chris Wilson67d97da2016-07-04 08:08:31 +01003051 container_of(work, typeof(*dev_priv), gt.idle_work.work);
Chris Wilson84a10742018-01-24 11:36:08 +00003052 unsigned int epoch = I915_EPOCH_INVALID;
Chris Wilson67d97da2016-07-04 08:08:31 +01003053 bool rearm_hangcheck;
3054
3055 if (!READ_ONCE(dev_priv->gt.awake))
3056 return;
3057
Chris Wilson4dfacb02018-05-31 09:22:43 +01003058 if (READ_ONCE(dev_priv->gt.active_requests))
3059 return;
3060
3061 /*
3062 * Flush out the last user context, leaving only the pinned
3063 * kernel context resident. When we are idling on the kernel_context,
3064 * no more new requests (with a context switch) are emitted and we
3065 * can finally rest. A consequence is that the idle work handler is
3066 * always called at least twice before idling (and if the system is
3067 * idle that implies a round trip through the retire worker).
3068 */
3069 mutex_lock(&dev_priv->drm.struct_mutex);
3070 i915_gem_switch_to_kernel_context(dev_priv);
3071 mutex_unlock(&dev_priv->drm.struct_mutex);
3072
3073 GEM_TRACE("active_requests=%d (after switch-to-kernel-context)\n",
3074 READ_ONCE(dev_priv->gt.active_requests));
3075
Imre Deak0cb56702016-11-07 11:20:04 +02003076 /*
3077 * Wait for last execlists context complete, but bail out in case a
Chris Wilsonffed7bd2018-03-01 10:33:38 +00003078 * new request is submitted. As we don't trust the hardware, we
3079 * continue on if the wait times out. This is necessary to allow
3080 * the machine to suspend even if the hardware dies, and we will
3081 * try to recover in resume (after depriving the hardware of power,
3082 * it may be in a better mmod).
Imre Deak0cb56702016-11-07 11:20:04 +02003083 */
Chris Wilsonffed7bd2018-03-01 10:33:38 +00003084 __wait_for(if (new_requests_since_last_retire(dev_priv)) return,
3085 intel_engines_are_idle(dev_priv),
3086 I915_IDLE_ENGINES_TIMEOUT * 1000,
3087 10, 500);
Chris Wilson67d97da2016-07-04 08:08:31 +01003088
3089 rearm_hangcheck =
3090 cancel_delayed_work_sync(&dev_priv->gpu_error.hangcheck_work);
3091
Chris Wilson5427f202017-10-23 22:32:34 +01003092 if (!mutex_trylock(&dev_priv->drm.struct_mutex)) {
Chris Wilson67d97da2016-07-04 08:08:31 +01003093 /* Currently busy, come back later */
3094 mod_delayed_work(dev_priv->wq,
3095 &dev_priv->gt.idle_work,
3096 msecs_to_jiffies(50));
3097 goto out_rearm;
3098 }
3099
Imre Deak93c97dc2016-11-07 11:20:03 +02003100 /*
3101 * New request retired after this work handler started, extend active
3102 * period until next instance of the work.
3103 */
Chris Wilson5427f202017-10-23 22:32:34 +01003104 if (new_requests_since_last_retire(dev_priv))
Imre Deak93c97dc2016-11-07 11:20:03 +02003105 goto out_unlock;
3106
Chris Wilsone4d20062018-04-06 16:51:44 +01003107 epoch = __i915_gem_park(dev_priv);
Chris Wilsonff320d62017-10-23 22:32:35 +01003108
Chris Wilson1934f5de2018-05-31 23:40:57 +01003109 assert_kernel_context_is_current(dev_priv);
3110
Chris Wilson67d97da2016-07-04 08:08:31 +01003111 rearm_hangcheck = false;
Chris Wilson67d97da2016-07-04 08:08:31 +01003112out_unlock:
Chris Wilson5427f202017-10-23 22:32:34 +01003113 mutex_unlock(&dev_priv->drm.struct_mutex);
Chris Wilson35c94182015-04-07 16:20:37 +01003114
Chris Wilson67d97da2016-07-04 08:08:31 +01003115out_rearm:
3116 if (rearm_hangcheck) {
3117 GEM_BUG_ON(!dev_priv->gt.awake);
3118 i915_queue_hangcheck(dev_priv);
Chris Wilson35c94182015-04-07 16:20:37 +01003119 }
Chris Wilson84a10742018-01-24 11:36:08 +00003120
3121 /*
3122 * When we are idle, it is an opportune time to reap our caches.
3123 * However, we have many objects that utilise RCU and the ordered
3124 * i915->wq that this work is executing on. To try and flush any
3125 * pending frees now we are idle, we first wait for an RCU grace
3126 * period, and then queue a task (that will run last on the wq) to
3127 * shrink and re-optimize the caches.
3128 */
3129 if (same_epoch(dev_priv, epoch)) {
3130 struct sleep_rcu_work *s = kmalloc(sizeof(*s), GFP_KERNEL);
3131 if (s) {
Chris Wilsona1db9c52018-11-08 09:21:01 +00003132 init_rcu_head(&s->rcu);
Chris Wilson84a10742018-01-24 11:36:08 +00003133 s->i915 = dev_priv;
3134 s->epoch = epoch;
3135 call_rcu(&s->rcu, __sleep_rcu);
3136 }
3137 }
Eric Anholt673a3942008-07-30 12:06:12 -07003138}
3139
Chris Wilsonb1f788c2016-08-04 07:52:45 +01003140void i915_gem_close_object(struct drm_gem_object *gem, struct drm_file *file)
3141{
Chris Wilsond1b48c12017-08-16 09:52:08 +01003142 struct drm_i915_private *i915 = to_i915(gem->dev);
Chris Wilsonb1f788c2016-08-04 07:52:45 +01003143 struct drm_i915_gem_object *obj = to_intel_bo(gem);
3144 struct drm_i915_file_private *fpriv = file->driver_priv;
Chris Wilsond1b48c12017-08-16 09:52:08 +01003145 struct i915_lut_handle *lut, *ln;
Chris Wilsonb1f788c2016-08-04 07:52:45 +01003146
Chris Wilsond1b48c12017-08-16 09:52:08 +01003147 mutex_lock(&i915->drm.struct_mutex);
3148
3149 list_for_each_entry_safe(lut, ln, &obj->lut_list, obj_link) {
3150 struct i915_gem_context *ctx = lut->ctx;
3151 struct i915_vma *vma;
3152
Chris Wilson432295d2017-08-22 12:05:15 +01003153 GEM_BUG_ON(ctx->file_priv == ERR_PTR(-EBADF));
Chris Wilsond1b48c12017-08-16 09:52:08 +01003154 if (ctx->file_priv != fpriv)
3155 continue;
3156
3157 vma = radix_tree_delete(&ctx->handles_vma, lut->handle);
Chris Wilson3ffff012017-08-22 12:05:17 +01003158 GEM_BUG_ON(vma->obj != obj);
3159
3160 /* We allow the process to have multiple handles to the same
3161 * vma, in the same fd namespace, by virtue of flink/open.
3162 */
3163 GEM_BUG_ON(!vma->open_count);
3164 if (!--vma->open_count && !i915_vma_is_ggtt(vma))
Chris Wilsonb1f788c2016-08-04 07:52:45 +01003165 i915_vma_close(vma);
Chris Wilsonf8a7fde2016-10-28 13:58:29 +01003166
Chris Wilsond1b48c12017-08-16 09:52:08 +01003167 list_del(&lut->obj_link);
3168 list_del(&lut->ctx_link);
Chris Wilson4ff4b442017-06-16 15:05:16 +01003169
Chris Wilsond1b48c12017-08-16 09:52:08 +01003170 kmem_cache_free(i915->luts, lut);
3171 __i915_gem_object_release_unless_active(obj);
Chris Wilsonf8a7fde2016-10-28 13:58:29 +01003172 }
Chris Wilsond1b48c12017-08-16 09:52:08 +01003173
3174 mutex_unlock(&i915->drm.struct_mutex);
Chris Wilsonb1f788c2016-08-04 07:52:45 +01003175}
3176
Chris Wilsone95433c72016-10-28 13:58:27 +01003177static unsigned long to_wait_timeout(s64 timeout_ns)
3178{
3179 if (timeout_ns < 0)
3180 return MAX_SCHEDULE_TIMEOUT;
3181
3182 if (timeout_ns == 0)
3183 return 0;
3184
3185 return nsecs_to_jiffies_timeout(timeout_ns);
3186}
3187
Ben Widawsky5816d642012-04-11 11:18:19 -07003188/**
Ben Widawsky23ba4fd2012-05-24 15:03:10 -07003189 * i915_gem_wait_ioctl - implements DRM_IOCTL_I915_GEM_WAIT
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +01003190 * @dev: drm device pointer
3191 * @data: ioctl data blob
3192 * @file: drm file pointer
Ben Widawsky23ba4fd2012-05-24 15:03:10 -07003193 *
3194 * Returns 0 if successful, else an error is returned with the remaining time in
3195 * the timeout parameter.
3196 * -ETIME: object is still busy after timeout
3197 * -ERESTARTSYS: signal interrupted the wait
3198 * -ENONENT: object doesn't exist
3199 * Also possible, but rare:
Chris Wilsonb8050142017-08-11 11:57:31 +01003200 * -EAGAIN: incomplete, restart syscall
Ben Widawsky23ba4fd2012-05-24 15:03:10 -07003201 * -ENOMEM: damn
3202 * -ENODEV: Internal IRQ fail
3203 * -E?: The add request failed
3204 *
3205 * The wait ioctl with a timeout of 0 reimplements the busy ioctl. With any
3206 * non-zero timeout parameter the wait ioctl will wait for the given number of
3207 * nanoseconds on an object becoming unbusy. Since the wait itself does so
3208 * without holding struct_mutex the object may become re-busied before this
3209 * function completes. A similar but shorter * race condition exists in the busy
3210 * ioctl
3211 */
3212int
3213i915_gem_wait_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
3214{
3215 struct drm_i915_gem_wait *args = data;
3216 struct drm_i915_gem_object *obj;
Chris Wilsone95433c72016-10-28 13:58:27 +01003217 ktime_t start;
3218 long ret;
Ben Widawsky23ba4fd2012-05-24 15:03:10 -07003219
Daniel Vetter11b5d512014-09-29 15:31:26 +02003220 if (args->flags != 0)
3221 return -EINVAL;
3222
Chris Wilson03ac0642016-07-20 13:31:51 +01003223 obj = i915_gem_object_lookup(file, args->bo_handle);
Chris Wilson033d5492016-08-05 10:14:17 +01003224 if (!obj)
Ben Widawsky23ba4fd2012-05-24 15:03:10 -07003225 return -ENOENT;
Chris Wilson033d5492016-08-05 10:14:17 +01003226
Chris Wilsone95433c72016-10-28 13:58:27 +01003227 start = ktime_get();
3228
3229 ret = i915_gem_object_wait(obj,
Chris Wilsone9eaf822018-10-01 15:47:55 +01003230 I915_WAIT_INTERRUPTIBLE |
3231 I915_WAIT_PRIORITY |
3232 I915_WAIT_ALL,
Chris Wilsone95433c72016-10-28 13:58:27 +01003233 to_wait_timeout(args->timeout_ns),
3234 to_rps_client(file));
3235
3236 if (args->timeout_ns > 0) {
3237 args->timeout_ns -= ktime_to_ns(ktime_sub(ktime_get(), start));
3238 if (args->timeout_ns < 0)
3239 args->timeout_ns = 0;
Chris Wilsonc1d20612017-02-16 12:54:41 +00003240
3241 /*
3242 * Apparently ktime isn't accurate enough and occasionally has a
3243 * bit of mismatch in the jiffies<->nsecs<->ktime loop. So patch
3244 * things up to make the test happy. We allow up to 1 jiffy.
3245 *
3246 * This is a regression from the timespec->ktime conversion.
3247 */
3248 if (ret == -ETIME && !nsecs_to_jiffies(args->timeout_ns))
3249 args->timeout_ns = 0;
Chris Wilsonb8050142017-08-11 11:57:31 +01003250
3251 /* Asked to wait beyond the jiffie/scheduler precision? */
3252 if (ret == -ETIME && args->timeout_ns)
3253 ret = -EAGAIN;
Ben Widawsky23ba4fd2012-05-24 15:03:10 -07003254 }
3255
Chris Wilsonf0cd5182016-10-28 13:58:43 +01003256 i915_gem_object_put(obj);
John Harrisonff865882014-11-24 18:49:28 +00003257 return ret;
Ben Widawsky23ba4fd2012-05-24 15:03:10 -07003258}
3259
Chris Wilson25112b62017-03-30 15:50:39 +01003260static int wait_for_engines(struct drm_i915_private *i915)
3261{
Chris Wilsonee42c002017-12-11 19:41:34 +00003262 if (wait_for(intel_engines_are_idle(i915), I915_IDLE_ENGINES_TIMEOUT)) {
Chris Wilson59e4b192017-12-11 19:41:35 +00003263 dev_err(i915->drm.dev,
3264 "Failed to idle engines, declaring wedged!\n");
Chris Wilson629820f2018-03-09 10:11:14 +00003265 GEM_TRACE_DUMP();
Chris Wilsoncad99462017-08-26 12:09:33 +01003266 i915_gem_set_wedged(i915);
3267 return -EIO;
Chris Wilson25112b62017-03-30 15:50:39 +01003268 }
3269
3270 return 0;
3271}
3272
Chris Wilson1e345562019-01-28 10:23:56 +00003273static long
3274wait_for_timelines(struct drm_i915_private *i915,
3275 unsigned int flags, long timeout)
3276{
3277 struct i915_gt_timelines *gt = &i915->gt.timelines;
3278 struct i915_timeline *tl;
3279
3280 if (!READ_ONCE(i915->gt.active_requests))
3281 return timeout;
3282
3283 mutex_lock(&gt->mutex);
Chris Wilson9407d3b2019-01-28 18:18:12 +00003284 list_for_each_entry(tl, &gt->active_list, link) {
Chris Wilson1e345562019-01-28 10:23:56 +00003285 struct i915_request *rq;
3286
Chris Wilson21950ee2019-02-05 13:00:05 +00003287 rq = i915_active_request_get_unlocked(&tl->last_request);
Chris Wilson1e345562019-01-28 10:23:56 +00003288 if (!rq)
3289 continue;
3290
3291 mutex_unlock(&gt->mutex);
3292
3293 /*
3294 * "Race-to-idle".
3295 *
3296 * Switching to the kernel context is often used a synchronous
3297 * step prior to idling, e.g. in suspend for flushing all
3298 * current operations to memory before sleeping. These we
3299 * want to complete as quickly as possible to avoid prolonged
3300 * stalls, so allow the gpu to boost to maximum clocks.
3301 */
3302 if (flags & I915_WAIT_FOR_IDLE_BOOST)
3303 gen6_rps_boost(rq, NULL);
3304
3305 timeout = i915_request_wait(rq, flags, timeout);
3306 i915_request_put(rq);
3307 if (timeout < 0)
3308 return timeout;
3309
3310 /* restart after reacquiring the lock */
3311 mutex_lock(&gt->mutex);
Chris Wilson9407d3b2019-01-28 18:18:12 +00003312 tl = list_entry(&gt->active_list, typeof(*tl), link);
Chris Wilson1e345562019-01-28 10:23:56 +00003313 }
3314 mutex_unlock(&gt->mutex);
3315
3316 return timeout;
3317}
3318
Chris Wilsonec625fb2018-07-09 13:20:42 +01003319int i915_gem_wait_for_idle(struct drm_i915_private *i915,
3320 unsigned int flags, long timeout)
Chris Wilson73cb9702016-10-28 13:58:46 +01003321{
Chris Wilsonec625fb2018-07-09 13:20:42 +01003322 GEM_TRACE("flags=%x (%s), timeout=%ld%s\n",
3323 flags, flags & I915_WAIT_LOCKED ? "locked" : "unlocked",
3324 timeout, timeout == MAX_SCHEDULE_TIMEOUT ? " (forever)" : "");
Chris Wilson09a4c022018-05-24 09:11:35 +01003325
Chris Wilson863e9fd2017-05-30 13:13:32 +01003326 /* If the device is asleep, we have no requests outstanding */
3327 if (!READ_ONCE(i915->gt.awake))
3328 return 0;
3329
Chris Wilson1e345562019-01-28 10:23:56 +00003330 timeout = wait_for_timelines(i915, flags, timeout);
3331 if (timeout < 0)
3332 return timeout;
3333
Chris Wilson9caa34a2016-11-11 14:58:08 +00003334 if (flags & I915_WAIT_LOCKED) {
Chris Wilsona89d1f92018-05-02 17:38:39 +01003335 int err;
Chris Wilson9caa34a2016-11-11 14:58:08 +00003336
3337 lockdep_assert_held(&i915->drm.struct_mutex);
3338
Chris Wilsonc1e63f62018-08-08 11:50:59 +01003339 if (GEM_SHOW_DEBUG() && !timeout) {
3340 /* Presume that timeout was non-zero to begin with! */
3341 dev_warn(&i915->drm.pdev->dev,
3342 "Missed idle-completion interrupt!\n");
3343 GEM_TRACE_DUMP();
3344 }
Chris Wilsona61b47f2018-06-27 12:53:34 +01003345
3346 err = wait_for_engines(i915);
3347 if (err)
3348 return err;
3349
Chris Wilsone61e0f52018-02-21 09:56:36 +00003350 i915_retire_requests(i915);
Chris Wilson09a4c022018-05-24 09:11:35 +01003351 GEM_BUG_ON(i915->gt.active_requests);
Chris Wilsona89d1f92018-05-02 17:38:39 +01003352 }
Chris Wilsona61b47f2018-06-27 12:53:34 +01003353
3354 return 0;
Daniel Vetter4df2faf2010-02-19 11:52:00 +01003355}
3356
Chris Wilson5a97bcc2017-02-22 11:40:46 +00003357static void __i915_gem_object_flush_for_display(struct drm_i915_gem_object *obj)
3358{
Chris Wilsone27ab732017-06-15 13:38:49 +01003359 /*
3360 * We manually flush the CPU domain so that we can override and
3361 * force the flush for the display, and perform it asyncrhonously.
3362 */
3363 flush_write_domain(obj, ~I915_GEM_DOMAIN_CPU);
3364 if (obj->cache_dirty)
3365 i915_gem_clflush_object(obj, I915_CLFLUSH_FORCE);
Christian Königc0a51fd2018-02-16 13:43:38 +01003366 obj->write_domain = 0;
Chris Wilson5a97bcc2017-02-22 11:40:46 +00003367}
3368
3369void i915_gem_object_flush_if_display(struct drm_i915_gem_object *obj)
3370{
Chris Wilsonbd3d2252017-10-13 21:26:14 +01003371 if (!READ_ONCE(obj->pin_global))
Chris Wilson5a97bcc2017-02-22 11:40:46 +00003372 return;
3373
3374 mutex_lock(&obj->base.dev->struct_mutex);
3375 __i915_gem_object_flush_for_display(obj);
3376 mutex_unlock(&obj->base.dev->struct_mutex);
3377}
3378
Eric Anholt2ef7eea2008-11-10 10:53:25 -08003379/**
Chris Wilsone22d8e32017-04-12 12:01:11 +01003380 * Moves a single object to the WC read, and possibly write domain.
3381 * @obj: object to act on
3382 * @write: ask for write access or read only
3383 *
3384 * This function returns when the move is complete, including waiting on
3385 * flushes to occur.
3386 */
3387int
3388i915_gem_object_set_to_wc_domain(struct drm_i915_gem_object *obj, bool write)
3389{
3390 int ret;
3391
3392 lockdep_assert_held(&obj->base.dev->struct_mutex);
3393
3394 ret = i915_gem_object_wait(obj,
3395 I915_WAIT_INTERRUPTIBLE |
3396 I915_WAIT_LOCKED |
3397 (write ? I915_WAIT_ALL : 0),
3398 MAX_SCHEDULE_TIMEOUT,
3399 NULL);
3400 if (ret)
3401 return ret;
3402
Christian Königc0a51fd2018-02-16 13:43:38 +01003403 if (obj->write_domain == I915_GEM_DOMAIN_WC)
Chris Wilsone22d8e32017-04-12 12:01:11 +01003404 return 0;
3405
3406 /* Flush and acquire obj->pages so that we are coherent through
3407 * direct access in memory with previous cached writes through
3408 * shmemfs and that our cache domain tracking remains valid.
3409 * For example, if the obj->filp was moved to swap without us
3410 * being notified and releasing the pages, we would mistakenly
3411 * continue to assume that the obj remained out of the CPU cached
3412 * domain.
3413 */
3414 ret = i915_gem_object_pin_pages(obj);
3415 if (ret)
3416 return ret;
3417
3418 flush_write_domain(obj, ~I915_GEM_DOMAIN_WC);
3419
3420 /* Serialise direct access to this object with the barriers for
3421 * coherent writes from the GPU, by effectively invalidating the
3422 * WC domain upon first access.
3423 */
Christian Königc0a51fd2018-02-16 13:43:38 +01003424 if ((obj->read_domains & I915_GEM_DOMAIN_WC) == 0)
Chris Wilsone22d8e32017-04-12 12:01:11 +01003425 mb();
3426
3427 /* It should now be out of any other write domains, and we can update
3428 * the domain values for our changes.
3429 */
Christian Königc0a51fd2018-02-16 13:43:38 +01003430 GEM_BUG_ON((obj->write_domain & ~I915_GEM_DOMAIN_WC) != 0);
3431 obj->read_domains |= I915_GEM_DOMAIN_WC;
Chris Wilsone22d8e32017-04-12 12:01:11 +01003432 if (write) {
Christian Königc0a51fd2018-02-16 13:43:38 +01003433 obj->read_domains = I915_GEM_DOMAIN_WC;
3434 obj->write_domain = I915_GEM_DOMAIN_WC;
Chris Wilsone22d8e32017-04-12 12:01:11 +01003435 obj->mm.dirty = true;
3436 }
3437
3438 i915_gem_object_unpin_pages(obj);
3439 return 0;
3440}
3441
3442/**
Eric Anholt2ef7eea2008-11-10 10:53:25 -08003443 * Moves a single object to the GTT read, and possibly write domain.
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +01003444 * @obj: object to act on
3445 * @write: ask for write access or read only
Eric Anholt2ef7eea2008-11-10 10:53:25 -08003446 *
3447 * This function returns when the move is complete, including waiting on
3448 * flushes to occur.
3449 */
Jesse Barnes79e53942008-11-07 14:24:08 -08003450int
Chris Wilson20217462010-11-23 15:26:33 +00003451i915_gem_object_set_to_gtt_domain(struct drm_i915_gem_object *obj, bool write)
Eric Anholt2ef7eea2008-11-10 10:53:25 -08003452{
Eric Anholte47c68e2008-11-14 13:35:19 -08003453 int ret;
Eric Anholt2ef7eea2008-11-10 10:53:25 -08003454
Chris Wilsone95433c72016-10-28 13:58:27 +01003455 lockdep_assert_held(&obj->base.dev->struct_mutex);
Chris Wilson4c7d62c2016-10-28 13:58:32 +01003456
Chris Wilsone95433c72016-10-28 13:58:27 +01003457 ret = i915_gem_object_wait(obj,
3458 I915_WAIT_INTERRUPTIBLE |
3459 I915_WAIT_LOCKED |
3460 (write ? I915_WAIT_ALL : 0),
3461 MAX_SCHEDULE_TIMEOUT,
3462 NULL);
Chris Wilson88241782011-01-07 17:09:48 +00003463 if (ret)
3464 return ret;
3465
Christian Königc0a51fd2018-02-16 13:43:38 +01003466 if (obj->write_domain == I915_GEM_DOMAIN_GTT)
Chris Wilsonc13d87e2016-07-20 09:21:15 +01003467 return 0;
3468
Chris Wilson43566de2015-01-02 16:29:29 +05303469 /* Flush and acquire obj->pages so that we are coherent through
3470 * direct access in memory with previous cached writes through
3471 * shmemfs and that our cache domain tracking remains valid.
3472 * For example, if the obj->filp was moved to swap without us
3473 * being notified and releasing the pages, we would mistakenly
3474 * continue to assume that the obj remained out of the CPU cached
3475 * domain.
3476 */
Chris Wilsona4f5ea62016-10-28 13:58:35 +01003477 ret = i915_gem_object_pin_pages(obj);
Chris Wilson43566de2015-01-02 16:29:29 +05303478 if (ret)
3479 return ret;
3480
Chris Wilsonef749212017-04-12 12:01:10 +01003481 flush_write_domain(obj, ~I915_GEM_DOMAIN_GTT);
Chris Wilson1c5d22f2009-08-25 11:15:50 +01003482
Chris Wilsond0a57782012-10-09 19:24:37 +01003483 /* Serialise direct access to this object with the barriers for
3484 * coherent writes from the GPU, by effectively invalidating the
3485 * GTT domain upon first access.
3486 */
Christian Königc0a51fd2018-02-16 13:43:38 +01003487 if ((obj->read_domains & I915_GEM_DOMAIN_GTT) == 0)
Chris Wilsond0a57782012-10-09 19:24:37 +01003488 mb();
3489
Eric Anholt2ef7eea2008-11-10 10:53:25 -08003490 /* It should now be out of any other write domains, and we can update
3491 * the domain values for our changes.
3492 */
Christian Königc0a51fd2018-02-16 13:43:38 +01003493 GEM_BUG_ON((obj->write_domain & ~I915_GEM_DOMAIN_GTT) != 0);
3494 obj->read_domains |= I915_GEM_DOMAIN_GTT;
Eric Anholte47c68e2008-11-14 13:35:19 -08003495 if (write) {
Christian Königc0a51fd2018-02-16 13:43:38 +01003496 obj->read_domains = I915_GEM_DOMAIN_GTT;
3497 obj->write_domain = I915_GEM_DOMAIN_GTT;
Chris Wilsona4f5ea62016-10-28 13:58:35 +01003498 obj->mm.dirty = true;
Eric Anholte47c68e2008-11-14 13:35:19 -08003499 }
3500
Chris Wilsona4f5ea62016-10-28 13:58:35 +01003501 i915_gem_object_unpin_pages(obj);
Eric Anholte47c68e2008-11-14 13:35:19 -08003502 return 0;
3503}
3504
Chris Wilsonef55f922015-10-09 14:11:27 +01003505/**
3506 * Changes the cache-level of an object across all VMA.
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +01003507 * @obj: object to act on
3508 * @cache_level: new cache level to set for the object
Chris Wilsonef55f922015-10-09 14:11:27 +01003509 *
3510 * After this function returns, the object will be in the new cache-level
3511 * across all GTT and the contents of the backing storage will be coherent,
3512 * with respect to the new cache-level. In order to keep the backing storage
3513 * coherent for all users, we only allow a single cache level to be set
3514 * globally on the object and prevent it from being changed whilst the
3515 * hardware is reading from the object. That is if the object is currently
3516 * on the scanout it will be set to uncached (or equivalent display
3517 * cache coherency) and all non-MOCS GPU access will also be uncached so
3518 * that all direct access to the scanout remains coherent.
3519 */
Chris Wilsone4ffd172011-04-04 09:44:39 +01003520int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj,
3521 enum i915_cache_level cache_level)
3522{
Chris Wilsonaa653a62016-08-04 07:52:27 +01003523 struct i915_vma *vma;
Chris Wilsona6a7cc42016-11-18 21:17:46 +00003524 int ret;
Chris Wilsone4ffd172011-04-04 09:44:39 +01003525
Chris Wilson4c7d62c2016-10-28 13:58:32 +01003526 lockdep_assert_held(&obj->base.dev->struct_mutex);
3527
Chris Wilsone4ffd172011-04-04 09:44:39 +01003528 if (obj->cache_level == cache_level)
Chris Wilsona6a7cc42016-11-18 21:17:46 +00003529 return 0;
Chris Wilsone4ffd172011-04-04 09:44:39 +01003530
Chris Wilsonef55f922015-10-09 14:11:27 +01003531 /* Inspect the list of currently bound VMA and unbind any that would
3532 * be invalid given the new cache-level. This is principally to
3533 * catch the issue of the CS prefetch crossing page boundaries and
3534 * reading an invalid PTE on older architectures.
3535 */
Chris Wilsonaa653a62016-08-04 07:52:27 +01003536restart:
Chris Wilson528cbd12019-01-28 10:23:54 +00003537 list_for_each_entry(vma, &obj->vma.list, obj_link) {
Chris Wilsonef55f922015-10-09 14:11:27 +01003538 if (!drm_mm_node_allocated(&vma->node))
3539 continue;
3540
Chris Wilson20dfbde2016-08-04 16:32:30 +01003541 if (i915_vma_is_pinned(vma)) {
Chris Wilsonef55f922015-10-09 14:11:27 +01003542 DRM_DEBUG("can not change the cache level of pinned objects\n");
3543 return -EBUSY;
3544 }
3545
Chris Wilson010e3e62017-12-06 12:49:13 +00003546 if (!i915_vma_is_closed(vma) &&
3547 i915_gem_valid_gtt_space(vma, cache_level))
Chris Wilsonaa653a62016-08-04 07:52:27 +01003548 continue;
3549
3550 ret = i915_vma_unbind(vma);
3551 if (ret)
3552 return ret;
3553
3554 /* As unbinding may affect other elements in the
3555 * obj->vma_list (due to side-effects from retiring
3556 * an active vma), play safe and restart the iterator.
3557 */
3558 goto restart;
Chris Wilson42d6ab42012-07-26 11:49:32 +01003559 }
3560
Chris Wilsonef55f922015-10-09 14:11:27 +01003561 /* We can reuse the existing drm_mm nodes but need to change the
3562 * cache-level on the PTE. We could simply unbind them all and
3563 * rebind with the correct cache-level on next use. However since
3564 * we already have a valid slot, dma mapping, pages etc, we may as
3565 * rewrite the PTE in the belief that doing so tramples upon less
3566 * state and so involves less work.
3567 */
Chris Wilson15717de2016-08-04 07:52:26 +01003568 if (obj->bind_count) {
Chris Wilsonef55f922015-10-09 14:11:27 +01003569 /* Before we change the PTE, the GPU must not be accessing it.
3570 * If we wait upon the object, we know that all the bound
3571 * VMA are no longer active.
3572 */
Chris Wilsone95433c72016-10-28 13:58:27 +01003573 ret = i915_gem_object_wait(obj,
3574 I915_WAIT_INTERRUPTIBLE |
3575 I915_WAIT_LOCKED |
3576 I915_WAIT_ALL,
3577 MAX_SCHEDULE_TIMEOUT,
3578 NULL);
Chris Wilsone4ffd172011-04-04 09:44:39 +01003579 if (ret)
3580 return ret;
3581
Tvrtko Ursulin0031fb92016-11-04 14:42:44 +00003582 if (!HAS_LLC(to_i915(obj->base.dev)) &&
3583 cache_level != I915_CACHE_NONE) {
Chris Wilsonef55f922015-10-09 14:11:27 +01003584 /* Access to snoopable pages through the GTT is
3585 * incoherent and on some machines causes a hard
3586 * lockup. Relinquish the CPU mmaping to force
3587 * userspace to refault in the pages and we can
3588 * then double check if the GTT mapping is still
3589 * valid for that pointer access.
3590 */
3591 i915_gem_release_mmap(obj);
Chris Wilsone4ffd172011-04-04 09:44:39 +01003592
Chris Wilsonef55f922015-10-09 14:11:27 +01003593 /* As we no longer need a fence for GTT access,
3594 * we can relinquish it now (and so prevent having
3595 * to steal a fence from someone else on the next
3596 * fence request). Note GPU activity would have
3597 * dropped the fence as all snoopable access is
3598 * supposed to be linear.
3599 */
Chris Wilsone2189dd2017-12-07 21:14:07 +00003600 for_each_ggtt_vma(vma, obj) {
Chris Wilson49ef5292016-08-18 17:17:00 +01003601 ret = i915_vma_put_fence(vma);
3602 if (ret)
3603 return ret;
3604 }
Chris Wilsonef55f922015-10-09 14:11:27 +01003605 } else {
3606 /* We either have incoherent backing store and
3607 * so no GTT access or the architecture is fully
3608 * coherent. In such cases, existing GTT mmaps
3609 * ignore the cache bit in the PTE and we can
3610 * rewrite it without confusing the GPU or having
3611 * to force userspace to fault back in its mmaps.
3612 */
Chris Wilsone4ffd172011-04-04 09:44:39 +01003613 }
3614
Chris Wilson528cbd12019-01-28 10:23:54 +00003615 list_for_each_entry(vma, &obj->vma.list, obj_link) {
Chris Wilsonef55f922015-10-09 14:11:27 +01003616 if (!drm_mm_node_allocated(&vma->node))
3617 continue;
3618
3619 ret = i915_vma_bind(vma, cache_level, PIN_UPDATE);
3620 if (ret)
3621 return ret;
3622 }
Chris Wilsone4ffd172011-04-04 09:44:39 +01003623 }
3624
Chris Wilson528cbd12019-01-28 10:23:54 +00003625 list_for_each_entry(vma, &obj->vma.list, obj_link)
Chris Wilson2c225692013-08-09 12:26:45 +01003626 vma->node.color = cache_level;
Chris Wilsonb8f55be2017-08-11 12:11:16 +01003627 i915_gem_object_set_cache_coherency(obj, cache_level);
Chris Wilsone27ab732017-06-15 13:38:49 +01003628 obj->cache_dirty = true; /* Always invalidate stale cachelines */
Chris Wilson2c225692013-08-09 12:26:45 +01003629
Chris Wilsone4ffd172011-04-04 09:44:39 +01003630 return 0;
3631}
3632
Ben Widawsky199adf42012-09-21 17:01:20 -07003633int i915_gem_get_caching_ioctl(struct drm_device *dev, void *data,
3634 struct drm_file *file)
Chris Wilsone6994ae2012-07-10 10:27:08 +01003635{
Ben Widawsky199adf42012-09-21 17:01:20 -07003636 struct drm_i915_gem_caching *args = data;
Chris Wilsone6994ae2012-07-10 10:27:08 +01003637 struct drm_i915_gem_object *obj;
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01003638 int err = 0;
Chris Wilsone6994ae2012-07-10 10:27:08 +01003639
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01003640 rcu_read_lock();
3641 obj = i915_gem_object_lookup_rcu(file, args->handle);
3642 if (!obj) {
3643 err = -ENOENT;
3644 goto out;
3645 }
Chris Wilsone6994ae2012-07-10 10:27:08 +01003646
Chris Wilson651d7942013-08-08 14:41:10 +01003647 switch (obj->cache_level) {
3648 case I915_CACHE_LLC:
3649 case I915_CACHE_L3_LLC:
3650 args->caching = I915_CACHING_CACHED;
3651 break;
3652
Chris Wilson4257d3b2013-08-08 14:41:11 +01003653 case I915_CACHE_WT:
3654 args->caching = I915_CACHING_DISPLAY;
3655 break;
3656
Chris Wilson651d7942013-08-08 14:41:10 +01003657 default:
3658 args->caching = I915_CACHING_NONE;
3659 break;
3660 }
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01003661out:
3662 rcu_read_unlock();
3663 return err;
Chris Wilsone6994ae2012-07-10 10:27:08 +01003664}
3665
Ben Widawsky199adf42012-09-21 17:01:20 -07003666int i915_gem_set_caching_ioctl(struct drm_device *dev, void *data,
3667 struct drm_file *file)
Chris Wilsone6994ae2012-07-10 10:27:08 +01003668{
Chris Wilson9c870d02016-10-24 13:42:15 +01003669 struct drm_i915_private *i915 = to_i915(dev);
Ben Widawsky199adf42012-09-21 17:01:20 -07003670 struct drm_i915_gem_caching *args = data;
Chris Wilsone6994ae2012-07-10 10:27:08 +01003671 struct drm_i915_gem_object *obj;
3672 enum i915_cache_level level;
Chris Wilsond65415d2017-01-19 08:22:10 +00003673 int ret = 0;
Chris Wilsone6994ae2012-07-10 10:27:08 +01003674
Ben Widawsky199adf42012-09-21 17:01:20 -07003675 switch (args->caching) {
3676 case I915_CACHING_NONE:
Chris Wilsone6994ae2012-07-10 10:27:08 +01003677 level = I915_CACHE_NONE;
3678 break;
Ben Widawsky199adf42012-09-21 17:01:20 -07003679 case I915_CACHING_CACHED:
Imre Deake5756c12015-08-14 18:43:30 +03003680 /*
3681 * Due to a HW issue on BXT A stepping, GPU stores via a
3682 * snooped mapping may leave stale data in a corresponding CPU
3683 * cacheline, whereas normally such cachelines would get
3684 * invalidated.
3685 */
Chris Wilson9c870d02016-10-24 13:42:15 +01003686 if (!HAS_LLC(i915) && !HAS_SNOOP(i915))
Imre Deake5756c12015-08-14 18:43:30 +03003687 return -ENODEV;
3688
Chris Wilsone6994ae2012-07-10 10:27:08 +01003689 level = I915_CACHE_LLC;
3690 break;
Chris Wilson4257d3b2013-08-08 14:41:11 +01003691 case I915_CACHING_DISPLAY:
Chris Wilson9c870d02016-10-24 13:42:15 +01003692 level = HAS_WT(i915) ? I915_CACHE_WT : I915_CACHE_NONE;
Chris Wilson4257d3b2013-08-08 14:41:11 +01003693 break;
Chris Wilsone6994ae2012-07-10 10:27:08 +01003694 default:
3695 return -EINVAL;
3696 }
3697
Chris Wilsond65415d2017-01-19 08:22:10 +00003698 obj = i915_gem_object_lookup(file, args->handle);
3699 if (!obj)
3700 return -ENOENT;
3701
Tina Zhanga03f3952017-11-14 10:25:13 +00003702 /*
3703 * The caching mode of proxy object is handled by its generator, and
3704 * not allowed to be changed by userspace.
3705 */
3706 if (i915_gem_object_is_proxy(obj)) {
3707 ret = -ENXIO;
3708 goto out;
3709 }
3710
Chris Wilsond65415d2017-01-19 08:22:10 +00003711 if (obj->cache_level == level)
3712 goto out;
3713
3714 ret = i915_gem_object_wait(obj,
3715 I915_WAIT_INTERRUPTIBLE,
3716 MAX_SCHEDULE_TIMEOUT,
3717 to_rps_client(file));
3718 if (ret)
3719 goto out;
3720
Ben Widawsky3bc29132012-09-26 16:15:20 -07003721 ret = i915_mutex_lock_interruptible(dev);
3722 if (ret)
Chris Wilsond65415d2017-01-19 08:22:10 +00003723 goto out;
Chris Wilsone6994ae2012-07-10 10:27:08 +01003724
3725 ret = i915_gem_object_set_cache_level(obj, level);
Chris Wilsone6994ae2012-07-10 10:27:08 +01003726 mutex_unlock(&dev->struct_mutex);
Chris Wilsond65415d2017-01-19 08:22:10 +00003727
3728out:
3729 i915_gem_object_put(obj);
Chris Wilsone6994ae2012-07-10 10:27:08 +01003730 return ret;
3731}
3732
Zhenyu Wangb9241ea2009-11-25 13:09:39 +08003733/*
Dhinakaran Pandiyan07bcd992018-03-06 19:34:18 -08003734 * Prepare buffer for display plane (scanout, cursors, etc). Can be called from
3735 * an uninterruptible phase (modesetting) and allows any flushes to be pipelined
3736 * (for pageflips). We only flush the caches while preparing the buffer for
3737 * display, the callers are responsible for frontbuffer flush.
Zhenyu Wangb9241ea2009-11-25 13:09:39 +08003738 */
Chris Wilson058d88c2016-08-15 10:49:06 +01003739struct i915_vma *
Chris Wilson2da3b9b2011-04-14 09:41:17 +01003740i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj,
3741 u32 alignment,
Chris Wilson59354852018-02-20 13:42:06 +00003742 const struct i915_ggtt_view *view,
3743 unsigned int flags)
Zhenyu Wangb9241ea2009-11-25 13:09:39 +08003744{
Chris Wilson058d88c2016-08-15 10:49:06 +01003745 struct i915_vma *vma;
Zhenyu Wangb9241ea2009-11-25 13:09:39 +08003746 int ret;
3747
Chris Wilson4c7d62c2016-10-28 13:58:32 +01003748 lockdep_assert_held(&obj->base.dev->struct_mutex);
3749
Chris Wilsonbd3d2252017-10-13 21:26:14 +01003750 /* Mark the global pin early so that we account for the
Chris Wilsoncc98b412013-08-09 12:25:09 +01003751 * display coherency whilst setting up the cache domains.
3752 */
Chris Wilsonbd3d2252017-10-13 21:26:14 +01003753 obj->pin_global++;
Chris Wilsoncc98b412013-08-09 12:25:09 +01003754
Eric Anholta7ef0642011-03-29 16:59:54 -07003755 /* The display engine is not coherent with the LLC cache on gen6. As
3756 * a result, we make sure that the pinning that is about to occur is
3757 * done with uncached PTEs. This is lowest common denominator for all
3758 * chipsets.
3759 *
3760 * However for gen6+, we could do better by using the GFDT bit instead
3761 * of uncaching, which would allow us to flush all the LLC-cached data
3762 * with that bit in the PTE to main memory with just one PIPE_CONTROL.
3763 */
Chris Wilson651d7942013-08-08 14:41:10 +01003764 ret = i915_gem_object_set_cache_level(obj,
Tvrtko Ursulin86527442016-10-13 11:03:00 +01003765 HAS_WT(to_i915(obj->base.dev)) ?
3766 I915_CACHE_WT : I915_CACHE_NONE);
Chris Wilson058d88c2016-08-15 10:49:06 +01003767 if (ret) {
3768 vma = ERR_PTR(ret);
Chris Wilsonbd3d2252017-10-13 21:26:14 +01003769 goto err_unpin_global;
Chris Wilson058d88c2016-08-15 10:49:06 +01003770 }
Eric Anholta7ef0642011-03-29 16:59:54 -07003771
Chris Wilson2da3b9b2011-04-14 09:41:17 +01003772 /* As the user may map the buffer once pinned in the display plane
3773 * (e.g. libkms for the bootup splash), we have to ensure that we
Chris Wilson2efb8132016-08-18 17:17:06 +01003774 * always use map_and_fenceable for all scanout buffers. However,
3775 * it may simply be too big to fit into mappable, in which case
3776 * put it anyway and hope that userspace can cope (but always first
3777 * try to preserve the existing ABI).
Chris Wilson2da3b9b2011-04-14 09:41:17 +01003778 */
Chris Wilson2efb8132016-08-18 17:17:06 +01003779 vma = ERR_PTR(-ENOSPC);
Chris Wilson59354852018-02-20 13:42:06 +00003780 if ((flags & PIN_MAPPABLE) == 0 &&
3781 (!view || view->type == I915_GGTT_VIEW_NORMAL))
Chris Wilson2efb8132016-08-18 17:17:06 +01003782 vma = i915_gem_object_ggtt_pin(obj, view, 0, alignment,
Chris Wilson59354852018-02-20 13:42:06 +00003783 flags |
3784 PIN_MAPPABLE |
3785 PIN_NONBLOCK);
3786 if (IS_ERR(vma))
Chris Wilson767a2222016-11-07 11:01:28 +00003787 vma = i915_gem_object_ggtt_pin(obj, view, 0, alignment, flags);
Chris Wilson058d88c2016-08-15 10:49:06 +01003788 if (IS_ERR(vma))
Chris Wilsonbd3d2252017-10-13 21:26:14 +01003789 goto err_unpin_global;
Chris Wilson2da3b9b2011-04-14 09:41:17 +01003790
Chris Wilsond8923dc2016-08-18 17:17:07 +01003791 vma->display_alignment = max_t(u64, vma->display_alignment, alignment);
3792
Chris Wilson5a97bcc2017-02-22 11:40:46 +00003793 __i915_gem_object_flush_for_display(obj);
Chris Wilsonb118c1e2010-05-27 13:18:14 +01003794
Chris Wilson2da3b9b2011-04-14 09:41:17 +01003795 /* It should now be out of any other write domains, and we can update
3796 * the domain values for our changes.
3797 */
Christian Königc0a51fd2018-02-16 13:43:38 +01003798 obj->read_domains |= I915_GEM_DOMAIN_GTT;
Zhenyu Wangb9241ea2009-11-25 13:09:39 +08003799
Chris Wilson058d88c2016-08-15 10:49:06 +01003800 return vma;
Chris Wilsoncc98b412013-08-09 12:25:09 +01003801
Chris Wilsonbd3d2252017-10-13 21:26:14 +01003802err_unpin_global:
3803 obj->pin_global--;
Chris Wilson058d88c2016-08-15 10:49:06 +01003804 return vma;
Chris Wilsoncc98b412013-08-09 12:25:09 +01003805}
3806
3807void
Chris Wilson058d88c2016-08-15 10:49:06 +01003808i915_gem_object_unpin_from_display_plane(struct i915_vma *vma)
Chris Wilsoncc98b412013-08-09 12:25:09 +01003809{
Chris Wilson49d73912016-11-29 09:50:08 +00003810 lockdep_assert_held(&vma->vm->i915->drm.struct_mutex);
Chris Wilson4c7d62c2016-10-28 13:58:32 +01003811
Chris Wilsonbd3d2252017-10-13 21:26:14 +01003812 if (WARN_ON(vma->obj->pin_global == 0))
Tvrtko Ursulin8a0c39b2015-04-13 11:50:09 +01003813 return;
3814
Chris Wilsonbd3d2252017-10-13 21:26:14 +01003815 if (--vma->obj->pin_global == 0)
Chris Wilsonf51455d2017-01-10 14:47:34 +00003816 vma->display_alignment = I915_GTT_MIN_ALIGNMENT;
Tvrtko Ursuline6617332015-03-23 11:10:33 +00003817
Chris Wilson383d5822016-08-18 17:17:08 +01003818 /* Bump the LRU to try and avoid premature eviction whilst flipping */
Chris Wilsonbefedbb2017-01-19 19:26:55 +00003819 i915_gem_object_bump_inactive_ggtt(vma->obj);
Chris Wilson383d5822016-08-18 17:17:08 +01003820
Chris Wilson058d88c2016-08-15 10:49:06 +01003821 i915_vma_unpin(vma);
Zhenyu Wangb9241ea2009-11-25 13:09:39 +08003822}
3823
Eric Anholte47c68e2008-11-14 13:35:19 -08003824/**
3825 * Moves a single object to the CPU read, and possibly write domain.
Tvrtko Ursulin14bb2c12016-06-03 14:02:17 +01003826 * @obj: object to act on
3827 * @write: requesting write or read-only access
Eric Anholte47c68e2008-11-14 13:35:19 -08003828 *
3829 * This function returns when the move is complete, including waiting on
3830 * flushes to occur.
3831 */
Chris Wilsondabdfe022012-03-26 10:10:27 +02003832int
Chris Wilson919926a2010-11-12 13:42:53 +00003833i915_gem_object_set_to_cpu_domain(struct drm_i915_gem_object *obj, bool write)
Eric Anholte47c68e2008-11-14 13:35:19 -08003834{
Eric Anholte47c68e2008-11-14 13:35:19 -08003835 int ret;
3836
Chris Wilsone95433c72016-10-28 13:58:27 +01003837 lockdep_assert_held(&obj->base.dev->struct_mutex);
Chris Wilson4c7d62c2016-10-28 13:58:32 +01003838
Chris Wilsone95433c72016-10-28 13:58:27 +01003839 ret = i915_gem_object_wait(obj,
3840 I915_WAIT_INTERRUPTIBLE |
3841 I915_WAIT_LOCKED |
3842 (write ? I915_WAIT_ALL : 0),
3843 MAX_SCHEDULE_TIMEOUT,
3844 NULL);
Chris Wilson88241782011-01-07 17:09:48 +00003845 if (ret)
3846 return ret;
3847
Chris Wilsonef749212017-04-12 12:01:10 +01003848 flush_write_domain(obj, ~I915_GEM_DOMAIN_CPU);
Eric Anholte47c68e2008-11-14 13:35:19 -08003849
Eric Anholte47c68e2008-11-14 13:35:19 -08003850 /* Flush the CPU cache if it's still invalid. */
Christian Königc0a51fd2018-02-16 13:43:38 +01003851 if ((obj->read_domains & I915_GEM_DOMAIN_CPU) == 0) {
Chris Wilson57822dc2017-02-22 11:40:48 +00003852 i915_gem_clflush_object(obj, I915_CLFLUSH_SYNC);
Christian Königc0a51fd2018-02-16 13:43:38 +01003853 obj->read_domains |= I915_GEM_DOMAIN_CPU;
Eric Anholte47c68e2008-11-14 13:35:19 -08003854 }
3855
3856 /* It should now be out of any other write domains, and we can update
3857 * the domain values for our changes.
3858 */
Christian Königc0a51fd2018-02-16 13:43:38 +01003859 GEM_BUG_ON(obj->write_domain & ~I915_GEM_DOMAIN_CPU);
Eric Anholte47c68e2008-11-14 13:35:19 -08003860
3861 /* If we're writing through the CPU, then the GPU read domains will
3862 * need to be invalidated at next use.
3863 */
Chris Wilsone27ab732017-06-15 13:38:49 +01003864 if (write)
3865 __start_cpu_write(obj);
Eric Anholt2ef7eea2008-11-10 10:53:25 -08003866
3867 return 0;
3868}
3869
Eric Anholt673a3942008-07-30 12:06:12 -07003870/* Throttle our rendering by waiting until the ring has completed our requests
3871 * emitted over 20 msec ago.
3872 *
Eric Anholtb9624422009-06-03 07:27:35 +00003873 * Note that if we were to use the current jiffies each time around the loop,
3874 * we wouldn't escape the function with any frames outstanding if the time to
3875 * render a frame was over 20ms.
3876 *
Eric Anholt673a3942008-07-30 12:06:12 -07003877 * This should get us reasonable parallelism between CPU and GPU but also
3878 * relatively low latency when blocking on a particular request to finish.
3879 */
3880static int
Chris Wilsonf787a5f2010-09-24 16:02:42 +01003881i915_gem_ring_throttle(struct drm_device *dev, struct drm_file *file)
Eric Anholt673a3942008-07-30 12:06:12 -07003882{
Chris Wilsonfac5e232016-07-04 11:34:36 +01003883 struct drm_i915_private *dev_priv = to_i915(dev);
Chris Wilsonf787a5f2010-09-24 16:02:42 +01003884 struct drm_i915_file_private *file_priv = file->driver_priv;
Chris Wilsond0bc54f2015-05-21 21:01:48 +01003885 unsigned long recent_enough = jiffies - DRM_I915_THROTTLE_JIFFIES;
Chris Wilsone61e0f52018-02-21 09:56:36 +00003886 struct i915_request *request, *target = NULL;
Chris Wilsone95433c72016-10-28 13:58:27 +01003887 long ret;
Eric Anholt673a3942008-07-30 12:06:12 -07003888
Chris Wilsonf4457ae2016-04-13 17:35:08 +01003889 /* ABI: return -EIO if already wedged */
3890 if (i915_terminally_wedged(&dev_priv->gpu_error))
3891 return -EIO;
Chris Wilsone110e8d2011-01-26 15:39:14 +00003892
Chris Wilson1c255952010-09-26 11:03:27 +01003893 spin_lock(&file_priv->mm.lock);
Chris Wilsonc8659ef2017-03-02 12:25:25 +00003894 list_for_each_entry(request, &file_priv->mm.request_list, client_link) {
Eric Anholtb9624422009-06-03 07:27:35 +00003895 if (time_after_eq(request->emitted_jiffies, recent_enough))
3896 break;
3897
Chris Wilsonc8659ef2017-03-02 12:25:25 +00003898 if (target) {
3899 list_del(&target->client_link);
3900 target->file_priv = NULL;
3901 }
John Harrisonfcfa423c2015-05-29 17:44:12 +01003902
John Harrison54fb2412014-11-24 18:49:27 +00003903 target = request;
Eric Anholtb9624422009-06-03 07:27:35 +00003904 }
John Harrisonff865882014-11-24 18:49:28 +00003905 if (target)
Chris Wilsone61e0f52018-02-21 09:56:36 +00003906 i915_request_get(target);
Chris Wilson1c255952010-09-26 11:03:27 +01003907 spin_unlock(&file_priv->mm.lock);
Chris Wilsonf787a5f2010-09-24 16:02:42 +01003908
John Harrison54fb2412014-11-24 18:49:27 +00003909 if (target == NULL)
Chris Wilsonf787a5f2010-09-24 16:02:42 +01003910 return 0;
3911
Chris Wilsone61e0f52018-02-21 09:56:36 +00003912 ret = i915_request_wait(target,
Chris Wilsone95433c72016-10-28 13:58:27 +01003913 I915_WAIT_INTERRUPTIBLE,
3914 MAX_SCHEDULE_TIMEOUT);
Chris Wilsone61e0f52018-02-21 09:56:36 +00003915 i915_request_put(target);
John Harrisonff865882014-11-24 18:49:28 +00003916
Chris Wilsone95433c72016-10-28 13:58:27 +01003917 return ret < 0 ? ret : 0;
Eric Anholt673a3942008-07-30 12:06:12 -07003918}
3919
Chris Wilson058d88c2016-08-15 10:49:06 +01003920struct i915_vma *
Joonas Lahtinenec7adb62015-03-16 14:11:13 +02003921i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
3922 const struct i915_ggtt_view *view,
Chris Wilson91b2db62016-08-04 16:32:23 +01003923 u64 size,
Chris Wilson2ffffd02016-08-04 16:32:22 +01003924 u64 alignment,
3925 u64 flags)
Joonas Lahtinenec7adb62015-03-16 14:11:13 +02003926{
Chris Wilsonad16d2e2016-10-13 09:55:04 +01003927 struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
Chris Wilson82ad6442018-06-05 16:37:58 +01003928 struct i915_address_space *vm = &dev_priv->ggtt.vm;
Chris Wilson59bfa122016-08-04 16:32:31 +01003929 struct i915_vma *vma;
3930 int ret;
Joonas Lahtinen72e96d642016-03-30 16:57:10 +03003931
Chris Wilson4c7d62c2016-10-28 13:58:32 +01003932 lockdep_assert_held(&obj->base.dev->struct_mutex);
3933
Chris Wilsonac87a6fd2018-02-20 13:42:05 +00003934 if (flags & PIN_MAPPABLE &&
3935 (!view || view->type == I915_GGTT_VIEW_NORMAL)) {
Chris Wilson43ae70d92017-10-09 09:44:01 +01003936 /* If the required space is larger than the available
3937 * aperture, we will not able to find a slot for the
3938 * object and unbinding the object now will be in
3939 * vain. Worse, doing so may cause us to ping-pong
3940 * the object in and out of the Global GTT and
3941 * waste a lot of cycles under the mutex.
3942 */
3943 if (obj->base.size > dev_priv->ggtt.mappable_end)
3944 return ERR_PTR(-E2BIG);
3945
3946 /* If NONBLOCK is set the caller is optimistically
3947 * trying to cache the full object within the mappable
3948 * aperture, and *must* have a fallback in place for
3949 * situations where we cannot bind the object. We
3950 * can be a little more lax here and use the fallback
3951 * more often to avoid costly migrations of ourselves
3952 * and other objects within the aperture.
3953 *
3954 * Half-the-aperture is used as a simple heuristic.
3955 * More interesting would to do search for a free
3956 * block prior to making the commitment to unbind.
3957 * That caters for the self-harm case, and with a
3958 * little more heuristics (e.g. NOFAULT, NOEVICT)
3959 * we could try to minimise harm to others.
3960 */
3961 if (flags & PIN_NONBLOCK &&
3962 obj->base.size > dev_priv->ggtt.mappable_end / 2)
3963 return ERR_PTR(-ENOSPC);
3964 }
3965
Chris Wilson718659a2017-01-16 15:21:28 +00003966 vma = i915_vma_instance(obj, vm, view);
Chris Wilsone0216b72017-01-19 19:26:57 +00003967 if (unlikely(IS_ERR(vma)))
Chris Wilson058d88c2016-08-15 10:49:06 +01003968 return vma;
Chris Wilson59bfa122016-08-04 16:32:31 +01003969
3970 if (i915_vma_misplaced(vma, size, alignment, flags)) {
Chris Wilson43ae70d92017-10-09 09:44:01 +01003971 if (flags & PIN_NONBLOCK) {
3972 if (i915_vma_is_pinned(vma) || i915_vma_is_active(vma))
3973 return ERR_PTR(-ENOSPC);
Chris Wilson59bfa122016-08-04 16:32:31 +01003974
Chris Wilson43ae70d92017-10-09 09:44:01 +01003975 if (flags & PIN_MAPPABLE &&
Chris Wilson944397f2017-01-09 16:16:11 +00003976 vma->fence_size > dev_priv->ggtt.mappable_end / 2)
Chris Wilsonad16d2e2016-10-13 09:55:04 +01003977 return ERR_PTR(-ENOSPC);
3978 }
3979
Chris Wilson59bfa122016-08-04 16:32:31 +01003980 WARN(i915_vma_is_pinned(vma),
3981 "bo is already pinned in ggtt with incorrect alignment:"
Chris Wilson05a20d02016-08-18 17:16:55 +01003982 " offset=%08x, req.alignment=%llx,"
3983 " req.map_and_fenceable=%d, vma->map_and_fenceable=%d\n",
3984 i915_ggtt_offset(vma), alignment,
Chris Wilson59bfa122016-08-04 16:32:31 +01003985 !!(flags & PIN_MAPPABLE),
Chris Wilson05a20d02016-08-18 17:16:55 +01003986 i915_vma_is_map_and_fenceable(vma));
Chris Wilson59bfa122016-08-04 16:32:31 +01003987 ret = i915_vma_unbind(vma);
3988 if (ret)
Chris Wilson058d88c2016-08-15 10:49:06 +01003989 return ERR_PTR(ret);
Chris Wilson59bfa122016-08-04 16:32:31 +01003990 }
3991
Chris Wilson058d88c2016-08-15 10:49:06 +01003992 ret = i915_vma_pin(vma, size, alignment, flags | PIN_GLOBAL);
3993 if (ret)
3994 return ERR_PTR(ret);
Joonas Lahtinenec7adb62015-03-16 14:11:13 +02003995
Chris Wilson058d88c2016-08-15 10:49:06 +01003996 return vma;
Eric Anholt673a3942008-07-30 12:06:12 -07003997}
3998
Chris Wilsonedf6b762016-08-09 09:23:33 +01003999static __always_inline unsigned int __busy_read_flag(unsigned int id)
Chris Wilson3fdc13c2016-08-05 10:14:18 +01004000{
4001 /* Note that we could alias engines in the execbuf API, but
4002 * that would be very unwise as it prevents userspace from
4003 * fine control over engine selection. Ahem.
4004 *
4005 * This should be something like EXEC_MAX_ENGINE instead of
4006 * I915_NUM_ENGINES.
4007 */
4008 BUILD_BUG_ON(I915_NUM_ENGINES > 16);
4009 return 0x10000 << id;
4010}
4011
4012static __always_inline unsigned int __busy_write_id(unsigned int id)
4013{
Chris Wilson70cb4722016-08-09 18:08:25 +01004014 /* The uABI guarantees an active writer is also amongst the read
4015 * engines. This would be true if we accessed the activity tracking
4016 * under the lock, but as we perform the lookup of the object and
4017 * its activity locklessly we can not guarantee that the last_write
4018 * being active implies that we have set the same engine flag from
4019 * last_read - hence we always set both read and write busy for
4020 * last_write.
4021 */
4022 return id | __busy_read_flag(id);
Chris Wilson3fdc13c2016-08-05 10:14:18 +01004023}
4024
Chris Wilsonedf6b762016-08-09 09:23:33 +01004025static __always_inline unsigned int
Chris Wilsond07f0e52016-10-28 13:58:44 +01004026__busy_set_if_active(const struct dma_fence *fence,
Chris Wilson3fdc13c2016-08-05 10:14:18 +01004027 unsigned int (*flag)(unsigned int id))
4028{
Chris Wilsone61e0f52018-02-21 09:56:36 +00004029 struct i915_request *rq;
Chris Wilson12555012016-08-16 09:50:40 +01004030
Chris Wilsond07f0e52016-10-28 13:58:44 +01004031 /* We have to check the current hw status of the fence as the uABI
4032 * guarantees forward progress. We could rely on the idle worker
4033 * to eventually flush us, but to minimise latency just ask the
4034 * hardware.
4035 *
4036 * Note we only report on the status of native fences.
4037 */
4038 if (!dma_fence_is_i915(fence))
Chris Wilson12555012016-08-16 09:50:40 +01004039 return 0;
4040
Chris Wilsond07f0e52016-10-28 13:58:44 +01004041 /* opencode to_request() in order to avoid const warnings */
Chris Wilsone61e0f52018-02-21 09:56:36 +00004042 rq = container_of(fence, struct i915_request, fence);
4043 if (i915_request_completed(rq))
Chris Wilsond07f0e52016-10-28 13:58:44 +01004044 return 0;
4045
Chris Wilson1d39f282017-04-11 13:43:06 +01004046 return flag(rq->engine->uabi_id);
Chris Wilson3fdc13c2016-08-05 10:14:18 +01004047}
4048
Chris Wilsonedf6b762016-08-09 09:23:33 +01004049static __always_inline unsigned int
Chris Wilsond07f0e52016-10-28 13:58:44 +01004050busy_check_reader(const struct dma_fence *fence)
Chris Wilson3fdc13c2016-08-05 10:14:18 +01004051{
Chris Wilsond07f0e52016-10-28 13:58:44 +01004052 return __busy_set_if_active(fence, __busy_read_flag);
Chris Wilson3fdc13c2016-08-05 10:14:18 +01004053}
4054
Chris Wilsonedf6b762016-08-09 09:23:33 +01004055static __always_inline unsigned int
Chris Wilsond07f0e52016-10-28 13:58:44 +01004056busy_check_writer(const struct dma_fence *fence)
Chris Wilson3fdc13c2016-08-05 10:14:18 +01004057{
Chris Wilsond07f0e52016-10-28 13:58:44 +01004058 if (!fence)
4059 return 0;
4060
4061 return __busy_set_if_active(fence, __busy_write_id);
Chris Wilson3fdc13c2016-08-05 10:14:18 +01004062}
4063
Eric Anholt673a3942008-07-30 12:06:12 -07004064int
Eric Anholt673a3942008-07-30 12:06:12 -07004065i915_gem_busy_ioctl(struct drm_device *dev, void *data,
Chris Wilson05394f32010-11-08 19:18:58 +00004066 struct drm_file *file)
Eric Anholt673a3942008-07-30 12:06:12 -07004067{
4068 struct drm_i915_gem_busy *args = data;
Chris Wilson05394f32010-11-08 19:18:58 +00004069 struct drm_i915_gem_object *obj;
Chris Wilsond07f0e52016-10-28 13:58:44 +01004070 struct reservation_object_list *list;
4071 unsigned int seq;
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004072 int err;
Eric Anholt673a3942008-07-30 12:06:12 -07004073
Chris Wilsond07f0e52016-10-28 13:58:44 +01004074 err = -ENOENT;
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004075 rcu_read_lock();
4076 obj = i915_gem_object_lookup_rcu(file, args->handle);
Chris Wilsond07f0e52016-10-28 13:58:44 +01004077 if (!obj)
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004078 goto out;
Chris Wilsond07f0e52016-10-28 13:58:44 +01004079
4080 /* A discrepancy here is that we do not report the status of
4081 * non-i915 fences, i.e. even though we may report the object as idle,
4082 * a call to set-domain may still stall waiting for foreign rendering.
4083 * This also means that wait-ioctl may report an object as busy,
4084 * where busy-ioctl considers it idle.
4085 *
4086 * We trade the ability to warn of foreign fences to report on which
4087 * i915 engines are active for the object.
4088 *
4089 * Alternatively, we can trade that extra information on read/write
4090 * activity with
4091 * args->busy =
4092 * !reservation_object_test_signaled_rcu(obj->resv, true);
4093 * to report the overall busyness. This is what the wait-ioctl does.
4094 *
4095 */
4096retry:
4097 seq = raw_read_seqcount(&obj->resv->seq);
4098
4099 /* Translate the exclusive fence to the READ *and* WRITE engine */
4100 args->busy = busy_check_writer(rcu_dereference(obj->resv->fence_excl));
4101
4102 /* Translate shared fences to READ set of engines */
4103 list = rcu_dereference(obj->resv->fence);
4104 if (list) {
4105 unsigned int shared_count = list->shared_count, i;
4106
4107 for (i = 0; i < shared_count; ++i) {
4108 struct dma_fence *fence =
4109 rcu_dereference(list->shared[i]);
4110
4111 args->busy |= busy_check_reader(fence);
4112 }
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004113 }
Zou Nan haid1b851f2010-05-21 09:08:57 +08004114
Chris Wilsond07f0e52016-10-28 13:58:44 +01004115 if (args->busy && read_seqcount_retry(&obj->resv->seq, seq))
4116 goto retry;
Chris Wilson426960b2016-01-15 16:51:46 +00004117
Chris Wilsond07f0e52016-10-28 13:58:44 +01004118 err = 0;
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004119out:
4120 rcu_read_unlock();
4121 return err;
Eric Anholt673a3942008-07-30 12:06:12 -07004122}
4123
4124int
4125i915_gem_throttle_ioctl(struct drm_device *dev, void *data,
4126 struct drm_file *file_priv)
4127{
Akshay Joshi0206e352011-08-16 15:34:10 -04004128 return i915_gem_ring_throttle(dev, file_priv);
Eric Anholt673a3942008-07-30 12:06:12 -07004129}
4130
Chris Wilson3ef94da2009-09-14 16:50:29 +01004131int
4132i915_gem_madvise_ioctl(struct drm_device *dev, void *data,
4133 struct drm_file *file_priv)
4134{
Chris Wilsonfac5e232016-07-04 11:34:36 +01004135 struct drm_i915_private *dev_priv = to_i915(dev);
Chris Wilson3ef94da2009-09-14 16:50:29 +01004136 struct drm_i915_gem_madvise *args = data;
Chris Wilson05394f32010-11-08 19:18:58 +00004137 struct drm_i915_gem_object *obj;
Chris Wilson1233e2d2016-10-28 13:58:37 +01004138 int err;
Chris Wilson3ef94da2009-09-14 16:50:29 +01004139
4140 switch (args->madv) {
4141 case I915_MADV_DONTNEED:
4142 case I915_MADV_WILLNEED:
4143 break;
4144 default:
4145 return -EINVAL;
4146 }
4147
Chris Wilson03ac0642016-07-20 13:31:51 +01004148 obj = i915_gem_object_lookup(file_priv, args->handle);
Chris Wilson1233e2d2016-10-28 13:58:37 +01004149 if (!obj)
4150 return -ENOENT;
4151
4152 err = mutex_lock_interruptible(&obj->mm.lock);
4153 if (err)
4154 goto out;
Chris Wilson3ef94da2009-09-14 16:50:29 +01004155
Chris Wilsonf1fa4f42017-10-13 21:26:13 +01004156 if (i915_gem_object_has_pages(obj) &&
Chris Wilson3e510a82016-08-05 10:14:23 +01004157 i915_gem_object_is_tiled(obj) &&
Daniel Vetter656bfa32014-11-20 09:26:30 +01004158 dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES) {
Chris Wilsonbc0629a2016-11-01 10:03:17 +00004159 if (obj->mm.madv == I915_MADV_WILLNEED) {
4160 GEM_BUG_ON(!obj->mm.quirked);
Chris Wilsona4f5ea62016-10-28 13:58:35 +01004161 __i915_gem_object_unpin_pages(obj);
Chris Wilsonbc0629a2016-11-01 10:03:17 +00004162 obj->mm.quirked = false;
4163 }
4164 if (args->madv == I915_MADV_WILLNEED) {
Chris Wilson2c3a3f42016-11-04 10:30:01 +00004165 GEM_BUG_ON(obj->mm.quirked);
Chris Wilsona4f5ea62016-10-28 13:58:35 +01004166 __i915_gem_object_pin_pages(obj);
Chris Wilsonbc0629a2016-11-01 10:03:17 +00004167 obj->mm.quirked = true;
4168 }
Daniel Vetter656bfa32014-11-20 09:26:30 +01004169 }
4170
Chris Wilsona4f5ea62016-10-28 13:58:35 +01004171 if (obj->mm.madv != __I915_MADV_PURGED)
4172 obj->mm.madv = args->madv;
Chris Wilson3ef94da2009-09-14 16:50:29 +01004173
Chris Wilson6c085a72012-08-20 11:40:46 +02004174 /* if the object is no longer attached, discard its backing storage */
Chris Wilsonf1fa4f42017-10-13 21:26:13 +01004175 if (obj->mm.madv == I915_MADV_DONTNEED &&
4176 !i915_gem_object_has_pages(obj))
Chris Wilson2d7ef392009-09-20 23:13:10 +01004177 i915_gem_object_truncate(obj);
4178
Chris Wilsona4f5ea62016-10-28 13:58:35 +01004179 args->retained = obj->mm.madv != __I915_MADV_PURGED;
Chris Wilson1233e2d2016-10-28 13:58:37 +01004180 mutex_unlock(&obj->mm.lock);
Chris Wilsonbb6baf72009-09-22 14:24:13 +01004181
Chris Wilson1233e2d2016-10-28 13:58:37 +01004182out:
Chris Wilsonf8c417c2016-07-20 13:31:53 +01004183 i915_gem_object_put(obj);
Chris Wilson1233e2d2016-10-28 13:58:37 +01004184 return err;
Chris Wilson3ef94da2009-09-14 16:50:29 +01004185}
4186
Chris Wilson5b8c8ae2016-11-16 19:07:04 +00004187static void
Chris Wilson21950ee2019-02-05 13:00:05 +00004188frontbuffer_retire(struct i915_active_request *active,
4189 struct i915_request *request)
Chris Wilson5b8c8ae2016-11-16 19:07:04 +00004190{
4191 struct drm_i915_gem_object *obj =
4192 container_of(active, typeof(*obj), frontbuffer_write);
4193
Chris Wilsond59b21e2017-02-22 11:40:49 +00004194 intel_fb_obj_flush(obj, ORIGIN_CS);
Chris Wilson5b8c8ae2016-11-16 19:07:04 +00004195}
4196
Chris Wilson37e680a2012-06-07 15:38:42 +01004197void i915_gem_object_init(struct drm_i915_gem_object *obj,
4198 const struct drm_i915_gem_object_ops *ops)
Chris Wilson0327d6b2012-08-11 15:41:06 +01004199{
Chris Wilson1233e2d2016-10-28 13:58:37 +01004200 mutex_init(&obj->mm.lock);
4201
Chris Wilson528cbd12019-01-28 10:23:54 +00004202 spin_lock_init(&obj->vma.lock);
4203 INIT_LIST_HEAD(&obj->vma.list);
4204
Chris Wilsond1b48c12017-08-16 09:52:08 +01004205 INIT_LIST_HEAD(&obj->lut_list);
Chris Wilson8d9d5742015-04-07 16:20:38 +01004206 INIT_LIST_HEAD(&obj->batch_pool_link);
Chris Wilson0327d6b2012-08-11 15:41:06 +01004207
Chris Wilson8811d612018-11-09 09:03:11 +00004208 init_rcu_head(&obj->rcu);
4209
Chris Wilson37e680a2012-06-07 15:38:42 +01004210 obj->ops = ops;
4211
Chris Wilsond07f0e52016-10-28 13:58:44 +01004212 reservation_object_init(&obj->__builtin_resv);
4213 obj->resv = &obj->__builtin_resv;
4214
Chris Wilson50349242016-08-18 17:17:04 +01004215 obj->frontbuffer_ggtt_origin = ORIGIN_GTT;
Chris Wilson21950ee2019-02-05 13:00:05 +00004216 i915_active_request_init(&obj->frontbuffer_write,
4217 NULL, frontbuffer_retire);
Chris Wilsona4f5ea62016-10-28 13:58:35 +01004218
4219 obj->mm.madv = I915_MADV_WILLNEED;
4220 INIT_RADIX_TREE(&obj->mm.get_page.radix, GFP_KERNEL | __GFP_NOWARN);
4221 mutex_init(&obj->mm.get_page.lock);
Chris Wilson0327d6b2012-08-11 15:41:06 +01004222
Dave Gordonf19ec8c2016-07-04 11:34:37 +01004223 i915_gem_info_add_obj(to_i915(obj->base.dev), obj->base.size);
Chris Wilson0327d6b2012-08-11 15:41:06 +01004224}
4225
Chris Wilson37e680a2012-06-07 15:38:42 +01004226static const struct drm_i915_gem_object_ops i915_gem_object_ops = {
Tvrtko Ursulin3599a912016-11-01 14:44:10 +00004227 .flags = I915_GEM_OBJECT_HAS_STRUCT_PAGE |
4228 I915_GEM_OBJECT_IS_SHRINKABLE,
Chris Wilson7c55e2c2017-03-07 12:03:38 +00004229
Chris Wilson37e680a2012-06-07 15:38:42 +01004230 .get_pages = i915_gem_object_get_pages_gtt,
4231 .put_pages = i915_gem_object_put_pages_gtt,
Chris Wilson7c55e2c2017-03-07 12:03:38 +00004232
4233 .pwrite = i915_gem_object_pwrite_gtt,
Chris Wilson37e680a2012-06-07 15:38:42 +01004234};
4235
Matthew Auld465c4032017-10-06 23:18:14 +01004236static int i915_gem_object_create_shmem(struct drm_device *dev,
4237 struct drm_gem_object *obj,
4238 size_t size)
4239{
4240 struct drm_i915_private *i915 = to_i915(dev);
4241 unsigned long flags = VM_NORESERVE;
4242 struct file *filp;
4243
4244 drm_gem_private_object_init(dev, obj, size);
4245
4246 if (i915->mm.gemfs)
4247 filp = shmem_file_setup_with_mnt(i915->mm.gemfs, "i915", size,
4248 flags);
4249 else
4250 filp = shmem_file_setup("i915", size, flags);
4251
4252 if (IS_ERR(filp))
4253 return PTR_ERR(filp);
4254
4255 obj->filp = filp;
4256
4257 return 0;
4258}
4259
Chris Wilsonb4bcbe22016-10-18 13:02:49 +01004260struct drm_i915_gem_object *
Tvrtko Ursulin12d79d72016-12-01 14:16:37 +00004261i915_gem_object_create(struct drm_i915_private *dev_priv, u64 size)
Daniel Vetterac52bc52010-04-09 19:05:06 +00004262{
Daniel Vetterc397b902010-04-09 19:05:07 +00004263 struct drm_i915_gem_object *obj;
Hugh Dickins5949eac2011-06-27 16:18:18 -07004264 struct address_space *mapping;
Chris Wilsonb8f55be2017-08-11 12:11:16 +01004265 unsigned int cache_level;
Daniel Vetter1a240d42012-11-29 22:18:51 +01004266 gfp_t mask;
Chris Wilsonfe3db792016-04-25 13:32:13 +01004267 int ret;
Daniel Vetterc397b902010-04-09 19:05:07 +00004268
Chris Wilsonb4bcbe22016-10-18 13:02:49 +01004269 /* There is a prevalence of the assumption that we fit the object's
4270 * page count inside a 32bit _signed_ variable. Let's document this and
4271 * catch if we ever need to fix it. In the meantime, if you do spot
4272 * such a local variable, please consider fixing!
4273 */
Tvrtko Ursulin7a3ee5d2017-03-30 17:31:30 +01004274 if (size >> PAGE_SHIFT > INT_MAX)
Chris Wilsonb4bcbe22016-10-18 13:02:49 +01004275 return ERR_PTR(-E2BIG);
4276
4277 if (overflows_type(size, obj->base.size))
4278 return ERR_PTR(-E2BIG);
4279
Tvrtko Ursulin187685c2016-12-01 14:16:36 +00004280 obj = i915_gem_object_alloc(dev_priv);
Daniel Vetterc397b902010-04-09 19:05:07 +00004281 if (obj == NULL)
Chris Wilsonfe3db792016-04-25 13:32:13 +01004282 return ERR_PTR(-ENOMEM);
Daniel Vetterc397b902010-04-09 19:05:07 +00004283
Matthew Auld465c4032017-10-06 23:18:14 +01004284 ret = i915_gem_object_create_shmem(&dev_priv->drm, &obj->base, size);
Chris Wilsonfe3db792016-04-25 13:32:13 +01004285 if (ret)
4286 goto fail;
Daniel Vetterc397b902010-04-09 19:05:07 +00004287
Chris Wilsonbed1ea92012-05-24 20:48:12 +01004288 mask = GFP_HIGHUSER | __GFP_RECLAIMABLE;
Jani Nikulac0f86832016-12-07 12:13:04 +02004289 if (IS_I965GM(dev_priv) || IS_I965G(dev_priv)) {
Chris Wilsonbed1ea92012-05-24 20:48:12 +01004290 /* 965gm cannot relocate objects above 4GiB. */
4291 mask &= ~__GFP_HIGHMEM;
4292 mask |= __GFP_DMA32;
4293 }
4294
Al Viro93c76a32015-12-04 23:45:44 -05004295 mapping = obj->base.filp->f_mapping;
Chris Wilsonbed1ea92012-05-24 20:48:12 +01004296 mapping_set_gfp_mask(mapping, mask);
Chris Wilson4846bf02017-06-09 12:03:46 +01004297 GEM_BUG_ON(!(mapping_gfp_mask(mapping) & __GFP_RECLAIM));
Hugh Dickins5949eac2011-06-27 16:18:18 -07004298
Chris Wilson37e680a2012-06-07 15:38:42 +01004299 i915_gem_object_init(obj, &i915_gem_object_ops);
Chris Wilson73aa8082010-09-30 11:46:12 +01004300
Christian Königc0a51fd2018-02-16 13:43:38 +01004301 obj->write_domain = I915_GEM_DOMAIN_CPU;
4302 obj->read_domains = I915_GEM_DOMAIN_CPU;
Daniel Vetterc397b902010-04-09 19:05:07 +00004303
Chris Wilsonb8f55be2017-08-11 12:11:16 +01004304 if (HAS_LLC(dev_priv))
Eugeni Dodonov3d29b842012-01-17 14:43:53 -02004305 /* On some devices, we can have the GPU use the LLC (the CPU
Eric Anholta1871112011-03-29 16:59:55 -07004306 * cache) for about a 10% performance improvement
4307 * compared to uncached. Graphics requests other than
4308 * display scanout are coherent with the CPU in
4309 * accessing this cache. This means in this mode we
4310 * don't need to clflush on the CPU side, and on the
4311 * GPU side we only need to flush internal caches to
4312 * get data visible to the CPU.
4313 *
4314 * However, we maintain the display planes as UC, and so
4315 * need to rebind when first used as such.
4316 */
Chris Wilsonb8f55be2017-08-11 12:11:16 +01004317 cache_level = I915_CACHE_LLC;
4318 else
4319 cache_level = I915_CACHE_NONE;
Eric Anholta1871112011-03-29 16:59:55 -07004320
Chris Wilsonb8f55be2017-08-11 12:11:16 +01004321 i915_gem_object_set_cache_coherency(obj, cache_level);
Chris Wilsone27ab732017-06-15 13:38:49 +01004322
Daniel Vetterd861e332013-07-24 23:25:03 +02004323 trace_i915_gem_object_create(obj);
4324
Chris Wilson05394f32010-11-08 19:18:58 +00004325 return obj;
Chris Wilsonfe3db792016-04-25 13:32:13 +01004326
4327fail:
4328 i915_gem_object_free(obj);
Chris Wilsonfe3db792016-04-25 13:32:13 +01004329 return ERR_PTR(ret);
Daniel Vetterac52bc52010-04-09 19:05:06 +00004330}
4331
Chris Wilson340fbd82014-05-22 09:16:52 +01004332static bool discard_backing_storage(struct drm_i915_gem_object *obj)
4333{
4334 /* If we are the last user of the backing storage (be it shmemfs
4335 * pages or stolen etc), we know that the pages are going to be
4336 * immediately released. In this case, we can then skip copying
4337 * back the contents from the GPU.
4338 */
4339
Chris Wilsona4f5ea62016-10-28 13:58:35 +01004340 if (obj->mm.madv != I915_MADV_WILLNEED)
Chris Wilson340fbd82014-05-22 09:16:52 +01004341 return false;
4342
4343 if (obj->base.filp == NULL)
4344 return true;
4345
4346 /* At first glance, this looks racy, but then again so would be
4347 * userspace racing mmap against close. However, the first external
4348 * reference to the filp can only be obtained through the
4349 * i915_gem_mmap_ioctl() which safeguards us against the user
4350 * acquiring such a reference whilst we are in the middle of
4351 * freeing the object.
4352 */
4353 return atomic_long_read(&obj->base.filp->f_count) == 1;
4354}
4355
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004356static void __i915_gem_free_objects(struct drm_i915_private *i915,
4357 struct llist_node *freed)
Chris Wilsonbe726152010-07-23 23:18:50 +01004358{
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004359 struct drm_i915_gem_object *obj, *on;
Chris Wilson538ef962019-01-14 14:21:18 +00004360 intel_wakeref_t wakeref;
Chris Wilsonbe726152010-07-23 23:18:50 +01004361
Chris Wilson538ef962019-01-14 14:21:18 +00004362 wakeref = intel_runtime_pm_get(i915);
Chris Wilsoncc731f52017-10-13 21:26:21 +01004363 llist_for_each_entry_safe(obj, on, freed, freed) {
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004364 struct i915_vma *vma, *vn;
Paulo Zanonif65c9162013-11-27 18:20:34 -02004365
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004366 trace_i915_gem_object_destroy(obj);
4367
Chris Wilsoncc731f52017-10-13 21:26:21 +01004368 mutex_lock(&i915->drm.struct_mutex);
4369
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004370 GEM_BUG_ON(i915_gem_object_is_active(obj));
Chris Wilson528cbd12019-01-28 10:23:54 +00004371 list_for_each_entry_safe(vma, vn, &obj->vma.list, obj_link) {
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004372 GEM_BUG_ON(i915_vma_is_active(vma));
4373 vma->flags &= ~I915_VMA_PIN_MASK;
Chris Wilson3365e222018-05-03 20:51:14 +01004374 i915_vma_destroy(vma);
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004375 }
Chris Wilson528cbd12019-01-28 10:23:54 +00004376 GEM_BUG_ON(!list_empty(&obj->vma.list));
4377 GEM_BUG_ON(!RB_EMPTY_ROOT(&obj->vma.tree));
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004378
Chris Wilsonf2123812017-10-16 12:40:37 +01004379 /* This serializes freeing with the shrinker. Since the free
4380 * is delayed, first by RCU then by the workqueue, we want the
4381 * shrinker to be able to free pages of unreferenced objects,
4382 * or else we may oom whilst there are plenty of deferred
4383 * freed objects.
4384 */
4385 if (i915_gem_object_has_pages(obj)) {
4386 spin_lock(&i915->mm.obj_lock);
4387 list_del_init(&obj->mm.link);
4388 spin_unlock(&i915->mm.obj_lock);
4389 }
4390
Chris Wilsoncc731f52017-10-13 21:26:21 +01004391 mutex_unlock(&i915->drm.struct_mutex);
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004392
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004393 GEM_BUG_ON(obj->bind_count);
Chris Wilsona65adaf2017-10-09 09:43:57 +01004394 GEM_BUG_ON(obj->userfault_count);
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004395 GEM_BUG_ON(atomic_read(&obj->frontbuffer_bits));
Chris Wilson67b48042017-08-22 12:05:16 +01004396 GEM_BUG_ON(!list_empty(&obj->lut_list));
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004397
4398 if (obj->ops->release)
4399 obj->ops->release(obj);
4400
4401 if (WARN_ON(i915_gem_object_has_pinned_pages(obj)))
4402 atomic_set(&obj->mm.pages_pin_count, 0);
Chris Wilson548625e2016-11-01 12:11:34 +00004403 __i915_gem_object_put_pages(obj, I915_MM_NORMAL);
Chris Wilsonf1fa4f42017-10-13 21:26:13 +01004404 GEM_BUG_ON(i915_gem_object_has_pages(obj));
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004405
4406 if (obj->base.import_attach)
4407 drm_prime_gem_destroy(&obj->base, NULL);
4408
Chris Wilsond07f0e52016-10-28 13:58:44 +01004409 reservation_object_fini(&obj->__builtin_resv);
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004410 drm_gem_object_release(&obj->base);
4411 i915_gem_info_remove_obj(i915, obj->base.size);
4412
4413 kfree(obj->bit_17);
4414 i915_gem_object_free(obj);
Chris Wilsoncc731f52017-10-13 21:26:21 +01004415
Chris Wilsonc9c704712018-02-19 22:06:31 +00004416 GEM_BUG_ON(!atomic_read(&i915->mm.free_count));
4417 atomic_dec(&i915->mm.free_count);
4418
Chris Wilsoncc731f52017-10-13 21:26:21 +01004419 if (on)
4420 cond_resched();
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004421 }
Chris Wilson538ef962019-01-14 14:21:18 +00004422 intel_runtime_pm_put(i915, wakeref);
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004423}
4424
4425static void i915_gem_flush_free_objects(struct drm_i915_private *i915)
4426{
4427 struct llist_node *freed;
4428
Chris Wilson87701b42017-10-13 21:26:20 +01004429 /* Free the oldest, most stale object to keep the free_list short */
4430 freed = NULL;
4431 if (!llist_empty(&i915->mm.free_list)) { /* quick test for hotpath */
4432 /* Only one consumer of llist_del_first() allowed */
4433 spin_lock(&i915->mm.free_lock);
4434 freed = llist_del_first(&i915->mm.free_list);
4435 spin_unlock(&i915->mm.free_lock);
4436 }
4437 if (unlikely(freed)) {
4438 freed->next = NULL;
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004439 __i915_gem_free_objects(i915, freed);
Chris Wilson87701b42017-10-13 21:26:20 +01004440 }
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004441}
4442
4443static void __i915_gem_free_work(struct work_struct *work)
4444{
4445 struct drm_i915_private *i915 =
4446 container_of(work, struct drm_i915_private, mm.free_work);
4447 struct llist_node *freed;
Chris Wilson26e12f82011-03-20 11:20:19 +00004448
Chris Wilson2ef1e722018-01-15 20:57:59 +00004449 /*
4450 * All file-owned VMA should have been released by this point through
Chris Wilsonb1f788c2016-08-04 07:52:45 +01004451 * i915_gem_close_object(), or earlier by i915_gem_context_close().
4452 * However, the object may also be bound into the global GTT (e.g.
4453 * older GPUs without per-process support, or for direct access through
4454 * the GTT either for the user or for scanout). Those VMA still need to
4455 * unbound now.
4456 */
Chris Wilson1488fc02012-04-24 15:47:31 +01004457
Chris Wilsonf991c492017-11-06 11:15:08 +00004458 spin_lock(&i915->mm.free_lock);
Chris Wilson5ad08be2017-04-07 11:25:51 +01004459 while ((freed = llist_del_all(&i915->mm.free_list))) {
Chris Wilsonf991c492017-11-06 11:15:08 +00004460 spin_unlock(&i915->mm.free_lock);
4461
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004462 __i915_gem_free_objects(i915, freed);
Chris Wilson5ad08be2017-04-07 11:25:51 +01004463 if (need_resched())
Chris Wilsonf991c492017-11-06 11:15:08 +00004464 return;
4465
4466 spin_lock(&i915->mm.free_lock);
Chris Wilson5ad08be2017-04-07 11:25:51 +01004467 }
Chris Wilsonf991c492017-11-06 11:15:08 +00004468 spin_unlock(&i915->mm.free_lock);
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004469}
4470
4471static void __i915_gem_free_object_rcu(struct rcu_head *head)
4472{
4473 struct drm_i915_gem_object *obj =
4474 container_of(head, typeof(*obj), rcu);
4475 struct drm_i915_private *i915 = to_i915(obj->base.dev);
4476
Chris Wilson2ef1e722018-01-15 20:57:59 +00004477 /*
Chris Wilson8811d612018-11-09 09:03:11 +00004478 * We reuse obj->rcu for the freed list, so we had better not treat
4479 * it like a rcu_head from this point forwards. And we expect all
4480 * objects to be freed via this path.
4481 */
4482 destroy_rcu_head(&obj->rcu);
4483
4484 /*
Chris Wilson2ef1e722018-01-15 20:57:59 +00004485 * Since we require blocking on struct_mutex to unbind the freed
4486 * object from the GPU before releasing resources back to the
4487 * system, we can not do that directly from the RCU callback (which may
4488 * be a softirq context), but must instead then defer that work onto a
4489 * kthread. We use the RCU callback rather than move the freed object
4490 * directly onto the work queue so that we can mix between using the
4491 * worker and performing frees directly from subsequent allocations for
4492 * crude but effective memory throttling.
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004493 */
4494 if (llist_add(&obj->freed, &i915->mm.free_list))
Chris Wilsonbeacbd12018-01-15 12:28:45 +00004495 queue_work(i915->wq, &i915->mm.free_work);
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004496}
4497
4498void i915_gem_free_object(struct drm_gem_object *gem_obj)
4499{
4500 struct drm_i915_gem_object *obj = to_intel_bo(gem_obj);
4501
Chris Wilsonbc0629a2016-11-01 10:03:17 +00004502 if (obj->mm.quirked)
4503 __i915_gem_object_unpin_pages(obj);
4504
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004505 if (discard_backing_storage(obj))
4506 obj->mm.madv = I915_MADV_DONTNEED;
Daniel Vettera071fa02014-06-18 23:28:09 +02004507
Chris Wilson2ef1e722018-01-15 20:57:59 +00004508 /*
4509 * Before we free the object, make sure any pure RCU-only
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004510 * read-side critical sections are complete, e.g.
4511 * i915_gem_busy_ioctl(). For the corresponding synchronized
4512 * lookup see i915_gem_object_lookup_rcu().
4513 */
Chris Wilsonc9c704712018-02-19 22:06:31 +00004514 atomic_inc(&to_i915(obj->base.dev)->mm.free_count);
Chris Wilsonfbbd37b2016-10-28 13:58:42 +01004515 call_rcu(&obj->rcu, __i915_gem_free_object_rcu);
Chris Wilsonbe726152010-07-23 23:18:50 +01004516}
4517
Chris Wilsonf8a7fde2016-10-28 13:58:29 +01004518void __i915_gem_object_release_unless_active(struct drm_i915_gem_object *obj)
4519{
4520 lockdep_assert_held(&obj->base.dev->struct_mutex);
4521
Chris Wilsond1b48c12017-08-16 09:52:08 +01004522 if (!i915_gem_object_has_active_reference(obj) &&
4523 i915_gem_object_is_active(obj))
Chris Wilsonf8a7fde2016-10-28 13:58:29 +01004524 i915_gem_object_set_active_reference(obj);
4525 else
4526 i915_gem_object_put(obj);
4527}
4528
Chris Wilson24145512017-01-24 11:01:35 +00004529void i915_gem_sanitize(struct drm_i915_private *i915)
4530{
Chris Wilson538ef962019-01-14 14:21:18 +00004531 intel_wakeref_t wakeref;
4532
Chris Wilsonc3160da2018-05-31 09:22:45 +01004533 GEM_TRACE("\n");
4534
Chris Wilson538ef962019-01-14 14:21:18 +00004535 wakeref = intel_runtime_pm_get(i915);
Chris Wilsonc3160da2018-05-31 09:22:45 +01004536 intel_uncore_forcewake_get(i915, FORCEWAKE_ALL);
4537
4538 /*
4539 * As we have just resumed the machine and woken the device up from
4540 * deep PCI sleep (presumably D3_cold), assume the HW has been reset
4541 * back to defaults, recovering from whatever wedged state we left it
4542 * in and so worth trying to use the device once more.
4543 */
Chris Wilson4dfacb02018-05-31 09:22:43 +01004544 if (i915_terminally_wedged(&i915->gpu_error))
Chris Wilsonf36325f2017-08-26 12:09:34 +01004545 i915_gem_unset_wedged(i915);
Chris Wilsonf36325f2017-08-26 12:09:34 +01004546
Chris Wilson24145512017-01-24 11:01:35 +00004547 /*
4548 * If we inherit context state from the BIOS or earlier occupants
4549 * of the GPU, the GPU may be in an inconsistent state when we
4550 * try to take over. The only way to remove the earlier state
4551 * is by resetting. However, resetting on earlier gen is tricky as
4552 * it may impact the display and we are uncertain about the stability
Joonas Lahtinenea117b82017-04-28 10:53:38 +03004553 * of the reset, so this could be applied to even earlier gen.
Chris Wilson24145512017-01-24 11:01:35 +00004554 */
Chris Wilson55277e12019-01-03 11:21:04 +00004555 intel_engines_sanitize(i915, false);
Chris Wilsonc3160da2018-05-31 09:22:45 +01004556
4557 intel_uncore_forcewake_put(i915, FORCEWAKE_ALL);
Chris Wilson538ef962019-01-14 14:21:18 +00004558 intel_runtime_pm_put(i915, wakeref);
Chris Wilsonc3160da2018-05-31 09:22:45 +01004559
Chris Wilsoneb8d0f52019-01-25 13:22:28 +00004560 mutex_lock(&i915->drm.struct_mutex);
Chris Wilson4dfacb02018-05-31 09:22:43 +01004561 i915_gem_contexts_lost(i915);
4562 mutex_unlock(&i915->drm.struct_mutex);
Chris Wilson24145512017-01-24 11:01:35 +00004563}
4564
Chris Wilsonbf061122018-07-09 14:02:04 +01004565int i915_gem_suspend(struct drm_i915_private *i915)
Eric Anholt673a3942008-07-30 12:06:12 -07004566{
Chris Wilson538ef962019-01-14 14:21:18 +00004567 intel_wakeref_t wakeref;
Chris Wilsondcff85c2016-08-05 10:14:11 +01004568 int ret;
Eric Anholt673a3942008-07-30 12:06:12 -07004569
Chris Wilson09a4c022018-05-24 09:11:35 +01004570 GEM_TRACE("\n");
4571
Chris Wilson538ef962019-01-14 14:21:18 +00004572 wakeref = intel_runtime_pm_get(i915);
Chris Wilsonbf061122018-07-09 14:02:04 +01004573 intel_suspend_gt_powersave(i915);
Chris Wilson54b4f682016-07-21 21:16:19 +01004574
Chris Wilsoneb8d0f52019-01-25 13:22:28 +00004575 flush_workqueue(i915->wq);
4576
Chris Wilsonbf061122018-07-09 14:02:04 +01004577 mutex_lock(&i915->drm.struct_mutex);
Chris Wilson5ab57c72016-07-15 14:56:20 +01004578
Chris Wilsonbf061122018-07-09 14:02:04 +01004579 /*
4580 * We have to flush all the executing contexts to main memory so
Chris Wilson5ab57c72016-07-15 14:56:20 +01004581 * that they can saved in the hibernation image. To ensure the last
4582 * context image is coherent, we have to switch away from it. That
Chris Wilsonbf061122018-07-09 14:02:04 +01004583 * leaves the i915->kernel_context still active when
Chris Wilson5ab57c72016-07-15 14:56:20 +01004584 * we actually suspend, and its image in memory may not match the GPU
4585 * state. Fortunately, the kernel_context is disposable and we do
4586 * not rely on its state.
4587 */
Chris Wilsonbf061122018-07-09 14:02:04 +01004588 if (!i915_terminally_wedged(&i915->gpu_error)) {
4589 ret = i915_gem_switch_to_kernel_context(i915);
Chris Wilsonecf73eb2017-11-30 10:29:51 +00004590 if (ret)
4591 goto err_unlock;
Chris Wilson5ab57c72016-07-15 14:56:20 +01004592
Chris Wilsonbf061122018-07-09 14:02:04 +01004593 ret = i915_gem_wait_for_idle(i915,
Chris Wilsonecf73eb2017-11-30 10:29:51 +00004594 I915_WAIT_INTERRUPTIBLE |
Chris Wilson06060352018-05-31 09:22:44 +01004595 I915_WAIT_LOCKED |
Chris Wilsonec625fb2018-07-09 13:20:42 +01004596 I915_WAIT_FOR_IDLE_BOOST,
4597 MAX_SCHEDULE_TIMEOUT);
Chris Wilsonecf73eb2017-11-30 10:29:51 +00004598 if (ret && ret != -EIO)
4599 goto err_unlock;
Chris Wilsonf7403342013-09-13 23:57:04 +01004600
Chris Wilsonbf061122018-07-09 14:02:04 +01004601 assert_kernel_context_is_current(i915);
Chris Wilsonecf73eb2017-11-30 10:29:51 +00004602 }
Chris Wilson01f8f332018-07-17 09:41:21 +01004603 i915_retire_requests(i915); /* ensure we flush after wedging */
4604
Chris Wilsonbf061122018-07-09 14:02:04 +01004605 mutex_unlock(&i915->drm.struct_mutex);
Chris Wilsoneb8d0f52019-01-25 13:22:28 +00004606 i915_reset_flush(i915);
Chris Wilson45c5f202013-10-16 11:50:01 +01004607
Chris Wilsoneb8d0f52019-01-25 13:22:28 +00004608 drain_delayed_work(&i915->gt.retire_work);
Chris Wilsonbdeb9782016-12-23 14:57:56 +00004609
Chris Wilsonbf061122018-07-09 14:02:04 +01004610 /*
4611 * As the idle_work is rearming if it detects a race, play safe and
Chris Wilsonbdeb9782016-12-23 14:57:56 +00004612 * repeat the flush until it is definitely idle.
4613 */
Chris Wilsonbf061122018-07-09 14:02:04 +01004614 drain_delayed_work(&i915->gt.idle_work);
Chris Wilsonbdeb9782016-12-23 14:57:56 +00004615
Chris Wilsoneb8d0f52019-01-25 13:22:28 +00004616 intel_uc_suspend(i915);
4617
Chris Wilsonbf061122018-07-09 14:02:04 +01004618 /*
4619 * Assert that we successfully flushed all the work and
Chris Wilsonbdcf1202014-11-25 11:56:33 +00004620 * reset the GPU back to its idle, low power state.
4621 */
Chris Wilsonbf061122018-07-09 14:02:04 +01004622 WARN_ON(i915->gt.awake);
4623 if (WARN_ON(!intel_engines_are_idle(i915)))
4624 i915_gem_set_wedged(i915); /* no hope, discard everything */
Chris Wilsonbdcf1202014-11-25 11:56:33 +00004625
Chris Wilson538ef962019-01-14 14:21:18 +00004626 intel_runtime_pm_put(i915, wakeref);
Chris Wilsonec92ad02018-05-31 09:22:46 +01004627 return 0;
4628
4629err_unlock:
Chris Wilsonbf061122018-07-09 14:02:04 +01004630 mutex_unlock(&i915->drm.struct_mutex);
Chris Wilson538ef962019-01-14 14:21:18 +00004631 intel_runtime_pm_put(i915, wakeref);
Chris Wilsonec92ad02018-05-31 09:22:46 +01004632 return ret;
4633}
4634
4635void i915_gem_suspend_late(struct drm_i915_private *i915)
4636{
Chris Wilson9776f472018-06-01 15:41:24 +01004637 struct drm_i915_gem_object *obj;
4638 struct list_head *phases[] = {
4639 &i915->mm.unbound_list,
4640 &i915->mm.bound_list,
4641 NULL
4642 }, **phase;
4643
Imre Deak1c777c52016-10-12 17:46:37 +03004644 /*
4645 * Neither the BIOS, ourselves or any other kernel
4646 * expects the system to be in execlists mode on startup,
4647 * so we need to reset the GPU back to legacy mode. And the only
4648 * known way to disable logical contexts is through a GPU reset.
4649 *
4650 * So in order to leave the system in a known default configuration,
4651 * always reset the GPU upon unload and suspend. Afterwards we then
4652 * clean up the GEM state tracking, flushing off the requests and
4653 * leaving the system in a known idle state.
4654 *
4655 * Note that is of the upmost importance that the GPU is idle and
4656 * all stray writes are flushed *before* we dismantle the backing
4657 * storage for the pinned objects.
4658 *
4659 * However, since we are uncertain that resetting the GPU on older
4660 * machines is a good idea, we don't - just in case it leaves the
4661 * machine in an unusable condition.
4662 */
Chris Wilsoncad99462017-08-26 12:09:33 +01004663
Chris Wilson9776f472018-06-01 15:41:24 +01004664 mutex_lock(&i915->drm.struct_mutex);
4665 for (phase = phases; *phase; phase++) {
4666 list_for_each_entry(obj, *phase, mm.link)
4667 WARN_ON(i915_gem_object_set_to_gtt_domain(obj, false));
4668 }
4669 mutex_unlock(&i915->drm.struct_mutex);
4670
Chris Wilsonec92ad02018-05-31 09:22:46 +01004671 intel_uc_sanitize(i915);
4672 i915_gem_sanitize(i915);
Eric Anholt673a3942008-07-30 12:06:12 -07004673}
4674
Chris Wilson37cd3302017-11-12 11:27:38 +00004675void i915_gem_resume(struct drm_i915_private *i915)
Chris Wilson5ab57c72016-07-15 14:56:20 +01004676{
Chris Wilson4dfacb02018-05-31 09:22:43 +01004677 GEM_TRACE("\n");
4678
Chris Wilson37cd3302017-11-12 11:27:38 +00004679 WARN_ON(i915->gt.awake);
Chris Wilson5ab57c72016-07-15 14:56:20 +01004680
Chris Wilson37cd3302017-11-12 11:27:38 +00004681 mutex_lock(&i915->drm.struct_mutex);
4682 intel_uncore_forcewake_get(i915, FORCEWAKE_ALL);
Imre Deak31ab49a2016-11-07 11:20:05 +02004683
Chris Wilson37cd3302017-11-12 11:27:38 +00004684 i915_gem_restore_gtt_mappings(i915);
4685 i915_gem_restore_fences(i915);
Chris Wilson5ab57c72016-07-15 14:56:20 +01004686
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00004687 /*
4688 * As we didn't flush the kernel context before suspend, we cannot
Chris Wilson5ab57c72016-07-15 14:56:20 +01004689 * guarantee that the context image is complete. So let's just reset
4690 * it and start again.
4691 */
Chris Wilson37cd3302017-11-12 11:27:38 +00004692 i915->gt.resume(i915);
Chris Wilson5ab57c72016-07-15 14:56:20 +01004693
Chris Wilson37cd3302017-11-12 11:27:38 +00004694 if (i915_gem_init_hw(i915))
4695 goto err_wedged;
4696
Michal Wajdeczko7cfca4a2018-03-02 11:15:49 +00004697 intel_uc_resume(i915);
Chris Wilson7469c622017-11-14 13:03:00 +00004698
Chris Wilson37cd3302017-11-12 11:27:38 +00004699 /* Always reload a context for powersaving. */
4700 if (i915_gem_switch_to_kernel_context(i915))
4701 goto err_wedged;
4702
4703out_unlock:
4704 intel_uncore_forcewake_put(i915, FORCEWAKE_ALL);
4705 mutex_unlock(&i915->drm.struct_mutex);
4706 return;
4707
4708err_wedged:
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00004709 if (!i915_terminally_wedged(&i915->gpu_error)) {
4710 DRM_ERROR("failed to re-initialize GPU, declaring wedged!\n");
4711 i915_gem_set_wedged(i915);
4712 }
Chris Wilson37cd3302017-11-12 11:27:38 +00004713 goto out_unlock;
Chris Wilson5ab57c72016-07-15 14:56:20 +01004714}
4715
Tvrtko Ursulinc6be6072016-11-16 08:55:31 +00004716void i915_gem_init_swizzling(struct drm_i915_private *dev_priv)
Daniel Vetterf691e2f2012-02-02 09:58:12 +01004717{
Tvrtko Ursulinc6be6072016-11-16 08:55:31 +00004718 if (INTEL_GEN(dev_priv) < 5 ||
Daniel Vetterf691e2f2012-02-02 09:58:12 +01004719 dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_NONE)
4720 return;
4721
4722 I915_WRITE(DISP_ARB_CTL, I915_READ(DISP_ARB_CTL) |
4723 DISP_TILE_SURFACE_SWIZZLING);
4724
Lucas De Marchicf819ef2018-12-12 10:10:43 -08004725 if (IS_GEN(dev_priv, 5))
Daniel Vetter11782b02012-01-31 16:47:55 +01004726 return;
4727
Daniel Vetterf691e2f2012-02-02 09:58:12 +01004728 I915_WRITE(TILECTL, I915_READ(TILECTL) | TILECTL_SWZCTL);
Lucas De Marchicf819ef2018-12-12 10:10:43 -08004729 if (IS_GEN(dev_priv, 6))
Daniel Vetter6b26c862012-04-24 14:04:12 +02004730 I915_WRITE(ARB_MODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_SNB));
Lucas De Marchicf819ef2018-12-12 10:10:43 -08004731 else if (IS_GEN(dev_priv, 7))
Daniel Vetter6b26c862012-04-24 14:04:12 +02004732 I915_WRITE(ARB_MODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_IVB));
Lucas De Marchicf819ef2018-12-12 10:10:43 -08004733 else if (IS_GEN(dev_priv, 8))
Ben Widawsky31a53362013-11-02 21:07:04 -07004734 I915_WRITE(GAMTARBMODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_BDW));
Ben Widawsky8782e262012-12-18 10:31:23 -08004735 else
4736 BUG();
Daniel Vetterf691e2f2012-02-02 09:58:12 +01004737}
Daniel Vettere21af882012-02-09 20:53:27 +01004738
Tvrtko Ursulin50a0bc92016-10-13 11:02:58 +01004739static void init_unused_ring(struct drm_i915_private *dev_priv, u32 base)
Ville Syrjälä81e7f202014-08-15 01:21:55 +03004740{
Ville Syrjälä81e7f202014-08-15 01:21:55 +03004741 I915_WRITE(RING_CTL(base), 0);
4742 I915_WRITE(RING_HEAD(base), 0);
4743 I915_WRITE(RING_TAIL(base), 0);
4744 I915_WRITE(RING_START(base), 0);
4745}
4746
Tvrtko Ursulin50a0bc92016-10-13 11:02:58 +01004747static void init_unused_rings(struct drm_i915_private *dev_priv)
Ville Syrjälä81e7f202014-08-15 01:21:55 +03004748{
Tvrtko Ursulin50a0bc92016-10-13 11:02:58 +01004749 if (IS_I830(dev_priv)) {
4750 init_unused_ring(dev_priv, PRB1_BASE);
4751 init_unused_ring(dev_priv, SRB0_BASE);
4752 init_unused_ring(dev_priv, SRB1_BASE);
4753 init_unused_ring(dev_priv, SRB2_BASE);
4754 init_unused_ring(dev_priv, SRB3_BASE);
Lucas De Marchicf819ef2018-12-12 10:10:43 -08004755 } else if (IS_GEN(dev_priv, 2)) {
Tvrtko Ursulin50a0bc92016-10-13 11:02:58 +01004756 init_unused_ring(dev_priv, SRB0_BASE);
4757 init_unused_ring(dev_priv, SRB1_BASE);
Lucas De Marchicf819ef2018-12-12 10:10:43 -08004758 } else if (IS_GEN(dev_priv, 3)) {
Tvrtko Ursulin50a0bc92016-10-13 11:02:58 +01004759 init_unused_ring(dev_priv, PRB1_BASE);
4760 init_unused_ring(dev_priv, PRB2_BASE);
Ville Syrjälä81e7f202014-08-15 01:21:55 +03004761 }
4762}
4763
Chris Wilson20a8a742017-02-08 14:30:31 +00004764static int __i915_gem_restart_engines(void *data)
Ben Widawsky4fc7c972013-02-08 11:49:24 -08004765{
Chris Wilson20a8a742017-02-08 14:30:31 +00004766 struct drm_i915_private *i915 = data;
Tvrtko Ursuline2f80392016-03-16 11:00:36 +00004767 struct intel_engine_cs *engine;
Akash Goel3b3f1652016-10-13 22:44:48 +05304768 enum intel_engine_id id;
Chris Wilson20a8a742017-02-08 14:30:31 +00004769 int err;
4770
4771 for_each_engine(engine, i915, id) {
4772 err = engine->init_hw(engine);
Chris Wilson8177e112018-02-07 11:15:45 +00004773 if (err) {
4774 DRM_ERROR("Failed to restart %s (%d)\n",
4775 engine->name, err);
Chris Wilson20a8a742017-02-08 14:30:31 +00004776 return err;
Chris Wilson8177e112018-02-07 11:15:45 +00004777 }
Chris Wilson20a8a742017-02-08 14:30:31 +00004778 }
4779
4780 return 0;
4781}
4782
4783int i915_gem_init_hw(struct drm_i915_private *dev_priv)
4784{
Chris Wilsond200cda2016-04-28 09:56:44 +01004785 int ret;
Ben Widawsky4fc7c972013-02-08 11:49:24 -08004786
Chris Wilsonde867c22016-10-25 13:16:02 +01004787 dev_priv->gt.last_init_time = ktime_get();
4788
Chris Wilson5e4f5182015-02-13 14:35:59 +00004789 /* Double layer security blanket, see i915_gem_init() */
4790 intel_uncore_forcewake_get(dev_priv, FORCEWAKE_ALL);
4791
Tvrtko Ursulin0031fb92016-11-04 14:42:44 +00004792 if (HAS_EDRAM(dev_priv) && INTEL_GEN(dev_priv) < 9)
Ben Widawsky05e21cc2013-07-04 11:02:04 -07004793 I915_WRITE(HSW_IDICR, I915_READ(HSW_IDICR) | IDIHASHMSK(0xf));
Ben Widawsky4fc7c972013-02-08 11:49:24 -08004794
Tvrtko Ursulin772c2a52016-10-13 11:03:01 +01004795 if (IS_HASWELL(dev_priv))
Tvrtko Ursulin50a0bc92016-10-13 11:02:58 +01004796 I915_WRITE(MI_PREDICATE_RESULT_2, IS_HSW_GT3(dev_priv) ?
Ville Syrjälä0bf21342013-11-29 14:56:12 +02004797 LOWER_SLICE_ENABLED : LOWER_SLICE_DISABLED);
Rodrigo Vivi94353732013-08-28 16:45:46 -03004798
Tvrtko Ursulin094304b2018-12-03 12:50:10 +00004799 /* Apply the GT workarounds... */
Tvrtko Ursulin25d140f2018-12-03 13:33:19 +00004800 intel_gt_apply_workarounds(dev_priv);
Tvrtko Ursulin094304b2018-12-03 12:50:10 +00004801 /* ...and determine whether they are sticking. */
4802 intel_gt_verify_workarounds(dev_priv, "init");
Oscar Mateo59b449d2018-04-10 09:12:47 -07004803
Tvrtko Ursulinc6be6072016-11-16 08:55:31 +00004804 i915_gem_init_swizzling(dev_priv);
Ben Widawsky4fc7c972013-02-08 11:49:24 -08004805
Daniel Vetterd5abdfd2014-11-20 09:45:19 +01004806 /*
4807 * At least 830 can leave some of the unused rings
4808 * "active" (ie. head != tail) after resume which
4809 * will prevent c3 entry. Makes sure all unused rings
4810 * are totally idle.
4811 */
Tvrtko Ursulin50a0bc92016-10-13 11:02:58 +01004812 init_unused_rings(dev_priv);
Daniel Vetterd5abdfd2014-11-20 09:45:19 +01004813
Dave Gordoned54c1a2016-01-19 19:02:54 +00004814 BUG_ON(!dev_priv->kernel_context);
Chris Wilson6f74b362017-10-15 15:37:25 +01004815 if (i915_terminally_wedged(&dev_priv->gpu_error)) {
4816 ret = -EIO;
4817 goto out;
4818 }
John Harrison90638cc2015-05-29 17:43:37 +01004819
Tvrtko Ursulinc6be6072016-11-16 08:55:31 +00004820 ret = i915_ppgtt_init_hw(dev_priv);
John Harrison4ad2fd82015-06-18 13:11:20 +01004821 if (ret) {
Chris Wilson8177e112018-02-07 11:15:45 +00004822 DRM_ERROR("Enabling PPGTT failed (%d)\n", ret);
John Harrison4ad2fd82015-06-18 13:11:20 +01004823 goto out;
4824 }
4825
Jackie Lif08e2032018-03-13 17:32:53 -07004826 ret = intel_wopcm_init_hw(&dev_priv->wopcm);
4827 if (ret) {
4828 DRM_ERROR("Enabling WOPCM failed (%d)\n", ret);
4829 goto out;
4830 }
4831
Michał Winiarski9bdc3572017-10-25 18:25:19 +01004832 /* We can't enable contexts until all firmware is loaded */
4833 ret = intel_uc_init_hw(dev_priv);
Chris Wilson8177e112018-02-07 11:15:45 +00004834 if (ret) {
4835 DRM_ERROR("Enabling uc failed (%d)\n", ret);
Michał Winiarski9bdc3572017-10-25 18:25:19 +01004836 goto out;
Chris Wilson8177e112018-02-07 11:15:45 +00004837 }
Michał Winiarski9bdc3572017-10-25 18:25:19 +01004838
Tvrtko Ursulinbf9e8422016-12-01 14:16:38 +00004839 intel_mocs_init_l3cc_table(dev_priv);
Peter Antoine0ccdacf2016-04-13 15:03:25 +01004840
Chris Wilson136109c2017-11-02 13:14:30 +00004841 /* Only when the HW is re-initialised, can we replay the requests */
4842 ret = __i915_gem_restart_engines(dev_priv);
Michal Wajdeczkob96f6eb2018-06-05 12:24:43 +00004843 if (ret)
4844 goto cleanup_uc;
Michał Winiarski60c0a662018-07-12 14:48:10 +02004845
Chris Wilson5e4f5182015-02-13 14:35:59 +00004846 intel_uncore_forcewake_put(dev_priv, FORCEWAKE_ALL);
Michał Winiarski60c0a662018-07-12 14:48:10 +02004847
4848 return 0;
Michal Wajdeczkob96f6eb2018-06-05 12:24:43 +00004849
4850cleanup_uc:
4851 intel_uc_fini_hw(dev_priv);
Michał Winiarski60c0a662018-07-12 14:48:10 +02004852out:
4853 intel_uncore_forcewake_put(dev_priv, FORCEWAKE_ALL);
4854
4855 return ret;
Zou Nan hai8187a2b2010-05-21 09:08:55 +08004856}
4857
Chris Wilsond2b4b972017-11-10 14:26:33 +00004858static int __intel_engines_record_defaults(struct drm_i915_private *i915)
4859{
4860 struct i915_gem_context *ctx;
4861 struct intel_engine_cs *engine;
4862 enum intel_engine_id id;
4863 int err;
4864
4865 /*
4866 * As we reset the gpu during very early sanitisation, the current
4867 * register state on the GPU should reflect its defaults values.
4868 * We load a context onto the hw (with restore-inhibit), then switch
4869 * over to a second context to save that default register state. We
4870 * can then prime every new context with that state so they all start
4871 * from the same default HW values.
4872 */
4873
4874 ctx = i915_gem_context_create_kernel(i915, 0);
4875 if (IS_ERR(ctx))
4876 return PTR_ERR(ctx);
4877
4878 for_each_engine(engine, i915, id) {
Chris Wilsone61e0f52018-02-21 09:56:36 +00004879 struct i915_request *rq;
Chris Wilsond2b4b972017-11-10 14:26:33 +00004880
Chris Wilsone61e0f52018-02-21 09:56:36 +00004881 rq = i915_request_alloc(engine, ctx);
Chris Wilsond2b4b972017-11-10 14:26:33 +00004882 if (IS_ERR(rq)) {
4883 err = PTR_ERR(rq);
4884 goto out_ctx;
4885 }
4886
Chris Wilson3fef5cda2017-11-20 10:20:02 +00004887 err = 0;
Chris Wilsond2b4b972017-11-10 14:26:33 +00004888 if (engine->init_context)
4889 err = engine->init_context(rq);
4890
Chris Wilson697b9a82018-06-12 11:51:35 +01004891 i915_request_add(rq);
Chris Wilsond2b4b972017-11-10 14:26:33 +00004892 if (err)
4893 goto err_active;
4894 }
4895
4896 err = i915_gem_switch_to_kernel_context(i915);
4897 if (err)
4898 goto err_active;
4899
Chris Wilson2621cef2018-07-09 13:20:43 +01004900 if (i915_gem_wait_for_idle(i915, I915_WAIT_LOCKED, HZ / 5)) {
4901 i915_gem_set_wedged(i915);
4902 err = -EIO; /* Caller will declare us wedged */
Chris Wilsond2b4b972017-11-10 14:26:33 +00004903 goto err_active;
Chris Wilson2621cef2018-07-09 13:20:43 +01004904 }
Chris Wilsond2b4b972017-11-10 14:26:33 +00004905
4906 assert_kernel_context_is_current(i915);
4907
Chris Wilson8e1cb322018-09-20 17:13:43 +01004908 /*
4909 * Immediately park the GPU so that we enable powersaving and
4910 * treat it as idle. The next time we issue a request, we will
4911 * unpark and start using the engine->pinned_default_state, otherwise
4912 * it is in limbo and an early reset may fail.
4913 */
4914 __i915_gem_park(i915);
4915
Chris Wilsond2b4b972017-11-10 14:26:33 +00004916 for_each_engine(engine, i915, id) {
4917 struct i915_vma *state;
Chris Wilson37d7c9c2018-09-14 13:35:03 +01004918 void *vaddr;
Chris Wilsond2b4b972017-11-10 14:26:33 +00004919
Chris Wilson666424a2018-09-14 13:35:04 +01004920 GEM_BUG_ON(to_intel_context(ctx, engine)->pin_count);
4921
Chris Wilsonab82a062018-04-30 14:15:01 +01004922 state = to_intel_context(ctx, engine)->state;
Chris Wilsond2b4b972017-11-10 14:26:33 +00004923 if (!state)
4924 continue;
4925
4926 /*
4927 * As we will hold a reference to the logical state, it will
4928 * not be torn down with the context, and importantly the
4929 * object will hold onto its vma (making it possible for a
4930 * stray GTT write to corrupt our defaults). Unmap the vma
4931 * from the GTT to prevent such accidents and reclaim the
4932 * space.
4933 */
4934 err = i915_vma_unbind(state);
4935 if (err)
4936 goto err_active;
4937
4938 err = i915_gem_object_set_to_cpu_domain(state->obj, false);
4939 if (err)
4940 goto err_active;
4941
4942 engine->default_state = i915_gem_object_get(state->obj);
Chris Wilson37d7c9c2018-09-14 13:35:03 +01004943
4944 /* Check we can acquire the image of the context state */
4945 vaddr = i915_gem_object_pin_map(engine->default_state,
Chris Wilson666424a2018-09-14 13:35:04 +01004946 I915_MAP_FORCE_WB);
Chris Wilson37d7c9c2018-09-14 13:35:03 +01004947 if (IS_ERR(vaddr)) {
4948 err = PTR_ERR(vaddr);
4949 goto err_active;
4950 }
4951
4952 i915_gem_object_unpin_map(engine->default_state);
Chris Wilsond2b4b972017-11-10 14:26:33 +00004953 }
4954
4955 if (IS_ENABLED(CONFIG_DRM_I915_DEBUG_GEM)) {
4956 unsigned int found = intel_engines_has_context_isolation(i915);
4957
4958 /*
4959 * Make sure that classes with multiple engine instances all
4960 * share the same basic configuration.
4961 */
4962 for_each_engine(engine, i915, id) {
4963 unsigned int bit = BIT(engine->uabi_class);
4964 unsigned int expected = engine->default_state ? bit : 0;
4965
4966 if ((found & bit) != expected) {
4967 DRM_ERROR("mismatching default context state for class %d on engine %s\n",
4968 engine->uabi_class, engine->name);
4969 }
4970 }
4971 }
4972
4973out_ctx:
4974 i915_gem_context_set_closed(ctx);
4975 i915_gem_context_put(ctx);
4976 return err;
4977
4978err_active:
4979 /*
4980 * If we have to abandon now, we expect the engines to be idle
4981 * and ready to be torn-down. First try to flush any remaining
4982 * request, ensure we are pointing at the kernel context and
4983 * then remove it.
4984 */
4985 if (WARN_ON(i915_gem_switch_to_kernel_context(i915)))
4986 goto out_ctx;
4987
Chris Wilsonec625fb2018-07-09 13:20:42 +01004988 if (WARN_ON(i915_gem_wait_for_idle(i915,
4989 I915_WAIT_LOCKED,
4990 MAX_SCHEDULE_TIMEOUT)))
Chris Wilsond2b4b972017-11-10 14:26:33 +00004991 goto out_ctx;
4992
4993 i915_gem_contexts_lost(i915);
4994 goto out_ctx;
4995}
4996
Chris Wilson51797492018-12-04 14:15:16 +00004997static int
4998i915_gem_init_scratch(struct drm_i915_private *i915, unsigned int size)
4999{
5000 struct drm_i915_gem_object *obj;
5001 struct i915_vma *vma;
5002 int ret;
5003
5004 obj = i915_gem_object_create_stolen(i915, size);
5005 if (!obj)
5006 obj = i915_gem_object_create_internal(i915, size);
5007 if (IS_ERR(obj)) {
5008 DRM_ERROR("Failed to allocate scratch page\n");
5009 return PTR_ERR(obj);
5010 }
5011
5012 vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
5013 if (IS_ERR(vma)) {
5014 ret = PTR_ERR(vma);
5015 goto err_unref;
5016 }
5017
5018 ret = i915_vma_pin(vma, 0, 0, PIN_GLOBAL | PIN_HIGH);
5019 if (ret)
5020 goto err_unref;
5021
5022 i915->gt.scratch = vma;
5023 return 0;
5024
5025err_unref:
5026 i915_gem_object_put(obj);
5027 return ret;
5028}
5029
5030static void i915_gem_fini_scratch(struct drm_i915_private *i915)
5031{
5032 i915_vma_unpin_and_release(&i915->gt.scratch, 0);
5033}
5034
Tvrtko Ursulinbf9e8422016-12-01 14:16:38 +00005035int i915_gem_init(struct drm_i915_private *dev_priv)
Chris Wilson1070a422012-04-24 15:47:41 +01005036{
Chris Wilson1070a422012-04-24 15:47:41 +01005037 int ret;
5038
Changbin Du52b24162018-05-08 17:07:05 +08005039 /* We need to fallback to 4K pages if host doesn't support huge gtt. */
5040 if (intel_vgpu_active(dev_priv) && !intel_vgpu_has_huge_gtt(dev_priv))
Matthew Auldda9fe3f32017-10-06 23:18:31 +01005041 mkwrite_device_info(dev_priv)->page_sizes =
5042 I915_GTT_PAGE_SIZE_4K;
5043
Chris Wilson94312822017-05-03 10:39:18 +01005044 dev_priv->mm.unordered_timeline = dma_fence_context_alloc(1);
Chris Wilson57822dc2017-02-22 11:40:48 +00005045
Chris Wilsonfb5c5512017-11-20 20:55:00 +00005046 if (HAS_LOGICAL_RING_CONTEXTS(dev_priv)) {
Chris Wilson821ed7d2016-09-09 14:11:53 +01005047 dev_priv->gt.resume = intel_lr_context_resume;
Tvrtko Ursulin117897f2016-03-16 11:00:40 +00005048 dev_priv->gt.cleanup_engine = intel_logical_ring_cleanup;
Chris Wilsonfb5c5512017-11-20 20:55:00 +00005049 } else {
5050 dev_priv->gt.resume = intel_legacy_submission_resume;
5051 dev_priv->gt.cleanup_engine = intel_engine_cleanup;
Oscar Mateoa83014d2014-07-24 17:04:21 +01005052 }
5053
Chris Wilson1e345562019-01-28 10:23:56 +00005054 i915_timelines_init(dev_priv);
5055
Chris Wilsonee487002017-11-22 17:26:21 +00005056 ret = i915_gem_init_userptr(dev_priv);
5057 if (ret)
5058 return ret;
5059
Sagar Arun Kamble70deead2018-01-24 21:16:58 +05305060 ret = intel_uc_init_misc(dev_priv);
Michał Winiarski3176ff42017-12-13 23:13:47 +01005061 if (ret)
5062 return ret;
5063
Michal Wajdeczkof7dc0152018-06-28 14:15:21 +00005064 ret = intel_wopcm_init(&dev_priv->wopcm);
5065 if (ret)
5066 goto err_uc_misc;
5067
Chris Wilson5e4f5182015-02-13 14:35:59 +00005068 /* This is just a security blanket to placate dragons.
5069 * On some systems, we very sporadically observe that the first TLBs
5070 * used by the CS may be stale, despite us poking the TLB reset. If
5071 * we hold the forcewake during initialisation these problems
5072 * just magically go away.
5073 */
Chris Wilsonee487002017-11-22 17:26:21 +00005074 mutex_lock(&dev_priv->drm.struct_mutex);
Chris Wilson5e4f5182015-02-13 14:35:59 +00005075 intel_uncore_forcewake_get(dev_priv, FORCEWAKE_ALL);
5076
Chris Wilsonf6b9d5ca2016-08-04 07:52:23 +01005077 ret = i915_gem_init_ggtt(dev_priv);
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005078 if (ret) {
5079 GEM_BUG_ON(ret == -EIO);
5080 goto err_unlock;
5081 }
Jesse Barnesd62b4892013-03-08 10:45:53 -08005082
Chris Wilson51797492018-12-04 14:15:16 +00005083 ret = i915_gem_init_scratch(dev_priv,
Lucas De Marchicf819ef2018-12-12 10:10:43 -08005084 IS_GEN(dev_priv, 2) ? SZ_256K : PAGE_SIZE);
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005085 if (ret) {
5086 GEM_BUG_ON(ret == -EIO);
5087 goto err_ggtt;
5088 }
Ben Widawsky2fa48d82013-12-06 14:11:04 -08005089
Chris Wilson51797492018-12-04 14:15:16 +00005090 ret = i915_gem_contexts_init(dev_priv);
5091 if (ret) {
5092 GEM_BUG_ON(ret == -EIO);
5093 goto err_scratch;
5094 }
5095
Tvrtko Ursulinbf9e8422016-12-01 14:16:38 +00005096 ret = intel_engines_init(dev_priv);
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005097 if (ret) {
5098 GEM_BUG_ON(ret == -EIO);
5099 goto err_context;
5100 }
Daniel Vetter53ca26c2012-04-26 23:28:03 +02005101
Chris Wilsonf58d13d2017-11-10 14:26:29 +00005102 intel_init_gt_powersave(dev_priv);
5103
Michał Winiarski61b5c152017-12-13 23:13:48 +01005104 ret = intel_uc_init(dev_priv);
Chris Wilsoncc6a8182017-11-10 14:26:30 +00005105 if (ret)
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005106 goto err_pm;
Chris Wilsoncc6a8182017-11-10 14:26:30 +00005107
Michał Winiarski61b5c152017-12-13 23:13:48 +01005108 ret = i915_gem_init_hw(dev_priv);
5109 if (ret)
5110 goto err_uc_init;
5111
Chris Wilsoncc6a8182017-11-10 14:26:30 +00005112 /*
5113 * Despite its name intel_init_clock_gating applies both display
5114 * clock gating workarounds; GT mmio workarounds and the occasional
5115 * GT power context workaround. Worse, sometimes it includes a context
5116 * register workaround which we need to apply before we record the
5117 * default HW state for all contexts.
5118 *
5119 * FIXME: break up the workarounds and apply them at the right time!
5120 */
5121 intel_init_clock_gating(dev_priv);
5122
Chris Wilsond2b4b972017-11-10 14:26:33 +00005123 ret = __intel_engines_record_defaults(dev_priv);
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005124 if (ret)
5125 goto err_init_hw;
5126
5127 if (i915_inject_load_failure()) {
5128 ret = -ENODEV;
5129 goto err_init_hw;
5130 }
5131
5132 if (i915_inject_load_failure()) {
5133 ret = -EIO;
5134 goto err_init_hw;
5135 }
5136
5137 intel_uncore_forcewake_put(dev_priv, FORCEWAKE_ALL);
5138 mutex_unlock(&dev_priv->drm.struct_mutex);
5139
5140 return 0;
5141
5142 /*
5143 * Unwinding is complicated by that we want to handle -EIO to mean
5144 * disable GPU submission but keep KMS alive. We want to mark the
5145 * HW as irrevisibly wedged, but keep enough state around that the
5146 * driver doesn't explode during runtime.
5147 */
5148err_init_hw:
Chris Wilson8571a052018-06-06 15:54:41 +01005149 mutex_unlock(&dev_priv->drm.struct_mutex);
5150
5151 WARN_ON(i915_gem_suspend(dev_priv));
5152 i915_gem_suspend_late(dev_priv);
5153
Chris Wilson8bcf9f72018-07-10 10:44:20 +01005154 i915_gem_drain_workqueue(dev_priv);
5155
Chris Wilson8571a052018-06-06 15:54:41 +01005156 mutex_lock(&dev_priv->drm.struct_mutex);
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005157 intel_uc_fini_hw(dev_priv);
Michał Winiarski61b5c152017-12-13 23:13:48 +01005158err_uc_init:
5159 intel_uc_fini(dev_priv);
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005160err_pm:
5161 if (ret != -EIO) {
5162 intel_cleanup_gt_powersave(dev_priv);
5163 i915_gem_cleanup_engines(dev_priv);
5164 }
5165err_context:
5166 if (ret != -EIO)
5167 i915_gem_contexts_fini(dev_priv);
Chris Wilson51797492018-12-04 14:15:16 +00005168err_scratch:
5169 i915_gem_fini_scratch(dev_priv);
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005170err_ggtt:
5171err_unlock:
5172 intel_uncore_forcewake_put(dev_priv, FORCEWAKE_ALL);
5173 mutex_unlock(&dev_priv->drm.struct_mutex);
5174
Michal Wajdeczkof7dc0152018-06-28 14:15:21 +00005175err_uc_misc:
Sagar Arun Kamble70deead2018-01-24 21:16:58 +05305176 intel_uc_fini_misc(dev_priv);
Sagar Arun Kambleda943b52018-01-10 18:24:16 +05305177
Chris Wilson1e345562019-01-28 10:23:56 +00005178 if (ret != -EIO) {
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005179 i915_gem_cleanup_userptr(dev_priv);
Chris Wilson1e345562019-01-28 10:23:56 +00005180 i915_timelines_fini(dev_priv);
5181 }
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005182
Chris Wilson60990322014-04-09 09:19:42 +01005183 if (ret == -EIO) {
Chris Wilson7ed43df2018-07-26 09:50:32 +01005184 mutex_lock(&dev_priv->drm.struct_mutex);
5185
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005186 /*
5187 * Allow engine initialisation to fail by marking the GPU as
Chris Wilson60990322014-04-09 09:19:42 +01005188 * wedged. But we only want to do this where the GPU is angry,
5189 * for all other failure, such as an allocation failure, bail.
5190 */
Chris Wilson6f74b362017-10-15 15:37:25 +01005191 if (!i915_terminally_wedged(&dev_priv->gpu_error)) {
Chris Wilson51c18bf2018-06-09 12:10:58 +01005192 i915_load_error(dev_priv,
5193 "Failed to initialize GPU, declaring it wedged!\n");
Chris Wilson6f74b362017-10-15 15:37:25 +01005194 i915_gem_set_wedged(dev_priv);
5195 }
Chris Wilson7ed43df2018-07-26 09:50:32 +01005196
5197 /* Minimal basic recovery for KMS */
5198 ret = i915_ggtt_enable_hw(dev_priv);
5199 i915_gem_restore_gtt_mappings(dev_priv);
5200 i915_gem_restore_fences(dev_priv);
5201 intel_init_clock_gating(dev_priv);
5202
5203 mutex_unlock(&dev_priv->drm.struct_mutex);
Chris Wilson1070a422012-04-24 15:47:41 +01005204 }
5205
Chris Wilson6ca9a2b2017-12-13 13:43:47 +00005206 i915_gem_drain_freed_objects(dev_priv);
Chris Wilson60990322014-04-09 09:19:42 +01005207 return ret;
Chris Wilson1070a422012-04-24 15:47:41 +01005208}
5209
Michal Wajdeczko8979187a2018-06-04 09:00:32 +00005210void i915_gem_fini(struct drm_i915_private *dev_priv)
5211{
5212 i915_gem_suspend_late(dev_priv);
Chris Wilson30b710842018-08-12 23:36:29 +01005213 intel_disable_gt_powersave(dev_priv);
Michal Wajdeczko8979187a2018-06-04 09:00:32 +00005214
5215 /* Flush any outstanding unpin_work. */
5216 i915_gem_drain_workqueue(dev_priv);
5217
5218 mutex_lock(&dev_priv->drm.struct_mutex);
5219 intel_uc_fini_hw(dev_priv);
5220 intel_uc_fini(dev_priv);
5221 i915_gem_cleanup_engines(dev_priv);
5222 i915_gem_contexts_fini(dev_priv);
Chris Wilson51797492018-12-04 14:15:16 +00005223 i915_gem_fini_scratch(dev_priv);
Michal Wajdeczko8979187a2018-06-04 09:00:32 +00005224 mutex_unlock(&dev_priv->drm.struct_mutex);
5225
Tvrtko Ursulin25d140f2018-12-03 13:33:19 +00005226 intel_wa_list_free(&dev_priv->gt_wa_list);
5227
Chris Wilson30b710842018-08-12 23:36:29 +01005228 intel_cleanup_gt_powersave(dev_priv);
5229
Michal Wajdeczko8979187a2018-06-04 09:00:32 +00005230 intel_uc_fini_misc(dev_priv);
5231 i915_gem_cleanup_userptr(dev_priv);
Chris Wilson1e345562019-01-28 10:23:56 +00005232 i915_timelines_fini(dev_priv);
Michal Wajdeczko8979187a2018-06-04 09:00:32 +00005233
5234 i915_gem_drain_freed_objects(dev_priv);
5235
5236 WARN_ON(!list_empty(&dev_priv->contexts.list));
5237}
5238
Chris Wilson24145512017-01-24 11:01:35 +00005239void i915_gem_init_mmio(struct drm_i915_private *i915)
5240{
5241 i915_gem_sanitize(i915);
5242}
5243
Zou Nan hai8187a2b2010-05-21 09:08:55 +08005244void
Tvrtko Ursulincb15d9f2016-12-01 14:16:39 +00005245i915_gem_cleanup_engines(struct drm_i915_private *dev_priv)
Zou Nan hai8187a2b2010-05-21 09:08:55 +08005246{
Tvrtko Ursuline2f80392016-03-16 11:00:36 +00005247 struct intel_engine_cs *engine;
Akash Goel3b3f1652016-10-13 22:44:48 +05305248 enum intel_engine_id id;
Zou Nan hai8187a2b2010-05-21 09:08:55 +08005249
Akash Goel3b3f1652016-10-13 22:44:48 +05305250 for_each_engine(engine, dev_priv, id)
Tvrtko Ursulin117897f2016-03-16 11:00:40 +00005251 dev_priv->gt.cleanup_engine(engine);
Zou Nan hai8187a2b2010-05-21 09:08:55 +08005252}
5253
Eric Anholt673a3942008-07-30 12:06:12 -07005254void
Imre Deak40ae4e12016-03-16 14:54:03 +02005255i915_gem_load_init_fences(struct drm_i915_private *dev_priv)
5256{
Chris Wilson49ef5292016-08-18 17:17:00 +01005257 int i;
Imre Deak40ae4e12016-03-16 14:54:03 +02005258
Tvrtko Ursulinc56b89f2018-02-09 21:58:46 +00005259 if (INTEL_GEN(dev_priv) >= 7 && !IS_VALLEYVIEW(dev_priv) &&
Imre Deak40ae4e12016-03-16 14:54:03 +02005260 !IS_CHERRYVIEW(dev_priv))
5261 dev_priv->num_fence_regs = 32;
Tvrtko Ursulinc56b89f2018-02-09 21:58:46 +00005262 else if (INTEL_GEN(dev_priv) >= 4 ||
Jani Nikula73f67aa2016-12-07 22:48:09 +02005263 IS_I945G(dev_priv) || IS_I945GM(dev_priv) ||
5264 IS_G33(dev_priv) || IS_PINEVIEW(dev_priv))
Imre Deak40ae4e12016-03-16 14:54:03 +02005265 dev_priv->num_fence_regs = 16;
5266 else
5267 dev_priv->num_fence_regs = 8;
5268
Chris Wilsonc0336662016-05-06 15:40:21 +01005269 if (intel_vgpu_active(dev_priv))
Imre Deak40ae4e12016-03-16 14:54:03 +02005270 dev_priv->num_fence_regs =
5271 I915_READ(vgtif_reg(avail_rs.fence_num));
5272
5273 /* Initialize fence registers to zero */
Chris Wilson49ef5292016-08-18 17:17:00 +01005274 for (i = 0; i < dev_priv->num_fence_regs; i++) {
5275 struct drm_i915_fence_reg *fence = &dev_priv->fence_regs[i];
5276
5277 fence->i915 = dev_priv;
5278 fence->id = i;
5279 list_add_tail(&fence->link, &dev_priv->mm.fence_list);
5280 }
Tvrtko Ursulin4362f4f2016-11-16 08:55:33 +00005281 i915_gem_restore_fences(dev_priv);
Imre Deak40ae4e12016-03-16 14:54:03 +02005282
Tvrtko Ursulin4362f4f2016-11-16 08:55:33 +00005283 i915_gem_detect_bit_6_swizzle(dev_priv);
Imre Deak40ae4e12016-03-16 14:54:03 +02005284}
5285
Chris Wilson9c52d1c2017-11-10 23:24:47 +00005286static void i915_gem_init__mm(struct drm_i915_private *i915)
5287{
5288 spin_lock_init(&i915->mm.object_stat_lock);
5289 spin_lock_init(&i915->mm.obj_lock);
5290 spin_lock_init(&i915->mm.free_lock);
5291
5292 init_llist_head(&i915->mm.free_list);
5293
5294 INIT_LIST_HEAD(&i915->mm.unbound_list);
5295 INIT_LIST_HEAD(&i915->mm.bound_list);
5296 INIT_LIST_HEAD(&i915->mm.fence_list);
5297 INIT_LIST_HEAD(&i915->mm.userfault_list);
5298
5299 INIT_WORK(&i915->mm.free_work, __i915_gem_free_work);
5300}
5301
Michal Wajdeczkoa0de9082018-03-23 12:34:49 +00005302int i915_gem_init_early(struct drm_i915_private *dev_priv)
Eric Anholt673a3942008-07-30 12:06:12 -07005303{
Tvrtko Ursulina9335682016-11-02 15:14:59 +00005304 int err = -ENOMEM;
Chris Wilson42dcedd2012-11-15 11:32:30 +00005305
Tvrtko Ursulina9335682016-11-02 15:14:59 +00005306 dev_priv->objects = KMEM_CACHE(drm_i915_gem_object, SLAB_HWCACHE_ALIGN);
5307 if (!dev_priv->objects)
Chris Wilson73cb9702016-10-28 13:58:46 +01005308 goto err_out;
Chris Wilson73cb9702016-10-28 13:58:46 +01005309
Tvrtko Ursulina9335682016-11-02 15:14:59 +00005310 dev_priv->vmas = KMEM_CACHE(i915_vma, SLAB_HWCACHE_ALIGN);
5311 if (!dev_priv->vmas)
Chris Wilson73cb9702016-10-28 13:58:46 +01005312 goto err_objects;
Chris Wilson73cb9702016-10-28 13:58:46 +01005313
Chris Wilsond1b48c12017-08-16 09:52:08 +01005314 dev_priv->luts = KMEM_CACHE(i915_lut_handle, 0);
5315 if (!dev_priv->luts)
5316 goto err_vmas;
5317
Chris Wilsone61e0f52018-02-21 09:56:36 +00005318 dev_priv->requests = KMEM_CACHE(i915_request,
Tvrtko Ursulina9335682016-11-02 15:14:59 +00005319 SLAB_HWCACHE_ALIGN |
5320 SLAB_RECLAIM_ACCOUNT |
Paul E. McKenney5f0d5a32017-01-18 02:53:44 -08005321 SLAB_TYPESAFE_BY_RCU);
Tvrtko Ursulina9335682016-11-02 15:14:59 +00005322 if (!dev_priv->requests)
Chris Wilsond1b48c12017-08-16 09:52:08 +01005323 goto err_luts;
Chris Wilson73cb9702016-10-28 13:58:46 +01005324
Chris Wilson52e54202016-11-14 20:41:02 +00005325 dev_priv->dependencies = KMEM_CACHE(i915_dependency,
5326 SLAB_HWCACHE_ALIGN |
5327 SLAB_RECLAIM_ACCOUNT);
5328 if (!dev_priv->dependencies)
5329 goto err_requests;
5330
Chris Wilsonc5cf9a92017-05-17 13:10:04 +01005331 dev_priv->priorities = KMEM_CACHE(i915_priolist, SLAB_HWCACHE_ALIGN);
5332 if (!dev_priv->priorities)
5333 goto err_dependencies;
5334
Chris Wilson643b4502018-04-30 14:15:03 +01005335 INIT_LIST_HEAD(&dev_priv->gt.active_rings);
Chris Wilson3365e222018-05-03 20:51:14 +01005336 INIT_LIST_HEAD(&dev_priv->gt.closed_vma);
Chris Wilson643b4502018-04-30 14:15:03 +01005337
Chris Wilson9c52d1c2017-11-10 23:24:47 +00005338 i915_gem_init__mm(dev_priv);
Chris Wilsonf2123812017-10-16 12:40:37 +01005339
Chris Wilson67d97da2016-07-04 08:08:31 +01005340 INIT_DELAYED_WORK(&dev_priv->gt.retire_work,
Eric Anholt673a3942008-07-30 12:06:12 -07005341 i915_gem_retire_work_handler);
Chris Wilson67d97da2016-07-04 08:08:31 +01005342 INIT_DELAYED_WORK(&dev_priv->gt.idle_work,
Chris Wilsonb29c19b2013-09-25 17:34:56 +01005343 i915_gem_idle_work_handler);
Chris Wilson1f15b762016-07-01 17:23:14 +01005344 init_waitqueue_head(&dev_priv->gpu_error.wait_queue);
Daniel Vetter1f83fee2012-11-15 17:17:22 +01005345 init_waitqueue_head(&dev_priv->gpu_error.reset_queue);
Chris Wilson18bb2bc2019-01-14 21:04:01 +00005346 mutex_init(&dev_priv->gpu_error.wedge_mutex);
Chris Wilson31169712009-09-14 16:50:28 +01005347
Joonas Lahtinen6f633402016-09-01 14:58:21 +03005348 atomic_set(&dev_priv->mm.bsd_engine_dispatch_index, 0);
5349
Chris Wilsonb5add952016-08-04 16:32:36 +01005350 spin_lock_init(&dev_priv->fb_tracking.lock);
Chris Wilson73cb9702016-10-28 13:58:46 +01005351
Matthew Auld465c4032017-10-06 23:18:14 +01005352 err = i915_gemfs_init(dev_priv);
5353 if (err)
5354 DRM_NOTE("Unable to create a private tmpfs mount, hugepage support will be disabled(%d).\n", err);
5355
Chris Wilson73cb9702016-10-28 13:58:46 +01005356 return 0;
5357
Chris Wilson52e54202016-11-14 20:41:02 +00005358err_dependencies:
5359 kmem_cache_destroy(dev_priv->dependencies);
Chris Wilson73cb9702016-10-28 13:58:46 +01005360err_requests:
5361 kmem_cache_destroy(dev_priv->requests);
Chris Wilsond1b48c12017-08-16 09:52:08 +01005362err_luts:
5363 kmem_cache_destroy(dev_priv->luts);
Chris Wilson73cb9702016-10-28 13:58:46 +01005364err_vmas:
5365 kmem_cache_destroy(dev_priv->vmas);
5366err_objects:
5367 kmem_cache_destroy(dev_priv->objects);
5368err_out:
5369 return err;
Eric Anholt673a3942008-07-30 12:06:12 -07005370}
Dave Airlie71acb5eb2008-12-30 20:31:46 +10005371
Michal Wajdeczkoa0de9082018-03-23 12:34:49 +00005372void i915_gem_cleanup_early(struct drm_i915_private *dev_priv)
Imre Deakd64aa092016-01-19 15:26:29 +02005373{
Chris Wilsonc4d4c1c2017-02-10 16:35:23 +00005374 i915_gem_drain_freed_objects(dev_priv);
Chris Wilsonc9c704712018-02-19 22:06:31 +00005375 GEM_BUG_ON(!llist_empty(&dev_priv->mm.free_list));
5376 GEM_BUG_ON(atomic_read(&dev_priv->mm.free_count));
Chris Wilsonc4d4c1c2017-02-10 16:35:23 +00005377 WARN_ON(dev_priv->mm.object_count);
Matthew Auldea84aa72016-11-17 21:04:11 +00005378
Chris Wilsonc5cf9a92017-05-17 13:10:04 +01005379 kmem_cache_destroy(dev_priv->priorities);
Chris Wilson52e54202016-11-14 20:41:02 +00005380 kmem_cache_destroy(dev_priv->dependencies);
Imre Deakd64aa092016-01-19 15:26:29 +02005381 kmem_cache_destroy(dev_priv->requests);
Chris Wilsond1b48c12017-08-16 09:52:08 +01005382 kmem_cache_destroy(dev_priv->luts);
Imre Deakd64aa092016-01-19 15:26:29 +02005383 kmem_cache_destroy(dev_priv->vmas);
5384 kmem_cache_destroy(dev_priv->objects);
Chris Wilson0eafec62016-08-04 16:32:41 +01005385
5386 /* And ensure that our DESTROY_BY_RCU slabs are truly destroyed */
5387 rcu_barrier();
Matthew Auld465c4032017-10-06 23:18:14 +01005388
5389 i915_gemfs_fini(dev_priv);
Imre Deakd64aa092016-01-19 15:26:29 +02005390}
5391
Chris Wilson6a800ea2016-09-21 14:51:07 +01005392int i915_gem_freeze(struct drm_i915_private *dev_priv)
5393{
Chris Wilsond0aa3012017-04-07 11:25:49 +01005394 /* Discard all purgeable objects, let userspace recover those as
5395 * required after resuming.
5396 */
Chris Wilson6a800ea2016-09-21 14:51:07 +01005397 i915_gem_shrink_all(dev_priv);
Chris Wilson6a800ea2016-09-21 14:51:07 +01005398
Chris Wilson6a800ea2016-09-21 14:51:07 +01005399 return 0;
5400}
5401
Chris Wilson95c778d2018-06-01 15:41:25 +01005402int i915_gem_freeze_late(struct drm_i915_private *i915)
Chris Wilson461fb992016-05-14 07:26:33 +01005403{
5404 struct drm_i915_gem_object *obj;
Chris Wilson7aab2d52016-09-09 20:02:18 +01005405 struct list_head *phases[] = {
Chris Wilson95c778d2018-06-01 15:41:25 +01005406 &i915->mm.unbound_list,
5407 &i915->mm.bound_list,
Chris Wilson7aab2d52016-09-09 20:02:18 +01005408 NULL
Chris Wilson95c778d2018-06-01 15:41:25 +01005409 }, **phase;
Chris Wilson461fb992016-05-14 07:26:33 +01005410
Chris Wilson95c778d2018-06-01 15:41:25 +01005411 /*
5412 * Called just before we write the hibernation image.
Chris Wilson461fb992016-05-14 07:26:33 +01005413 *
5414 * We need to update the domain tracking to reflect that the CPU
5415 * will be accessing all the pages to create and restore from the
5416 * hibernation, and so upon restoration those pages will be in the
5417 * CPU domain.
5418 *
5419 * To make sure the hibernation image contains the latest state,
5420 * we update that state just before writing out the image.
Chris Wilson7aab2d52016-09-09 20:02:18 +01005421 *
5422 * To try and reduce the hibernation image, we manually shrink
Chris Wilsond0aa3012017-04-07 11:25:49 +01005423 * the objects as well, see i915_gem_freeze()
Chris Wilson461fb992016-05-14 07:26:33 +01005424 */
5425
Chris Wilson95c778d2018-06-01 15:41:25 +01005426 i915_gem_shrink(i915, -1UL, NULL, I915_SHRINK_UNBOUND);
5427 i915_gem_drain_freed_objects(i915);
Chris Wilson461fb992016-05-14 07:26:33 +01005428
Chris Wilson95c778d2018-06-01 15:41:25 +01005429 mutex_lock(&i915->drm.struct_mutex);
5430 for (phase = phases; *phase; phase++) {
5431 list_for_each_entry(obj, *phase, mm.link)
5432 WARN_ON(i915_gem_object_set_to_cpu_domain(obj, true));
Chris Wilson461fb992016-05-14 07:26:33 +01005433 }
Chris Wilson95c778d2018-06-01 15:41:25 +01005434 mutex_unlock(&i915->drm.struct_mutex);
Chris Wilson461fb992016-05-14 07:26:33 +01005435
5436 return 0;
5437}
5438
Chris Wilsonf787a5f2010-09-24 16:02:42 +01005439void i915_gem_release(struct drm_device *dev, struct drm_file *file)
Eric Anholtb9624422009-06-03 07:27:35 +00005440{
Chris Wilsonf787a5f2010-09-24 16:02:42 +01005441 struct drm_i915_file_private *file_priv = file->driver_priv;
Chris Wilsone61e0f52018-02-21 09:56:36 +00005442 struct i915_request *request;
Eric Anholtb9624422009-06-03 07:27:35 +00005443
5444 /* Clean up our request list when the client is going away, so that
5445 * later retire_requests won't dereference our soon-to-be-gone
5446 * file_priv.
5447 */
Chris Wilson1c255952010-09-26 11:03:27 +01005448 spin_lock(&file_priv->mm.lock);
Chris Wilsonc8659ef2017-03-02 12:25:25 +00005449 list_for_each_entry(request, &file_priv->mm.request_list, client_link)
Chris Wilsonf787a5f2010-09-24 16:02:42 +01005450 request->file_priv = NULL;
Chris Wilson1c255952010-09-26 11:03:27 +01005451 spin_unlock(&file_priv->mm.lock);
Chris Wilsonb29c19b2013-09-25 17:34:56 +01005452}
5453
Chris Wilson829a0af2017-06-20 12:05:45 +01005454int i915_gem_open(struct drm_i915_private *i915, struct drm_file *file)
Chris Wilsonb29c19b2013-09-25 17:34:56 +01005455{
5456 struct drm_i915_file_private *file_priv;
Ben Widawskye422b8882013-12-06 14:10:58 -08005457 int ret;
Chris Wilsonb29c19b2013-09-25 17:34:56 +01005458
Chris Wilsonc4c29d72016-11-09 10:45:07 +00005459 DRM_DEBUG("\n");
Chris Wilsonb29c19b2013-09-25 17:34:56 +01005460
5461 file_priv = kzalloc(sizeof(*file_priv), GFP_KERNEL);
5462 if (!file_priv)
5463 return -ENOMEM;
5464
5465 file->driver_priv = file_priv;
Chris Wilson829a0af2017-06-20 12:05:45 +01005466 file_priv->dev_priv = i915;
Chris Wilsonab0e7ff2014-02-25 17:11:24 +02005467 file_priv->file = file;
Chris Wilsonb29c19b2013-09-25 17:34:56 +01005468
5469 spin_lock_init(&file_priv->mm.lock);
5470 INIT_LIST_HEAD(&file_priv->mm.request_list);
Chris Wilsonb29c19b2013-09-25 17:34:56 +01005471
Chris Wilsonc80ff162016-07-27 09:07:27 +01005472 file_priv->bsd_engine = -1;
Mika Kuoppala14921f32018-06-15 13:44:29 +03005473 file_priv->hang_timestamp = jiffies;
Tvrtko Ursulinde1add32016-01-15 15:12:50 +00005474
Chris Wilson829a0af2017-06-20 12:05:45 +01005475 ret = i915_gem_context_open(i915, file);
Ben Widawskye422b8882013-12-06 14:10:58 -08005476 if (ret)
5477 kfree(file_priv);
Chris Wilsonb29c19b2013-09-25 17:34:56 +01005478
Ben Widawskye422b8882013-12-06 14:10:58 -08005479 return ret;
Chris Wilsonb29c19b2013-09-25 17:34:56 +01005480}
5481
Daniel Vetterb680c372014-09-19 18:27:27 +02005482/**
5483 * i915_gem_track_fb - update frontbuffer tracking
Geliang Tangd9072a32015-09-15 05:58:44 -07005484 * @old: current GEM buffer for the frontbuffer slots
5485 * @new: new GEM buffer for the frontbuffer slots
5486 * @frontbuffer_bits: bitmask of frontbuffer slots
Daniel Vetterb680c372014-09-19 18:27:27 +02005487 *
5488 * This updates the frontbuffer tracking bits @frontbuffer_bits by clearing them
5489 * from @old and setting them in @new. Both @old and @new can be NULL.
5490 */
Daniel Vettera071fa02014-06-18 23:28:09 +02005491void i915_gem_track_fb(struct drm_i915_gem_object *old,
5492 struct drm_i915_gem_object *new,
5493 unsigned frontbuffer_bits)
5494{
Chris Wilsonfaf5bf02016-08-04 16:32:37 +01005495 /* Control of individual bits within the mask are guarded by
5496 * the owning plane->mutex, i.e. we can never see concurrent
5497 * manipulation of individual bits. But since the bitfield as a whole
5498 * is updated using RMW, we need to use atomics in order to update
5499 * the bits.
5500 */
5501 BUILD_BUG_ON(INTEL_FRONTBUFFER_BITS_PER_PIPE * I915_MAX_PIPES >
Chris Wilson74f6e182018-09-26 11:47:07 +01005502 BITS_PER_TYPE(atomic_t));
Chris Wilsonfaf5bf02016-08-04 16:32:37 +01005503
Daniel Vettera071fa02014-06-18 23:28:09 +02005504 if (old) {
Chris Wilsonfaf5bf02016-08-04 16:32:37 +01005505 WARN_ON(!(atomic_read(&old->frontbuffer_bits) & frontbuffer_bits));
5506 atomic_andnot(frontbuffer_bits, &old->frontbuffer_bits);
Daniel Vettera071fa02014-06-18 23:28:09 +02005507 }
5508
5509 if (new) {
Chris Wilsonfaf5bf02016-08-04 16:32:37 +01005510 WARN_ON(atomic_read(&new->frontbuffer_bits) & frontbuffer_bits);
5511 atomic_or(frontbuffer_bits, &new->frontbuffer_bits);
Daniel Vettera071fa02014-06-18 23:28:09 +02005512 }
5513}
5514
Dave Gordonea702992015-07-09 19:29:02 +01005515/* Allocate a new GEM object and fill it with the supplied data */
5516struct drm_i915_gem_object *
Tvrtko Ursulin12d79d72016-12-01 14:16:37 +00005517i915_gem_object_create_from_data(struct drm_i915_private *dev_priv,
Dave Gordonea702992015-07-09 19:29:02 +01005518 const void *data, size_t size)
5519{
5520 struct drm_i915_gem_object *obj;
Chris Wilsonbe062fa2017-03-17 19:46:48 +00005521 struct file *file;
5522 size_t offset;
5523 int err;
Dave Gordonea702992015-07-09 19:29:02 +01005524
Tvrtko Ursulin12d79d72016-12-01 14:16:37 +00005525 obj = i915_gem_object_create(dev_priv, round_up(size, PAGE_SIZE));
Chris Wilsonfe3db792016-04-25 13:32:13 +01005526 if (IS_ERR(obj))
Dave Gordonea702992015-07-09 19:29:02 +01005527 return obj;
5528
Christian Königc0a51fd2018-02-16 13:43:38 +01005529 GEM_BUG_ON(obj->write_domain != I915_GEM_DOMAIN_CPU);
Dave Gordonea702992015-07-09 19:29:02 +01005530
Chris Wilsonbe062fa2017-03-17 19:46:48 +00005531 file = obj->base.filp;
5532 offset = 0;
5533 do {
5534 unsigned int len = min_t(typeof(size), size, PAGE_SIZE);
5535 struct page *page;
5536 void *pgdata, *vaddr;
Dave Gordonea702992015-07-09 19:29:02 +01005537
Chris Wilsonbe062fa2017-03-17 19:46:48 +00005538 err = pagecache_write_begin(file, file->f_mapping,
5539 offset, len, 0,
5540 &page, &pgdata);
5541 if (err < 0)
5542 goto fail;
Dave Gordonea702992015-07-09 19:29:02 +01005543
Chris Wilsonbe062fa2017-03-17 19:46:48 +00005544 vaddr = kmap(page);
5545 memcpy(vaddr, data, len);
5546 kunmap(page);
5547
5548 err = pagecache_write_end(file, file->f_mapping,
5549 offset, len, len,
5550 page, pgdata);
5551 if (err < 0)
5552 goto fail;
5553
5554 size -= len;
5555 data += len;
5556 offset += len;
5557 } while (size);
Dave Gordonea702992015-07-09 19:29:02 +01005558
5559 return obj;
5560
5561fail:
Chris Wilsonf8c417c2016-07-20 13:31:53 +01005562 i915_gem_object_put(obj);
Chris Wilsonbe062fa2017-03-17 19:46:48 +00005563 return ERR_PTR(err);
Dave Gordonea702992015-07-09 19:29:02 +01005564}
Chris Wilson96d77632016-10-28 13:58:33 +01005565
5566struct scatterlist *
5567i915_gem_object_get_sg(struct drm_i915_gem_object *obj,
5568 unsigned int n,
5569 unsigned int *offset)
5570{
Chris Wilsona4f5ea62016-10-28 13:58:35 +01005571 struct i915_gem_object_page_iter *iter = &obj->mm.get_page;
Chris Wilson96d77632016-10-28 13:58:33 +01005572 struct scatterlist *sg;
5573 unsigned int idx, count;
5574
5575 might_sleep();
5576 GEM_BUG_ON(n >= obj->base.size >> PAGE_SHIFT);
Chris Wilsona4f5ea62016-10-28 13:58:35 +01005577 GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
Chris Wilson96d77632016-10-28 13:58:33 +01005578
5579 /* As we iterate forward through the sg, we record each entry in a
5580 * radixtree for quick repeated (backwards) lookups. If we have seen
5581 * this index previously, we will have an entry for it.
5582 *
5583 * Initial lookup is O(N), but this is amortized to O(1) for
5584 * sequential page access (where each new request is consecutive
5585 * to the previous one). Repeated lookups are O(lg(obj->base.size)),
5586 * i.e. O(1) with a large constant!
5587 */
5588 if (n < READ_ONCE(iter->sg_idx))
5589 goto lookup;
5590
5591 mutex_lock(&iter->lock);
5592
5593 /* We prefer to reuse the last sg so that repeated lookup of this
5594 * (or the subsequent) sg are fast - comparing against the last
5595 * sg is faster than going through the radixtree.
5596 */
5597
5598 sg = iter->sg_pos;
5599 idx = iter->sg_idx;
5600 count = __sg_page_count(sg);
5601
5602 while (idx + count <= n) {
Matthew Wilcox3159f942017-11-03 13:30:42 -04005603 void *entry;
5604 unsigned long i;
Chris Wilson96d77632016-10-28 13:58:33 +01005605 int ret;
5606
5607 /* If we cannot allocate and insert this entry, or the
5608 * individual pages from this range, cancel updating the
5609 * sg_idx so that on this lookup we are forced to linearly
5610 * scan onwards, but on future lookups we will try the
5611 * insertion again (in which case we need to be careful of
5612 * the error return reporting that we have already inserted
5613 * this index).
5614 */
5615 ret = radix_tree_insert(&iter->radix, idx, sg);
5616 if (ret && ret != -EEXIST)
5617 goto scan;
5618
Matthew Wilcox3159f942017-11-03 13:30:42 -04005619 entry = xa_mk_value(idx);
Chris Wilson96d77632016-10-28 13:58:33 +01005620 for (i = 1; i < count; i++) {
Matthew Wilcox3159f942017-11-03 13:30:42 -04005621 ret = radix_tree_insert(&iter->radix, idx + i, entry);
Chris Wilson96d77632016-10-28 13:58:33 +01005622 if (ret && ret != -EEXIST)
5623 goto scan;
5624 }
5625
5626 idx += count;
5627 sg = ____sg_next(sg);
5628 count = __sg_page_count(sg);
5629 }
5630
5631scan:
5632 iter->sg_pos = sg;
5633 iter->sg_idx = idx;
5634
5635 mutex_unlock(&iter->lock);
5636
5637 if (unlikely(n < idx)) /* insertion completed by another thread */
5638 goto lookup;
5639
5640 /* In case we failed to insert the entry into the radixtree, we need
5641 * to look beyond the current sg.
5642 */
5643 while (idx + count <= n) {
5644 idx += count;
5645 sg = ____sg_next(sg);
5646 count = __sg_page_count(sg);
5647 }
5648
5649 *offset = n - idx;
5650 return sg;
5651
5652lookup:
5653 rcu_read_lock();
5654
5655 sg = radix_tree_lookup(&iter->radix, n);
5656 GEM_BUG_ON(!sg);
5657
5658 /* If this index is in the middle of multi-page sg entry,
Matthew Wilcox3159f942017-11-03 13:30:42 -04005659 * the radix tree will contain a value entry that points
Chris Wilson96d77632016-10-28 13:58:33 +01005660 * to the start of that range. We will return the pointer to
5661 * the base page and the offset of this page within the
5662 * sg entry's range.
5663 */
5664 *offset = 0;
Matthew Wilcox3159f942017-11-03 13:30:42 -04005665 if (unlikely(xa_is_value(sg))) {
5666 unsigned long base = xa_to_value(sg);
Chris Wilson96d77632016-10-28 13:58:33 +01005667
5668 sg = radix_tree_lookup(&iter->radix, base);
5669 GEM_BUG_ON(!sg);
5670
5671 *offset = n - base;
5672 }
5673
5674 rcu_read_unlock();
5675
5676 return sg;
5677}
5678
5679struct page *
5680i915_gem_object_get_page(struct drm_i915_gem_object *obj, unsigned int n)
5681{
5682 struct scatterlist *sg;
5683 unsigned int offset;
5684
5685 GEM_BUG_ON(!i915_gem_object_has_struct_page(obj));
5686
5687 sg = i915_gem_object_get_sg(obj, n, &offset);
5688 return nth_page(sg_page(sg), offset);
5689}
5690
5691/* Like i915_gem_object_get_page(), but mark the returned page dirty */
5692struct page *
5693i915_gem_object_get_dirty_page(struct drm_i915_gem_object *obj,
5694 unsigned int n)
5695{
5696 struct page *page;
5697
5698 page = i915_gem_object_get_page(obj, n);
Chris Wilsona4f5ea62016-10-28 13:58:35 +01005699 if (!obj->mm.dirty)
Chris Wilson96d77632016-10-28 13:58:33 +01005700 set_page_dirty(page);
5701
5702 return page;
5703}
5704
5705dma_addr_t
5706i915_gem_object_get_dma_address(struct drm_i915_gem_object *obj,
5707 unsigned long n)
5708{
5709 struct scatterlist *sg;
5710 unsigned int offset;
5711
5712 sg = i915_gem_object_get_sg(obj, n, &offset);
5713 return sg_dma_address(sg) + (offset << PAGE_SHIFT);
5714}
Chris Wilson935a2f72017-02-13 17:15:13 +00005715
Chris Wilson8eeb7902017-07-26 19:16:01 +01005716int i915_gem_object_attach_phys(struct drm_i915_gem_object *obj, int align)
5717{
5718 struct sg_table *pages;
5719 int err;
5720
5721 if (align > obj->base.size)
5722 return -EINVAL;
5723
5724 if (obj->ops == &i915_gem_phys_ops)
5725 return 0;
5726
5727 if (obj->ops != &i915_gem_object_ops)
5728 return -EINVAL;
5729
5730 err = i915_gem_object_unbind(obj);
5731 if (err)
5732 return err;
5733
5734 mutex_lock(&obj->mm.lock);
5735
5736 if (obj->mm.madv != I915_MADV_WILLNEED) {
5737 err = -EFAULT;
5738 goto err_unlock;
5739 }
5740
5741 if (obj->mm.quirked) {
5742 err = -EFAULT;
5743 goto err_unlock;
5744 }
5745
5746 if (obj->mm.mapping) {
5747 err = -EBUSY;
5748 goto err_unlock;
5749 }
5750
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01005751 pages = __i915_gem_object_unset_pages(obj);
Chris Wilsonf2123812017-10-16 12:40:37 +01005752
Chris Wilson8eeb7902017-07-26 19:16:01 +01005753 obj->ops = &i915_gem_phys_ops;
5754
Chris Wilson8fb6a5d2017-07-26 19:16:02 +01005755 err = ____i915_gem_object_get_pages(obj);
Chris Wilson8eeb7902017-07-26 19:16:01 +01005756 if (err)
5757 goto err_xfer;
5758
5759 /* Perma-pin (until release) the physical set of pages */
5760 __i915_gem_object_pin_pages(obj);
5761
5762 if (!IS_ERR_OR_NULL(pages))
5763 i915_gem_object_ops.put_pages(obj, pages);
5764 mutex_unlock(&obj->mm.lock);
5765 return 0;
5766
5767err_xfer:
5768 obj->ops = &i915_gem_object_ops;
Chris Wilsonacd1c1e2018-06-11 08:55:32 +01005769 if (!IS_ERR_OR_NULL(pages)) {
5770 unsigned int sg_page_sizes = i915_sg_page_sizes(pages->sgl);
5771
5772 __i915_gem_object_set_pages(obj, pages, sg_page_sizes);
5773 }
Chris Wilson8eeb7902017-07-26 19:16:01 +01005774err_unlock:
5775 mutex_unlock(&obj->mm.lock);
5776 return err;
5777}
5778
Chris Wilson935a2f72017-02-13 17:15:13 +00005779#if IS_ENABLED(CONFIG_DRM_I915_SELFTEST)
5780#include "selftests/scatterlist.c"
Chris Wilson66d9cb52017-02-13 17:15:17 +00005781#include "selftests/mock_gem_device.c"
Chris Wilson44653982017-02-13 17:15:20 +00005782#include "selftests/huge_gem_object.c"
Matthew Auld40498662017-10-06 23:18:29 +01005783#include "selftests/huge_pages.c"
Chris Wilson8335fd62017-02-13 17:15:28 +00005784#include "selftests/i915_gem_object.c"
Chris Wilson17059452017-02-13 17:15:32 +00005785#include "selftests/i915_gem_coherency.c"
Chris Wilson3f51b7e12018-08-30 14:48:06 +01005786#include "selftests/i915_gem.c"
Chris Wilson935a2f72017-02-13 17:15:13 +00005787#endif