Merge branch 'irq-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel...
[deliverable/linux.git] / fs / xfs / xfs_icache.c
1 /*
2 * Copyright (c) 2000-2005 Silicon Graphics, Inc.
3 * All Rights Reserved.
4 *
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of the GNU General Public License as
7 * published by the Free Software Foundation.
8 *
9 * This program is distributed in the hope that it would be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
13 *
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write the Free Software Foundation,
16 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
17 */
18 #include "xfs.h"
19 #include "xfs_fs.h"
20 #include "xfs_format.h"
21 #include "xfs_log_format.h"
22 #include "xfs_trans_resv.h"
23 #include "xfs_sb.h"
24 #include "xfs_mount.h"
25 #include "xfs_inode.h"
26 #include "xfs_error.h"
27 #include "xfs_trans.h"
28 #include "xfs_trans_priv.h"
29 #include "xfs_inode_item.h"
30 #include "xfs_quota.h"
31 #include "xfs_trace.h"
32 #include "xfs_icache.h"
33 #include "xfs_bmap_util.h"
34 #include "xfs_dquot_item.h"
35 #include "xfs_dquot.h"
36
37 #include <linux/kthread.h>
38 #include <linux/freezer.h>
39
40 STATIC void __xfs_inode_clear_reclaim_tag(struct xfs_mount *mp,
41 struct xfs_perag *pag, struct xfs_inode *ip);
42
43 /*
44 * Allocate and initialise an xfs_inode.
45 */
46 struct xfs_inode *
47 xfs_inode_alloc(
48 struct xfs_mount *mp,
49 xfs_ino_t ino)
50 {
51 struct xfs_inode *ip;
52
53 /*
54 * if this didn't occur in transactions, we could use
55 * KM_MAYFAIL and return NULL here on ENOMEM. Set the
56 * code up to do this anyway.
57 */
58 ip = kmem_zone_alloc(xfs_inode_zone, KM_SLEEP);
59 if (!ip)
60 return NULL;
61 if (inode_init_always(mp->m_super, VFS_I(ip))) {
62 kmem_zone_free(xfs_inode_zone, ip);
63 return NULL;
64 }
65
66 XFS_STATS_INC(vn_active);
67 ASSERT(atomic_read(&ip->i_pincount) == 0);
68 ASSERT(!spin_is_locked(&ip->i_flags_lock));
69 ASSERT(!xfs_isiflocked(ip));
70 ASSERT(ip->i_ino == 0);
71
72 mrlock_init(&ip->i_iolock, MRLOCK_BARRIER, "xfsio", ip->i_ino);
73
74 /* initialise the xfs inode */
75 ip->i_ino = ino;
76 ip->i_mount = mp;
77 memset(&ip->i_imap, 0, sizeof(struct xfs_imap));
78 ip->i_afp = NULL;
79 memset(&ip->i_df, 0, sizeof(xfs_ifork_t));
80 ip->i_flags = 0;
81 ip->i_delayed_blks = 0;
82 memset(&ip->i_d, 0, sizeof(xfs_icdinode_t));
83
84 return ip;
85 }
86
87 STATIC void
88 xfs_inode_free_callback(
89 struct rcu_head *head)
90 {
91 struct inode *inode = container_of(head, struct inode, i_rcu);
92 struct xfs_inode *ip = XFS_I(inode);
93
94 kmem_zone_free(xfs_inode_zone, ip);
95 }
96
97 void
98 xfs_inode_free(
99 struct xfs_inode *ip)
100 {
101 switch (ip->i_d.di_mode & S_IFMT) {
102 case S_IFREG:
103 case S_IFDIR:
104 case S_IFLNK:
105 xfs_idestroy_fork(ip, XFS_DATA_FORK);
106 break;
107 }
108
109 if (ip->i_afp)
110 xfs_idestroy_fork(ip, XFS_ATTR_FORK);
111
112 if (ip->i_itemp) {
113 ASSERT(!(ip->i_itemp->ili_item.li_flags & XFS_LI_IN_AIL));
114 xfs_inode_item_destroy(ip);
115 ip->i_itemp = NULL;
116 }
117
118 /*
119 * Because we use RCU freeing we need to ensure the inode always
120 * appears to be reclaimed with an invalid inode number when in the
121 * free state. The ip->i_flags_lock provides the barrier against lookup
122 * races.
123 */
124 spin_lock(&ip->i_flags_lock);
125 ip->i_flags = XFS_IRECLAIM;
126 ip->i_ino = 0;
127 spin_unlock(&ip->i_flags_lock);
128
129 /* asserts to verify all state is correct here */
130 ASSERT(atomic_read(&ip->i_pincount) == 0);
131 ASSERT(!xfs_isiflocked(ip));
132 XFS_STATS_DEC(vn_active);
133
134 call_rcu(&VFS_I(ip)->i_rcu, xfs_inode_free_callback);
135 }
136
137 /*
138 * Check the validity of the inode we just found it the cache
139 */
140 static int
141 xfs_iget_cache_hit(
142 struct xfs_perag *pag,
143 struct xfs_inode *ip,
144 xfs_ino_t ino,
145 int flags,
146 int lock_flags) __releases(RCU)
147 {
148 struct inode *inode = VFS_I(ip);
149 struct xfs_mount *mp = ip->i_mount;
150 int error;
151
152 /*
153 * check for re-use of an inode within an RCU grace period due to the
154 * radix tree nodes not being updated yet. We monitor for this by
155 * setting the inode number to zero before freeing the inode structure.
156 * If the inode has been reallocated and set up, then the inode number
157 * will not match, so check for that, too.
158 */
159 spin_lock(&ip->i_flags_lock);
160 if (ip->i_ino != ino) {
161 trace_xfs_iget_skip(ip);
162 XFS_STATS_INC(xs_ig_frecycle);
163 error = -EAGAIN;
164 goto out_error;
165 }
166
167
168 /*
169 * If we are racing with another cache hit that is currently
170 * instantiating this inode or currently recycling it out of
171 * reclaimabe state, wait for the initialisation to complete
172 * before continuing.
173 *
174 * XXX(hch): eventually we should do something equivalent to
175 * wait_on_inode to wait for these flags to be cleared
176 * instead of polling for it.
177 */
178 if (ip->i_flags & (XFS_INEW|XFS_IRECLAIM)) {
179 trace_xfs_iget_skip(ip);
180 XFS_STATS_INC(xs_ig_frecycle);
181 error = -EAGAIN;
182 goto out_error;
183 }
184
185 /*
186 * If lookup is racing with unlink return an error immediately.
187 */
188 if (ip->i_d.di_mode == 0 && !(flags & XFS_IGET_CREATE)) {
189 error = -ENOENT;
190 goto out_error;
191 }
192
193 /*
194 * If IRECLAIMABLE is set, we've torn down the VFS inode already.
195 * Need to carefully get it back into useable state.
196 */
197 if (ip->i_flags & XFS_IRECLAIMABLE) {
198 trace_xfs_iget_reclaim(ip);
199
200 /*
201 * We need to set XFS_IRECLAIM to prevent xfs_reclaim_inode
202 * from stomping over us while we recycle the inode. We can't
203 * clear the radix tree reclaimable tag yet as it requires
204 * pag_ici_lock to be held exclusive.
205 */
206 ip->i_flags |= XFS_IRECLAIM;
207
208 spin_unlock(&ip->i_flags_lock);
209 rcu_read_unlock();
210
211 error = inode_init_always(mp->m_super, inode);
212 if (error) {
213 /*
214 * Re-initializing the inode failed, and we are in deep
215 * trouble. Try to re-add it to the reclaim list.
216 */
217 rcu_read_lock();
218 spin_lock(&ip->i_flags_lock);
219
220 ip->i_flags &= ~(XFS_INEW | XFS_IRECLAIM);
221 ASSERT(ip->i_flags & XFS_IRECLAIMABLE);
222 trace_xfs_iget_reclaim_fail(ip);
223 goto out_error;
224 }
225
226 spin_lock(&pag->pag_ici_lock);
227 spin_lock(&ip->i_flags_lock);
228
229 /*
230 * Clear the per-lifetime state in the inode as we are now
231 * effectively a new inode and need to return to the initial
232 * state before reuse occurs.
233 */
234 ip->i_flags &= ~XFS_IRECLAIM_RESET_FLAGS;
235 ip->i_flags |= XFS_INEW;
236 __xfs_inode_clear_reclaim_tag(mp, pag, ip);
237 inode->i_state = I_NEW;
238
239 ASSERT(!rwsem_is_locked(&ip->i_iolock.mr_lock));
240 mrlock_init(&ip->i_iolock, MRLOCK_BARRIER, "xfsio", ip->i_ino);
241
242 spin_unlock(&ip->i_flags_lock);
243 spin_unlock(&pag->pag_ici_lock);
244 } else {
245 /* If the VFS inode is being torn down, pause and try again. */
246 if (!igrab(inode)) {
247 trace_xfs_iget_skip(ip);
248 error = -EAGAIN;
249 goto out_error;
250 }
251
252 /* We've got a live one. */
253 spin_unlock(&ip->i_flags_lock);
254 rcu_read_unlock();
255 trace_xfs_iget_hit(ip);
256 }
257
258 if (lock_flags != 0)
259 xfs_ilock(ip, lock_flags);
260
261 xfs_iflags_clear(ip, XFS_ISTALE | XFS_IDONTCACHE);
262 XFS_STATS_INC(xs_ig_found);
263
264 return 0;
265
266 out_error:
267 spin_unlock(&ip->i_flags_lock);
268 rcu_read_unlock();
269 return error;
270 }
271
272
273 static int
274 xfs_iget_cache_miss(
275 struct xfs_mount *mp,
276 struct xfs_perag *pag,
277 xfs_trans_t *tp,
278 xfs_ino_t ino,
279 struct xfs_inode **ipp,
280 int flags,
281 int lock_flags)
282 {
283 struct xfs_inode *ip;
284 int error;
285 xfs_agino_t agino = XFS_INO_TO_AGINO(mp, ino);
286 int iflags;
287
288 ip = xfs_inode_alloc(mp, ino);
289 if (!ip)
290 return -ENOMEM;
291
292 error = xfs_iread(mp, tp, ip, flags);
293 if (error)
294 goto out_destroy;
295
296 trace_xfs_iget_miss(ip);
297
298 if ((ip->i_d.di_mode == 0) && !(flags & XFS_IGET_CREATE)) {
299 error = -ENOENT;
300 goto out_destroy;
301 }
302
303 /*
304 * Preload the radix tree so we can insert safely under the
305 * write spinlock. Note that we cannot sleep inside the preload
306 * region. Since we can be called from transaction context, don't
307 * recurse into the file system.
308 */
309 if (radix_tree_preload(GFP_NOFS)) {
310 error = -EAGAIN;
311 goto out_destroy;
312 }
313
314 /*
315 * Because the inode hasn't been added to the radix-tree yet it can't
316 * be found by another thread, so we can do the non-sleeping lock here.
317 */
318 if (lock_flags) {
319 if (!xfs_ilock_nowait(ip, lock_flags))
320 BUG();
321 }
322
323 /*
324 * These values must be set before inserting the inode into the radix
325 * tree as the moment it is inserted a concurrent lookup (allowed by the
326 * RCU locking mechanism) can find it and that lookup must see that this
327 * is an inode currently under construction (i.e. that XFS_INEW is set).
328 * The ip->i_flags_lock that protects the XFS_INEW flag forms the
329 * memory barrier that ensures this detection works correctly at lookup
330 * time.
331 */
332 iflags = XFS_INEW;
333 if (flags & XFS_IGET_DONTCACHE)
334 iflags |= XFS_IDONTCACHE;
335 ip->i_udquot = NULL;
336 ip->i_gdquot = NULL;
337 ip->i_pdquot = NULL;
338 xfs_iflags_set(ip, iflags);
339
340 /* insert the new inode */
341 spin_lock(&pag->pag_ici_lock);
342 error = radix_tree_insert(&pag->pag_ici_root, agino, ip);
343 if (unlikely(error)) {
344 WARN_ON(error != -EEXIST);
345 XFS_STATS_INC(xs_ig_dup);
346 error = -EAGAIN;
347 goto out_preload_end;
348 }
349 spin_unlock(&pag->pag_ici_lock);
350 radix_tree_preload_end();
351
352 *ipp = ip;
353 return 0;
354
355 out_preload_end:
356 spin_unlock(&pag->pag_ici_lock);
357 radix_tree_preload_end();
358 if (lock_flags)
359 xfs_iunlock(ip, lock_flags);
360 out_destroy:
361 __destroy_inode(VFS_I(ip));
362 xfs_inode_free(ip);
363 return error;
364 }
365
366 /*
367 * Look up an inode by number in the given file system.
368 * The inode is looked up in the cache held in each AG.
369 * If the inode is found in the cache, initialise the vfs inode
370 * if necessary.
371 *
372 * If it is not in core, read it in from the file system's device,
373 * add it to the cache and initialise the vfs inode.
374 *
375 * The inode is locked according to the value of the lock_flags parameter.
376 * This flag parameter indicates how and if the inode's IO lock and inode lock
377 * should be taken.
378 *
379 * mp -- the mount point structure for the current file system. It points
380 * to the inode hash table.
381 * tp -- a pointer to the current transaction if there is one. This is
382 * simply passed through to the xfs_iread() call.
383 * ino -- the number of the inode desired. This is the unique identifier
384 * within the file system for the inode being requested.
385 * lock_flags -- flags indicating how to lock the inode. See the comment
386 * for xfs_ilock() for a list of valid values.
387 */
388 int
389 xfs_iget(
390 xfs_mount_t *mp,
391 xfs_trans_t *tp,
392 xfs_ino_t ino,
393 uint flags,
394 uint lock_flags,
395 xfs_inode_t **ipp)
396 {
397 xfs_inode_t *ip;
398 int error;
399 xfs_perag_t *pag;
400 xfs_agino_t agino;
401
402 /*
403 * xfs_reclaim_inode() uses the ILOCK to ensure an inode
404 * doesn't get freed while it's being referenced during a
405 * radix tree traversal here. It assumes this function
406 * aqcuires only the ILOCK (and therefore it has no need to
407 * involve the IOLOCK in this synchronization).
408 */
409 ASSERT((lock_flags & (XFS_IOLOCK_EXCL | XFS_IOLOCK_SHARED)) == 0);
410
411 /* reject inode numbers outside existing AGs */
412 if (!ino || XFS_INO_TO_AGNO(mp, ino) >= mp->m_sb.sb_agcount)
413 return -EINVAL;
414
415 /* get the perag structure and ensure that it's inode capable */
416 pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ino));
417 agino = XFS_INO_TO_AGINO(mp, ino);
418
419 again:
420 error = 0;
421 rcu_read_lock();
422 ip = radix_tree_lookup(&pag->pag_ici_root, agino);
423
424 if (ip) {
425 error = xfs_iget_cache_hit(pag, ip, ino, flags, lock_flags);
426 if (error)
427 goto out_error_or_again;
428 } else {
429 rcu_read_unlock();
430 XFS_STATS_INC(xs_ig_missed);
431
432 error = xfs_iget_cache_miss(mp, pag, tp, ino, &ip,
433 flags, lock_flags);
434 if (error)
435 goto out_error_or_again;
436 }
437 xfs_perag_put(pag);
438
439 *ipp = ip;
440
441 /*
442 * If we have a real type for an on-disk inode, we can setup the inode
443 * now. If it's a new inode being created, xfs_ialloc will handle it.
444 */
445 if (xfs_iflags_test(ip, XFS_INEW) && ip->i_d.di_mode != 0)
446 xfs_setup_existing_inode(ip);
447 return 0;
448
449 out_error_or_again:
450 if (error == -EAGAIN) {
451 delay(1);
452 goto again;
453 }
454 xfs_perag_put(pag);
455 return error;
456 }
457
458 /*
459 * The inode lookup is done in batches to keep the amount of lock traffic and
460 * radix tree lookups to a minimum. The batch size is a trade off between
461 * lookup reduction and stack usage. This is in the reclaim path, so we can't
462 * be too greedy.
463 */
464 #define XFS_LOOKUP_BATCH 32
465
466 STATIC int
467 xfs_inode_ag_walk_grab(
468 struct xfs_inode *ip)
469 {
470 struct inode *inode = VFS_I(ip);
471
472 ASSERT(rcu_read_lock_held());
473
474 /*
475 * check for stale RCU freed inode
476 *
477 * If the inode has been reallocated, it doesn't matter if it's not in
478 * the AG we are walking - we are walking for writeback, so if it
479 * passes all the "valid inode" checks and is dirty, then we'll write
480 * it back anyway. If it has been reallocated and still being
481 * initialised, the XFS_INEW check below will catch it.
482 */
483 spin_lock(&ip->i_flags_lock);
484 if (!ip->i_ino)
485 goto out_unlock_noent;
486
487 /* avoid new or reclaimable inodes. Leave for reclaim code to flush */
488 if (__xfs_iflags_test(ip, XFS_INEW | XFS_IRECLAIMABLE | XFS_IRECLAIM))
489 goto out_unlock_noent;
490 spin_unlock(&ip->i_flags_lock);
491
492 /* nothing to sync during shutdown */
493 if (XFS_FORCED_SHUTDOWN(ip->i_mount))
494 return -EFSCORRUPTED;
495
496 /* If we can't grab the inode, it must on it's way to reclaim. */
497 if (!igrab(inode))
498 return -ENOENT;
499
500 /* inode is valid */
501 return 0;
502
503 out_unlock_noent:
504 spin_unlock(&ip->i_flags_lock);
505 return -ENOENT;
506 }
507
508 STATIC int
509 xfs_inode_ag_walk(
510 struct xfs_mount *mp,
511 struct xfs_perag *pag,
512 int (*execute)(struct xfs_inode *ip, int flags,
513 void *args),
514 int flags,
515 void *args,
516 int tag)
517 {
518 uint32_t first_index;
519 int last_error = 0;
520 int skipped;
521 int done;
522 int nr_found;
523
524 restart:
525 done = 0;
526 skipped = 0;
527 first_index = 0;
528 nr_found = 0;
529 do {
530 struct xfs_inode *batch[XFS_LOOKUP_BATCH];
531 int error = 0;
532 int i;
533
534 rcu_read_lock();
535
536 if (tag == -1)
537 nr_found = radix_tree_gang_lookup(&pag->pag_ici_root,
538 (void **)batch, first_index,
539 XFS_LOOKUP_BATCH);
540 else
541 nr_found = radix_tree_gang_lookup_tag(
542 &pag->pag_ici_root,
543 (void **) batch, first_index,
544 XFS_LOOKUP_BATCH, tag);
545
546 if (!nr_found) {
547 rcu_read_unlock();
548 break;
549 }
550
551 /*
552 * Grab the inodes before we drop the lock. if we found
553 * nothing, nr == 0 and the loop will be skipped.
554 */
555 for (i = 0; i < nr_found; i++) {
556 struct xfs_inode *ip = batch[i];
557
558 if (done || xfs_inode_ag_walk_grab(ip))
559 batch[i] = NULL;
560
561 /*
562 * Update the index for the next lookup. Catch
563 * overflows into the next AG range which can occur if
564 * we have inodes in the last block of the AG and we
565 * are currently pointing to the last inode.
566 *
567 * Because we may see inodes that are from the wrong AG
568 * due to RCU freeing and reallocation, only update the
569 * index if it lies in this AG. It was a race that lead
570 * us to see this inode, so another lookup from the
571 * same index will not find it again.
572 */
573 if (XFS_INO_TO_AGNO(mp, ip->i_ino) != pag->pag_agno)
574 continue;
575 first_index = XFS_INO_TO_AGINO(mp, ip->i_ino + 1);
576 if (first_index < XFS_INO_TO_AGINO(mp, ip->i_ino))
577 done = 1;
578 }
579
580 /* unlock now we've grabbed the inodes. */
581 rcu_read_unlock();
582
583 for (i = 0; i < nr_found; i++) {
584 if (!batch[i])
585 continue;
586 error = execute(batch[i], flags, args);
587 IRELE(batch[i]);
588 if (error == -EAGAIN) {
589 skipped++;
590 continue;
591 }
592 if (error && last_error != -EFSCORRUPTED)
593 last_error = error;
594 }
595
596 /* bail out if the filesystem is corrupted. */
597 if (error == -EFSCORRUPTED)
598 break;
599
600 cond_resched();
601
602 } while (nr_found && !done);
603
604 if (skipped) {
605 delay(1);
606 goto restart;
607 }
608 return last_error;
609 }
610
611 /*
612 * Background scanning to trim post-EOF preallocated space. This is queued
613 * based on the 'speculative_prealloc_lifetime' tunable (5m by default).
614 */
615 STATIC void
616 xfs_queue_eofblocks(
617 struct xfs_mount *mp)
618 {
619 rcu_read_lock();
620 if (radix_tree_tagged(&mp->m_perag_tree, XFS_ICI_EOFBLOCKS_TAG))
621 queue_delayed_work(mp->m_eofblocks_workqueue,
622 &mp->m_eofblocks_work,
623 msecs_to_jiffies(xfs_eofb_secs * 1000));
624 rcu_read_unlock();
625 }
626
627 void
628 xfs_eofblocks_worker(
629 struct work_struct *work)
630 {
631 struct xfs_mount *mp = container_of(to_delayed_work(work),
632 struct xfs_mount, m_eofblocks_work);
633 xfs_icache_free_eofblocks(mp, NULL);
634 xfs_queue_eofblocks(mp);
635 }
636
637 int
638 xfs_inode_ag_iterator(
639 struct xfs_mount *mp,
640 int (*execute)(struct xfs_inode *ip, int flags,
641 void *args),
642 int flags,
643 void *args)
644 {
645 struct xfs_perag *pag;
646 int error = 0;
647 int last_error = 0;
648 xfs_agnumber_t ag;
649
650 ag = 0;
651 while ((pag = xfs_perag_get(mp, ag))) {
652 ag = pag->pag_agno + 1;
653 error = xfs_inode_ag_walk(mp, pag, execute, flags, args, -1);
654 xfs_perag_put(pag);
655 if (error) {
656 last_error = error;
657 if (error == -EFSCORRUPTED)
658 break;
659 }
660 }
661 return last_error;
662 }
663
664 int
665 xfs_inode_ag_iterator_tag(
666 struct xfs_mount *mp,
667 int (*execute)(struct xfs_inode *ip, int flags,
668 void *args),
669 int flags,
670 void *args,
671 int tag)
672 {
673 struct xfs_perag *pag;
674 int error = 0;
675 int last_error = 0;
676 xfs_agnumber_t ag;
677
678 ag = 0;
679 while ((pag = xfs_perag_get_tag(mp, ag, tag))) {
680 ag = pag->pag_agno + 1;
681 error = xfs_inode_ag_walk(mp, pag, execute, flags, args, tag);
682 xfs_perag_put(pag);
683 if (error) {
684 last_error = error;
685 if (error == -EFSCORRUPTED)
686 break;
687 }
688 }
689 return last_error;
690 }
691
692 /*
693 * Queue a new inode reclaim pass if there are reclaimable inodes and there
694 * isn't a reclaim pass already in progress. By default it runs every 5s based
695 * on the xfs periodic sync default of 30s. Perhaps this should have it's own
696 * tunable, but that can be done if this method proves to be ineffective or too
697 * aggressive.
698 */
699 static void
700 xfs_reclaim_work_queue(
701 struct xfs_mount *mp)
702 {
703
704 rcu_read_lock();
705 if (radix_tree_tagged(&mp->m_perag_tree, XFS_ICI_RECLAIM_TAG)) {
706 queue_delayed_work(mp->m_reclaim_workqueue, &mp->m_reclaim_work,
707 msecs_to_jiffies(xfs_syncd_centisecs / 6 * 10));
708 }
709 rcu_read_unlock();
710 }
711
712 /*
713 * This is a fast pass over the inode cache to try to get reclaim moving on as
714 * many inodes as possible in a short period of time. It kicks itself every few
715 * seconds, as well as being kicked by the inode cache shrinker when memory
716 * goes low. It scans as quickly as possible avoiding locked inodes or those
717 * already being flushed, and once done schedules a future pass.
718 */
719 void
720 xfs_reclaim_worker(
721 struct work_struct *work)
722 {
723 struct xfs_mount *mp = container_of(to_delayed_work(work),
724 struct xfs_mount, m_reclaim_work);
725
726 xfs_reclaim_inodes(mp, SYNC_TRYLOCK);
727 xfs_reclaim_work_queue(mp);
728 }
729
730 static void
731 __xfs_inode_set_reclaim_tag(
732 struct xfs_perag *pag,
733 struct xfs_inode *ip)
734 {
735 radix_tree_tag_set(&pag->pag_ici_root,
736 XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino),
737 XFS_ICI_RECLAIM_TAG);
738
739 if (!pag->pag_ici_reclaimable) {
740 /* propagate the reclaim tag up into the perag radix tree */
741 spin_lock(&ip->i_mount->m_perag_lock);
742 radix_tree_tag_set(&ip->i_mount->m_perag_tree,
743 XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino),
744 XFS_ICI_RECLAIM_TAG);
745 spin_unlock(&ip->i_mount->m_perag_lock);
746
747 /* schedule periodic background inode reclaim */
748 xfs_reclaim_work_queue(ip->i_mount);
749
750 trace_xfs_perag_set_reclaim(ip->i_mount, pag->pag_agno,
751 -1, _RET_IP_);
752 }
753 pag->pag_ici_reclaimable++;
754 }
755
756 /*
757 * We set the inode flag atomically with the radix tree tag.
758 * Once we get tag lookups on the radix tree, this inode flag
759 * can go away.
760 */
761 void
762 xfs_inode_set_reclaim_tag(
763 xfs_inode_t *ip)
764 {
765 struct xfs_mount *mp = ip->i_mount;
766 struct xfs_perag *pag;
767
768 pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
769 spin_lock(&pag->pag_ici_lock);
770 spin_lock(&ip->i_flags_lock);
771 __xfs_inode_set_reclaim_tag(pag, ip);
772 __xfs_iflags_set(ip, XFS_IRECLAIMABLE);
773 spin_unlock(&ip->i_flags_lock);
774 spin_unlock(&pag->pag_ici_lock);
775 xfs_perag_put(pag);
776 }
777
778 STATIC void
779 __xfs_inode_clear_reclaim(
780 xfs_perag_t *pag,
781 xfs_inode_t *ip)
782 {
783 pag->pag_ici_reclaimable--;
784 if (!pag->pag_ici_reclaimable) {
785 /* clear the reclaim tag from the perag radix tree */
786 spin_lock(&ip->i_mount->m_perag_lock);
787 radix_tree_tag_clear(&ip->i_mount->m_perag_tree,
788 XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino),
789 XFS_ICI_RECLAIM_TAG);
790 spin_unlock(&ip->i_mount->m_perag_lock);
791 trace_xfs_perag_clear_reclaim(ip->i_mount, pag->pag_agno,
792 -1, _RET_IP_);
793 }
794 }
795
796 STATIC void
797 __xfs_inode_clear_reclaim_tag(
798 xfs_mount_t *mp,
799 xfs_perag_t *pag,
800 xfs_inode_t *ip)
801 {
802 radix_tree_tag_clear(&pag->pag_ici_root,
803 XFS_INO_TO_AGINO(mp, ip->i_ino), XFS_ICI_RECLAIM_TAG);
804 __xfs_inode_clear_reclaim(pag, ip);
805 }
806
807 /*
808 * Grab the inode for reclaim exclusively.
809 * Return 0 if we grabbed it, non-zero otherwise.
810 */
811 STATIC int
812 xfs_reclaim_inode_grab(
813 struct xfs_inode *ip,
814 int flags)
815 {
816 ASSERT(rcu_read_lock_held());
817
818 /* quick check for stale RCU freed inode */
819 if (!ip->i_ino)
820 return 1;
821
822 /*
823 * If we are asked for non-blocking operation, do unlocked checks to
824 * see if the inode already is being flushed or in reclaim to avoid
825 * lock traffic.
826 */
827 if ((flags & SYNC_TRYLOCK) &&
828 __xfs_iflags_test(ip, XFS_IFLOCK | XFS_IRECLAIM))
829 return 1;
830
831 /*
832 * The radix tree lock here protects a thread in xfs_iget from racing
833 * with us starting reclaim on the inode. Once we have the
834 * XFS_IRECLAIM flag set it will not touch us.
835 *
836 * Due to RCU lookup, we may find inodes that have been freed and only
837 * have XFS_IRECLAIM set. Indeed, we may see reallocated inodes that
838 * aren't candidates for reclaim at all, so we must check the
839 * XFS_IRECLAIMABLE is set first before proceeding to reclaim.
840 */
841 spin_lock(&ip->i_flags_lock);
842 if (!__xfs_iflags_test(ip, XFS_IRECLAIMABLE) ||
843 __xfs_iflags_test(ip, XFS_IRECLAIM)) {
844 /* not a reclaim candidate. */
845 spin_unlock(&ip->i_flags_lock);
846 return 1;
847 }
848 __xfs_iflags_set(ip, XFS_IRECLAIM);
849 spin_unlock(&ip->i_flags_lock);
850 return 0;
851 }
852
853 /*
854 * Inodes in different states need to be treated differently. The following
855 * table lists the inode states and the reclaim actions necessary:
856 *
857 * inode state iflush ret required action
858 * --------------- ---------- ---------------
859 * bad - reclaim
860 * shutdown EIO unpin and reclaim
861 * clean, unpinned 0 reclaim
862 * stale, unpinned 0 reclaim
863 * clean, pinned(*) 0 requeue
864 * stale, pinned EAGAIN requeue
865 * dirty, async - requeue
866 * dirty, sync 0 reclaim
867 *
868 * (*) dgc: I don't think the clean, pinned state is possible but it gets
869 * handled anyway given the order of checks implemented.
870 *
871 * Also, because we get the flush lock first, we know that any inode that has
872 * been flushed delwri has had the flush completed by the time we check that
873 * the inode is clean.
874 *
875 * Note that because the inode is flushed delayed write by AIL pushing, the
876 * flush lock may already be held here and waiting on it can result in very
877 * long latencies. Hence for sync reclaims, where we wait on the flush lock,
878 * the caller should push the AIL first before trying to reclaim inodes to
879 * minimise the amount of time spent waiting. For background relaim, we only
880 * bother to reclaim clean inodes anyway.
881 *
882 * Hence the order of actions after gaining the locks should be:
883 * bad => reclaim
884 * shutdown => unpin and reclaim
885 * pinned, async => requeue
886 * pinned, sync => unpin
887 * stale => reclaim
888 * clean => reclaim
889 * dirty, async => requeue
890 * dirty, sync => flush, wait and reclaim
891 */
892 STATIC int
893 xfs_reclaim_inode(
894 struct xfs_inode *ip,
895 struct xfs_perag *pag,
896 int sync_mode)
897 {
898 struct xfs_buf *bp = NULL;
899 int error;
900
901 restart:
902 error = 0;
903 xfs_ilock(ip, XFS_ILOCK_EXCL);
904 if (!xfs_iflock_nowait(ip)) {
905 if (!(sync_mode & SYNC_WAIT))
906 goto out;
907 xfs_iflock(ip);
908 }
909
910 if (XFS_FORCED_SHUTDOWN(ip->i_mount)) {
911 xfs_iunpin_wait(ip);
912 xfs_iflush_abort(ip, false);
913 goto reclaim;
914 }
915 if (xfs_ipincount(ip)) {
916 if (!(sync_mode & SYNC_WAIT))
917 goto out_ifunlock;
918 xfs_iunpin_wait(ip);
919 }
920 if (xfs_iflags_test(ip, XFS_ISTALE))
921 goto reclaim;
922 if (xfs_inode_clean(ip))
923 goto reclaim;
924
925 /*
926 * Never flush out dirty data during non-blocking reclaim, as it would
927 * just contend with AIL pushing trying to do the same job.
928 */
929 if (!(sync_mode & SYNC_WAIT))
930 goto out_ifunlock;
931
932 /*
933 * Now we have an inode that needs flushing.
934 *
935 * Note that xfs_iflush will never block on the inode buffer lock, as
936 * xfs_ifree_cluster() can lock the inode buffer before it locks the
937 * ip->i_lock, and we are doing the exact opposite here. As a result,
938 * doing a blocking xfs_imap_to_bp() to get the cluster buffer would
939 * result in an ABBA deadlock with xfs_ifree_cluster().
940 *
941 * As xfs_ifree_cluser() must gather all inodes that are active in the
942 * cache to mark them stale, if we hit this case we don't actually want
943 * to do IO here - we want the inode marked stale so we can simply
944 * reclaim it. Hence if we get an EAGAIN error here, just unlock the
945 * inode, back off and try again. Hopefully the next pass through will
946 * see the stale flag set on the inode.
947 */
948 error = xfs_iflush(ip, &bp);
949 if (error == -EAGAIN) {
950 xfs_iunlock(ip, XFS_ILOCK_EXCL);
951 /* backoff longer than in xfs_ifree_cluster */
952 delay(2);
953 goto restart;
954 }
955
956 if (!error) {
957 error = xfs_bwrite(bp);
958 xfs_buf_relse(bp);
959 }
960
961 xfs_iflock(ip);
962 reclaim:
963 xfs_ifunlock(ip);
964 xfs_iunlock(ip, XFS_ILOCK_EXCL);
965
966 XFS_STATS_INC(xs_ig_reclaims);
967 /*
968 * Remove the inode from the per-AG radix tree.
969 *
970 * Because radix_tree_delete won't complain even if the item was never
971 * added to the tree assert that it's been there before to catch
972 * problems with the inode life time early on.
973 */
974 spin_lock(&pag->pag_ici_lock);
975 if (!radix_tree_delete(&pag->pag_ici_root,
976 XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino)))
977 ASSERT(0);
978 __xfs_inode_clear_reclaim(pag, ip);
979 spin_unlock(&pag->pag_ici_lock);
980
981 /*
982 * Here we do an (almost) spurious inode lock in order to coordinate
983 * with inode cache radix tree lookups. This is because the lookup
984 * can reference the inodes in the cache without taking references.
985 *
986 * We make that OK here by ensuring that we wait until the inode is
987 * unlocked after the lookup before we go ahead and free it.
988 */
989 xfs_ilock(ip, XFS_ILOCK_EXCL);
990 xfs_qm_dqdetach(ip);
991 xfs_iunlock(ip, XFS_ILOCK_EXCL);
992
993 xfs_inode_free(ip);
994 return error;
995
996 out_ifunlock:
997 xfs_ifunlock(ip);
998 out:
999 xfs_iflags_clear(ip, XFS_IRECLAIM);
1000 xfs_iunlock(ip, XFS_ILOCK_EXCL);
1001 /*
1002 * We could return -EAGAIN here to make reclaim rescan the inode tree in
1003 * a short while. However, this just burns CPU time scanning the tree
1004 * waiting for IO to complete and the reclaim work never goes back to
1005 * the idle state. Instead, return 0 to let the next scheduled
1006 * background reclaim attempt to reclaim the inode again.
1007 */
1008 return 0;
1009 }
1010
1011 /*
1012 * Walk the AGs and reclaim the inodes in them. Even if the filesystem is
1013 * corrupted, we still want to try to reclaim all the inodes. If we don't,
1014 * then a shut down during filesystem unmount reclaim walk leak all the
1015 * unreclaimed inodes.
1016 */
1017 STATIC int
1018 xfs_reclaim_inodes_ag(
1019 struct xfs_mount *mp,
1020 int flags,
1021 int *nr_to_scan)
1022 {
1023 struct xfs_perag *pag;
1024 int error = 0;
1025 int last_error = 0;
1026 xfs_agnumber_t ag;
1027 int trylock = flags & SYNC_TRYLOCK;
1028 int skipped;
1029
1030 restart:
1031 ag = 0;
1032 skipped = 0;
1033 while ((pag = xfs_perag_get_tag(mp, ag, XFS_ICI_RECLAIM_TAG))) {
1034 unsigned long first_index = 0;
1035 int done = 0;
1036 int nr_found = 0;
1037
1038 ag = pag->pag_agno + 1;
1039
1040 if (trylock) {
1041 if (!mutex_trylock(&pag->pag_ici_reclaim_lock)) {
1042 skipped++;
1043 xfs_perag_put(pag);
1044 continue;
1045 }
1046 first_index = pag->pag_ici_reclaim_cursor;
1047 } else
1048 mutex_lock(&pag->pag_ici_reclaim_lock);
1049
1050 do {
1051 struct xfs_inode *batch[XFS_LOOKUP_BATCH];
1052 int i;
1053
1054 rcu_read_lock();
1055 nr_found = radix_tree_gang_lookup_tag(
1056 &pag->pag_ici_root,
1057 (void **)batch, first_index,
1058 XFS_LOOKUP_BATCH,
1059 XFS_ICI_RECLAIM_TAG);
1060 if (!nr_found) {
1061 done = 1;
1062 rcu_read_unlock();
1063 break;
1064 }
1065
1066 /*
1067 * Grab the inodes before we drop the lock. if we found
1068 * nothing, nr == 0 and the loop will be skipped.
1069 */
1070 for (i = 0; i < nr_found; i++) {
1071 struct xfs_inode *ip = batch[i];
1072
1073 if (done || xfs_reclaim_inode_grab(ip, flags))
1074 batch[i] = NULL;
1075
1076 /*
1077 * Update the index for the next lookup. Catch
1078 * overflows into the next AG range which can
1079 * occur if we have inodes in the last block of
1080 * the AG and we are currently pointing to the
1081 * last inode.
1082 *
1083 * Because we may see inodes that are from the
1084 * wrong AG due to RCU freeing and
1085 * reallocation, only update the index if it
1086 * lies in this AG. It was a race that lead us
1087 * to see this inode, so another lookup from
1088 * the same index will not find it again.
1089 */
1090 if (XFS_INO_TO_AGNO(mp, ip->i_ino) !=
1091 pag->pag_agno)
1092 continue;
1093 first_index = XFS_INO_TO_AGINO(mp, ip->i_ino + 1);
1094 if (first_index < XFS_INO_TO_AGINO(mp, ip->i_ino))
1095 done = 1;
1096 }
1097
1098 /* unlock now we've grabbed the inodes. */
1099 rcu_read_unlock();
1100
1101 for (i = 0; i < nr_found; i++) {
1102 if (!batch[i])
1103 continue;
1104 error = xfs_reclaim_inode(batch[i], pag, flags);
1105 if (error && last_error != -EFSCORRUPTED)
1106 last_error = error;
1107 }
1108
1109 *nr_to_scan -= XFS_LOOKUP_BATCH;
1110
1111 cond_resched();
1112
1113 } while (nr_found && !done && *nr_to_scan > 0);
1114
1115 if (trylock && !done)
1116 pag->pag_ici_reclaim_cursor = first_index;
1117 else
1118 pag->pag_ici_reclaim_cursor = 0;
1119 mutex_unlock(&pag->pag_ici_reclaim_lock);
1120 xfs_perag_put(pag);
1121 }
1122
1123 /*
1124 * if we skipped any AG, and we still have scan count remaining, do
1125 * another pass this time using blocking reclaim semantics (i.e
1126 * waiting on the reclaim locks and ignoring the reclaim cursors). This
1127 * ensure that when we get more reclaimers than AGs we block rather
1128 * than spin trying to execute reclaim.
1129 */
1130 if (skipped && (flags & SYNC_WAIT) && *nr_to_scan > 0) {
1131 trylock = 0;
1132 goto restart;
1133 }
1134 return last_error;
1135 }
1136
1137 int
1138 xfs_reclaim_inodes(
1139 xfs_mount_t *mp,
1140 int mode)
1141 {
1142 int nr_to_scan = INT_MAX;
1143
1144 return xfs_reclaim_inodes_ag(mp, mode, &nr_to_scan);
1145 }
1146
1147 /*
1148 * Scan a certain number of inodes for reclaim.
1149 *
1150 * When called we make sure that there is a background (fast) inode reclaim in
1151 * progress, while we will throttle the speed of reclaim via doing synchronous
1152 * reclaim of inodes. That means if we come across dirty inodes, we wait for
1153 * them to be cleaned, which we hope will not be very long due to the
1154 * background walker having already kicked the IO off on those dirty inodes.
1155 */
1156 long
1157 xfs_reclaim_inodes_nr(
1158 struct xfs_mount *mp,
1159 int nr_to_scan)
1160 {
1161 /* kick background reclaimer and push the AIL */
1162 xfs_reclaim_work_queue(mp);
1163 xfs_ail_push_all(mp->m_ail);
1164
1165 return xfs_reclaim_inodes_ag(mp, SYNC_TRYLOCK | SYNC_WAIT, &nr_to_scan);
1166 }
1167
1168 /*
1169 * Return the number of reclaimable inodes in the filesystem for
1170 * the shrinker to determine how much to reclaim.
1171 */
1172 int
1173 xfs_reclaim_inodes_count(
1174 struct xfs_mount *mp)
1175 {
1176 struct xfs_perag *pag;
1177 xfs_agnumber_t ag = 0;
1178 int reclaimable = 0;
1179
1180 while ((pag = xfs_perag_get_tag(mp, ag, XFS_ICI_RECLAIM_TAG))) {
1181 ag = pag->pag_agno + 1;
1182 reclaimable += pag->pag_ici_reclaimable;
1183 xfs_perag_put(pag);
1184 }
1185 return reclaimable;
1186 }
1187
1188 STATIC int
1189 xfs_inode_match_id(
1190 struct xfs_inode *ip,
1191 struct xfs_eofblocks *eofb)
1192 {
1193 if ((eofb->eof_flags & XFS_EOF_FLAGS_UID) &&
1194 !uid_eq(VFS_I(ip)->i_uid, eofb->eof_uid))
1195 return 0;
1196
1197 if ((eofb->eof_flags & XFS_EOF_FLAGS_GID) &&
1198 !gid_eq(VFS_I(ip)->i_gid, eofb->eof_gid))
1199 return 0;
1200
1201 if ((eofb->eof_flags & XFS_EOF_FLAGS_PRID) &&
1202 xfs_get_projid(ip) != eofb->eof_prid)
1203 return 0;
1204
1205 return 1;
1206 }
1207
1208 /*
1209 * A union-based inode filtering algorithm. Process the inode if any of the
1210 * criteria match. This is for global/internal scans only.
1211 */
1212 STATIC int
1213 xfs_inode_match_id_union(
1214 struct xfs_inode *ip,
1215 struct xfs_eofblocks *eofb)
1216 {
1217 if ((eofb->eof_flags & XFS_EOF_FLAGS_UID) &&
1218 uid_eq(VFS_I(ip)->i_uid, eofb->eof_uid))
1219 return 1;
1220
1221 if ((eofb->eof_flags & XFS_EOF_FLAGS_GID) &&
1222 gid_eq(VFS_I(ip)->i_gid, eofb->eof_gid))
1223 return 1;
1224
1225 if ((eofb->eof_flags & XFS_EOF_FLAGS_PRID) &&
1226 xfs_get_projid(ip) == eofb->eof_prid)
1227 return 1;
1228
1229 return 0;
1230 }
1231
1232 STATIC int
1233 xfs_inode_free_eofblocks(
1234 struct xfs_inode *ip,
1235 int flags,
1236 void *args)
1237 {
1238 int ret;
1239 struct xfs_eofblocks *eofb = args;
1240 bool need_iolock = true;
1241 int match;
1242
1243 ASSERT(!eofb || (eofb && eofb->eof_scan_owner != 0));
1244
1245 if (!xfs_can_free_eofblocks(ip, false)) {
1246 /* inode could be preallocated or append-only */
1247 trace_xfs_inode_free_eofblocks_invalid(ip);
1248 xfs_inode_clear_eofblocks_tag(ip);
1249 return 0;
1250 }
1251
1252 /*
1253 * If the mapping is dirty the operation can block and wait for some
1254 * time. Unless we are waiting, skip it.
1255 */
1256 if (!(flags & SYNC_WAIT) &&
1257 mapping_tagged(VFS_I(ip)->i_mapping, PAGECACHE_TAG_DIRTY))
1258 return 0;
1259
1260 if (eofb) {
1261 if (eofb->eof_flags & XFS_EOF_FLAGS_UNION)
1262 match = xfs_inode_match_id_union(ip, eofb);
1263 else
1264 match = xfs_inode_match_id(ip, eofb);
1265 if (!match)
1266 return 0;
1267
1268 /* skip the inode if the file size is too small */
1269 if (eofb->eof_flags & XFS_EOF_FLAGS_MINFILESIZE &&
1270 XFS_ISIZE(ip) < eofb->eof_min_file_size)
1271 return 0;
1272
1273 /*
1274 * A scan owner implies we already hold the iolock. Skip it in
1275 * xfs_free_eofblocks() to avoid deadlock. This also eliminates
1276 * the possibility of EAGAIN being returned.
1277 */
1278 if (eofb->eof_scan_owner == ip->i_ino)
1279 need_iolock = false;
1280 }
1281
1282 ret = xfs_free_eofblocks(ip->i_mount, ip, need_iolock);
1283
1284 /* don't revisit the inode if we're not waiting */
1285 if (ret == -EAGAIN && !(flags & SYNC_WAIT))
1286 ret = 0;
1287
1288 return ret;
1289 }
1290
1291 int
1292 xfs_icache_free_eofblocks(
1293 struct xfs_mount *mp,
1294 struct xfs_eofblocks *eofb)
1295 {
1296 int flags = SYNC_TRYLOCK;
1297
1298 if (eofb && (eofb->eof_flags & XFS_EOF_FLAGS_SYNC))
1299 flags = SYNC_WAIT;
1300
1301 return xfs_inode_ag_iterator_tag(mp, xfs_inode_free_eofblocks, flags,
1302 eofb, XFS_ICI_EOFBLOCKS_TAG);
1303 }
1304
1305 /*
1306 * Run eofblocks scans on the quotas applicable to the inode. For inodes with
1307 * multiple quotas, we don't know exactly which quota caused an allocation
1308 * failure. We make a best effort by including each quota under low free space
1309 * conditions (less than 1% free space) in the scan.
1310 */
1311 int
1312 xfs_inode_free_quota_eofblocks(
1313 struct xfs_inode *ip)
1314 {
1315 int scan = 0;
1316 struct xfs_eofblocks eofb = {0};
1317 struct xfs_dquot *dq;
1318
1319 ASSERT(xfs_isilocked(ip, XFS_IOLOCK_EXCL));
1320
1321 /*
1322 * Set the scan owner to avoid a potential livelock. Otherwise, the scan
1323 * can repeatedly trylock on the inode we're currently processing. We
1324 * run a sync scan to increase effectiveness and use the union filter to
1325 * cover all applicable quotas in a single scan.
1326 */
1327 eofb.eof_scan_owner = ip->i_ino;
1328 eofb.eof_flags = XFS_EOF_FLAGS_UNION|XFS_EOF_FLAGS_SYNC;
1329
1330 if (XFS_IS_UQUOTA_ENFORCED(ip->i_mount)) {
1331 dq = xfs_inode_dquot(ip, XFS_DQ_USER);
1332 if (dq && xfs_dquot_lowsp(dq)) {
1333 eofb.eof_uid = VFS_I(ip)->i_uid;
1334 eofb.eof_flags |= XFS_EOF_FLAGS_UID;
1335 scan = 1;
1336 }
1337 }
1338
1339 if (XFS_IS_GQUOTA_ENFORCED(ip->i_mount)) {
1340 dq = xfs_inode_dquot(ip, XFS_DQ_GROUP);
1341 if (dq && xfs_dquot_lowsp(dq)) {
1342 eofb.eof_gid = VFS_I(ip)->i_gid;
1343 eofb.eof_flags |= XFS_EOF_FLAGS_GID;
1344 scan = 1;
1345 }
1346 }
1347
1348 if (scan)
1349 xfs_icache_free_eofblocks(ip->i_mount, &eofb);
1350
1351 return scan;
1352 }
1353
1354 void
1355 xfs_inode_set_eofblocks_tag(
1356 xfs_inode_t *ip)
1357 {
1358 struct xfs_mount *mp = ip->i_mount;
1359 struct xfs_perag *pag;
1360 int tagged;
1361
1362 pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
1363 spin_lock(&pag->pag_ici_lock);
1364 trace_xfs_inode_set_eofblocks_tag(ip);
1365
1366 tagged = radix_tree_tagged(&pag->pag_ici_root,
1367 XFS_ICI_EOFBLOCKS_TAG);
1368 radix_tree_tag_set(&pag->pag_ici_root,
1369 XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino),
1370 XFS_ICI_EOFBLOCKS_TAG);
1371 if (!tagged) {
1372 /* propagate the eofblocks tag up into the perag radix tree */
1373 spin_lock(&ip->i_mount->m_perag_lock);
1374 radix_tree_tag_set(&ip->i_mount->m_perag_tree,
1375 XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino),
1376 XFS_ICI_EOFBLOCKS_TAG);
1377 spin_unlock(&ip->i_mount->m_perag_lock);
1378
1379 /* kick off background trimming */
1380 xfs_queue_eofblocks(ip->i_mount);
1381
1382 trace_xfs_perag_set_eofblocks(ip->i_mount, pag->pag_agno,
1383 -1, _RET_IP_);
1384 }
1385
1386 spin_unlock(&pag->pag_ici_lock);
1387 xfs_perag_put(pag);
1388 }
1389
1390 void
1391 xfs_inode_clear_eofblocks_tag(
1392 xfs_inode_t *ip)
1393 {
1394 struct xfs_mount *mp = ip->i_mount;
1395 struct xfs_perag *pag;
1396
1397 pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
1398 spin_lock(&pag->pag_ici_lock);
1399 trace_xfs_inode_clear_eofblocks_tag(ip);
1400
1401 radix_tree_tag_clear(&pag->pag_ici_root,
1402 XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino),
1403 XFS_ICI_EOFBLOCKS_TAG);
1404 if (!radix_tree_tagged(&pag->pag_ici_root, XFS_ICI_EOFBLOCKS_TAG)) {
1405 /* clear the eofblocks tag from the perag radix tree */
1406 spin_lock(&ip->i_mount->m_perag_lock);
1407 radix_tree_tag_clear(&ip->i_mount->m_perag_tree,
1408 XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino),
1409 XFS_ICI_EOFBLOCKS_TAG);
1410 spin_unlock(&ip->i_mount->m_perag_lock);
1411 trace_xfs_perag_clear_eofblocks(ip->i_mount, pag->pag_agno,
1412 -1, _RET_IP_);
1413 }
1414
1415 spin_unlock(&pag->pag_ici_lock);
1416 xfs_perag_put(pag);
1417 }
1418
This page took 0.063324 seconds and 5 git commands to generate.