1 // SPDX-License-Identifier: GPL-2.0
3 * Copyright (c) 2000-2005 Silicon Graphics, Inc.
10 #include "xfs_shared.h"
11 #include "xfs_format.h"
12 #include "xfs_log_format.h"
13 #include "xfs_trans_resv.h"
15 #include "xfs_mount.h"
16 #include "xfs_inode.h"
17 #include "xfs_trans.h"
18 #include "xfs_quota.h"
20 #include "xfs_icache.h"
25 struct xfs_qoff_logitem
**qoffstartp
,
30 struct xfs_qoff_logitem
*qoffi
;
32 error
= xfs_trans_alloc(mp
, &M_RES(mp
)->tr_qm_quotaoff
, 0, 0, 0, &tp
);
36 qoffi
= xfs_trans_get_qoff_item(tp
, NULL
, flags
& XFS_ALL_QUOTA_ACCT
);
37 xfs_trans_log_quotaoff_item(tp
, qoffi
);
39 spin_lock(&mp
->m_sb_lock
);
40 mp
->m_sb
.sb_qflags
= (mp
->m_qflags
& ~(flags
)) & XFS_MOUNT_QUOTA_ALL
;
41 spin_unlock(&mp
->m_sb_lock
);
46 * We have to make sure that the transaction is secure on disk before we
47 * return and actually stop quota accounting. So, make it synchronous.
48 * We don't care about quotoff's performance.
50 xfs_trans_set_sync(tp
);
51 error
= xfs_trans_commit(tp
);
61 xfs_qm_log_quotaoff_end(
63 struct xfs_qoff_logitem
**startqoff
,
68 struct xfs_qoff_logitem
*qoffi
;
70 error
= xfs_trans_alloc(mp
, &M_RES(mp
)->tr_qm_equotaoff
, 0, 0, 0, &tp
);
74 qoffi
= xfs_trans_get_qoff_item(tp
, *startqoff
,
75 flags
& XFS_ALL_QUOTA_ACCT
);
76 xfs_trans_log_quotaoff_item(tp
, qoffi
);
80 * We have to make sure that the transaction is secure on disk before we
81 * return and actually stop quota accounting. So, make it synchronous.
82 * We don't care about quotoff's performance.
84 xfs_trans_set_sync(tp
);
85 return xfs_trans_commit(tp
);
89 * Turn off quota accounting and/or enforcement for all udquots and/or
90 * gdquots. Called only at unmount time.
92 * This assumes that there are no dquots of this file system cached
93 * incore, and modifies the ondisk dquot directly. Therefore, for example,
94 * it is an error to call this twice, without purging the cache.
97 xfs_qm_scall_quotaoff(
101 struct xfs_quotainfo
*q
= mp
->m_quotainfo
;
104 uint inactivate_flags
;
105 struct xfs_qoff_logitem
*qoffstart
= NULL
;
108 * No file system can have quotas enabled on disk but not in core.
109 * Note that quota utilities (like quotaoff) _expect_
110 * errno == -EEXIST here.
112 if ((mp
->m_qflags
& flags
) == 0)
116 flags
&= (XFS_ALL_QUOTA_ACCT
| XFS_ALL_QUOTA_ENFD
);
119 * We don't want to deal with two quotaoffs messing up each other,
120 * so we're going to serialize it. quotaoff isn't exactly a performance
122 * If quotaoff, then we must be dealing with the root filesystem.
125 mutex_lock(&q
->qi_quotaofflock
);
128 * If we're just turning off quota enforcement, change mp and go.
130 if ((flags
& XFS_ALL_QUOTA_ACCT
) == 0) {
131 mp
->m_qflags
&= ~(flags
);
133 spin_lock(&mp
->m_sb_lock
);
134 mp
->m_sb
.sb_qflags
= mp
->m_qflags
;
135 spin_unlock(&mp
->m_sb_lock
);
136 mutex_unlock(&q
->qi_quotaofflock
);
138 /* XXX what to do if error ? Revert back to old vals incore ? */
139 return xfs_sync_sb(mp
, false);
143 inactivate_flags
= 0;
145 * If accounting is off, we must turn enforcement off, clear the
146 * quota 'CHKD' certificate to make it known that we have to
147 * do a quotacheck the next time this quota is turned on.
149 if (flags
& XFS_UQUOTA_ACCT
) {
150 dqtype
|= XFS_QMOPT_UQUOTA
;
151 flags
|= (XFS_UQUOTA_CHKD
| XFS_UQUOTA_ENFD
);
152 inactivate_flags
|= XFS_UQUOTA_ACTIVE
;
154 if (flags
& XFS_GQUOTA_ACCT
) {
155 dqtype
|= XFS_QMOPT_GQUOTA
;
156 flags
|= (XFS_GQUOTA_CHKD
| XFS_GQUOTA_ENFD
);
157 inactivate_flags
|= XFS_GQUOTA_ACTIVE
;
159 if (flags
& XFS_PQUOTA_ACCT
) {
160 dqtype
|= XFS_QMOPT_PQUOTA
;
161 flags
|= (XFS_PQUOTA_CHKD
| XFS_PQUOTA_ENFD
);
162 inactivate_flags
|= XFS_PQUOTA_ACTIVE
;
166 * Nothing to do? Don't complain. This happens when we're just
167 * turning off quota enforcement.
169 if ((mp
->m_qflags
& flags
) == 0)
173 * Write the LI_QUOTAOFF log record, and do SB changes atomically,
174 * and synchronously. If we fail to write, we should abort the
175 * operation as it cannot be recovered safely if we crash.
177 error
= xfs_qm_log_quotaoff(mp
, &qoffstart
, flags
);
182 * Next we clear the XFS_MOUNT_*DQ_ACTIVE bit(s) in the mount struct
183 * to take care of the race between dqget and quotaoff. We don't take
184 * any special locks to reset these bits. All processes need to check
185 * these bits *after* taking inode lock(s) to see if the particular
186 * quota type is in the process of being turned off. If *ACTIVE, it is
187 * guaranteed that all dquot structures and all quotainode ptrs will all
188 * stay valid as long as that inode is kept locked.
190 * There is no turning back after this.
192 mp
->m_qflags
&= ~inactivate_flags
;
195 * Give back all the dquot reference(s) held by inodes.
196 * Here we go thru every single incore inode in this file system, and
197 * do a dqrele on the i_udquot/i_gdquot that it may have.
198 * Essentially, as long as somebody has an inode locked, this guarantees
199 * that quotas will not be turned off. This is handy because in a
200 * transaction once we lock the inode(s) and check for quotaon, we can
201 * depend on the quota inodes (and other things) being valid as long as
202 * we keep the lock(s).
204 xfs_qm_dqrele_all_inodes(mp
, flags
);
207 * Next we make the changes in the quota flag in the mount struct.
208 * This isn't protected by a particular lock directly, because we
209 * don't want to take a mrlock every time we depend on quotas being on.
211 mp
->m_qflags
&= ~flags
;
214 * Go through all the dquots of this file system and purge them,
215 * according to what was turned off.
217 xfs_qm_dqpurge_all(mp
, dqtype
);
220 * Transactions that had started before ACTIVE state bit was cleared
221 * could have logged many dquots, so they'd have higher LSNs than
222 * the first QUOTAOFF log record does. If we happen to crash when
223 * the tail of the log has gone past the QUOTAOFF record, but
224 * before the last dquot modification, those dquots __will__
225 * recover, and that's not good.
227 * So, we have QUOTAOFF start and end logitems; the start
228 * logitem won't get overwritten until the end logitem appears...
230 error
= xfs_qm_log_quotaoff_end(mp
, &qoffstart
, flags
);
232 /* We're screwed now. Shutdown is the only option. */
233 xfs_force_shutdown(mp
, SHUTDOWN_CORRUPT_INCORE
);
238 * If all quotas are completely turned off, close shop.
240 if (mp
->m_qflags
== 0) {
241 mutex_unlock(&q
->qi_quotaofflock
);
242 xfs_qm_destroy_quotainfo(mp
);
247 * Release our quotainode references if we don't need them anymore.
249 if ((dqtype
& XFS_QMOPT_UQUOTA
) && q
->qi_uquotaip
) {
250 xfs_irele(q
->qi_uquotaip
);
251 q
->qi_uquotaip
= NULL
;
253 if ((dqtype
& XFS_QMOPT_GQUOTA
) && q
->qi_gquotaip
) {
254 xfs_irele(q
->qi_gquotaip
);
255 q
->qi_gquotaip
= NULL
;
257 if ((dqtype
& XFS_QMOPT_PQUOTA
) && q
->qi_pquotaip
) {
258 xfs_irele(q
->qi_pquotaip
);
259 q
->qi_pquotaip
= NULL
;
263 if (error
&& qoffstart
)
264 xfs_qm_qoff_logitem_relse(qoffstart
);
265 mutex_unlock(&q
->qi_quotaofflock
);
270 xfs_qm_scall_trunc_qfile(
271 struct xfs_mount
*mp
,
274 struct xfs_inode
*ip
;
275 struct xfs_trans
*tp
;
278 if (ino
== NULLFSINO
)
281 error
= xfs_iget(mp
, NULL
, ino
, 0, 0, &ip
);
285 xfs_ilock(ip
, XFS_IOLOCK_EXCL
);
287 error
= xfs_trans_alloc(mp
, &M_RES(mp
)->tr_itruncate
, 0, 0, 0, &tp
);
289 xfs_iunlock(ip
, XFS_IOLOCK_EXCL
);
293 xfs_ilock(ip
, XFS_ILOCK_EXCL
);
294 xfs_trans_ijoin(tp
, ip
, 0);
297 xfs_trans_log_inode(tp
, ip
, XFS_ILOG_CORE
);
299 error
= xfs_itruncate_extents(&tp
, ip
, XFS_DATA_FORK
, 0);
301 xfs_trans_cancel(tp
);
305 ASSERT(ip
->i_d
.di_nextents
== 0);
307 xfs_trans_ichgtime(tp
, ip
, XFS_ICHGTIME_MOD
| XFS_ICHGTIME_CHG
);
308 error
= xfs_trans_commit(tp
);
311 xfs_iunlock(ip
, XFS_ILOCK_EXCL
| XFS_IOLOCK_EXCL
);
318 xfs_qm_scall_trunc_qfiles(
324 if (!xfs_sb_version_hasquota(&mp
->m_sb
) || flags
== 0 ||
325 (flags
& ~XFS_DQ_ALLTYPES
)) {
326 xfs_debug(mp
, "%s: flags=%x m_qflags=%x",
327 __func__
, flags
, mp
->m_qflags
);
331 if (flags
& XFS_DQ_USER
) {
332 error
= xfs_qm_scall_trunc_qfile(mp
, mp
->m_sb
.sb_uquotino
);
336 if (flags
& XFS_DQ_GROUP
) {
337 error
= xfs_qm_scall_trunc_qfile(mp
, mp
->m_sb
.sb_gquotino
);
341 if (flags
& XFS_DQ_PROJ
)
342 error
= xfs_qm_scall_trunc_qfile(mp
, mp
->m_sb
.sb_pquotino
);
348 * Switch on (a given) quota enforcement for a filesystem. This takes
349 * effect immediately.
350 * (Switching on quota accounting must be done at mount time.)
353 xfs_qm_scall_quotaon(
360 flags
&= (XFS_ALL_QUOTA_ACCT
| XFS_ALL_QUOTA_ENFD
);
362 * Switching on quota accounting must be done at mount time.
364 flags
&= ~(XFS_ALL_QUOTA_ACCT
);
367 xfs_debug(mp
, "%s: zero flags, m_qflags=%x",
368 __func__
, mp
->m_qflags
);
373 * Can't enforce without accounting. We check the superblock
374 * qflags here instead of m_qflags because rootfs can have
375 * quota acct on ondisk without m_qflags' knowing.
377 if (((mp
->m_sb
.sb_qflags
& XFS_UQUOTA_ACCT
) == 0 &&
378 (flags
& XFS_UQUOTA_ENFD
)) ||
379 ((mp
->m_sb
.sb_qflags
& XFS_GQUOTA_ACCT
) == 0 &&
380 (flags
& XFS_GQUOTA_ENFD
)) ||
381 ((mp
->m_sb
.sb_qflags
& XFS_PQUOTA_ACCT
) == 0 &&
382 (flags
& XFS_PQUOTA_ENFD
))) {
384 "%s: Can't enforce without acct, flags=%x sbflags=%x",
385 __func__
, flags
, mp
->m_sb
.sb_qflags
);
389 * If everything's up to-date incore, then don't waste time.
391 if ((mp
->m_qflags
& flags
) == flags
)
395 * Change sb_qflags on disk but not incore mp->qflags
396 * if this is the root filesystem.
398 spin_lock(&mp
->m_sb_lock
);
399 qf
= mp
->m_sb
.sb_qflags
;
400 mp
->m_sb
.sb_qflags
= qf
| flags
;
401 spin_unlock(&mp
->m_sb_lock
);
404 * There's nothing to change if it's the same.
406 if ((qf
& flags
) == flags
)
409 error
= xfs_sync_sb(mp
, false);
413 * If we aren't trying to switch on quota enforcement, we are done.
415 if (((mp
->m_sb
.sb_qflags
& XFS_UQUOTA_ACCT
) !=
416 (mp
->m_qflags
& XFS_UQUOTA_ACCT
)) ||
417 ((mp
->m_sb
.sb_qflags
& XFS_PQUOTA_ACCT
) !=
418 (mp
->m_qflags
& XFS_PQUOTA_ACCT
)) ||
419 ((mp
->m_sb
.sb_qflags
& XFS_GQUOTA_ACCT
) !=
420 (mp
->m_qflags
& XFS_GQUOTA_ACCT
)))
423 if (! XFS_IS_QUOTA_RUNNING(mp
))
427 * Switch on quota enforcement in core.
429 mutex_lock(&mp
->m_quotainfo
->qi_quotaofflock
);
430 mp
->m_qflags
|= (flags
& XFS_ALL_QUOTA_ENFD
);
431 mutex_unlock(&mp
->m_quotainfo
->qi_quotaofflock
);
436 #define XFS_QC_MASK \
437 (QC_LIMIT_MASK | QC_TIMER_MASK | QC_WARNS_MASK)
440 * Adjust quota limits, and start/stop timers accordingly.
443 xfs_qm_scall_setqlim(
444 struct xfs_mount
*mp
,
447 struct qc_dqblk
*newlim
)
449 struct xfs_quotainfo
*q
= mp
->m_quotainfo
;
450 struct xfs_disk_dquot
*ddq
;
451 struct xfs_dquot
*dqp
;
452 struct xfs_trans
*tp
;
453 struct xfs_def_quota
*defq
;
455 xfs_qcnt_t hard
, soft
;
457 if (newlim
->d_fieldmask
& ~XFS_QC_MASK
)
459 if ((newlim
->d_fieldmask
& XFS_QC_MASK
) == 0)
463 * We don't want to race with a quotaoff so take the quotaoff lock.
464 * We don't hold an inode lock, so there's nothing else to stop
465 * a quotaoff from happening.
467 mutex_lock(&q
->qi_quotaofflock
);
470 * Get the dquot (locked) before we start, as we need to do a
471 * transaction to allocate it if it doesn't exist. Once we have the
472 * dquot, unlock it so we can start the next transaction safely. We hold
473 * a reference to the dquot, so it's safe to do this unlock/lock without
474 * it being reclaimed in the mean time.
476 error
= xfs_qm_dqget(mp
, id
, type
, true, &dqp
);
478 ASSERT(error
!= -ENOENT
);
482 defq
= xfs_get_defquota(dqp
, q
);
485 error
= xfs_trans_alloc(mp
, &M_RES(mp
)->tr_qm_setqlim
, 0, 0, 0, &tp
);
490 xfs_trans_dqjoin(tp
, dqp
);
494 * Make sure that hardlimits are >= soft limits before changing.
496 hard
= (newlim
->d_fieldmask
& QC_SPC_HARD
) ?
497 (xfs_qcnt_t
) XFS_B_TO_FSB(mp
, newlim
->d_spc_hardlimit
) :
498 be64_to_cpu(ddq
->d_blk_hardlimit
);
499 soft
= (newlim
->d_fieldmask
& QC_SPC_SOFT
) ?
500 (xfs_qcnt_t
) XFS_B_TO_FSB(mp
, newlim
->d_spc_softlimit
) :
501 be64_to_cpu(ddq
->d_blk_softlimit
);
502 if (hard
== 0 || hard
>= soft
) {
503 ddq
->d_blk_hardlimit
= cpu_to_be64(hard
);
504 ddq
->d_blk_softlimit
= cpu_to_be64(soft
);
505 xfs_dquot_set_prealloc_limits(dqp
);
507 defq
->bhardlimit
= hard
;
508 defq
->bsoftlimit
= soft
;
511 xfs_debug(mp
, "blkhard %Ld < blksoft %Ld", hard
, soft
);
513 hard
= (newlim
->d_fieldmask
& QC_RT_SPC_HARD
) ?
514 (xfs_qcnt_t
) XFS_B_TO_FSB(mp
, newlim
->d_rt_spc_hardlimit
) :
515 be64_to_cpu(ddq
->d_rtb_hardlimit
);
516 soft
= (newlim
->d_fieldmask
& QC_RT_SPC_SOFT
) ?
517 (xfs_qcnt_t
) XFS_B_TO_FSB(mp
, newlim
->d_rt_spc_softlimit
) :
518 be64_to_cpu(ddq
->d_rtb_softlimit
);
519 if (hard
== 0 || hard
>= soft
) {
520 ddq
->d_rtb_hardlimit
= cpu_to_be64(hard
);
521 ddq
->d_rtb_softlimit
= cpu_to_be64(soft
);
523 defq
->rtbhardlimit
= hard
;
524 defq
->rtbsoftlimit
= soft
;
527 xfs_debug(mp
, "rtbhard %Ld < rtbsoft %Ld", hard
, soft
);
530 hard
= (newlim
->d_fieldmask
& QC_INO_HARD
) ?
531 (xfs_qcnt_t
) newlim
->d_ino_hardlimit
:
532 be64_to_cpu(ddq
->d_ino_hardlimit
);
533 soft
= (newlim
->d_fieldmask
& QC_INO_SOFT
) ?
534 (xfs_qcnt_t
) newlim
->d_ino_softlimit
:
535 be64_to_cpu(ddq
->d_ino_softlimit
);
536 if (hard
== 0 || hard
>= soft
) {
537 ddq
->d_ino_hardlimit
= cpu_to_be64(hard
);
538 ddq
->d_ino_softlimit
= cpu_to_be64(soft
);
540 defq
->ihardlimit
= hard
;
541 defq
->isoftlimit
= soft
;
544 xfs_debug(mp
, "ihard %Ld < isoft %Ld", hard
, soft
);
548 * Update warnings counter(s) if requested
550 if (newlim
->d_fieldmask
& QC_SPC_WARNS
)
551 ddq
->d_bwarns
= cpu_to_be16(newlim
->d_spc_warns
);
552 if (newlim
->d_fieldmask
& QC_INO_WARNS
)
553 ddq
->d_iwarns
= cpu_to_be16(newlim
->d_ino_warns
);
554 if (newlim
->d_fieldmask
& QC_RT_SPC_WARNS
)
555 ddq
->d_rtbwarns
= cpu_to_be16(newlim
->d_rt_spc_warns
);
559 * Timelimits for the super user set the relative time
560 * the other users can be over quota for this file system.
561 * If it is zero a default is used. Ditto for the default
562 * soft and hard limit values (already done, above), and
565 if (newlim
->d_fieldmask
& QC_SPC_TIMER
) {
566 q
->qi_btimelimit
= newlim
->d_spc_timer
;
567 ddq
->d_btimer
= cpu_to_be32(newlim
->d_spc_timer
);
569 if (newlim
->d_fieldmask
& QC_INO_TIMER
) {
570 q
->qi_itimelimit
= newlim
->d_ino_timer
;
571 ddq
->d_itimer
= cpu_to_be32(newlim
->d_ino_timer
);
573 if (newlim
->d_fieldmask
& QC_RT_SPC_TIMER
) {
574 q
->qi_rtbtimelimit
= newlim
->d_rt_spc_timer
;
575 ddq
->d_rtbtimer
= cpu_to_be32(newlim
->d_rt_spc_timer
);
577 if (newlim
->d_fieldmask
& QC_SPC_WARNS
)
578 q
->qi_bwarnlimit
= newlim
->d_spc_warns
;
579 if (newlim
->d_fieldmask
& QC_INO_WARNS
)
580 q
->qi_iwarnlimit
= newlim
->d_ino_warns
;
581 if (newlim
->d_fieldmask
& QC_RT_SPC_WARNS
)
582 q
->qi_rtbwarnlimit
= newlim
->d_rt_spc_warns
;
585 * If the user is now over quota, start the timelimit.
586 * The user will not be 'warned'.
587 * Note that we keep the timers ticking, whether enforcement
588 * is on or off. We don't really want to bother with iterating
589 * over all ondisk dquots and turning the timers on/off.
591 xfs_qm_adjust_dqtimers(mp
, ddq
);
593 dqp
->dq_flags
|= XFS_DQ_DIRTY
;
594 xfs_trans_log_dquot(tp
, dqp
);
596 error
= xfs_trans_commit(tp
);
601 mutex_unlock(&q
->qi_quotaofflock
);
605 /* Fill out the quota context. */
607 xfs_qm_scall_getquota_fill_qc(
608 struct xfs_mount
*mp
,
610 const struct xfs_dquot
*dqp
,
611 struct qc_dqblk
*dst
)
613 memset(dst
, 0, sizeof(*dst
));
614 dst
->d_spc_hardlimit
=
615 XFS_FSB_TO_B(mp
, be64_to_cpu(dqp
->q_core
.d_blk_hardlimit
));
616 dst
->d_spc_softlimit
=
617 XFS_FSB_TO_B(mp
, be64_to_cpu(dqp
->q_core
.d_blk_softlimit
));
618 dst
->d_ino_hardlimit
= be64_to_cpu(dqp
->q_core
.d_ino_hardlimit
);
619 dst
->d_ino_softlimit
= be64_to_cpu(dqp
->q_core
.d_ino_softlimit
);
620 dst
->d_space
= XFS_FSB_TO_B(mp
, dqp
->q_res_bcount
);
621 dst
->d_ino_count
= dqp
->q_res_icount
;
622 dst
->d_spc_timer
= be32_to_cpu(dqp
->q_core
.d_btimer
);
623 dst
->d_ino_timer
= be32_to_cpu(dqp
->q_core
.d_itimer
);
624 dst
->d_ino_warns
= be16_to_cpu(dqp
->q_core
.d_iwarns
);
625 dst
->d_spc_warns
= be16_to_cpu(dqp
->q_core
.d_bwarns
);
626 dst
->d_rt_spc_hardlimit
=
627 XFS_FSB_TO_B(mp
, be64_to_cpu(dqp
->q_core
.d_rtb_hardlimit
));
628 dst
->d_rt_spc_softlimit
=
629 XFS_FSB_TO_B(mp
, be64_to_cpu(dqp
->q_core
.d_rtb_softlimit
));
630 dst
->d_rt_space
= XFS_FSB_TO_B(mp
, dqp
->q_res_rtbcount
);
631 dst
->d_rt_spc_timer
= be32_to_cpu(dqp
->q_core
.d_rtbtimer
);
632 dst
->d_rt_spc_warns
= be16_to_cpu(dqp
->q_core
.d_rtbwarns
);
635 * Internally, we don't reset all the timers when quota enforcement
636 * gets turned off. No need to confuse the user level code,
637 * so return zeroes in that case.
639 if ((!XFS_IS_UQUOTA_ENFORCED(mp
) &&
640 dqp
->q_core
.d_flags
== XFS_DQ_USER
) ||
641 (!XFS_IS_GQUOTA_ENFORCED(mp
) &&
642 dqp
->q_core
.d_flags
== XFS_DQ_GROUP
) ||
643 (!XFS_IS_PQUOTA_ENFORCED(mp
) &&
644 dqp
->q_core
.d_flags
== XFS_DQ_PROJ
)) {
645 dst
->d_spc_timer
= 0;
646 dst
->d_ino_timer
= 0;
647 dst
->d_rt_spc_timer
= 0;
651 if (((XFS_IS_UQUOTA_ENFORCED(mp
) && type
== XFS_DQ_USER
) ||
652 (XFS_IS_GQUOTA_ENFORCED(mp
) && type
== XFS_DQ_GROUP
) ||
653 (XFS_IS_PQUOTA_ENFORCED(mp
) && type
== XFS_DQ_PROJ
)) &&
654 dqp
->q_core
.d_id
!= 0) {
655 if ((dst
->d_space
> dst
->d_spc_softlimit
) &&
656 (dst
->d_spc_softlimit
> 0)) {
657 ASSERT(dst
->d_spc_timer
!= 0);
659 if ((dst
->d_ino_count
> dst
->d_ino_softlimit
) &&
660 (dst
->d_ino_softlimit
> 0)) {
661 ASSERT(dst
->d_ino_timer
!= 0);
667 /* Return the quota information for the dquot matching id. */
669 xfs_qm_scall_getquota(
670 struct xfs_mount
*mp
,
673 struct qc_dqblk
*dst
)
675 struct xfs_dquot
*dqp
;
679 * Try to get the dquot. We don't want it allocated on disk, so don't
680 * set doalloc. If it doesn't exist, we'll get ENOENT back.
682 error
= xfs_qm_dqget(mp
, id
, type
, false, &dqp
);
687 * If everything's NULL, this dquot doesn't quite exist as far as
688 * our utility programs are concerned.
690 if (XFS_IS_DQUOT_UNINITIALIZED(dqp
)) {
695 xfs_qm_scall_getquota_fill_qc(mp
, type
, dqp
, dst
);
703 * Return the quota information for the first initialized dquot whose id
704 * is at least as high as id.
707 xfs_qm_scall_getquota_next(
708 struct xfs_mount
*mp
,
711 struct qc_dqblk
*dst
)
713 struct xfs_dquot
*dqp
;
716 error
= xfs_qm_dqget_next(mp
, *id
, type
, &dqp
);
720 /* Fill in the ID we actually read from disk */
721 *id
= be32_to_cpu(dqp
->q_core
.d_id
);
723 xfs_qm_scall_getquota_fill_qc(mp
, type
, dqp
, dst
);
731 struct xfs_inode
*ip
,
735 /* skip quota inodes */
736 if (ip
== ip
->i_mount
->m_quotainfo
->qi_uquotaip
||
737 ip
== ip
->i_mount
->m_quotainfo
->qi_gquotaip
||
738 ip
== ip
->i_mount
->m_quotainfo
->qi_pquotaip
) {
739 ASSERT(ip
->i_udquot
== NULL
);
740 ASSERT(ip
->i_gdquot
== NULL
);
741 ASSERT(ip
->i_pdquot
== NULL
);
745 xfs_ilock(ip
, XFS_ILOCK_EXCL
);
746 if ((flags
& XFS_UQUOTA_ACCT
) && ip
->i_udquot
) {
747 xfs_qm_dqrele(ip
->i_udquot
);
750 if ((flags
& XFS_GQUOTA_ACCT
) && ip
->i_gdquot
) {
751 xfs_qm_dqrele(ip
->i_gdquot
);
754 if ((flags
& XFS_PQUOTA_ACCT
) && ip
->i_pdquot
) {
755 xfs_qm_dqrele(ip
->i_pdquot
);
758 xfs_iunlock(ip
, XFS_ILOCK_EXCL
);
764 * Go thru all the inodes in the file system, releasing their dquots.
766 * Note that the mount structure gets modified to indicate that quotas are off
767 * AFTER this, in the case of quotaoff.
770 xfs_qm_dqrele_all_inodes(
771 struct xfs_mount
*mp
,
774 ASSERT(mp
->m_quotainfo
);
775 xfs_inode_ag_iterator_flags(mp
, xfs_dqrele_inode
, flags
, NULL
,
776 XFS_AGITER_INEW_WAIT
);