2 * Copyright (c) 2000-2002 Silicon Graphics, Inc.
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of the GNU General Public License as
7 * published by the Free Software Foundation.
9 * This program is distributed in the hope that it would be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write the Free Software Foundation,
16 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
21 #include "xfs_trans.h"
24 #include "xfs_alloc.h"
25 #include "xfs_quota.h"
26 #include "xfs_mount.h"
27 #include "xfs_bmap_btree.h"
28 #include "xfs_inode.h"
29 #include "xfs_itable.h"
31 #include "xfs_rtalloc.h"
32 #include "xfs_error.h"
34 #include "xfs_buf_item.h"
35 #include "xfs_trans_priv.h"
38 STATIC
void xfs_trans_alloc_dqinfo(xfs_trans_t
*);
41 * Add the locked dquot to the transaction.
42 * The dquot must be locked, and it cannot be associated with any
50 ASSERT(dqp
->q_transp
!= tp
);
51 ASSERT(XFS_DQ_IS_LOCKED(dqp
));
52 ASSERT(dqp
->q_logitem
.qli_dquot
== dqp
);
55 * Get a log_item_desc to point at the new item.
57 xfs_trans_add_item(tp
, &dqp
->q_logitem
.qli_item
);
60 * Initialize d_transp so we can later determine if this dquot is
61 * associated with this transaction.
68 * This is called to mark the dquot as needing
69 * to be logged when the transaction is committed. The dquot must
70 * already be associated with the given transaction.
71 * Note that it marks the entire transaction as dirty. In the ordinary
72 * case, this gets called via xfs_trans_commit, after the transaction
73 * is already dirty. However, there's nothing stop this from getting
74 * called directly, as done by xfs_qm_scall_setqlim. Hence, the TRANS_DIRTY
82 ASSERT(dqp
->q_transp
== tp
);
83 ASSERT(XFS_DQ_IS_LOCKED(dqp
));
85 tp
->t_flags
|= XFS_TRANS_DIRTY
;
86 dqp
->q_logitem
.qli_item
.li_desc
->lid_flags
|= XFS_LID_DIRTY
;
90 * Carry forward whatever is left of the quota blk reservation to
91 * the spanky new transaction
100 xfs_dqtrx_t
*oqa
, *nqa
;
105 xfs_trans_alloc_dqinfo(ntp
);
106 oqa
= otp
->t_dqinfo
->dqa_usrdquots
;
107 nqa
= ntp
->t_dqinfo
->dqa_usrdquots
;
110 * Because the quota blk reservation is carried forward,
111 * it is also necessary to carry forward the DQ_DIRTY flag.
113 if(otp
->t_flags
& XFS_TRANS_DQ_DIRTY
)
114 ntp
->t_flags
|= XFS_TRANS_DQ_DIRTY
;
116 for (j
= 0; j
< 2; j
++) {
117 for (i
= 0; i
< XFS_QM_TRANS_MAXDQS
; i
++) {
118 if (oqa
[i
].qt_dquot
== NULL
)
123 nq
->qt_dquot
= oq
->qt_dquot
;
124 nq
->qt_bcount_delta
= nq
->qt_icount_delta
= 0;
125 nq
->qt_rtbcount_delta
= 0;
128 * Transfer whatever is left of the reservations.
130 nq
->qt_blk_res
= oq
->qt_blk_res
- oq
->qt_blk_res_used
;
131 oq
->qt_blk_res
= oq
->qt_blk_res_used
;
133 nq
->qt_rtblk_res
= oq
->qt_rtblk_res
-
134 oq
->qt_rtblk_res_used
;
135 oq
->qt_rtblk_res
= oq
->qt_rtblk_res_used
;
137 nq
->qt_ino_res
= oq
->qt_ino_res
- oq
->qt_ino_res_used
;
138 oq
->qt_ino_res
= oq
->qt_ino_res_used
;
141 oqa
= otp
->t_dqinfo
->dqa_grpdquots
;
142 nqa
= ntp
->t_dqinfo
->dqa_grpdquots
;
147 * Wrap around mod_dquot to account for both user and group quotas.
150 xfs_trans_mod_dquot_byino(
156 xfs_mount_t
*mp
= tp
->t_mountp
;
158 if (!XFS_IS_QUOTA_RUNNING(mp
) ||
159 !XFS_IS_QUOTA_ON(mp
) ||
160 ip
->i_ino
== mp
->m_sb
.sb_uquotino
||
161 ip
->i_ino
== mp
->m_sb
.sb_gquotino
)
164 if (tp
->t_dqinfo
== NULL
)
165 xfs_trans_alloc_dqinfo(tp
);
167 if (XFS_IS_UQUOTA_ON(mp
) && ip
->i_udquot
)
168 (void) xfs_trans_mod_dquot(tp
, ip
->i_udquot
, field
, delta
);
169 if (XFS_IS_OQUOTA_ON(mp
) && ip
->i_gdquot
)
170 (void) xfs_trans_mod_dquot(tp
, ip
->i_gdquot
, field
, delta
);
181 qa
= XFS_QM_ISUDQ(dqp
) ?
182 tp
->t_dqinfo
->dqa_usrdquots
: tp
->t_dqinfo
->dqa_grpdquots
;
184 for (i
= 0; i
< XFS_QM_TRANS_MAXDQS
; i
++) {
185 if (qa
[i
].qt_dquot
== NULL
||
186 qa
[i
].qt_dquot
== dqp
)
194 * Make the changes in the transaction structure.
195 * The moral equivalent to xfs_trans_mod_sb().
196 * We don't touch any fields in the dquot, so we don't care
197 * if it's locked or not (most of the time it won't be).
209 ASSERT(XFS_IS_QUOTA_RUNNING(tp
->t_mountp
));
212 if (tp
->t_dqinfo
== NULL
)
213 xfs_trans_alloc_dqinfo(tp
);
215 * Find either the first free slot or the slot that belongs
218 qtrx
= xfs_trans_get_dqtrx(tp
, dqp
);
220 if (qtrx
->qt_dquot
== NULL
)
221 qtrx
->qt_dquot
= dqp
;
226 * regular disk blk reservation
228 case XFS_TRANS_DQ_RES_BLKS
:
229 qtrx
->qt_blk_res
+= (ulong
)delta
;
235 case XFS_TRANS_DQ_RES_INOS
:
236 qtrx
->qt_ino_res
+= (ulong
)delta
;
242 case XFS_TRANS_DQ_BCOUNT
:
243 if (qtrx
->qt_blk_res
&& delta
> 0) {
244 qtrx
->qt_blk_res_used
+= (ulong
)delta
;
245 ASSERT(qtrx
->qt_blk_res
>= qtrx
->qt_blk_res_used
);
247 qtrx
->qt_bcount_delta
+= delta
;
250 case XFS_TRANS_DQ_DELBCOUNT
:
251 qtrx
->qt_delbcnt_delta
+= delta
;
257 case XFS_TRANS_DQ_ICOUNT
:
258 if (qtrx
->qt_ino_res
&& delta
> 0) {
259 qtrx
->qt_ino_res_used
+= (ulong
)delta
;
260 ASSERT(qtrx
->qt_ino_res
>= qtrx
->qt_ino_res_used
);
262 qtrx
->qt_icount_delta
+= delta
;
268 case XFS_TRANS_DQ_RES_RTBLKS
:
269 qtrx
->qt_rtblk_res
+= (ulong
)delta
;
275 case XFS_TRANS_DQ_RTBCOUNT
:
276 if (qtrx
->qt_rtblk_res
&& delta
> 0) {
277 qtrx
->qt_rtblk_res_used
+= (ulong
)delta
;
278 ASSERT(qtrx
->qt_rtblk_res
>= qtrx
->qt_rtblk_res_used
);
280 qtrx
->qt_rtbcount_delta
+= delta
;
283 case XFS_TRANS_DQ_DELRTBCOUNT
:
284 qtrx
->qt_delrtb_delta
+= delta
;
290 tp
->t_flags
|= XFS_TRANS_DQ_DIRTY
;
295 * Given an array of dqtrx structures, lock all the dquots associated
296 * and join them to the transaction, provided they have been modified.
297 * We know that the highest number of dquots (of one type - usr OR grp),
298 * involved in a transaction is 2 and that both usr and grp combined - 3.
299 * So, we don't attempt to make this very generic.
302 xfs_trans_dqlockedjoin(
306 ASSERT(q
[0].qt_dquot
!= NULL
);
307 if (q
[1].qt_dquot
== NULL
) {
308 xfs_dqlock(q
[0].qt_dquot
);
309 xfs_trans_dqjoin(tp
, q
[0].qt_dquot
);
311 ASSERT(XFS_QM_TRANS_MAXDQS
== 2);
312 xfs_dqlock2(q
[0].qt_dquot
, q
[1].qt_dquot
);
313 xfs_trans_dqjoin(tp
, q
[0].qt_dquot
);
314 xfs_trans_dqjoin(tp
, q
[1].qt_dquot
);
320 * Called by xfs_trans_commit() and similar in spirit to
321 * xfs_trans_apply_sb_deltas().
322 * Go thru all the dquots belonging to this transaction and modify the
323 * INCORE dquot to reflect the actual usages.
324 * Unreserve just the reservations done by this transaction.
325 * dquot is still left locked at exit.
328 xfs_trans_apply_dquot_deltas(
333 xfs_dqtrx_t
*qtrx
, *qa
;
338 if (!(tp
->t_flags
& XFS_TRANS_DQ_DIRTY
))
341 ASSERT(tp
->t_dqinfo
);
342 qa
= tp
->t_dqinfo
->dqa_usrdquots
;
343 for (j
= 0; j
< 2; j
++) {
344 if (qa
[0].qt_dquot
== NULL
) {
345 qa
= tp
->t_dqinfo
->dqa_grpdquots
;
350 * Lock all of the dquots and join them to the transaction.
352 xfs_trans_dqlockedjoin(tp
, qa
);
354 for (i
= 0; i
< XFS_QM_TRANS_MAXDQS
; i
++) {
357 * The array of dquots is filled
358 * sequentially, not sparsely.
360 if ((dqp
= qtrx
->qt_dquot
) == NULL
)
363 ASSERT(XFS_DQ_IS_LOCKED(dqp
));
364 ASSERT(dqp
->q_transp
== tp
);
367 * adjust the actual number of blocks used
372 * The issue here is - sometimes we don't make a blkquota
373 * reservation intentionally to be fair to users
374 * (when the amount is small). On the other hand,
375 * delayed allocs do make reservations, but that's
376 * outside of a transaction, so we have no
377 * idea how much was really reserved.
378 * So, here we've accumulated delayed allocation blks and
379 * non-delay blks. The assumption is that the
380 * delayed ones are always reserved (outside of a
381 * transaction), and the others may or may not have
382 * quota reservations.
384 totalbdelta
= qtrx
->qt_bcount_delta
+
385 qtrx
->qt_delbcnt_delta
;
386 totalrtbdelta
= qtrx
->qt_rtbcount_delta
+
387 qtrx
->qt_delrtb_delta
;
390 ASSERT(be64_to_cpu(d
->d_bcount
) >=
393 if (totalrtbdelta
< 0)
394 ASSERT(be64_to_cpu(d
->d_rtbcount
) >=
397 if (qtrx
->qt_icount_delta
< 0)
398 ASSERT(be64_to_cpu(d
->d_icount
) >=
399 -qtrx
->qt_icount_delta
);
402 be64_add_cpu(&d
->d_bcount
, (xfs_qcnt_t
)totalbdelta
);
404 if (qtrx
->qt_icount_delta
)
405 be64_add_cpu(&d
->d_icount
, (xfs_qcnt_t
)qtrx
->qt_icount_delta
);
408 be64_add_cpu(&d
->d_rtbcount
, (xfs_qcnt_t
)totalrtbdelta
);
411 * Get any default limits in use.
412 * Start/reset the timer(s) if needed.
415 xfs_qm_adjust_dqlimits(tp
->t_mountp
, d
);
416 xfs_qm_adjust_dqtimers(tp
->t_mountp
, d
);
419 dqp
->dq_flags
|= XFS_DQ_DIRTY
;
421 * add this to the list of items to get logged
423 xfs_trans_log_dquot(tp
, dqp
);
425 * Take off what's left of the original reservation.
426 * In case of delayed allocations, there's no
427 * reservation that a transaction structure knows of.
429 if (qtrx
->qt_blk_res
!= 0) {
430 if (qtrx
->qt_blk_res
!= qtrx
->qt_blk_res_used
) {
431 if (qtrx
->qt_blk_res
>
432 qtrx
->qt_blk_res_used
)
433 dqp
->q_res_bcount
-= (xfs_qcnt_t
)
435 qtrx
->qt_blk_res_used
);
437 dqp
->q_res_bcount
-= (xfs_qcnt_t
)
438 (qtrx
->qt_blk_res_used
-
443 * These blks were never reserved, either inside
444 * a transaction or outside one (in a delayed
445 * allocation). Also, this isn't always a
446 * negative number since we sometimes
447 * deliberately skip quota reservations.
449 if (qtrx
->qt_bcount_delta
) {
451 (xfs_qcnt_t
)qtrx
->qt_bcount_delta
;
455 * Adjust the RT reservation.
457 if (qtrx
->qt_rtblk_res
!= 0) {
458 if (qtrx
->qt_rtblk_res
!= qtrx
->qt_rtblk_res_used
) {
459 if (qtrx
->qt_rtblk_res
>
460 qtrx
->qt_rtblk_res_used
)
461 dqp
->q_res_rtbcount
-= (xfs_qcnt_t
)
462 (qtrx
->qt_rtblk_res
-
463 qtrx
->qt_rtblk_res_used
);
465 dqp
->q_res_rtbcount
-= (xfs_qcnt_t
)
466 (qtrx
->qt_rtblk_res_used
-
470 if (qtrx
->qt_rtbcount_delta
)
471 dqp
->q_res_rtbcount
+=
472 (xfs_qcnt_t
)qtrx
->qt_rtbcount_delta
;
476 * Adjust the inode reservation.
478 if (qtrx
->qt_ino_res
!= 0) {
479 ASSERT(qtrx
->qt_ino_res
>=
480 qtrx
->qt_ino_res_used
);
481 if (qtrx
->qt_ino_res
> qtrx
->qt_ino_res_used
)
482 dqp
->q_res_icount
-= (xfs_qcnt_t
)
484 qtrx
->qt_ino_res_used
);
486 if (qtrx
->qt_icount_delta
)
488 (xfs_qcnt_t
)qtrx
->qt_icount_delta
;
491 ASSERT(dqp
->q_res_bcount
>=
492 be64_to_cpu(dqp
->q_core
.d_bcount
));
493 ASSERT(dqp
->q_res_icount
>=
494 be64_to_cpu(dqp
->q_core
.d_icount
));
495 ASSERT(dqp
->q_res_rtbcount
>=
496 be64_to_cpu(dqp
->q_core
.d_rtbcount
));
499 * Do the group quotas next
501 qa
= tp
->t_dqinfo
->dqa_grpdquots
;
506 * Release the reservations, and adjust the dquots accordingly.
507 * This is called only when the transaction is being aborted. If by
508 * any chance we have done dquot modifications incore (ie. deltas) already,
509 * we simply throw those away, since that's the expected behavior
510 * when a transaction is curtailed without a commit.
513 xfs_trans_unreserve_and_mod_dquots(
518 xfs_dqtrx_t
*qtrx
, *qa
;
521 if (!tp
->t_dqinfo
|| !(tp
->t_flags
& XFS_TRANS_DQ_DIRTY
))
524 qa
= tp
->t_dqinfo
->dqa_usrdquots
;
526 for (j
= 0; j
< 2; j
++) {
527 for (i
= 0; i
< XFS_QM_TRANS_MAXDQS
; i
++) {
530 * We assume that the array of dquots is filled
531 * sequentially, not sparsely.
533 if ((dqp
= qtrx
->qt_dquot
) == NULL
)
536 * Unreserve the original reservation. We don't care
537 * about the number of blocks used field, or deltas.
538 * Also we don't bother to zero the fields.
541 if (qtrx
->qt_blk_res
) {
545 (xfs_qcnt_t
)qtrx
->qt_blk_res
;
547 if (qtrx
->qt_ino_res
) {
553 (xfs_qcnt_t
)qtrx
->qt_ino_res
;
556 if (qtrx
->qt_rtblk_res
) {
561 dqp
->q_res_rtbcount
-=
562 (xfs_qcnt_t
)qtrx
->qt_rtblk_res
;
568 qa
= tp
->t_dqinfo
->dqa_grpdquots
;
574 struct xfs_mount
*mp
,
575 struct xfs_dquot
*dqp
,
578 /* no warnings for project quotas - we just return ENOSPC later */
579 if (dqp
->dq_flags
& XFS_DQ_PROJ
)
581 quota_send_warning(make_kqid(&init_user_ns
,
582 (dqp
->dq_flags
& XFS_DQ_USER
) ?
584 be32_to_cpu(dqp
->q_core
.d_id
)),
585 mp
->m_super
->s_dev
, type
);
589 * This reserves disk blocks and inodes against a dquot.
590 * Flags indicate if the dquot is to be locked here and also
591 * if the blk reservation is for RT or regular blocks.
592 * Sending in XFS_QMOPT_FORCE_RES flag skips the quota check.
603 xfs_qcnt_t hardlimit
;
604 xfs_qcnt_t softlimit
;
606 xfs_qwarncnt_t warns
;
607 xfs_qwarncnt_t warnlimit
;
608 xfs_qcnt_t total_count
;
609 xfs_qcnt_t
*resbcountp
;
610 xfs_quotainfo_t
*q
= mp
->m_quotainfo
;
615 if (flags
& XFS_TRANS_DQ_RES_BLKS
) {
616 hardlimit
= be64_to_cpu(dqp
->q_core
.d_blk_hardlimit
);
618 hardlimit
= q
->qi_bhardlimit
;
619 softlimit
= be64_to_cpu(dqp
->q_core
.d_blk_softlimit
);
621 softlimit
= q
->qi_bsoftlimit
;
622 timer
= be32_to_cpu(dqp
->q_core
.d_btimer
);
623 warns
= be16_to_cpu(dqp
->q_core
.d_bwarns
);
624 warnlimit
= dqp
->q_mount
->m_quotainfo
->qi_bwarnlimit
;
625 resbcountp
= &dqp
->q_res_bcount
;
627 ASSERT(flags
& XFS_TRANS_DQ_RES_RTBLKS
);
628 hardlimit
= be64_to_cpu(dqp
->q_core
.d_rtb_hardlimit
);
630 hardlimit
= q
->qi_rtbhardlimit
;
631 softlimit
= be64_to_cpu(dqp
->q_core
.d_rtb_softlimit
);
633 softlimit
= q
->qi_rtbsoftlimit
;
634 timer
= be32_to_cpu(dqp
->q_core
.d_rtbtimer
);
635 warns
= be16_to_cpu(dqp
->q_core
.d_rtbwarns
);
636 warnlimit
= dqp
->q_mount
->m_quotainfo
->qi_rtbwarnlimit
;
637 resbcountp
= &dqp
->q_res_rtbcount
;
640 if ((flags
& XFS_QMOPT_FORCE_RES
) == 0 &&
642 ((XFS_IS_UQUOTA_ENFORCED(dqp
->q_mount
) && XFS_QM_ISUDQ(dqp
)) ||
643 (XFS_IS_OQUOTA_ENFORCED(dqp
->q_mount
) &&
644 (XFS_QM_ISPDQ(dqp
) || XFS_QM_ISGDQ(dqp
))))) {
647 * dquot is locked already. See if we'd go over the
648 * hardlimit or exceed the timelimit if we allocate
651 total_count
= *resbcountp
+ nblks
;
652 if (hardlimit
&& total_count
> hardlimit
) {
653 xfs_quota_warn(mp
, dqp
, QUOTA_NL_BHARDWARN
);
656 if (softlimit
&& total_count
> softlimit
) {
657 if ((timer
!= 0 && get_seconds() > timer
) ||
658 (warns
!= 0 && warns
>= warnlimit
)) {
659 xfs_quota_warn(mp
, dqp
,
660 QUOTA_NL_BSOFTLONGWARN
);
664 xfs_quota_warn(mp
, dqp
, QUOTA_NL_BSOFTWARN
);
668 total_count
= be64_to_cpu(dqp
->q_core
.d_icount
) + ninos
;
669 timer
= be32_to_cpu(dqp
->q_core
.d_itimer
);
670 warns
= be16_to_cpu(dqp
->q_core
.d_iwarns
);
671 warnlimit
= dqp
->q_mount
->m_quotainfo
->qi_iwarnlimit
;
672 hardlimit
= be64_to_cpu(dqp
->q_core
.d_ino_hardlimit
);
674 hardlimit
= q
->qi_ihardlimit
;
675 softlimit
= be64_to_cpu(dqp
->q_core
.d_ino_softlimit
);
677 softlimit
= q
->qi_isoftlimit
;
679 if (hardlimit
&& total_count
> hardlimit
) {
680 xfs_quota_warn(mp
, dqp
, QUOTA_NL_IHARDWARN
);
683 if (softlimit
&& total_count
> softlimit
) {
684 if ((timer
!= 0 && get_seconds() > timer
) ||
685 (warns
!= 0 && warns
>= warnlimit
)) {
686 xfs_quota_warn(mp
, dqp
,
687 QUOTA_NL_ISOFTLONGWARN
);
690 xfs_quota_warn(mp
, dqp
, QUOTA_NL_ISOFTWARN
);
696 * Change the reservation, but not the actual usage.
697 * Note that q_res_bcount = q_core.d_bcount + resv
699 (*resbcountp
) += (xfs_qcnt_t
)nblks
;
701 dqp
->q_res_icount
+= (xfs_qcnt_t
)ninos
;
704 * note the reservation amt in the trans struct too,
705 * so that the transaction knows how much was reserved by
706 * it against this particular dquot.
707 * We don't do this when we are reserving for a delayed allocation,
708 * because we don't have the luxury of a transaction envelope then.
711 ASSERT(tp
->t_dqinfo
);
712 ASSERT(flags
& XFS_QMOPT_RESBLK_MASK
);
714 xfs_trans_mod_dquot(tp
, dqp
,
715 flags
& XFS_QMOPT_RESBLK_MASK
,
718 xfs_trans_mod_dquot(tp
, dqp
,
719 XFS_TRANS_DQ_RES_INOS
,
722 ASSERT(dqp
->q_res_bcount
>= be64_to_cpu(dqp
->q_core
.d_bcount
));
723 ASSERT(dqp
->q_res_rtbcount
>= be64_to_cpu(dqp
->q_core
.d_rtbcount
));
724 ASSERT(dqp
->q_res_icount
>= be64_to_cpu(dqp
->q_core
.d_icount
));
731 if (flags
& XFS_QMOPT_ENOSPC
)
738 * Given dquot(s), make disk block and/or inode reservations against them.
739 * The fact that this does the reservation against both the usr and
740 * grp/prj quotas is important, because this follows a both-or-nothing
743 * flags = XFS_QMOPT_FORCE_RES evades limit enforcement. Used by chown.
744 * XFS_QMOPT_ENOSPC returns ENOSPC not EDQUOT. Used by pquota.
745 * XFS_TRANS_DQ_RES_BLKS reserves regular disk blocks
746 * XFS_TRANS_DQ_RES_RTBLKS reserves realtime disk blocks
747 * dquots are unlocked on return, if they were not locked by caller.
750 xfs_trans_reserve_quota_bydquots(
759 int resvd
= 0, error
;
761 if (!XFS_IS_QUOTA_RUNNING(mp
) || !XFS_IS_QUOTA_ON(mp
))
764 if (tp
&& tp
->t_dqinfo
== NULL
)
765 xfs_trans_alloc_dqinfo(tp
);
767 ASSERT(flags
& XFS_QMOPT_RESBLK_MASK
);
770 error
= xfs_trans_dqresv(tp
, mp
, udqp
, nblks
, ninos
,
771 (flags
& ~XFS_QMOPT_ENOSPC
));
778 error
= xfs_trans_dqresv(tp
, mp
, gdqp
, nblks
, ninos
, flags
);
781 * can't do it, so backout previous reservation
784 flags
|= XFS_QMOPT_FORCE_RES
;
785 xfs_trans_dqresv(tp
, mp
, udqp
,
786 -nblks
, -ninos
, flags
);
793 * Didn't change anything critical, so, no need to log
800 * Lock the dquot and change the reservation if we can.
801 * This doesn't change the actual usage, just the reservation.
802 * The inode sent in is locked.
805 xfs_trans_reserve_quota_nblks(
806 struct xfs_trans
*tp
,
807 struct xfs_inode
*ip
,
812 struct xfs_mount
*mp
= ip
->i_mount
;
814 if (!XFS_IS_QUOTA_RUNNING(mp
) || !XFS_IS_QUOTA_ON(mp
))
816 if (XFS_IS_PQUOTA_ON(mp
))
817 flags
|= XFS_QMOPT_ENOSPC
;
819 ASSERT(ip
->i_ino
!= mp
->m_sb
.sb_uquotino
);
820 ASSERT(ip
->i_ino
!= mp
->m_sb
.sb_gquotino
);
822 ASSERT(xfs_isilocked(ip
, XFS_ILOCK_EXCL
));
823 ASSERT((flags
& ~(XFS_QMOPT_FORCE_RES
| XFS_QMOPT_ENOSPC
)) ==
824 XFS_TRANS_DQ_RES_RTBLKS
||
825 (flags
& ~(XFS_QMOPT_FORCE_RES
| XFS_QMOPT_ENOSPC
)) ==
826 XFS_TRANS_DQ_RES_BLKS
);
829 * Reserve nblks against these dquots, with trans as the mediator.
831 return xfs_trans_reserve_quota_bydquots(tp
, mp
,
832 ip
->i_udquot
, ip
->i_gdquot
,
833 nblks
, ninos
, flags
);
837 * This routine is called to allocate a quotaoff log item.
840 xfs_trans_get_qoff_item(
842 xfs_qoff_logitem_t
*startqoff
,
845 xfs_qoff_logitem_t
*q
;
849 q
= xfs_qm_qoff_logitem_init(tp
->t_mountp
, startqoff
, flags
);
853 * Get a log_item_desc to point at the new item.
855 xfs_trans_add_item(tp
, &q
->qql_item
);
861 * This is called to mark the quotaoff logitem as needing
862 * to be logged when the transaction is committed. The logitem must
863 * already be associated with the given transaction.
866 xfs_trans_log_quotaoff_item(
868 xfs_qoff_logitem_t
*qlp
)
870 tp
->t_flags
|= XFS_TRANS_DIRTY
;
871 qlp
->qql_item
.li_desc
->lid_flags
|= XFS_LID_DIRTY
;
875 xfs_trans_alloc_dqinfo(
878 tp
->t_dqinfo
= kmem_zone_zalloc(xfs_qm_dqtrxzone
, KM_SLEEP
);
882 xfs_trans_free_dqinfo(
887 kmem_zone_free(xfs_qm_dqtrxzone
, tp
->t_dqinfo
);