2 * Copyright (c) 2000-2002 Silicon Graphics, Inc.
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of the GNU General Public License as
7 * published by the Free Software Foundation.
9 * This program is distributed in the hope that it would be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write the Free Software Foundation,
16 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
20 #include "xfs_shared.h"
21 #include "xfs_format.h"
22 #include "xfs_log_format.h"
23 #include "xfs_trans_resv.h"
24 #include "xfs_mount.h"
25 #include "xfs_inode.h"
26 #include "xfs_error.h"
27 #include "xfs_trans.h"
28 #include "xfs_trans_priv.h"
29 #include "xfs_quota.h"
32 STATIC
void xfs_trans_alloc_dqinfo(xfs_trans_t
*);
35 * Add the locked dquot to the transaction.
36 * The dquot must be locked, and it cannot be associated with any
44 ASSERT(dqp
->q_transp
!= tp
);
45 ASSERT(XFS_DQ_IS_LOCKED(dqp
));
46 ASSERT(dqp
->q_logitem
.qli_dquot
== dqp
);
49 * Get a log_item_desc to point at the new item.
51 xfs_trans_add_item(tp
, &dqp
->q_logitem
.qli_item
);
54 * Initialize d_transp so we can later determine if this dquot is
55 * associated with this transaction.
62 * This is called to mark the dquot as needing
63 * to be logged when the transaction is committed. The dquot must
64 * already be associated with the given transaction.
65 * Note that it marks the entire transaction as dirty. In the ordinary
66 * case, this gets called via xfs_trans_commit, after the transaction
67 * is already dirty. However, there's nothing stop this from getting
68 * called directly, as done by xfs_qm_scall_setqlim. Hence, the TRANS_DIRTY
76 ASSERT(dqp
->q_transp
== tp
);
77 ASSERT(XFS_DQ_IS_LOCKED(dqp
));
79 tp
->t_flags
|= XFS_TRANS_DIRTY
;
80 dqp
->q_logitem
.qli_item
.li_desc
->lid_flags
|= XFS_LID_DIRTY
;
84 * Carry forward whatever is left of the quota blk reservation to
85 * the spanky new transaction
94 xfs_dqtrx_t
*oqa
, *nqa
;
100 xfs_trans_alloc_dqinfo(ntp
);
103 * Because the quota blk reservation is carried forward,
104 * it is also necessary to carry forward the DQ_DIRTY flag.
106 if (otp
->t_flags
& XFS_TRANS_DQ_DIRTY
)
107 ntp
->t_flags
|= XFS_TRANS_DQ_DIRTY
;
109 for (j
= 0; j
< XFS_QM_TRANS_DQTYPES
; j
++) {
110 oqa
= otp
->t_dqinfo
->dqs
[j
];
111 nqa
= ntp
->t_dqinfo
->dqs
[j
];
112 for (i
= 0; i
< XFS_QM_TRANS_MAXDQS
; i
++) {
115 if (oqa
[i
].qt_dquot
== NULL
)
120 if (oq
->qt_blk_res
&& oq
->qt_bcount_delta
> 0)
121 blk_res_used
= oq
->qt_bcount_delta
;
123 nq
->qt_dquot
= oq
->qt_dquot
;
124 nq
->qt_bcount_delta
= nq
->qt_icount_delta
= 0;
125 nq
->qt_rtbcount_delta
= 0;
128 * Transfer whatever is left of the reservations.
130 nq
->qt_blk_res
= oq
->qt_blk_res
- blk_res_used
;
131 oq
->qt_blk_res
= blk_res_used
;
133 nq
->qt_rtblk_res
= oq
->qt_rtblk_res
-
134 oq
->qt_rtblk_res_used
;
135 oq
->qt_rtblk_res
= oq
->qt_rtblk_res_used
;
137 nq
->qt_ino_res
= oq
->qt_ino_res
- oq
->qt_ino_res_used
;
138 oq
->qt_ino_res
= oq
->qt_ino_res_used
;
145 * Wrap around mod_dquot to account for both user and group quotas.
148 xfs_trans_mod_dquot_byino(
154 xfs_mount_t
*mp
= tp
->t_mountp
;
156 if (!XFS_IS_QUOTA_RUNNING(mp
) ||
157 !XFS_IS_QUOTA_ON(mp
) ||
158 xfs_is_quota_inode(&mp
->m_sb
, ip
->i_ino
))
161 if (tp
->t_dqinfo
== NULL
)
162 xfs_trans_alloc_dqinfo(tp
);
164 if (XFS_IS_UQUOTA_ON(mp
) && ip
->i_udquot
)
165 (void) xfs_trans_mod_dquot(tp
, ip
->i_udquot
, field
, delta
);
166 if (XFS_IS_GQUOTA_ON(mp
) && ip
->i_gdquot
)
167 (void) xfs_trans_mod_dquot(tp
, ip
->i_gdquot
, field
, delta
);
168 if (XFS_IS_PQUOTA_ON(mp
) && ip
->i_pdquot
)
169 (void) xfs_trans_mod_dquot(tp
, ip
->i_pdquot
, field
, delta
);
172 STATIC
struct xfs_dqtrx
*
174 struct xfs_trans
*tp
,
175 struct xfs_dquot
*dqp
)
178 struct xfs_dqtrx
*qa
;
180 if (XFS_QM_ISUDQ(dqp
))
181 qa
= tp
->t_dqinfo
->dqs
[XFS_QM_TRANS_USR
];
182 else if (XFS_QM_ISGDQ(dqp
))
183 qa
= tp
->t_dqinfo
->dqs
[XFS_QM_TRANS_GRP
];
184 else if (XFS_QM_ISPDQ(dqp
))
185 qa
= tp
->t_dqinfo
->dqs
[XFS_QM_TRANS_PRJ
];
189 for (i
= 0; i
< XFS_QM_TRANS_MAXDQS
; i
++) {
190 if (qa
[i
].qt_dquot
== NULL
||
191 qa
[i
].qt_dquot
== dqp
)
199 * Make the changes in the transaction structure.
200 * The moral equivalent to xfs_trans_mod_sb().
201 * We don't touch any fields in the dquot, so we don't care
202 * if it's locked or not (most of the time it won't be).
214 ASSERT(XFS_IS_QUOTA_RUNNING(tp
->t_mountp
));
217 if (tp
->t_dqinfo
== NULL
)
218 xfs_trans_alloc_dqinfo(tp
);
220 * Find either the first free slot or the slot that belongs
223 qtrx
= xfs_trans_get_dqtrx(tp
, dqp
);
225 if (qtrx
->qt_dquot
== NULL
)
226 qtrx
->qt_dquot
= dqp
;
231 * regular disk blk reservation
233 case XFS_TRANS_DQ_RES_BLKS
:
234 qtrx
->qt_blk_res
+= (ulong
)delta
;
240 case XFS_TRANS_DQ_RES_INOS
:
241 qtrx
->qt_ino_res
+= (ulong
)delta
;
247 case XFS_TRANS_DQ_BCOUNT
:
248 qtrx
->qt_bcount_delta
+= delta
;
251 case XFS_TRANS_DQ_DELBCOUNT
:
252 qtrx
->qt_delbcnt_delta
+= delta
;
258 case XFS_TRANS_DQ_ICOUNT
:
259 if (qtrx
->qt_ino_res
&& delta
> 0) {
260 qtrx
->qt_ino_res_used
+= (ulong
)delta
;
261 ASSERT(qtrx
->qt_ino_res
>= qtrx
->qt_ino_res_used
);
263 qtrx
->qt_icount_delta
+= delta
;
269 case XFS_TRANS_DQ_RES_RTBLKS
:
270 qtrx
->qt_rtblk_res
+= (ulong
)delta
;
276 case XFS_TRANS_DQ_RTBCOUNT
:
277 if (qtrx
->qt_rtblk_res
&& delta
> 0) {
278 qtrx
->qt_rtblk_res_used
+= (ulong
)delta
;
279 ASSERT(qtrx
->qt_rtblk_res
>= qtrx
->qt_rtblk_res_used
);
281 qtrx
->qt_rtbcount_delta
+= delta
;
284 case XFS_TRANS_DQ_DELRTBCOUNT
:
285 qtrx
->qt_delrtb_delta
+= delta
;
291 tp
->t_flags
|= XFS_TRANS_DQ_DIRTY
;
296 * Given an array of dqtrx structures, lock all the dquots associated and join
297 * them to the transaction, provided they have been modified. We know that the
298 * highest number of dquots of one type - usr, grp and prj - involved in a
299 * transaction is 3 so we don't need to make this very generic.
302 xfs_trans_dqlockedjoin(
306 ASSERT(q
[0].qt_dquot
!= NULL
);
307 if (q
[1].qt_dquot
== NULL
) {
308 xfs_dqlock(q
[0].qt_dquot
);
309 xfs_trans_dqjoin(tp
, q
[0].qt_dquot
);
311 ASSERT(XFS_QM_TRANS_MAXDQS
== 2);
312 xfs_dqlock2(q
[0].qt_dquot
, q
[1].qt_dquot
);
313 xfs_trans_dqjoin(tp
, q
[0].qt_dquot
);
314 xfs_trans_dqjoin(tp
, q
[1].qt_dquot
);
320 * Called by xfs_trans_commit() and similar in spirit to
321 * xfs_trans_apply_sb_deltas().
322 * Go thru all the dquots belonging to this transaction and modify the
323 * INCORE dquot to reflect the actual usages.
324 * Unreserve just the reservations done by this transaction.
325 * dquot is still left locked at exit.
328 xfs_trans_apply_dquot_deltas(
329 struct xfs_trans
*tp
)
332 struct xfs_dquot
*dqp
;
333 struct xfs_dqtrx
*qtrx
, *qa
;
334 struct xfs_disk_dquot
*d
;
338 if (!(tp
->t_flags
& XFS_TRANS_DQ_DIRTY
))
341 ASSERT(tp
->t_dqinfo
);
342 for (j
= 0; j
< XFS_QM_TRANS_DQTYPES
; j
++) {
343 qa
= tp
->t_dqinfo
->dqs
[j
];
344 if (qa
[0].qt_dquot
== NULL
)
348 * Lock all of the dquots and join them to the transaction.
350 xfs_trans_dqlockedjoin(tp
, qa
);
352 for (i
= 0; i
< XFS_QM_TRANS_MAXDQS
; i
++) {
355 * The array of dquots is filled
356 * sequentially, not sparsely.
358 if ((dqp
= qtrx
->qt_dquot
) == NULL
)
361 ASSERT(XFS_DQ_IS_LOCKED(dqp
));
362 ASSERT(dqp
->q_transp
== tp
);
365 * adjust the actual number of blocks used
370 * The issue here is - sometimes we don't make a blkquota
371 * reservation intentionally to be fair to users
372 * (when the amount is small). On the other hand,
373 * delayed allocs do make reservations, but that's
374 * outside of a transaction, so we have no
375 * idea how much was really reserved.
376 * So, here we've accumulated delayed allocation blks and
377 * non-delay blks. The assumption is that the
378 * delayed ones are always reserved (outside of a
379 * transaction), and the others may or may not have
380 * quota reservations.
382 totalbdelta
= qtrx
->qt_bcount_delta
+
383 qtrx
->qt_delbcnt_delta
;
384 totalrtbdelta
= qtrx
->qt_rtbcount_delta
+
385 qtrx
->qt_delrtb_delta
;
388 ASSERT(be64_to_cpu(d
->d_bcount
) >=
391 if (totalrtbdelta
< 0)
392 ASSERT(be64_to_cpu(d
->d_rtbcount
) >=
395 if (qtrx
->qt_icount_delta
< 0)
396 ASSERT(be64_to_cpu(d
->d_icount
) >=
397 -qtrx
->qt_icount_delta
);
400 be64_add_cpu(&d
->d_bcount
, (xfs_qcnt_t
)totalbdelta
);
402 if (qtrx
->qt_icount_delta
)
403 be64_add_cpu(&d
->d_icount
, (xfs_qcnt_t
)qtrx
->qt_icount_delta
);
406 be64_add_cpu(&d
->d_rtbcount
, (xfs_qcnt_t
)totalrtbdelta
);
409 * Get any default limits in use.
410 * Start/reset the timer(s) if needed.
413 xfs_qm_adjust_dqlimits(tp
->t_mountp
, dqp
);
414 xfs_qm_adjust_dqtimers(tp
->t_mountp
, d
);
417 dqp
->dq_flags
|= XFS_DQ_DIRTY
;
419 * add this to the list of items to get logged
421 xfs_trans_log_dquot(tp
, dqp
);
423 * Take off what's left of the original reservation.
424 * In case of delayed allocations, there's no
425 * reservation that a transaction structure knows of.
427 if (qtrx
->qt_blk_res
!= 0) {
428 ulong blk_res_used
= 0;
430 if (qtrx
->qt_bcount_delta
> 0)
431 blk_res_used
= qtrx
->qt_bcount_delta
;
433 if (qtrx
->qt_blk_res
!= blk_res_used
) {
434 if (qtrx
->qt_blk_res
> blk_res_used
)
435 dqp
->q_res_bcount
-= (xfs_qcnt_t
)
439 dqp
->q_res_bcount
-= (xfs_qcnt_t
)
445 * These blks were never reserved, either inside
446 * a transaction or outside one (in a delayed
447 * allocation). Also, this isn't always a
448 * negative number since we sometimes
449 * deliberately skip quota reservations.
451 if (qtrx
->qt_bcount_delta
) {
453 (xfs_qcnt_t
)qtrx
->qt_bcount_delta
;
457 * Adjust the RT reservation.
459 if (qtrx
->qt_rtblk_res
!= 0) {
460 if (qtrx
->qt_rtblk_res
!= qtrx
->qt_rtblk_res_used
) {
461 if (qtrx
->qt_rtblk_res
>
462 qtrx
->qt_rtblk_res_used
)
463 dqp
->q_res_rtbcount
-= (xfs_qcnt_t
)
464 (qtrx
->qt_rtblk_res
-
465 qtrx
->qt_rtblk_res_used
);
467 dqp
->q_res_rtbcount
-= (xfs_qcnt_t
)
468 (qtrx
->qt_rtblk_res_used
-
472 if (qtrx
->qt_rtbcount_delta
)
473 dqp
->q_res_rtbcount
+=
474 (xfs_qcnt_t
)qtrx
->qt_rtbcount_delta
;
478 * Adjust the inode reservation.
480 if (qtrx
->qt_ino_res
!= 0) {
481 ASSERT(qtrx
->qt_ino_res
>=
482 qtrx
->qt_ino_res_used
);
483 if (qtrx
->qt_ino_res
> qtrx
->qt_ino_res_used
)
484 dqp
->q_res_icount
-= (xfs_qcnt_t
)
486 qtrx
->qt_ino_res_used
);
488 if (qtrx
->qt_icount_delta
)
490 (xfs_qcnt_t
)qtrx
->qt_icount_delta
;
493 ASSERT(dqp
->q_res_bcount
>=
494 be64_to_cpu(dqp
->q_core
.d_bcount
));
495 ASSERT(dqp
->q_res_icount
>=
496 be64_to_cpu(dqp
->q_core
.d_icount
));
497 ASSERT(dqp
->q_res_rtbcount
>=
498 be64_to_cpu(dqp
->q_core
.d_rtbcount
));
504 * Release the reservations, and adjust the dquots accordingly.
505 * This is called only when the transaction is being aborted. If by
506 * any chance we have done dquot modifications incore (ie. deltas) already,
507 * we simply throw those away, since that's the expected behavior
508 * when a transaction is curtailed without a commit.
511 xfs_trans_unreserve_and_mod_dquots(
516 xfs_dqtrx_t
*qtrx
, *qa
;
519 if (!tp
->t_dqinfo
|| !(tp
->t_flags
& XFS_TRANS_DQ_DIRTY
))
522 for (j
= 0; j
< XFS_QM_TRANS_DQTYPES
; j
++) {
523 qa
= tp
->t_dqinfo
->dqs
[j
];
525 for (i
= 0; i
< XFS_QM_TRANS_MAXDQS
; i
++) {
528 * We assume that the array of dquots is filled
529 * sequentially, not sparsely.
531 if ((dqp
= qtrx
->qt_dquot
) == NULL
)
534 * Unreserve the original reservation. We don't care
535 * about the number of blocks used field, or deltas.
536 * Also we don't bother to zero the fields.
539 if (qtrx
->qt_blk_res
) {
543 (xfs_qcnt_t
)qtrx
->qt_blk_res
;
545 if (qtrx
->qt_ino_res
) {
551 (xfs_qcnt_t
)qtrx
->qt_ino_res
;
554 if (qtrx
->qt_rtblk_res
) {
559 dqp
->q_res_rtbcount
-=
560 (xfs_qcnt_t
)qtrx
->qt_rtblk_res
;
571 struct xfs_mount
*mp
,
572 struct xfs_dquot
*dqp
,
575 /* no warnings for project quotas - we just return ENOSPC later */
576 if (dqp
->dq_flags
& XFS_DQ_PROJ
)
578 quota_send_warning(make_kqid(&init_user_ns
,
579 (dqp
->dq_flags
& XFS_DQ_USER
) ?
581 be32_to_cpu(dqp
->q_core
.d_id
)),
582 mp
->m_super
->s_dev
, type
);
586 * This reserves disk blocks and inodes against a dquot.
587 * Flags indicate if the dquot is to be locked here and also
588 * if the blk reservation is for RT or regular blocks.
589 * Sending in XFS_QMOPT_FORCE_RES flag skips the quota check.
600 xfs_qcnt_t hardlimit
;
601 xfs_qcnt_t softlimit
;
603 xfs_qwarncnt_t warns
;
604 xfs_qwarncnt_t warnlimit
;
605 xfs_qcnt_t total_count
;
606 xfs_qcnt_t
*resbcountp
;
607 xfs_quotainfo_t
*q
= mp
->m_quotainfo
;
612 if (flags
& XFS_TRANS_DQ_RES_BLKS
) {
613 hardlimit
= be64_to_cpu(dqp
->q_core
.d_blk_hardlimit
);
615 hardlimit
= q
->qi_bhardlimit
;
616 softlimit
= be64_to_cpu(dqp
->q_core
.d_blk_softlimit
);
618 softlimit
= q
->qi_bsoftlimit
;
619 timer
= be32_to_cpu(dqp
->q_core
.d_btimer
);
620 warns
= be16_to_cpu(dqp
->q_core
.d_bwarns
);
621 warnlimit
= dqp
->q_mount
->m_quotainfo
->qi_bwarnlimit
;
622 resbcountp
= &dqp
->q_res_bcount
;
624 ASSERT(flags
& XFS_TRANS_DQ_RES_RTBLKS
);
625 hardlimit
= be64_to_cpu(dqp
->q_core
.d_rtb_hardlimit
);
627 hardlimit
= q
->qi_rtbhardlimit
;
628 softlimit
= be64_to_cpu(dqp
->q_core
.d_rtb_softlimit
);
630 softlimit
= q
->qi_rtbsoftlimit
;
631 timer
= be32_to_cpu(dqp
->q_core
.d_rtbtimer
);
632 warns
= be16_to_cpu(dqp
->q_core
.d_rtbwarns
);
633 warnlimit
= dqp
->q_mount
->m_quotainfo
->qi_rtbwarnlimit
;
634 resbcountp
= &dqp
->q_res_rtbcount
;
637 if ((flags
& XFS_QMOPT_FORCE_RES
) == 0 &&
639 ((XFS_IS_UQUOTA_ENFORCED(dqp
->q_mount
) && XFS_QM_ISUDQ(dqp
)) ||
640 (XFS_IS_GQUOTA_ENFORCED(dqp
->q_mount
) && XFS_QM_ISGDQ(dqp
)) ||
641 (XFS_IS_PQUOTA_ENFORCED(dqp
->q_mount
) && XFS_QM_ISPDQ(dqp
)))) {
644 * dquot is locked already. See if we'd go over the
645 * hardlimit or exceed the timelimit if we allocate
648 total_count
= *resbcountp
+ nblks
;
649 if (hardlimit
&& total_count
> hardlimit
) {
650 xfs_quota_warn(mp
, dqp
, QUOTA_NL_BHARDWARN
);
653 if (softlimit
&& total_count
> softlimit
) {
654 if ((timer
!= 0 && get_seconds() > timer
) ||
655 (warns
!= 0 && warns
>= warnlimit
)) {
656 xfs_quota_warn(mp
, dqp
,
657 QUOTA_NL_BSOFTLONGWARN
);
661 xfs_quota_warn(mp
, dqp
, QUOTA_NL_BSOFTWARN
);
665 total_count
= be64_to_cpu(dqp
->q_core
.d_icount
) + ninos
;
666 timer
= be32_to_cpu(dqp
->q_core
.d_itimer
);
667 warns
= be16_to_cpu(dqp
->q_core
.d_iwarns
);
668 warnlimit
= dqp
->q_mount
->m_quotainfo
->qi_iwarnlimit
;
669 hardlimit
= be64_to_cpu(dqp
->q_core
.d_ino_hardlimit
);
671 hardlimit
= q
->qi_ihardlimit
;
672 softlimit
= be64_to_cpu(dqp
->q_core
.d_ino_softlimit
);
674 softlimit
= q
->qi_isoftlimit
;
676 if (hardlimit
&& total_count
> hardlimit
) {
677 xfs_quota_warn(mp
, dqp
, QUOTA_NL_IHARDWARN
);
680 if (softlimit
&& total_count
> softlimit
) {
681 if ((timer
!= 0 && get_seconds() > timer
) ||
682 (warns
!= 0 && warns
>= warnlimit
)) {
683 xfs_quota_warn(mp
, dqp
,
684 QUOTA_NL_ISOFTLONGWARN
);
687 xfs_quota_warn(mp
, dqp
, QUOTA_NL_ISOFTWARN
);
693 * Change the reservation, but not the actual usage.
694 * Note that q_res_bcount = q_core.d_bcount + resv
696 (*resbcountp
) += (xfs_qcnt_t
)nblks
;
698 dqp
->q_res_icount
+= (xfs_qcnt_t
)ninos
;
701 * note the reservation amt in the trans struct too,
702 * so that the transaction knows how much was reserved by
703 * it against this particular dquot.
704 * We don't do this when we are reserving for a delayed allocation,
705 * because we don't have the luxury of a transaction envelope then.
708 ASSERT(tp
->t_dqinfo
);
709 ASSERT(flags
& XFS_QMOPT_RESBLK_MASK
);
711 xfs_trans_mod_dquot(tp
, dqp
,
712 flags
& XFS_QMOPT_RESBLK_MASK
,
715 xfs_trans_mod_dquot(tp
, dqp
,
716 XFS_TRANS_DQ_RES_INOS
,
719 ASSERT(dqp
->q_res_bcount
>= be64_to_cpu(dqp
->q_core
.d_bcount
));
720 ASSERT(dqp
->q_res_rtbcount
>= be64_to_cpu(dqp
->q_core
.d_rtbcount
));
721 ASSERT(dqp
->q_res_icount
>= be64_to_cpu(dqp
->q_core
.d_icount
));
728 if (flags
& XFS_QMOPT_ENOSPC
)
735 * Given dquot(s), make disk block and/or inode reservations against them.
736 * The fact that this does the reservation against user, group and
737 * project quotas is important, because this follows a all-or-nothing
740 * flags = XFS_QMOPT_FORCE_RES evades limit enforcement. Used by chown.
741 * XFS_QMOPT_ENOSPC returns ENOSPC not EDQUOT. Used by pquota.
742 * XFS_TRANS_DQ_RES_BLKS reserves regular disk blocks
743 * XFS_TRANS_DQ_RES_RTBLKS reserves realtime disk blocks
744 * dquots are unlocked on return, if they were not locked by caller.
747 xfs_trans_reserve_quota_bydquots(
748 struct xfs_trans
*tp
,
749 struct xfs_mount
*mp
,
750 struct xfs_dquot
*udqp
,
751 struct xfs_dquot
*gdqp
,
752 struct xfs_dquot
*pdqp
,
759 if (!XFS_IS_QUOTA_RUNNING(mp
) || !XFS_IS_QUOTA_ON(mp
))
762 if (tp
&& tp
->t_dqinfo
== NULL
)
763 xfs_trans_alloc_dqinfo(tp
);
765 ASSERT(flags
& XFS_QMOPT_RESBLK_MASK
);
768 error
= xfs_trans_dqresv(tp
, mp
, udqp
, nblks
, ninos
,
769 (flags
& ~XFS_QMOPT_ENOSPC
));
775 error
= xfs_trans_dqresv(tp
, mp
, gdqp
, nblks
, ninos
, flags
);
781 error
= xfs_trans_dqresv(tp
, mp
, pdqp
, nblks
, ninos
, flags
);
787 * Didn't change anything critical, so, no need to log
792 flags
|= XFS_QMOPT_FORCE_RES
;
794 xfs_trans_dqresv(tp
, mp
, gdqp
, -nblks
, -ninos
, flags
);
796 flags
|= XFS_QMOPT_FORCE_RES
;
798 xfs_trans_dqresv(tp
, mp
, udqp
, -nblks
, -ninos
, flags
);
804 * Lock the dquot and change the reservation if we can.
805 * This doesn't change the actual usage, just the reservation.
806 * The inode sent in is locked.
809 xfs_trans_reserve_quota_nblks(
810 struct xfs_trans
*tp
,
811 struct xfs_inode
*ip
,
816 struct xfs_mount
*mp
= ip
->i_mount
;
818 if (!XFS_IS_QUOTA_RUNNING(mp
) || !XFS_IS_QUOTA_ON(mp
))
820 if (XFS_IS_PQUOTA_ON(mp
))
821 flags
|= XFS_QMOPT_ENOSPC
;
823 ASSERT(!xfs_is_quota_inode(&mp
->m_sb
, ip
->i_ino
));
825 ASSERT(xfs_isilocked(ip
, XFS_ILOCK_EXCL
));
826 ASSERT((flags
& ~(XFS_QMOPT_FORCE_RES
| XFS_QMOPT_ENOSPC
)) ==
827 XFS_TRANS_DQ_RES_RTBLKS
||
828 (flags
& ~(XFS_QMOPT_FORCE_RES
| XFS_QMOPT_ENOSPC
)) ==
829 XFS_TRANS_DQ_RES_BLKS
);
832 * Reserve nblks against these dquots, with trans as the mediator.
834 return xfs_trans_reserve_quota_bydquots(tp
, mp
,
835 ip
->i_udquot
, ip
->i_gdquot
,
837 nblks
, ninos
, flags
);
841 * This routine is called to allocate a quotaoff log item.
844 xfs_trans_get_qoff_item(
846 xfs_qoff_logitem_t
*startqoff
,
849 xfs_qoff_logitem_t
*q
;
853 q
= xfs_qm_qoff_logitem_init(tp
->t_mountp
, startqoff
, flags
);
857 * Get a log_item_desc to point at the new item.
859 xfs_trans_add_item(tp
, &q
->qql_item
);
865 * This is called to mark the quotaoff logitem as needing
866 * to be logged when the transaction is committed. The logitem must
867 * already be associated with the given transaction.
870 xfs_trans_log_quotaoff_item(
872 xfs_qoff_logitem_t
*qlp
)
874 tp
->t_flags
|= XFS_TRANS_DIRTY
;
875 qlp
->qql_item
.li_desc
->lid_flags
|= XFS_LID_DIRTY
;
879 xfs_trans_alloc_dqinfo(
882 tp
->t_dqinfo
= kmem_zone_zalloc(xfs_qm_dqtrxzone
, KM_SLEEP
);
886 xfs_trans_free_dqinfo(
891 kmem_zone_free(xfs_qm_dqtrxzone
, tp
->t_dqinfo
);