struct xfs_disk_dquot *d = &dq->q_core;
struct xfs_quotainfo *qi = mp->m_quotainfo;
xfs_fileoff_t offset;
- unsigned long long bcount;
- unsigned long long icount;
- unsigned long long rcount;
xfs_ino_t fs_icount;
int error = 0;
xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
/* Check the resource counts. */
- bcount = be64_to_cpu(d->d_bcount);
- icount = be64_to_cpu(d->d_icount);
- rcount = be64_to_cpu(d->d_rtbcount);
fs_icount = percpu_counter_sum(&mp->m_icount);
/*
* if there are no quota limits.
*/
if (xfs_sb_version_hasreflink(&mp->m_sb)) {
- if (mp->m_sb.sb_dblocks < bcount)
+ if (mp->m_sb.sb_dblocks < dq->q_blk.count)
xchk_fblock_set_warning(sc, XFS_DATA_FORK,
offset);
} else {
- if (mp->m_sb.sb_dblocks < bcount)
+ if (mp->m_sb.sb_dblocks < dq->q_blk.count)
xchk_fblock_set_corrupt(sc, XFS_DATA_FORK,
offset);
}
- if (icount > fs_icount || rcount > mp->m_sb.sb_rblocks)
+ if (dq->q_ino.count > fs_icount || dq->q_rtb.count > mp->m_sb.sb_rblocks)
xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
/*
goto out;
if (dq->q_blk.hardlimit != 0 &&
- bcount > dq->q_blk.hardlimit)
+ dq->q_blk.count > dq->q_blk.hardlimit)
xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
if (dq->q_ino.hardlimit != 0 &&
- icount > dq->q_ino.hardlimit)
+ dq->q_ino.count > dq->q_ino.hardlimit)
xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
if (dq->q_rtb.hardlimit != 0 &&
- rcount > dq->q_rtb.hardlimit)
+ dq->q_rtb.count > dq->q_rtb.hardlimit)
xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
out:
if (!d->d_btimer) {
if ((dq->q_blk.softlimit &&
- (be64_to_cpu(d->d_bcount) > dq->q_blk.softlimit)) ||
+ (dq->q_blk.count > dq->q_blk.softlimit)) ||
(dq->q_blk.hardlimit &&
- (be64_to_cpu(d->d_bcount) > dq->q_blk.hardlimit))) {
+ (dq->q_blk.count > dq->q_blk.hardlimit))) {
d->d_btimer = cpu_to_be32(ktime_get_real_seconds() +
defq->btimelimit);
} else {
}
} else {
if ((!dq->q_blk.softlimit ||
- (be64_to_cpu(d->d_bcount) <= dq->q_blk.softlimit)) &&
+ (dq->q_blk.count <= dq->q_blk.softlimit)) &&
(!dq->q_blk.hardlimit ||
- (be64_to_cpu(d->d_bcount) <= dq->q_blk.hardlimit))) {
+ (dq->q_blk.count <= dq->q_blk.hardlimit))) {
d->d_btimer = 0;
}
}
if (!d->d_itimer) {
if ((dq->q_ino.softlimit &&
- (be64_to_cpu(d->d_icount) > dq->q_ino.softlimit)) ||
+ (dq->q_ino.count > dq->q_ino.softlimit)) ||
(dq->q_ino.hardlimit &&
- (be64_to_cpu(d->d_icount) > dq->q_ino.hardlimit))) {
+ (dq->q_ino.count > dq->q_ino.hardlimit))) {
d->d_itimer = cpu_to_be32(ktime_get_real_seconds() +
defq->itimelimit);
} else {
}
} else {
if ((!dq->q_ino.softlimit ||
- (be64_to_cpu(d->d_icount) <= dq->q_ino.softlimit)) &&
+ (dq->q_ino.count <= dq->q_ino.softlimit)) &&
(!dq->q_ino.hardlimit ||
- (be64_to_cpu(d->d_icount) <= dq->q_ino.hardlimit))) {
+ (dq->q_ino.count <= dq->q_ino.hardlimit))) {
d->d_itimer = 0;
}
}
if (!d->d_rtbtimer) {
if ((dq->q_rtb.softlimit &&
- (be64_to_cpu(d->d_rtbcount) > dq->q_rtb.softlimit)) ||
+ (dq->q_rtb.count > dq->q_rtb.softlimit)) ||
(dq->q_rtb.hardlimit &&
- (be64_to_cpu(d->d_rtbcount) > dq->q_rtb.hardlimit))) {
+ (dq->q_rtb.count > dq->q_rtb.hardlimit))) {
d->d_rtbtimer = cpu_to_be32(ktime_get_real_seconds() +
defq->rtbtimelimit);
} else {
}
} else {
if ((!dq->q_rtb.softlimit ||
- (be64_to_cpu(d->d_rtbcount) <= dq->q_rtb.softlimit)) &&
+ (dq->q_rtb.count <= dq->q_rtb.softlimit)) &&
(!dq->q_rtb.hardlimit ||
- (be64_to_cpu(d->d_rtbcount) <= dq->q_rtb.hardlimit))) {
+ (dq->q_rtb.count <= dq->q_rtb.hardlimit))) {
d->d_rtbtimer = 0;
}
}
dqp->q_rtb.hardlimit = be64_to_cpu(ddqp->d_rtb_hardlimit);
dqp->q_rtb.softlimit = be64_to_cpu(ddqp->d_rtb_softlimit);
+ dqp->q_blk.count = be64_to_cpu(ddqp->d_bcount);
+ dqp->q_ino.count = be64_to_cpu(ddqp->d_icount);
+ dqp->q_rtb.count = be64_to_cpu(ddqp->d_rtbcount);
+
/*
* Reservation counters are defined as reservation plus current usage
* to avoid having to add every time.
*/
- dqp->q_blk.reserved = be64_to_cpu(ddqp->d_bcount);
- dqp->q_ino.reserved = be64_to_cpu(ddqp->d_icount);
- dqp->q_rtb.reserved = be64_to_cpu(ddqp->d_rtbcount);
+ dqp->q_blk.reserved = dqp->q_blk.count;
+ dqp->q_ino.reserved = dqp->q_ino.count;
+ dqp->q_rtb.reserved = dqp->q_rtb.count;
/* initialize the dquot speculative prealloc thresholds */
xfs_dquot_set_prealloc_limits(dqp);
ddqp->d_ino_softlimit = cpu_to_be64(dqp->q_ino.softlimit);
ddqp->d_rtb_hardlimit = cpu_to_be64(dqp->q_rtb.hardlimit);
ddqp->d_rtb_softlimit = cpu_to_be64(dqp->q_rtb.softlimit);
+
+ ddqp->d_bcount = cpu_to_be64(dqp->q_blk.count);
+ ddqp->d_icount = cpu_to_be64(dqp->q_ino.count);
+ ddqp->d_rtbcount = cpu_to_be64(dqp->q_rtb.count);
}
/* Allocate and initialize the dquot buffer for this in-core dquot. */
if (dqp->q_id == 0)
return NULL;
- if (dqp->q_blk.softlimit &&
- be64_to_cpu(ddq->d_bcount) > dqp->q_blk.softlimit &&
+ if (dqp->q_blk.softlimit && dqp->q_blk.count > dqp->q_blk.softlimit &&
!ddq->d_btimer)
return __this_address;
- if (dqp->q_ino.softlimit &&
- be64_to_cpu(ddq->d_icount) > dqp->q_ino.softlimit &&
+ if (dqp->q_ino.softlimit && dqp->q_ino.count > dqp->q_ino.softlimit &&
!ddq->d_itimer)
return __this_address;
- if (dqp->q_rtb.softlimit &&
- be64_to_cpu(ddq->d_rtbcount) > dqp->q_rtb.softlimit &&
+ if (dqp->q_rtb.softlimit && dqp->q_rtb.count > dqp->q_rtb.softlimit &&
!ddq->d_rtbtimer)
return __this_address;
/* Total resources allocated and reserved. */
xfs_qcnt_t reserved;
+ /* Total resources allocated. */
+ xfs_qcnt_t count;
+
/* Absolute and preferred limits. */
xfs_qcnt_t hardlimit;
xfs_qcnt_t softlimit;
* Adjust the inode count and the block count to reflect this inode's
* resource usage.
*/
- be64_add_cpu(&dqp->q_core.d_icount, 1);
+ dqp->q_ino.count++;
dqp->q_ino.reserved++;
if (nblks) {
- be64_add_cpu(&dqp->q_core.d_bcount, nblks);
+ dqp->q_blk.count += nblks;
dqp->q_blk.reserved += nblks;
}
if (rtblks) {
- be64_add_cpu(&dqp->q_core.d_rtbcount, rtblks);
+ dqp->q_rtb.count += rtblks;
dqp->q_rtb.reserved += rtblks;
}
!dqp->q_rtb.softlimit && \
!dqp->q_ino.hardlimit && \
!dqp->q_ino.softlimit && \
- !dqp->q_core.d_bcount && \
- !dqp->q_core.d_rtbcount && \
- !dqp->q_core.d_icount)
+ !dqp->q_blk.count && \
+ !dqp->q_rtb.count && \
+ !dqp->q_ino.count)
/* Defaults for each quota type: time limits, warn limits, usage limits */
struct xfs_def_quota {
__entry->flags = dqp->dq_flags | dqp->q_flags;
__entry->nrefs = dqp->q_nrefs;
__entry->res_bcount = dqp->q_blk.reserved;
- __entry->bcount = be64_to_cpu(dqp->q_core.d_bcount);
- __entry->icount = be64_to_cpu(dqp->q_core.d_icount);
+ __entry->bcount = dqp->q_blk.count;
+ __entry->icount = dqp->q_ino.count;
__entry->blk_hardlimit = dqp->q_blk.hardlimit;
__entry->blk_softlimit = dqp->q_blk.softlimit;
__entry->ino_hardlimit = dqp->q_ino.hardlimit;
int i, j;
struct xfs_dquot *dqp;
struct xfs_dqtrx *qtrx, *qa;
- struct xfs_disk_dquot *d;
int64_t totalbdelta;
int64_t totalrtbdelta;
/*
* adjust the actual number of blocks used
*/
- d = &dqp->q_core;
/*
* The issue here is - sometimes we don't make a blkquota
qtrx->qt_delrtb_delta;
#ifdef DEBUG
if (totalbdelta < 0)
- ASSERT(be64_to_cpu(d->d_bcount) >=
- -totalbdelta);
+ ASSERT(dqp->q_blk.count >= -totalbdelta);
if (totalrtbdelta < 0)
- ASSERT(be64_to_cpu(d->d_rtbcount) >=
- -totalrtbdelta);
+ ASSERT(dqp->q_rtb.count >= -totalrtbdelta);
if (qtrx->qt_icount_delta < 0)
- ASSERT(be64_to_cpu(d->d_icount) >=
- -qtrx->qt_icount_delta);
+ ASSERT(dqp->q_ino.count >= -qtrx->qt_icount_delta);
#endif
if (totalbdelta)
- be64_add_cpu(&d->d_bcount, (xfs_qcnt_t)totalbdelta);
+ dqp->q_blk.count += totalbdelta;
if (qtrx->qt_icount_delta)
- be64_add_cpu(&d->d_icount, (xfs_qcnt_t)qtrx->qt_icount_delta);
+ dqp->q_ino.count += qtrx->qt_icount_delta;
if (totalrtbdelta)
- be64_add_cpu(&d->d_rtbcount, (xfs_qcnt_t)totalrtbdelta);
+ dqp->q_rtb.count += totalrtbdelta;
/*
* Get any default limits in use.
(xfs_qcnt_t)qtrx->qt_icount_delta;
}
- ASSERT(dqp->q_blk.reserved >=
- be64_to_cpu(dqp->q_core.d_bcount));
- ASSERT(dqp->q_ino.reserved >=
- be64_to_cpu(dqp->q_core.d_icount));
- ASSERT(dqp->q_rtb.reserved >=
- be64_to_cpu(dqp->q_core.d_rtbcount));
+ ASSERT(dqp->q_blk.reserved >= dqp->q_blk.count);
+ ASSERT(dqp->q_ino.reserved >= dqp->q_ino.count);
+ ASSERT(dqp->q_rtb.reserved >= dqp->q_rtb.count);
}
}
}
/*
* Change the reservation, but not the actual usage.
- * Note that q_blk.reserved = q_core.d_bcount + resv
+ * Note that q_blk.reserved = q_blk.count + resv
*/
(*resbcountp) += (xfs_qcnt_t)nblks;
if (ninos != 0)
XFS_TRANS_DQ_RES_INOS,
ninos);
}
- ASSERT(dqp->q_blk.reserved >= be64_to_cpu(dqp->q_core.d_bcount));
- ASSERT(dqp->q_rtb.reserved >= be64_to_cpu(dqp->q_core.d_rtbcount));
- ASSERT(dqp->q_ino.reserved >= be64_to_cpu(dqp->q_core.d_icount));
+ ASSERT(dqp->q_blk.reserved >= dqp->q_blk.count);
+ ASSERT(dqp->q_rtb.reserved >= dqp->q_rtb.count);
+ ASSERT(dqp->q_ino.reserved >= dqp->q_ino.count);
xfs_dqunlock(dqp);
return 0;