Linux-libre 5.3.12-gnu
[librecmc/linux-libre.git] / fs / xfs / scrub / quota.c
1 // SPDX-License-Identifier: GPL-2.0+
2 /*
3  * Copyright (C) 2017 Oracle.  All Rights Reserved.
4  * Author: Darrick J. Wong <darrick.wong@oracle.com>
5  */
6 #include "xfs.h"
7 #include "xfs_fs.h"
8 #include "xfs_shared.h"
9 #include "xfs_format.h"
10 #include "xfs_trans_resv.h"
11 #include "xfs_mount.h"
12 #include "xfs_log_format.h"
13 #include "xfs_trans.h"
14 #include "xfs_inode.h"
15 #include "xfs_quota.h"
16 #include "xfs_qm.h"
17 #include "scrub/scrub.h"
18 #include "scrub/common.h"
19
20 /* Convert a scrub type code to a DQ flag, or return 0 if error. */
21 static inline uint
22 xchk_quota_to_dqtype(
23         struct xfs_scrub        *sc)
24 {
25         switch (sc->sm->sm_type) {
26         case XFS_SCRUB_TYPE_UQUOTA:
27                 return XFS_DQ_USER;
28         case XFS_SCRUB_TYPE_GQUOTA:
29                 return XFS_DQ_GROUP;
30         case XFS_SCRUB_TYPE_PQUOTA:
31                 return XFS_DQ_PROJ;
32         default:
33                 return 0;
34         }
35 }
36
37 /* Set us up to scrub a quota. */
38 int
39 xchk_setup_quota(
40         struct xfs_scrub        *sc,
41         struct xfs_inode        *ip)
42 {
43         uint                    dqtype;
44         int                     error;
45
46         if (!XFS_IS_QUOTA_RUNNING(sc->mp) || !XFS_IS_QUOTA_ON(sc->mp))
47                 return -ENOENT;
48
49         dqtype = xchk_quota_to_dqtype(sc);
50         if (dqtype == 0)
51                 return -EINVAL;
52         sc->flags |= XCHK_HAS_QUOTAOFFLOCK;
53         mutex_lock(&sc->mp->m_quotainfo->qi_quotaofflock);
54         if (!xfs_this_quota_on(sc->mp, dqtype))
55                 return -ENOENT;
56         error = xchk_setup_fs(sc, ip);
57         if (error)
58                 return error;
59         sc->ip = xfs_quota_inode(sc->mp, dqtype);
60         xfs_ilock(sc->ip, XFS_ILOCK_EXCL);
61         sc->ilock_flags = XFS_ILOCK_EXCL;
62         return 0;
63 }
64
65 /* Quotas. */
66
67 struct xchk_quota_info {
68         struct xfs_scrub        *sc;
69         xfs_dqid_t              last_id;
70 };
71
72 /* Scrub the fields in an individual quota item. */
73 STATIC int
74 xchk_quota_item(
75         struct xfs_dquot        *dq,
76         uint                    dqtype,
77         void                    *priv)
78 {
79         struct xchk_quota_info  *sqi = priv;
80         struct xfs_scrub        *sc = sqi->sc;
81         struct xfs_mount        *mp = sc->mp;
82         struct xfs_disk_dquot   *d = &dq->q_core;
83         struct xfs_quotainfo    *qi = mp->m_quotainfo;
84         xfs_fileoff_t           offset;
85         unsigned long long      bsoft;
86         unsigned long long      isoft;
87         unsigned long long      rsoft;
88         unsigned long long      bhard;
89         unsigned long long      ihard;
90         unsigned long long      rhard;
91         unsigned long long      bcount;
92         unsigned long long      icount;
93         unsigned long long      rcount;
94         xfs_ino_t               fs_icount;
95         xfs_dqid_t              id = be32_to_cpu(d->d_id);
96
97         /*
98          * Except for the root dquot, the actual dquot we got must either have
99          * the same or higher id as we saw before.
100          */
101         offset = id / qi->qi_dqperchunk;
102         if (id && id <= sqi->last_id)
103                 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
104
105         sqi->last_id = id;
106
107         /* Did we get the dquot type we wanted? */
108         if (dqtype != (d->d_flags & XFS_DQ_ALLTYPES))
109                 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
110
111         if (d->d_pad0 != cpu_to_be32(0) || d->d_pad != cpu_to_be16(0))
112                 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
113
114         /* Check the limits. */
115         bhard = be64_to_cpu(d->d_blk_hardlimit);
116         ihard = be64_to_cpu(d->d_ino_hardlimit);
117         rhard = be64_to_cpu(d->d_rtb_hardlimit);
118
119         bsoft = be64_to_cpu(d->d_blk_softlimit);
120         isoft = be64_to_cpu(d->d_ino_softlimit);
121         rsoft = be64_to_cpu(d->d_rtb_softlimit);
122
123         /*
124          * Warn if the hard limits are larger than the fs.
125          * Administrators can do this, though in production this seems
126          * suspect, which is why we flag it for review.
127          *
128          * Complain about corruption if the soft limit is greater than
129          * the hard limit.
130          */
131         if (bhard > mp->m_sb.sb_dblocks)
132                 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
133         if (bsoft > bhard)
134                 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
135
136         if (ihard > M_IGEO(mp)->maxicount)
137                 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
138         if (isoft > ihard)
139                 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
140
141         if (rhard > mp->m_sb.sb_rblocks)
142                 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
143         if (rsoft > rhard)
144                 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
145
146         /* Check the resource counts. */
147         bcount = be64_to_cpu(d->d_bcount);
148         icount = be64_to_cpu(d->d_icount);
149         rcount = be64_to_cpu(d->d_rtbcount);
150         fs_icount = percpu_counter_sum(&mp->m_icount);
151
152         /*
153          * Check that usage doesn't exceed physical limits.  However, on
154          * a reflink filesystem we're allowed to exceed physical space
155          * if there are no quota limits.
156          */
157         if (xfs_sb_version_hasreflink(&mp->m_sb)) {
158                 if (mp->m_sb.sb_dblocks < bcount)
159                         xchk_fblock_set_warning(sc, XFS_DATA_FORK,
160                                         offset);
161         } else {
162                 if (mp->m_sb.sb_dblocks < bcount)
163                         xchk_fblock_set_corrupt(sc, XFS_DATA_FORK,
164                                         offset);
165         }
166         if (icount > fs_icount || rcount > mp->m_sb.sb_rblocks)
167                 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
168
169         /*
170          * We can violate the hard limits if the admin suddenly sets a
171          * lower limit than the actual usage.  However, we flag it for
172          * admin review.
173          */
174         if (id != 0 && bhard != 0 && bcount > bhard)
175                 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
176         if (id != 0 && ihard != 0 && icount > ihard)
177                 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
178         if (id != 0 && rhard != 0 && rcount > rhard)
179                 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
180
181         return 0;
182 }
183
184 /* Check the quota's data fork. */
185 STATIC int
186 xchk_quota_data_fork(
187         struct xfs_scrub        *sc)
188 {
189         struct xfs_bmbt_irec    irec = { 0 };
190         struct xfs_iext_cursor  icur;
191         struct xfs_quotainfo    *qi = sc->mp->m_quotainfo;
192         struct xfs_ifork        *ifp;
193         xfs_fileoff_t           max_dqid_off;
194         int                     error = 0;
195
196         /* Invoke the fork scrubber. */
197         error = xchk_metadata_inode_forks(sc);
198         if (error || (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT))
199                 return error;
200
201         /* Check for data fork problems that apply only to quota files. */
202         max_dqid_off = ((xfs_dqid_t)-1) / qi->qi_dqperchunk;
203         ifp = XFS_IFORK_PTR(sc->ip, XFS_DATA_FORK);
204         for_each_xfs_iext(ifp, &icur, &irec) {
205                 if (xchk_should_terminate(sc, &error))
206                         break;
207                 /*
208                  * delalloc extents or blocks mapped above the highest
209                  * quota id shouldn't happen.
210                  */
211                 if (isnullstartblock(irec.br_startblock) ||
212                     irec.br_startoff > max_dqid_off ||
213                     irec.br_startoff + irec.br_blockcount - 1 > max_dqid_off) {
214                         xchk_fblock_set_corrupt(sc, XFS_DATA_FORK,
215                                         irec.br_startoff);
216                         break;
217                 }
218         }
219
220         return error;
221 }
222
223 /* Scrub all of a quota type's items. */
224 int
225 xchk_quota(
226         struct xfs_scrub        *sc)
227 {
228         struct xchk_quota_info  sqi;
229         struct xfs_mount        *mp = sc->mp;
230         struct xfs_quotainfo    *qi = mp->m_quotainfo;
231         uint                    dqtype;
232         int                     error = 0;
233
234         dqtype = xchk_quota_to_dqtype(sc);
235
236         /* Look for problem extents. */
237         error = xchk_quota_data_fork(sc);
238         if (error)
239                 goto out;
240         if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
241                 goto out;
242
243         /*
244          * Check all the quota items.  Now that we've checked the quota inode
245          * data fork we have to drop ILOCK_EXCL to use the regular dquot
246          * functions.
247          */
248         xfs_iunlock(sc->ip, sc->ilock_flags);
249         sc->ilock_flags = 0;
250         sqi.sc = sc;
251         sqi.last_id = 0;
252         error = xfs_qm_dqiterate(mp, dqtype, xchk_quota_item, &sqi);
253         sc->ilock_flags = XFS_ILOCK_EXCL;
254         xfs_ilock(sc->ip, sc->ilock_flags);
255         if (!xchk_fblock_process_error(sc, XFS_DATA_FORK,
256                         sqi.last_id * qi->qi_dqperchunk, &error))
257                 goto out;
258
259 out:
260         return error;
261 }