mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2025-01-09 22:50:41 +00:00
b4b52b881c
Hi Linus, This is my very first pull-request. I've been working full-time as a kernel developer for more than two years now. During this time I've been fixing bugs reported by Coverity all over the tree and, as part of my work, I'm also contributing to the KSPP. My work in the kernel community has been supervised by Greg KH and Kees Cook. OK. So, after the quick introduction above, please, pull the following patches that mark switch cases where we are expecting to fall through. These patches are part of the ongoing efforts to enable -Wimplicit-fallthrough. They have been ignored for a long time (most of them more than 3 months, even after pinging multiple times), which is the reason why I've created this tree. Most of them have been baking in linux-next for a whole development cycle. And with Stephen Rothwell's help, we've had linux-next nag-emails going out for newly introduced code that triggers -Wimplicit-fallthrough to avoid gaining more of these cases while we work to remove the ones that are already present. I'm happy to let you know that we are getting close to completing this work. Currently, there are only 32 of 2311 of these cases left to be addressed in linux-next. I'm auditing every case; I take a look into the code and analyze it in order to determine if I'm dealing with an actual bug or a false positive, as explained here: https://lore.kernel.org/lkml/c2fad584-1705-a5f2-d63c-824e9b96cf50@embeddedor.com/ While working on this, I've found and fixed the following missing break/return bugs, some of them introduced more than 5 years ago: 84242b82d81c54e009a2aaa74d3d9eff70babf56 7850b51b6c21812be96d0200b74cff1f40587d98 5e420fe635813e5746b296cfc8fff4853ae205a2 09186e503486da4a17f16f2f7c679e6e3e2a32f4 b5be853181a8d4a6e20f2073ccd273d6280cad88 7264235ee74f51d26fbdf97bf98c6102a460484f cc5034a5d293dd620484d1d836aa16c6764a1c8c 479826cc86118e0d87e5cefb3df5b748e0480924 5340f23df8fe27a270af3fa1a93cd07293d23dd9 df997abeebadaa4824271009e2d2b526a70a11cb 2f10d823739680d2477ce34437e8a08a53117f40 307b00c5e695857ca92fc6a4b8ab6c48f988a1b1 5d25ff7a544889bc4b749fda31778d6a18dddbcb a7ed5b3e7dca197de4da6273940a7ca6d1d756a1 c24bfa8f21b59283580043dada19a6e943b6e426 ad0eaee6195db1db1749dd46b9e6f4466793d178 9ba8376ce1e2cbf4ce44f7e4bee1d0648e10d594 dc586a60a11d0260308db1bebe788ad8973e2729 a8e9b186f153a44690ad0363a56716e7077ad28c 4e57562b4846e42cd1c2e556f0ece18c1154e116 60747828eac28836b49bed214399b0c972f19df3 c5b974bee9d2ceae4c441ae5a01e498c2674e100 cc44ba91166beb78f9cb29d5e3d41c0a2d0a7329 2c930e3d0aed1505e86e0928d323df5027817740 Once this work is finish, we'll be able to universally enable "-Wimplicit-fallthrough" to avoid any of these kinds of bugs from entering the kernel again. Thanks Signed-off-by: Gustavo A. R. Silva <gustavo@embeddedor.com> -----BEGIN PGP SIGNATURE----- iQIzBAABCAAdFiEEkmRahXBSurMIg1YvRwW0y0cG2zEFAlzQR2IACgkQRwW0y0cG 2zEJbQ//X930OcBtT/9DRW4XL1Jeq0Mjssz/GLX2Vpup5CwwcTROG65no80Zezf/ yQRWnUjGX0OBv/fmUK32/nTxI/7k7NkmIXJHe0HiEF069GEENB7FT6tfDzIPjU8M qQkB8NsSUWJs3IH6BVynb/9MGE1VpGBDbYk7CBZRtRJT1RMM+3kQPucgiZMgUBPo Yd9zKwn4i/8tcOCli++EUdQ29ukMoY2R3qpK4LftdX9sXLKZBWNwQbiCwSkjnvJK I6FDiA7RaWH2wWGlL7BpN5RrvAXp3z8QN/JZnivIGt4ijtAyxFUL/9KOEgQpBQN2 6TBRhfTQFM73NCyzLgGLNzvd8awem1rKGSBBUvevaPbgesgM+Of65wmmTQRhFNCt A7+e286X1GiK3aNcjUKrByKWm7x590EWmDzmpmICxNPdt5DHQ6EkmvBdNjnxCMrO aGA24l78tBN09qN45LR7wtHYuuyR0Jt9bCmeQZmz7+x3ICDHi/+Gw7XPN/eM9+T6 lZbbINiYUyZVxOqwzkYDCsdv9+kUvu3e4rPs20NERWRpV8FEvBIyMjXAg6NAMTue K+ikkyMBxCvyw+NMimHJwtD7ho4FkLPcoeXb2ZGJTRHixiZAEtF1RaQ7dA05Q/SL gbSc0DgLZeHlLBT+BSWC2Z8SDnoIhQFXW49OmuACwCUC68NHKps= =k30z -----END PGP SIGNATURE----- Merge tag 'Wimplicit-fallthrough-5.2-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/gustavoars/linux Pull Wimplicit-fallthrough updates from Gustavo A. R. Silva: "Mark switch cases where we are expecting to fall through. This is part of the ongoing efforts to enable -Wimplicit-fallthrough. Most of them have been baking in linux-next for a whole development cycle. And with Stephen Rothwell's help, we've had linux-next nag-emails going out for newly introduced code that triggers -Wimplicit-fallthrough to avoid gaining more of these cases while we work to remove the ones that are already present. We are getting close to completing this work. Currently, there are only 32 of 2311 of these cases left to be addressed in linux-next. I'm auditing every case; I take a look into the code and analyze it in order to determine if I'm dealing with an actual bug or a false positive, as explained here: https://lore.kernel.org/lkml/c2fad584-1705-a5f2-d63c-824e9b96cf50@embeddedor.com/ While working on this, I've found and fixed the several missing break/return bugs, some of them introduced more than 5 years ago. Once this work is finished, we'll be able to universally enable "-Wimplicit-fallthrough" to avoid any of these kinds of bugs from entering the kernel again" * tag 'Wimplicit-fallthrough-5.2-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/gustavoars/linux: (27 commits) memstick: mark expected switch fall-throughs drm/nouveau/nvkm: mark expected switch fall-throughs NFC: st21nfca: Fix fall-through warnings NFC: pn533: mark expected switch fall-throughs block: Mark expected switch fall-throughs ASN.1: mark expected switch fall-through lib/cmdline.c: mark expected switch fall-throughs lib: zstd: Mark expected switch fall-throughs scsi: sym53c8xx_2: sym_nvram: Mark expected switch fall-through scsi: sym53c8xx_2: sym_hipd: mark expected switch fall-throughs scsi: ppa: mark expected switch fall-through scsi: osst: mark expected switch fall-throughs scsi: lpfc: lpfc_scsi: Mark expected switch fall-throughs scsi: lpfc: lpfc_nvme: Mark expected switch fall-through scsi: lpfc: lpfc_nportdisc: Mark expected switch fall-through scsi: lpfc: lpfc_hbadisc: Mark expected switch fall-throughs scsi: lpfc: lpfc_els: Mark expected switch fall-throughs scsi: lpfc: lpfc_ct: Mark expected switch fall-throughs scsi: imm: mark expected switch fall-throughs scsi: csiostor: csio_wr: mark expected switch fall-through ...
607 lines
17 KiB
C
607 lines
17 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* linux/fs/ufs/util.h
|
|
*
|
|
* Copyright (C) 1998
|
|
* Daniel Pirkl <daniel.pirkl@email.cz>
|
|
* Charles University, Faculty of Mathematics and Physics
|
|
*/
|
|
|
|
#include <linux/buffer_head.h>
|
|
#include <linux/fs.h>
|
|
#include "swab.h"
|
|
|
|
|
|
/*
|
|
* some useful macros
|
|
*/
|
|
#define in_range(b,first,len) ((b)>=(first)&&(b)<(first)+(len))
|
|
|
|
/*
|
|
* functions used for retyping
|
|
*/
|
|
static inline struct ufs_buffer_head *UCPI_UBH(struct ufs_cg_private_info *cpi)
|
|
{
|
|
return &cpi->c_ubh;
|
|
}
|
|
static inline struct ufs_buffer_head *USPI_UBH(struct ufs_sb_private_info *spi)
|
|
{
|
|
return &spi->s_ubh;
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
* macros used for accessing structures
|
|
*/
|
|
static inline s32
|
|
ufs_get_fs_state(struct super_block *sb, struct ufs_super_block_first *usb1,
|
|
struct ufs_super_block_third *usb3)
|
|
{
|
|
switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
|
|
case UFS_ST_SUNOS:
|
|
if (fs32_to_cpu(sb, usb3->fs_postblformat) == UFS_42POSTBLFMT)
|
|
return fs32_to_cpu(sb, usb1->fs_u0.fs_sun.fs_state);
|
|
/* Fall Through - to UFS_ST_SUN */
|
|
case UFS_ST_SUN:
|
|
return fs32_to_cpu(sb, usb3->fs_un2.fs_sun.fs_state);
|
|
case UFS_ST_SUNx86:
|
|
return fs32_to_cpu(sb, usb1->fs_u1.fs_sunx86.fs_state);
|
|
case UFS_ST_44BSD:
|
|
default:
|
|
return fs32_to_cpu(sb, usb3->fs_un2.fs_44.fs_state);
|
|
}
|
|
}
|
|
|
|
static inline void
|
|
ufs_set_fs_state(struct super_block *sb, struct ufs_super_block_first *usb1,
|
|
struct ufs_super_block_third *usb3, s32 value)
|
|
{
|
|
switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
|
|
case UFS_ST_SUNOS:
|
|
if (fs32_to_cpu(sb, usb3->fs_postblformat) == UFS_42POSTBLFMT) {
|
|
usb1->fs_u0.fs_sun.fs_state = cpu_to_fs32(sb, value);
|
|
break;
|
|
}
|
|
/* Fall Through - to UFS_ST_SUN */
|
|
case UFS_ST_SUN:
|
|
usb3->fs_un2.fs_sun.fs_state = cpu_to_fs32(sb, value);
|
|
break;
|
|
case UFS_ST_SUNx86:
|
|
usb1->fs_u1.fs_sunx86.fs_state = cpu_to_fs32(sb, value);
|
|
break;
|
|
case UFS_ST_44BSD:
|
|
usb3->fs_un2.fs_44.fs_state = cpu_to_fs32(sb, value);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static inline u32
|
|
ufs_get_fs_npsect(struct super_block *sb, struct ufs_super_block_first *usb1,
|
|
struct ufs_super_block_third *usb3)
|
|
{
|
|
if ((UFS_SB(sb)->s_flags & UFS_ST_MASK) == UFS_ST_SUNx86)
|
|
return fs32_to_cpu(sb, usb3->fs_un2.fs_sunx86.fs_npsect);
|
|
else
|
|
return fs32_to_cpu(sb, usb1->fs_u1.fs_sun.fs_npsect);
|
|
}
|
|
|
|
static inline u64
|
|
ufs_get_fs_qbmask(struct super_block *sb, struct ufs_super_block_third *usb3)
|
|
{
|
|
__fs64 tmp;
|
|
|
|
switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
|
|
case UFS_ST_SUNOS:
|
|
case UFS_ST_SUN:
|
|
((__fs32 *)&tmp)[0] = usb3->fs_un2.fs_sun.fs_qbmask[0];
|
|
((__fs32 *)&tmp)[1] = usb3->fs_un2.fs_sun.fs_qbmask[1];
|
|
break;
|
|
case UFS_ST_SUNx86:
|
|
((__fs32 *)&tmp)[0] = usb3->fs_un2.fs_sunx86.fs_qbmask[0];
|
|
((__fs32 *)&tmp)[1] = usb3->fs_un2.fs_sunx86.fs_qbmask[1];
|
|
break;
|
|
case UFS_ST_44BSD:
|
|
((__fs32 *)&tmp)[0] = usb3->fs_un2.fs_44.fs_qbmask[0];
|
|
((__fs32 *)&tmp)[1] = usb3->fs_un2.fs_44.fs_qbmask[1];
|
|
break;
|
|
}
|
|
|
|
return fs64_to_cpu(sb, tmp);
|
|
}
|
|
|
|
static inline u64
|
|
ufs_get_fs_qfmask(struct super_block *sb, struct ufs_super_block_third *usb3)
|
|
{
|
|
__fs64 tmp;
|
|
|
|
switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
|
|
case UFS_ST_SUNOS:
|
|
case UFS_ST_SUN:
|
|
((__fs32 *)&tmp)[0] = usb3->fs_un2.fs_sun.fs_qfmask[0];
|
|
((__fs32 *)&tmp)[1] = usb3->fs_un2.fs_sun.fs_qfmask[1];
|
|
break;
|
|
case UFS_ST_SUNx86:
|
|
((__fs32 *)&tmp)[0] = usb3->fs_un2.fs_sunx86.fs_qfmask[0];
|
|
((__fs32 *)&tmp)[1] = usb3->fs_un2.fs_sunx86.fs_qfmask[1];
|
|
break;
|
|
case UFS_ST_44BSD:
|
|
((__fs32 *)&tmp)[0] = usb3->fs_un2.fs_44.fs_qfmask[0];
|
|
((__fs32 *)&tmp)[1] = usb3->fs_un2.fs_44.fs_qfmask[1];
|
|
break;
|
|
}
|
|
|
|
return fs64_to_cpu(sb, tmp);
|
|
}
|
|
|
|
static inline u16
|
|
ufs_get_de_namlen(struct super_block *sb, struct ufs_dir_entry *de)
|
|
{
|
|
if ((UFS_SB(sb)->s_flags & UFS_DE_MASK) == UFS_DE_OLD)
|
|
return fs16_to_cpu(sb, de->d_u.d_namlen);
|
|
else
|
|
return de->d_u.d_44.d_namlen; /* XXX this seems wrong */
|
|
}
|
|
|
|
static inline void
|
|
ufs_set_de_namlen(struct super_block *sb, struct ufs_dir_entry *de, u16 value)
|
|
{
|
|
if ((UFS_SB(sb)->s_flags & UFS_DE_MASK) == UFS_DE_OLD)
|
|
de->d_u.d_namlen = cpu_to_fs16(sb, value);
|
|
else
|
|
de->d_u.d_44.d_namlen = value; /* XXX this seems wrong */
|
|
}
|
|
|
|
static inline void
|
|
ufs_set_de_type(struct super_block *sb, struct ufs_dir_entry *de, int mode)
|
|
{
|
|
if ((UFS_SB(sb)->s_flags & UFS_DE_MASK) != UFS_DE_44BSD)
|
|
return;
|
|
|
|
/*
|
|
* TODO turn this into a table lookup
|
|
*/
|
|
switch (mode & S_IFMT) {
|
|
case S_IFSOCK:
|
|
de->d_u.d_44.d_type = DT_SOCK;
|
|
break;
|
|
case S_IFLNK:
|
|
de->d_u.d_44.d_type = DT_LNK;
|
|
break;
|
|
case S_IFREG:
|
|
de->d_u.d_44.d_type = DT_REG;
|
|
break;
|
|
case S_IFBLK:
|
|
de->d_u.d_44.d_type = DT_BLK;
|
|
break;
|
|
case S_IFDIR:
|
|
de->d_u.d_44.d_type = DT_DIR;
|
|
break;
|
|
case S_IFCHR:
|
|
de->d_u.d_44.d_type = DT_CHR;
|
|
break;
|
|
case S_IFIFO:
|
|
de->d_u.d_44.d_type = DT_FIFO;
|
|
break;
|
|
default:
|
|
de->d_u.d_44.d_type = DT_UNKNOWN;
|
|
}
|
|
}
|
|
|
|
static inline u32
|
|
ufs_get_inode_uid(struct super_block *sb, struct ufs_inode *inode)
|
|
{
|
|
switch (UFS_SB(sb)->s_flags & UFS_UID_MASK) {
|
|
case UFS_UID_44BSD:
|
|
return fs32_to_cpu(sb, inode->ui_u3.ui_44.ui_uid);
|
|
case UFS_UID_EFT:
|
|
if (inode->ui_u1.oldids.ui_suid == 0xFFFF)
|
|
return fs32_to_cpu(sb, inode->ui_u3.ui_sun.ui_uid);
|
|
/* Fall through */
|
|
default:
|
|
return fs16_to_cpu(sb, inode->ui_u1.oldids.ui_suid);
|
|
}
|
|
}
|
|
|
|
static inline void
|
|
ufs_set_inode_uid(struct super_block *sb, struct ufs_inode *inode, u32 value)
|
|
{
|
|
switch (UFS_SB(sb)->s_flags & UFS_UID_MASK) {
|
|
case UFS_UID_44BSD:
|
|
inode->ui_u3.ui_44.ui_uid = cpu_to_fs32(sb, value);
|
|
inode->ui_u1.oldids.ui_suid = cpu_to_fs16(sb, value);
|
|
break;
|
|
case UFS_UID_EFT:
|
|
inode->ui_u3.ui_sun.ui_uid = cpu_to_fs32(sb, value);
|
|
if (value > 0xFFFF)
|
|
value = 0xFFFF;
|
|
/* Fall through */
|
|
default:
|
|
inode->ui_u1.oldids.ui_suid = cpu_to_fs16(sb, value);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static inline u32
|
|
ufs_get_inode_gid(struct super_block *sb, struct ufs_inode *inode)
|
|
{
|
|
switch (UFS_SB(sb)->s_flags & UFS_UID_MASK) {
|
|
case UFS_UID_44BSD:
|
|
return fs32_to_cpu(sb, inode->ui_u3.ui_44.ui_gid);
|
|
case UFS_UID_EFT:
|
|
if (inode->ui_u1.oldids.ui_sgid == 0xFFFF)
|
|
return fs32_to_cpu(sb, inode->ui_u3.ui_sun.ui_gid);
|
|
/* Fall through */
|
|
default:
|
|
return fs16_to_cpu(sb, inode->ui_u1.oldids.ui_sgid);
|
|
}
|
|
}
|
|
|
|
static inline void
|
|
ufs_set_inode_gid(struct super_block *sb, struct ufs_inode *inode, u32 value)
|
|
{
|
|
switch (UFS_SB(sb)->s_flags & UFS_UID_MASK) {
|
|
case UFS_UID_44BSD:
|
|
inode->ui_u3.ui_44.ui_gid = cpu_to_fs32(sb, value);
|
|
inode->ui_u1.oldids.ui_sgid = cpu_to_fs16(sb, value);
|
|
break;
|
|
case UFS_UID_EFT:
|
|
inode->ui_u3.ui_sun.ui_gid = cpu_to_fs32(sb, value);
|
|
if (value > 0xFFFF)
|
|
value = 0xFFFF;
|
|
/* Fall through */
|
|
default:
|
|
inode->ui_u1.oldids.ui_sgid = cpu_to_fs16(sb, value);
|
|
break;
|
|
}
|
|
}
|
|
|
|
extern dev_t ufs_get_inode_dev(struct super_block *, struct ufs_inode_info *);
|
|
extern void ufs_set_inode_dev(struct super_block *, struct ufs_inode_info *, dev_t);
|
|
extern int ufs_prepare_chunk(struct page *page, loff_t pos, unsigned len);
|
|
|
|
/*
|
|
* These functions manipulate ufs buffers
|
|
*/
|
|
#define ubh_bread(sb,fragment,size) _ubh_bread_(uspi,sb,fragment,size)
|
|
extern struct ufs_buffer_head * _ubh_bread_(struct ufs_sb_private_info *, struct super_block *, u64 , u64);
|
|
extern struct ufs_buffer_head * ubh_bread_uspi(struct ufs_sb_private_info *, struct super_block *, u64, u64);
|
|
extern void ubh_brelse (struct ufs_buffer_head *);
|
|
extern void ubh_brelse_uspi (struct ufs_sb_private_info *);
|
|
extern void ubh_mark_buffer_dirty (struct ufs_buffer_head *);
|
|
extern void ubh_mark_buffer_uptodate (struct ufs_buffer_head *, int);
|
|
extern void ubh_sync_block(struct ufs_buffer_head *);
|
|
extern void ubh_bforget (struct ufs_buffer_head *);
|
|
extern int ubh_buffer_dirty (struct ufs_buffer_head *);
|
|
#define ubh_ubhcpymem(mem,ubh,size) _ubh_ubhcpymem_(uspi,mem,ubh,size)
|
|
extern void _ubh_ubhcpymem_(struct ufs_sb_private_info *, unsigned char *, struct ufs_buffer_head *, unsigned);
|
|
#define ubh_memcpyubh(ubh,mem,size) _ubh_memcpyubh_(uspi,ubh,mem,size)
|
|
extern void _ubh_memcpyubh_(struct ufs_sb_private_info *, struct ufs_buffer_head *, unsigned char *, unsigned);
|
|
|
|
/* This functions works with cache pages*/
|
|
extern struct page *ufs_get_locked_page(struct address_space *mapping,
|
|
pgoff_t index);
|
|
static inline void ufs_put_locked_page(struct page *page)
|
|
{
|
|
unlock_page(page);
|
|
put_page(page);
|
|
}
|
|
|
|
|
|
/*
|
|
* macros and inline function to get important structures from ufs_sb_private_info
|
|
*/
|
|
|
|
static inline void *get_usb_offset(struct ufs_sb_private_info *uspi,
|
|
unsigned int offset)
|
|
{
|
|
unsigned int index;
|
|
|
|
index = offset >> uspi->s_fshift;
|
|
offset &= ~uspi->s_fmask;
|
|
return uspi->s_ubh.bh[index]->b_data + offset;
|
|
}
|
|
|
|
#define ubh_get_usb_first(uspi) \
|
|
((struct ufs_super_block_first *)get_usb_offset((uspi), 0))
|
|
|
|
#define ubh_get_usb_second(uspi) \
|
|
((struct ufs_super_block_second *)get_usb_offset((uspi), UFS_SECTOR_SIZE))
|
|
|
|
#define ubh_get_usb_third(uspi) \
|
|
((struct ufs_super_block_third *)get_usb_offset((uspi), 2*UFS_SECTOR_SIZE))
|
|
|
|
|
|
#define ubh_get_ucg(ubh) \
|
|
((struct ufs_cylinder_group *)((ubh)->bh[0]->b_data))
|
|
|
|
|
|
/*
|
|
* Extract byte from ufs_buffer_head
|
|
* Extract the bits for a block from a map inside ufs_buffer_head
|
|
*/
|
|
#define ubh_get_addr8(ubh,begin) \
|
|
((u8*)(ubh)->bh[(begin) >> uspi->s_fshift]->b_data + \
|
|
((begin) & ~uspi->s_fmask))
|
|
|
|
#define ubh_get_addr16(ubh,begin) \
|
|
(((__fs16*)((ubh)->bh[(begin) >> (uspi->s_fshift-1)]->b_data)) + \
|
|
((begin) & ((uspi->fsize>>1) - 1)))
|
|
|
|
#define ubh_get_addr32(ubh,begin) \
|
|
(((__fs32*)((ubh)->bh[(begin) >> (uspi->s_fshift-2)]->b_data)) + \
|
|
((begin) & ((uspi->s_fsize>>2) - 1)))
|
|
|
|
#define ubh_get_addr64(ubh,begin) \
|
|
(((__fs64*)((ubh)->bh[(begin) >> (uspi->s_fshift-3)]->b_data)) + \
|
|
((begin) & ((uspi->s_fsize>>3) - 1)))
|
|
|
|
#define ubh_get_addr ubh_get_addr8
|
|
|
|
static inline void *ubh_get_data_ptr(struct ufs_sb_private_info *uspi,
|
|
struct ufs_buffer_head *ubh,
|
|
u64 blk)
|
|
{
|
|
if (uspi->fs_magic == UFS2_MAGIC)
|
|
return ubh_get_addr64(ubh, blk);
|
|
else
|
|
return ubh_get_addr32(ubh, blk);
|
|
}
|
|
|
|
#define ubh_blkmap(ubh,begin,bit) \
|
|
((*ubh_get_addr(ubh, (begin) + ((bit) >> 3)) >> ((bit) & 7)) & (0xff >> (UFS_MAXFRAG - uspi->s_fpb)))
|
|
|
|
static inline u64
|
|
ufs_freefrags(struct ufs_sb_private_info *uspi)
|
|
{
|
|
return ufs_blkstofrags(uspi->cs_total.cs_nbfree) +
|
|
uspi->cs_total.cs_nffree;
|
|
}
|
|
|
|
/*
|
|
* Macros to access cylinder group array structures
|
|
*/
|
|
#define ubh_cg_blktot(ucpi,cylno) \
|
|
(*((__fs32*)ubh_get_addr(UCPI_UBH(ucpi), (ucpi)->c_btotoff + ((cylno) << 2))))
|
|
|
|
#define ubh_cg_blks(ucpi,cylno,rpos) \
|
|
(*((__fs16*)ubh_get_addr(UCPI_UBH(ucpi), \
|
|
(ucpi)->c_boff + (((cylno) * uspi->s_nrpos + (rpos)) << 1 ))))
|
|
|
|
/*
|
|
* Bitmap operations
|
|
* These functions work like classical bitmap operations.
|
|
* The difference is that we don't have the whole bitmap
|
|
* in one contiguous chunk of memory, but in several buffers.
|
|
* The parameters of each function are super_block, ufs_buffer_head and
|
|
* position of the beginning of the bitmap.
|
|
*/
|
|
#define ubh_setbit(ubh,begin,bit) \
|
|
(*ubh_get_addr(ubh, (begin) + ((bit) >> 3)) |= (1 << ((bit) & 7)))
|
|
|
|
#define ubh_clrbit(ubh,begin,bit) \
|
|
(*ubh_get_addr (ubh, (begin) + ((bit) >> 3)) &= ~(1 << ((bit) & 7)))
|
|
|
|
#define ubh_isset(ubh,begin,bit) \
|
|
(*ubh_get_addr (ubh, (begin) + ((bit) >> 3)) & (1 << ((bit) & 7)))
|
|
|
|
#define ubh_isclr(ubh,begin,bit) (!ubh_isset(ubh,begin,bit))
|
|
|
|
#define ubh_find_first_zero_bit(ubh,begin,size) _ubh_find_next_zero_bit_(uspi,ubh,begin,size,0)
|
|
|
|
#define ubh_find_next_zero_bit(ubh,begin,size,offset) _ubh_find_next_zero_bit_(uspi,ubh,begin,size,offset)
|
|
static inline unsigned _ubh_find_next_zero_bit_(
|
|
struct ufs_sb_private_info * uspi, struct ufs_buffer_head * ubh,
|
|
unsigned begin, unsigned size, unsigned offset)
|
|
{
|
|
unsigned base, count, pos;
|
|
|
|
size -= offset;
|
|
begin <<= 3;
|
|
offset += begin;
|
|
base = offset >> uspi->s_bpfshift;
|
|
offset &= uspi->s_bpfmask;
|
|
for (;;) {
|
|
count = min_t(unsigned int, size + offset, uspi->s_bpf);
|
|
size -= count - offset;
|
|
pos = find_next_zero_bit_le(ubh->bh[base]->b_data, count, offset);
|
|
if (pos < count || !size)
|
|
break;
|
|
base++;
|
|
offset = 0;
|
|
}
|
|
return (base << uspi->s_bpfshift) + pos - begin;
|
|
}
|
|
|
|
static inline unsigned find_last_zero_bit (unsigned char * bitmap,
|
|
unsigned size, unsigned offset)
|
|
{
|
|
unsigned bit, i;
|
|
unsigned char * mapp;
|
|
unsigned char map;
|
|
|
|
mapp = bitmap + (size >> 3);
|
|
map = *mapp--;
|
|
bit = 1 << (size & 7);
|
|
for (i = size; i > offset; i--) {
|
|
if ((map & bit) == 0)
|
|
break;
|
|
if ((i & 7) != 0) {
|
|
bit >>= 1;
|
|
} else {
|
|
map = *mapp--;
|
|
bit = 1 << 7;
|
|
}
|
|
}
|
|
return i;
|
|
}
|
|
|
|
#define ubh_find_last_zero_bit(ubh,begin,size,offset) _ubh_find_last_zero_bit_(uspi,ubh,begin,size,offset)
|
|
static inline unsigned _ubh_find_last_zero_bit_(
|
|
struct ufs_sb_private_info * uspi, struct ufs_buffer_head * ubh,
|
|
unsigned begin, unsigned start, unsigned end)
|
|
{
|
|
unsigned base, count, pos, size;
|
|
|
|
size = start - end;
|
|
begin <<= 3;
|
|
start += begin;
|
|
base = start >> uspi->s_bpfshift;
|
|
start &= uspi->s_bpfmask;
|
|
for (;;) {
|
|
count = min_t(unsigned int,
|
|
size + (uspi->s_bpf - start), uspi->s_bpf)
|
|
- (uspi->s_bpf - start);
|
|
size -= count;
|
|
pos = find_last_zero_bit (ubh->bh[base]->b_data,
|
|
start, start - count);
|
|
if (pos > start - count || !size)
|
|
break;
|
|
base--;
|
|
start = uspi->s_bpf;
|
|
}
|
|
return (base << uspi->s_bpfshift) + pos - begin;
|
|
}
|
|
|
|
#define ubh_isblockclear(ubh,begin,block) (!_ubh_isblockset_(uspi,ubh,begin,block))
|
|
|
|
#define ubh_isblockset(ubh,begin,block) _ubh_isblockset_(uspi,ubh,begin,block)
|
|
static inline int _ubh_isblockset_(struct ufs_sb_private_info * uspi,
|
|
struct ufs_buffer_head * ubh, unsigned begin, unsigned block)
|
|
{
|
|
u8 mask;
|
|
switch (uspi->s_fpb) {
|
|
case 8:
|
|
return (*ubh_get_addr (ubh, begin + block) == 0xff);
|
|
case 4:
|
|
mask = 0x0f << ((block & 0x01) << 2);
|
|
return (*ubh_get_addr (ubh, begin + (block >> 1)) & mask) == mask;
|
|
case 2:
|
|
mask = 0x03 << ((block & 0x03) << 1);
|
|
return (*ubh_get_addr (ubh, begin + (block >> 2)) & mask) == mask;
|
|
case 1:
|
|
mask = 0x01 << (block & 0x07);
|
|
return (*ubh_get_addr (ubh, begin + (block >> 3)) & mask) == mask;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
#define ubh_clrblock(ubh,begin,block) _ubh_clrblock_(uspi,ubh,begin,block)
|
|
static inline void _ubh_clrblock_(struct ufs_sb_private_info * uspi,
|
|
struct ufs_buffer_head * ubh, unsigned begin, unsigned block)
|
|
{
|
|
switch (uspi->s_fpb) {
|
|
case 8:
|
|
*ubh_get_addr (ubh, begin + block) = 0x00;
|
|
return;
|
|
case 4:
|
|
*ubh_get_addr (ubh, begin + (block >> 1)) &= ~(0x0f << ((block & 0x01) << 2));
|
|
return;
|
|
case 2:
|
|
*ubh_get_addr (ubh, begin + (block >> 2)) &= ~(0x03 << ((block & 0x03) << 1));
|
|
return;
|
|
case 1:
|
|
*ubh_get_addr (ubh, begin + (block >> 3)) &= ~(0x01 << ((block & 0x07)));
|
|
return;
|
|
}
|
|
}
|
|
|
|
#define ubh_setblock(ubh,begin,block) _ubh_setblock_(uspi,ubh,begin,block)
|
|
static inline void _ubh_setblock_(struct ufs_sb_private_info * uspi,
|
|
struct ufs_buffer_head * ubh, unsigned begin, unsigned block)
|
|
{
|
|
switch (uspi->s_fpb) {
|
|
case 8:
|
|
*ubh_get_addr(ubh, begin + block) = 0xff;
|
|
return;
|
|
case 4:
|
|
*ubh_get_addr(ubh, begin + (block >> 1)) |= (0x0f << ((block & 0x01) << 2));
|
|
return;
|
|
case 2:
|
|
*ubh_get_addr(ubh, begin + (block >> 2)) |= (0x03 << ((block & 0x03) << 1));
|
|
return;
|
|
case 1:
|
|
*ubh_get_addr(ubh, begin + (block >> 3)) |= (0x01 << ((block & 0x07)));
|
|
return;
|
|
}
|
|
}
|
|
|
|
static inline void ufs_fragacct (struct super_block * sb, unsigned blockmap,
|
|
__fs32 * fraglist, int cnt)
|
|
{
|
|
struct ufs_sb_private_info * uspi;
|
|
unsigned fragsize, pos;
|
|
|
|
uspi = UFS_SB(sb)->s_uspi;
|
|
|
|
fragsize = 0;
|
|
for (pos = 0; pos < uspi->s_fpb; pos++) {
|
|
if (blockmap & (1 << pos)) {
|
|
fragsize++;
|
|
}
|
|
else if (fragsize > 0) {
|
|
fs32_add(sb, &fraglist[fragsize], cnt);
|
|
fragsize = 0;
|
|
}
|
|
}
|
|
if (fragsize > 0 && fragsize < uspi->s_fpb)
|
|
fs32_add(sb, &fraglist[fragsize], cnt);
|
|
}
|
|
|
|
static inline void *ufs_get_direct_data_ptr(struct ufs_sb_private_info *uspi,
|
|
struct ufs_inode_info *ufsi,
|
|
unsigned blk)
|
|
{
|
|
BUG_ON(blk > UFS_TIND_BLOCK);
|
|
return uspi->fs_magic == UFS2_MAGIC ?
|
|
(void *)&ufsi->i_u1.u2_i_data[blk] :
|
|
(void *)&ufsi->i_u1.i_data[blk];
|
|
}
|
|
|
|
static inline u64 ufs_data_ptr_to_cpu(struct super_block *sb, void *p)
|
|
{
|
|
return UFS_SB(sb)->s_uspi->fs_magic == UFS2_MAGIC ?
|
|
fs64_to_cpu(sb, *(__fs64 *)p) :
|
|
fs32_to_cpu(sb, *(__fs32 *)p);
|
|
}
|
|
|
|
static inline void ufs_cpu_to_data_ptr(struct super_block *sb, void *p, u64 val)
|
|
{
|
|
if (UFS_SB(sb)->s_uspi->fs_magic == UFS2_MAGIC)
|
|
*(__fs64 *)p = cpu_to_fs64(sb, val);
|
|
else
|
|
*(__fs32 *)p = cpu_to_fs32(sb, val);
|
|
}
|
|
|
|
static inline void ufs_data_ptr_clear(struct ufs_sb_private_info *uspi,
|
|
void *p)
|
|
{
|
|
if (uspi->fs_magic == UFS2_MAGIC)
|
|
*(__fs64 *)p = 0;
|
|
else
|
|
*(__fs32 *)p = 0;
|
|
}
|
|
|
|
static inline int ufs_is_data_ptr_zero(struct ufs_sb_private_info *uspi,
|
|
void *p)
|
|
{
|
|
if (uspi->fs_magic == UFS2_MAGIC)
|
|
return *(__fs64 *)p == 0;
|
|
else
|
|
return *(__fs32 *)p == 0;
|
|
}
|
|
|
|
static inline __fs32 ufs_get_seconds(struct super_block *sbp)
|
|
{
|
|
time64_t now = ktime_get_real_seconds();
|
|
|
|
/* Signed 32-bit interpretation wraps around in 2038, which
|
|
* happens in ufs1 inode stamps but not ufs2 using 64-bits
|
|
* stamps. For superblock and blockgroup, let's assume
|
|
* unsigned 32-bit stamps, which are good until y2106.
|
|
* Wrap around rather than clamp here to make the dirty
|
|
* file system detection work in the superblock stamp.
|
|
*/
|
|
return cpu_to_fs32(sbp, lower_32_bits(now));
|
|
}
|