fs/buffer.c | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-)
The maximum IO size that a block device can read as a single block is
based on the min folio order and not the PAGE_SIZE as we have bs > ps
support for block devices[1].
Calculate the upper limit based on the on min folio order.
[1] https://lore.kernel.org/linux-block/20250221223823.1680616-1-mcgrof@kernel.org/
Signed-off-by: Pankaj Raghav <p.raghav@samsung.com>
---
I found this while I was adding bs > ps support to ext4. Ext4 uses this
routine to read the superblock.
fs/buffer.c | 5 +++--
1 file changed, 3 insertions(+), 2 deletions(-)
diff --git a/fs/buffer.c b/fs/buffer.c
index 8cf4a1dc481e..98f90da69a0a 100644
--- a/fs/buffer.c
+++ b/fs/buffer.c
@@ -1121,10 +1121,11 @@ __getblk_slow(struct block_device *bdev, sector_t block,
unsigned size, gfp_t gfp)
{
bool blocking = gfpflags_allow_blocking(gfp);
+ int blocklog = PAGE_SHIFT + mapping_min_folio_order(bdev->bd_mapping);
/* Size must be multiple of hard sectorsize */
- if (unlikely(size & (bdev_logical_block_size(bdev)-1) ||
- (size < 512 || size > PAGE_SIZE))) {
+ if (unlikely(size & (bdev_logical_block_size(bdev) - 1) ||
+ (size < 512 || size > (1U << blocklog)))) {
printk(KERN_ERR "getblk(): invalid block size %d requested\n",
size);
printk(KERN_ERR "logical block size: %d\n",
base-commit: e04c78d86a9699d136910cfc0bdcf01087e3267e
--
2.49.0
On Wed 18-06-25 11:17:10, Pankaj Raghav wrote: > The maximum IO size that a block device can read as a single block is > based on the min folio order and not the PAGE_SIZE as we have bs > ps > support for block devices[1]. > > Calculate the upper limit based on the on min folio order. > > [1] https://lore.kernel.org/linux-block/20250221223823.1680616-1-mcgrof@kernel.org/ > > Signed-off-by: Pankaj Raghav <p.raghav@samsung.com> ... > --- > I found this while I was adding bs > ps support to ext4. Ext4 uses this > routine to read the superblock. > > fs/buffer.c | 5 +++-- > 1 file changed, 3 insertions(+), 2 deletions(-) > > diff --git a/fs/buffer.c b/fs/buffer.c > index 8cf4a1dc481e..98f90da69a0a 100644 > --- a/fs/buffer.c > +++ b/fs/buffer.c > @@ -1121,10 +1121,11 @@ __getblk_slow(struct block_device *bdev, sector_t block, > unsigned size, gfp_t gfp) > { > bool blocking = gfpflags_allow_blocking(gfp); > + int blocklog = PAGE_SHIFT + mapping_min_folio_order(bdev->bd_mapping); > > /* Size must be multiple of hard sectorsize */ > - if (unlikely(size & (bdev_logical_block_size(bdev)-1) || > - (size < 512 || size > PAGE_SIZE))) { > + if (unlikely(size & (bdev_logical_block_size(bdev) - 1) || > + (size < 512 || size > (1U << blocklog)))) { So this doesn't quite make sense to me. Shouldn't it be capped from above by PAGE_SIZE << mapping_max_folio_order(bdev->bd_mapping)? Honza > printk(KERN_ERR "getblk(): invalid block size %d requested\n", > size); > printk(KERN_ERR "logical block size: %d\n", > > base-commit: e04c78d86a9699d136910cfc0bdcf01087e3267e > -- > 2.49.0 > -- Jan Kara <jack@suse.com> SUSE Labs, CR
> > diff --git a/fs/buffer.c b/fs/buffer.c > > index 8cf4a1dc481e..98f90da69a0a 100644 > > --- a/fs/buffer.c > > +++ b/fs/buffer.c > > @@ -1121,10 +1121,11 @@ __getblk_slow(struct block_device *bdev, sector_t block, > > unsigned size, gfp_t gfp) > > { > > bool blocking = gfpflags_allow_blocking(gfp); > > + int blocklog = PAGE_SHIFT + mapping_min_folio_order(bdev->bd_mapping); > > > > /* Size must be multiple of hard sectorsize */ > > - if (unlikely(size & (bdev_logical_block_size(bdev)-1) || > > - (size < 512 || size > PAGE_SIZE))) { > > + if (unlikely(size & (bdev_logical_block_size(bdev) - 1) || > > + (size < 512 || size > (1U << blocklog)))) { > > So this doesn't quite make sense to me. Shouldn't it be capped from above > by PAGE_SIZE << mapping_max_folio_order(bdev->bd_mapping)? This __getblk_slow() function is used to read a block from a block device and fill the page cache along with creating buffer heads. I think the reason we have this check is to make sure the size, which is block size is within the limits from 512 (SECTOR_SIZE) to upper limit on block size. That upper limit on block size was PAGE_SIZE before the lbs support in block devices, but now the upper limit of block size is mapping_min_folio_order. We set that in set_blocksize(). So a single block cannot be bigger than (PAGE_SIZE << mapping_min_folio_order). I hope that makes sense. -- Pankaj Raghav
On Wed 18-06-25 21:50:56, Pankaj Raghav (Samsung) wrote: > > > diff --git a/fs/buffer.c b/fs/buffer.c > > > index 8cf4a1dc481e..98f90da69a0a 100644 > > > --- a/fs/buffer.c > > > +++ b/fs/buffer.c > > > @@ -1121,10 +1121,11 @@ __getblk_slow(struct block_device *bdev, sector_t block, > > > unsigned size, gfp_t gfp) > > > { > > > bool blocking = gfpflags_allow_blocking(gfp); > > > + int blocklog = PAGE_SHIFT + mapping_min_folio_order(bdev->bd_mapping); > > > > > > /* Size must be multiple of hard sectorsize */ > > > - if (unlikely(size & (bdev_logical_block_size(bdev)-1) || > > > - (size < 512 || size > PAGE_SIZE))) { > > > + if (unlikely(size & (bdev_logical_block_size(bdev) - 1) || > > > + (size < 512 || size > (1U << blocklog)))) { > > > > So this doesn't quite make sense to me. Shouldn't it be capped from above > > by PAGE_SIZE << mapping_max_folio_order(bdev->bd_mapping)? > > This __getblk_slow() function is used to read a block from a block > device and fill the page cache along with creating buffer heads. > > I think the reason we have this check is to make sure the size, which is > block size is within the limits from 512 (SECTOR_SIZE) to upper limit on block size. > > That upper limit on block size was PAGE_SIZE before the lbs support in > block devices, but now the upper limit of block size is mapping_min_folio_order. > We set that in set_blocksize(). So a single block cannot be bigger than > (PAGE_SIZE << mapping_min_folio_order). Ah, right. Thanks for explanation. Feel free to add: Reviewed-by: Jan Kara <jack@suse.cz> Honza -- Jan Kara <jack@suse.com> SUSE Labs, CR
© 2016 - 2025 Red Hat, Inc.