From nobody Fri May 3 21:43:20 2024 Delivered-To: importer@patchew.org Received-SPF: none (zohomail.com: 192.237.175.120 is neither permitted nor denied by domain of lists.xenproject.org) client-ip=192.237.175.120; envelope-from=xen-devel-bounces@lists.xenproject.org; helo=lists.xenproject.org; Authentication-Results: mx.zohomail.com; dkim=fail; spf=none (zohomail.com: 192.237.175.120 is neither permitted nor denied by domain of lists.xenproject.org) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=fail(p=none dis=none) header.from=redhat.com ARC-Seal: i=1; a=rsa-sha256; t=1583886253; cv=none; d=zohomail.com; s=zohoarc; b=aQKEqZNGPBdrjZN2tbbDeji8QgikD8Tg7JB3mnQgXYMCbbLIy3Fk2+i+EW/j/sjs+Hb4k/Cr9ylzX7tPyim0dcaejzCHGgQ6oBoR70HieAYHxfD7jeaf6c/UqZKzTRYvoAeF0ENnLfDv4sUqmhc++nnyqTyf8JpF5vpmZ4LPKL4= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1583886253; h=Content-Type:Content-Transfer-Encoding:Cc:Date:From:List-Subscribe:List-Post:List-Id:List-Help:List-Unsubscribe:MIME-Version:Message-ID:Sender:Subject:To; bh=ATpbzdw7cZQEHAUNbjKm+k6wH5CG6GAr1gH8qB9eQys=; b=E54MpOMYiZJejFpuhaJWjoIVHTfapWEhAe0ZVIGblKyzW67RQPRVBVPZkx9Ep7HP8JySAcClP4saG3SdtVnMGQJB94Ae+vHp9p7O9T0aatBXtLY1XLg1Y5v546fPOVteC9xYRJdQQTFwheazIRCvzhEl2fMr4OjtrdFG6/W+l+4= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail; spf=none (zohomail.com: 192.237.175.120 is neither permitted nor denied by domain of lists.xenproject.org) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.xenproject.org (lists.xenproject.org [192.237.175.120]) by mx.zohomail.com with SMTPS id 1583886253176579.3163384979605; Tue, 10 Mar 2020 17:24:13 -0700 (PDT) Received: from localhost ([127.0.0.1] helo=lists.xenproject.org) by lists.xenproject.org with esmtp (Exim 4.89) (envelope-from ) id 1jBp9W-0002Nl-74; Wed, 11 Mar 2020 00:23:14 +0000 Received: from all-amaz-eas1.inumbo.com ([34.197.232.57] helo=us1-amaz-eas2.inumbo.com) by lists.xenproject.org with esmtp (Exim 4.89) (envelope-from ) id 1jBp9U-0002Ng-Ov for xen-devel@lists.xenproject.org; Wed, 11 Mar 2020 00:23:12 +0000 Received: from us-smtp-1.mimecast.com (unknown [205.139.110.61]) by us1-amaz-eas2.inumbo.com (Halon) with ESMTP id 7d1f1918-632e-11ea-ae8c-12813bfff9fa; Wed, 11 Mar 2020 00:23:10 +0000 (UTC) Received: from mail-wr1-f69.google.com (mail-wr1-f69.google.com [209.85.221.69]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-402-jFon7Y2GN3Sf0a1N-7REgA-1; Tue, 10 Mar 2020 20:23:08 -0400 Received: by mail-wr1-f69.google.com with SMTP id 31so96851wrq.0 for ; Tue, 10 Mar 2020 17:23:08 -0700 (PDT) Received: from raver.teknoraver.net (net-93-144-9-178.cust.vodafonedsl.it. [93.144.9.178]) by smtp.gmail.com with ESMTPSA id c2sm6285481wma.39.2020.03.10.17.23.05 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 10 Mar 2020 17:23:05 -0700 (PDT) X-Inumbo-ID: 7d1f1918-632e-11ea-ae8c-12813bfff9fa DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1583886190; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding; bh=wmOVNWU5sk2f0X2YNiN/aFt2nLXVZ8y5oHh3vhFdAik=; b=ELDwxQSivJ5ITa0la6lmtU335US6CZEDHKqqIhRCGx7rMNwWXC5F2+7tXoklg4X/JEewRI CZo2fjcjF6xh+MvktUmtF6j7yWTS8HhTXDEgzE1qy7BHt+1P1tWPMMudL4rZJ8/IqMHXHL BqRGu7haDl60/AxATo+tr8h07JCEf/I= X-MC-Unique: jFon7Y2GN3Sf0a1N-7REgA-1 X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:mime-version :content-transfer-encoding; bh=AF0phWtBzaybkf0W4o0XLko38kCzMiwJ9FW0oNdt4Ec=; b=qlLgf2XCqLvDAfdmwpcWL7RFZHML86i6JTDMjdgdu/bMkdB9T3BLEMUdrby98GzfOd lznRm93YiKUdsLRON5nIOGEoWG0B4K7HlInn37T+9sGwL9TuD2Ca75zS2DLhjNZV5raa 2aCNNJZpR1VWbl8+IUhJPiocA1vQCDC8jWSR1uHEFlhafwk0qTxrAo6r9zggfBczABaC ZVuYuN+4ZoMnrzGHXpC1Wj9u+eKnbPAkGXjirigqi9GYcAlmg/gwyFkJOQiJuDo+vRw7 G+uHXGjW+Jkh+maL3pnRdtxild+4XSmv63jt/EZ3iCgGuhDHqfd7gxdIH0N84agwUPBI +L+w== X-Gm-Message-State: ANhLgQ2rRPhQere/otKYxaqduf86mrnM/umpZTetOstBsq61SJkZ1JQa eVxCd6FYCXCaPtomaUB47B0nxwkdxIbVBauk+0SQlTum3HmXuQUDHYOKn/oQNb4JHoviuO3o79C eCP5V3m6accgFV2JgDIHS9ZVCVOY= X-Received: by 2002:a05:600c:24b:: with SMTP id 11mr135685wmj.1.1583886186684; Tue, 10 Mar 2020 17:23:06 -0700 (PDT) X-Google-Smtp-Source: ADFU+vskya3VIJ2O35SM2qD5qkRNEFkTaIMR/iZ1Olb+WBpvswMPPj2jGybAw7/Dtv1byE5BePKdLg== X-Received: by 2002:a05:600c:24b:: with SMTP id 11mr135633wmj.1.1583886186002; Tue, 10 Mar 2020 17:23:06 -0700 (PDT) From: Matteo Croce To: linux-block@vger.kernel.org, linux-kernel@vger.kernel.org, linux-nvdimm@lists.01.org, linux-bcache@vger.kernel.org, linux-raid@vger.kernel.org, linux-mmc@vger.kernel.org, xen-devel@lists.xenproject.org, linux-scsi@vger.kernel.org, linux-nfs@vger.kernel.org Date: Wed, 11 Mar 2020 01:22:54 +0100 Message-Id: <20200311002254.121365-1-mcroce@redhat.com> X-Mailer: git-send-email 2.24.1 MIME-Version: 1.0 X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Subject: [Xen-devel] [PATCH v3] block: refactor duplicated macros X-BeenThere: xen-devel@lists.xenproject.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: Xen developer discussion List-Unsubscribe: , List-Post: List-Help: List-Subscribe: , Cc: Jens Axboe , Ulf Hansson , "James E.J. Bottomley" , Guoqing Jiang , Song Liu , Dan Williams , Anna Schumaker Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: quoted-printable Errors-To: xen-devel-bounces@lists.xenproject.org Sender: "Xen-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) The macros PAGE_SECTORS, PAGE_SECTORS_SHIFT and SECTOR_MASK are defined several times in different flavours across the whole tree. Define them just once in a common header. While at it, replace replace "PAGE_SHIFT - 9" with "PAGE_SECTORS_SHIFT" too and rename SECTOR_MASK to PAGE_SECTORS_MASK. Signed-off-by: Matteo Croce Acked-by: Coly Li Reviewed-by: Martin K. Petersen --- v3: As Guoqing Jiang suggested, replace "PAGE_SHIFT - 9" with "PAGE_SECTORS_SHI= FT" v2: As Dan Williams suggested: #define PAGE_SECTORS_MASK (~(PAGE_SECTORS - 1)) block/blk-lib.c | 2 +- block/blk-settings.c | 4 ++-- block/partition-generic.c | 2 +- drivers/block/brd.c | 3 --- drivers/block/null_blk_main.c | 14 +++++--------- drivers/block/zram/zram_drv.c | 8 ++++---- drivers/block/zram/zram_drv.h | 2 -- drivers/dax/super.c | 2 +- drivers/md/bcache/util.h | 2 -- drivers/md/dm-bufio.c | 6 +++--- drivers/md/dm-integrity.c | 10 +++++----- drivers/md/dm-table.c | 2 +- drivers/md/md.c | 4 ++-- drivers/md/raid1.c | 2 +- drivers/md/raid10.c | 2 +- drivers/md/raid5-cache.c | 10 +++++----- drivers/md/raid5.h | 2 +- drivers/mmc/core/host.c | 3 ++- drivers/nvme/host/fc.c | 2 +- drivers/nvme/target/loop.c | 2 +- drivers/scsi/xen-scsifront.c | 4 ++-- fs/erofs/internal.h | 2 +- fs/ext2/dir.c | 2 +- fs/iomap/buffered-io.c | 2 +- fs/libfs.c | 2 +- fs/nfs/blocklayout/blocklayout.h | 2 -- fs/nilfs2/dir.c | 2 +- include/linux/blkdev.h | 4 ++++ include/linux/device-mapper.h | 1 - mm/page_io.c | 4 ++-- mm/swapfile.c | 12 ++++++------ 31 files changed, 56 insertions(+), 65 deletions(-) diff --git a/block/blk-lib.c b/block/blk-lib.c index 5f2c429d4378..f5e705d307e0 100644 --- a/block/blk-lib.c +++ b/block/blk-lib.c @@ -260,7 +260,7 @@ static int __blkdev_issue_write_zeroes(struct block_dev= ice *bdev, */ static unsigned int __blkdev_sectors_to_bio_pages(sector_t nr_sects) { - sector_t pages =3D DIV_ROUND_UP_SECTOR_T(nr_sects, PAGE_SIZE / 512); + sector_t pages =3D DIV_ROUND_UP_SECTOR_T(nr_sects, PAGE_SECTORS); =20 return min(pages, (sector_t)BIO_MAX_PAGES); } diff --git a/block/blk-settings.c b/block/blk-settings.c index c8eda2e7b91e..e29933581ac0 100644 --- a/block/blk-settings.c +++ b/block/blk-settings.c @@ -186,7 +186,7 @@ void blk_queue_max_hw_sectors(struct request_queue *q, = unsigned int max_hw_secto unsigned int max_sectors; =20 if ((max_hw_sectors << 9) < PAGE_SIZE) { - max_hw_sectors =3D 1 << (PAGE_SHIFT - 9); + max_hw_sectors =3D 1 << PAGE_SECTORS_SHIFT; printk(KERN_INFO "%s: set to minimum %d\n", __func__, max_hw_sectors); } @@ -195,7 +195,7 @@ void blk_queue_max_hw_sectors(struct request_queue *q, = unsigned int max_hw_secto max_sectors =3D min_not_zero(max_hw_sectors, limits->max_dev_sectors); max_sectors =3D min_t(unsigned int, max_sectors, BLK_DEF_MAX_SECTORS); limits->max_sectors =3D max_sectors; - q->backing_dev_info->io_pages =3D max_sectors >> (PAGE_SHIFT - 9); + q->backing_dev_info->io_pages =3D max_sectors >> PAGE_SECTORS_SHIFT; } EXPORT_SYMBOL(blk_queue_max_hw_sectors); =20 diff --git a/block/partition-generic.c b/block/partition-generic.c index 564fae77711d..5f06b7a6c119 100644 --- a/block/partition-generic.c +++ b/block/partition-generic.c @@ -612,7 +612,7 @@ unsigned char *read_dev_sector(struct block_device *bde= v, sector_t n, Sector *p) if (PageError(page)) goto fail; p->v =3D page; - return (unsigned char *)page_address(page) + ((n & ((1 << (PAGE_SHIFT -= 9)) - 1)) << 9); + return (unsigned char *)page_address(page) + ((n & ((1 << PAGE_SECTORS_= SHIFT) - 1)) << 9); fail: put_page(page); } diff --git a/drivers/block/brd.c b/drivers/block/brd.c index 220c5e18aba0..33e2cbe11400 100644 --- a/drivers/block/brd.c +++ b/drivers/block/brd.c @@ -25,9 +25,6 @@ =20 #include =20 -#define PAGE_SECTORS_SHIFT (PAGE_SHIFT - SECTOR_SHIFT) -#define PAGE_SECTORS (1 << PAGE_SECTORS_SHIFT) - /* * Each block ramdisk device has a radix_tree brd_pages of pages that stor= es * the pages containing the block device's contents. A brd page's ->index = is diff --git a/drivers/block/null_blk_main.c b/drivers/block/null_blk_main.c index 16510795e377..b5ebde97bf0b 100644 --- a/drivers/block/null_blk_main.c +++ b/drivers/block/null_blk_main.c @@ -11,10 +11,6 @@ #include #include "null_blk.h" =20 -#define PAGE_SECTORS_SHIFT (PAGE_SHIFT - SECTOR_SHIFT) -#define PAGE_SECTORS (1 << PAGE_SECTORS_SHIFT) -#define SECTOR_MASK (PAGE_SECTORS - 1) - #define FREE_BATCH 16 =20 #define TICKS_PER_SEC 50ULL @@ -721,7 +717,7 @@ static void null_free_sector(struct nullb *nullb, secto= r_t sector, =20 root =3D is_cache ? &nullb->dev->cache : &nullb->dev->data; idx =3D sector >> PAGE_SECTORS_SHIFT; - sector_bit =3D (sector & SECTOR_MASK); + sector_bit =3D (sector & PAGE_SECTORS_MASK); =20 t_page =3D radix_tree_lookup(root, idx); if (t_page) { @@ -792,7 +788,7 @@ static struct nullb_page *__null_lookup_page(struct nul= lb *nullb, struct radix_tree_root *root; =20 idx =3D sector >> PAGE_SECTORS_SHIFT; - sector_bit =3D (sector & SECTOR_MASK); + sector_bit =3D (sector & PAGE_SECTORS_MASK); =20 root =3D is_cache ? &nullb->dev->cache : &nullb->dev->data; t_page =3D radix_tree_lookup(root, idx); @@ -967,7 +963,7 @@ static int copy_to_nullb(struct nullb *nullb, struct pa= ge *source, if (null_cache_active(nullb) && !is_fua) null_make_cache_space(nullb, PAGE_SIZE); =20 - offset =3D (sector & SECTOR_MASK) << SECTOR_SHIFT; + offset =3D (sector & PAGE_SECTORS_MASK) << SECTOR_SHIFT; t_page =3D null_insert_page(nullb, sector, !null_cache_active(nullb) || is_fua); if (!t_page) @@ -979,7 +975,7 @@ static int copy_to_nullb(struct nullb *nullb, struct pa= ge *source, kunmap_atomic(dst); kunmap_atomic(src); =20 - __set_bit(sector & SECTOR_MASK, t_page->bitmap); + __set_bit(sector & PAGE_SECTORS_MASK, t_page->bitmap); =20 if (is_fua) null_free_sector(nullb, sector, true); @@ -1001,7 +997,7 @@ static int copy_from_nullb(struct nullb *nullb, struct= page *dest, while (count < n) { temp =3D min_t(size_t, nullb->dev->blocksize, n - count); =20 - offset =3D (sector & SECTOR_MASK) << SECTOR_SHIFT; + offset =3D (sector & PAGE_SECTORS_MASK) << SECTOR_SHIFT; t_page =3D null_lookup_page(nullb, sector, false, !null_cache_active(nullb)); =20 diff --git a/drivers/block/zram/zram_drv.c b/drivers/block/zram/zram_drv.c index 1bdb5793842b..725bd33f302d 100644 --- a/drivers/block/zram/zram_drv.c +++ b/drivers/block/zram/zram_drv.c @@ -1548,9 +1548,9 @@ static void __zram_make_request(struct zram *zram, st= ruct bio *bio) struct bio_vec bvec; struct bvec_iter iter; =20 - index =3D bio->bi_iter.bi_sector >> SECTORS_PER_PAGE_SHIFT; + index =3D bio->bi_iter.bi_sector >> PAGE_SECTORS_SHIFT; offset =3D (bio->bi_iter.bi_sector & - (SECTORS_PER_PAGE - 1)) << SECTOR_SHIFT; + PAGE_SECTORS_MASK) << SECTOR_SHIFT; =20 switch (bio_op(bio)) { case REQ_OP_DISCARD: @@ -1643,8 +1643,8 @@ static int zram_rw_page(struct block_device *bdev, se= ctor_t sector, goto out; } =20 - index =3D sector >> SECTORS_PER_PAGE_SHIFT; - offset =3D (sector & (SECTORS_PER_PAGE - 1)) << SECTOR_SHIFT; + index =3D sector >> PAGE_SECTORS_SHIFT; + offset =3D (sector & PAGE_SECTORS_MASK) << SECTOR_SHIFT; =20 bv.bv_page =3D page; bv.bv_len =3D PAGE_SIZE; diff --git a/drivers/block/zram/zram_drv.h b/drivers/block/zram/zram_drv.h index f2fd46daa760..12309175d55e 100644 --- a/drivers/block/zram/zram_drv.h +++ b/drivers/block/zram/zram_drv.h @@ -21,8 +21,6 @@ =20 #include "zcomp.h" =20 -#define SECTORS_PER_PAGE_SHIFT (PAGE_SHIFT - SECTOR_SHIFT) -#define SECTORS_PER_PAGE (1 << SECTORS_PER_PAGE_SHIFT) #define ZRAM_LOGICAL_BLOCK_SHIFT 12 #define ZRAM_LOGICAL_BLOCK_SIZE (1 << ZRAM_LOGICAL_BLOCK_SHIFT) #define ZRAM_SECTOR_PER_LOGICAL_BLOCK \ diff --git a/drivers/dax/super.c b/drivers/dax/super.c index 0aa4b6bc5101..7f7672f72085 100644 --- a/drivers/dax/super.c +++ b/drivers/dax/super.c @@ -92,7 +92,7 @@ bool __generic_fsdax_supported(struct dax_device *dax_dev, return false; } =20 - last_page =3D PFN_DOWN((start + sectors - 1) * 512) * PAGE_SIZE / 512; + last_page =3D PFN_DOWN((start + sectors - 1) * 512) * PAGE_SECTORS; err =3D bdev_dax_pgoff(bdev, last_page, PAGE_SIZE, &pgoff_end); if (err) { pr_debug("%s: error: unaligned partition for dax\n", diff --git a/drivers/md/bcache/util.h b/drivers/md/bcache/util.h index c029f7443190..55196e0f37c3 100644 --- a/drivers/md/bcache/util.h +++ b/drivers/md/bcache/util.h @@ -15,8 +15,6 @@ =20 #include "closure.h" =20 -#define PAGE_SECTORS (PAGE_SIZE / 512) - struct closure; =20 #ifdef CONFIG_BCACHE_DEBUG diff --git a/drivers/md/dm-bufio.c b/drivers/md/dm-bufio.c index 2d519c223562..f4496ce0d598 100644 --- a/drivers/md/dm-bufio.c +++ b/drivers/md/dm-bufio.c @@ -384,7 +384,7 @@ static void *alloc_buffer_data(struct dm_bufio_client *= c, gfp_t gfp_mask, gfp_mask & __GFP_NORETRY) { *data_mode =3D DATA_MODE_GET_FREE_PAGES; return (void *)__get_free_pages(gfp_mask, - c->sectors_per_block_bits - (PAGE_SHIFT - SECTOR_SHIFT)); + c->sectors_per_block_bits - PAGE_SECTORS_SHIFT); } =20 *data_mode =3D DATA_MODE_VMALLOC; @@ -422,7 +422,7 @@ static void free_buffer_data(struct dm_bufio_client *c, =20 case DATA_MODE_GET_FREE_PAGES: free_pages((unsigned long)data, - c->sectors_per_block_bits - (PAGE_SHIFT - SECTOR_SHIFT)); + c->sectors_per_block_bits - PAGE_SECTORS_SHIFT); break; =20 case DATA_MODE_VMALLOC: @@ -597,7 +597,7 @@ static void use_bio(struct dm_buffer *b, int rw, sector= _t sector, unsigned vec_size, len; =20 vec_size =3D b->c->block_size >> PAGE_SHIFT; - if (unlikely(b->c->sectors_per_block_bits < PAGE_SHIFT - SECTOR_SHIFT)) + if (unlikely(b->c->sectors_per_block_bits < PAGE_SECTORS_SHIFT)) vec_size +=3D 2; =20 bio =3D bio_kmalloc(GFP_NOWAIT | __GFP_NORETRY | __GFP_NOWARN, vec_size); diff --git a/drivers/md/dm-integrity.c b/drivers/md/dm-integrity.c index b225b3e445fa..4e60cda465cc 100644 --- a/drivers/md/dm-integrity.c +++ b/drivers/md/dm-integrity.c @@ -652,7 +652,7 @@ static void page_list_location(struct dm_integrity_c *i= c, unsigned section, unsi =20 sector =3D section * ic->journal_section_sectors + offset; =20 - *pl_index =3D sector >> (PAGE_SHIFT - SECTOR_SHIFT); + *pl_index =3D sector >> PAGE_SECTORS_SHIFT; *pl_offset =3D (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); } =20 @@ -951,7 +951,7 @@ static void rw_journal_sectors(struct dm_integrity_c *i= c, int op, int op_flags, return; } =20 - pl_index =3D sector >> (PAGE_SHIFT - SECTOR_SHIFT); + pl_index =3D sector >> PAGE_SECTORS_SHIFT; pl_offset =3D (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); =20 io_req.bi_op =3D op; @@ -1072,7 +1072,7 @@ static void copy_from_journal(struct dm_integrity_c *= ic, unsigned section, unsig =20 sector =3D section * ic->journal_section_sectors + JOURNAL_BLOCK_SECTORS = + offset; =20 - pl_index =3D sector >> (PAGE_SHIFT - SECTOR_SHIFT); + pl_index =3D sector >> PAGE_SECTORS_SHIFT; pl_offset =3D (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); =20 io_req.bi_op =3D REQ_OP_WRITE; @@ -3343,7 +3343,7 @@ static int create_journal(struct dm_integrity_c *ic, = char **error) ic->commit_ids[3] =3D cpu_to_le64(0x4444444444444444ULL); =20 journal_pages =3D roundup((__u64)ic->journal_sections * ic->journal_secti= on_sectors, - PAGE_SIZE >> SECTOR_SHIFT) >> (PAGE_SHIFT - SECTOR_SHIFT); + PAGE_SIZE >> SECTOR_SHIFT) >> PAGE_SECTORS_SHIFT; journal_desc_size =3D journal_pages * sizeof(struct page_list); if (journal_pages >=3D totalram_pages() - totalhigh_pages() || journal_de= sc_size > ULONG_MAX) { *error =3D "Journal doesn't fit into memory"; @@ -4075,7 +4075,7 @@ static int dm_integrity_ctr(struct dm_target *ti, uns= igned argc, char **argv) spin_lock_init(&bbs->bio_queue_lock); =20 sector =3D i * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT); - pl_index =3D sector >> (PAGE_SHIFT - SECTOR_SHIFT); + pl_index =3D sector >> PAGE_SECTORS_SHIFT; pl_offset =3D (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); =20 bbs->bitmap =3D lowmem_page_address(ic->journal[pl_index].page) + pl_of= fset; diff --git a/drivers/md/dm-table.c b/drivers/md/dm-table.c index 0a2cc197f62b..56000202492e 100644 --- a/drivers/md/dm-table.c +++ b/drivers/md/dm-table.c @@ -1964,7 +1964,7 @@ void dm_table_set_restrictions(struct dm_table *t, st= ruct request_queue *q, #endif =20 /* Allow reads to exceed readahead limits */ - q->backing_dev_info->io_pages =3D limits->max_sectors >> (PAGE_SHIFT - 9); + q->backing_dev_info->io_pages =3D limits->max_sectors >> PAGE_SECTORS_SHI= FT; } =20 unsigned int dm_table_get_num_targets(struct dm_table *t) diff --git a/drivers/md/md.c b/drivers/md/md.c index 469f551863be..b28f9390608f 100644 --- a/drivers/md/md.c +++ b/drivers/md/md.c @@ -1734,7 +1734,7 @@ static int super_1_load(struct md_rdev *rdev, struct = md_rdev *refdev, int minor_ __le64 *bbp; int i; int sectors =3D le16_to_cpu(sb->bblog_size); - if (sectors > (PAGE_SIZE / 512)) + if (sectors > PAGE_SECTORS) return -EINVAL; offset =3D le32_to_cpu(sb->bblog_offset); if (offset =3D=3D 0) @@ -8733,7 +8733,7 @@ void md_do_sync(struct md_thread *thread) /* * Tune reconstruction: */ - window =3D 32 * (PAGE_SIZE / 512); + window =3D 32 * PAGE_SECTORS; pr_debug("md: using %dk window, over a total of %lluk.\n", window/2, (unsigned long long)max_sectors/2); =20 diff --git a/drivers/md/raid1.c b/drivers/md/raid1.c index cd810e195086..fd70fc62ad11 100644 --- a/drivers/md/raid1.c +++ b/drivers/md/raid1.c @@ -2129,7 +2129,7 @@ static void process_checks(struct r1bio *r1_bio) int vcnt; =20 /* Fix variable parts of all bios */ - vcnt =3D (r1_bio->sectors + PAGE_SIZE / 512 - 1) >> (PAGE_SHIFT - 9); + vcnt =3D (r1_bio->sectors + PAGE_SECTORS - 1) >> PAGE_SECTORS_SHIFT; for (i =3D 0; i < conf->raid_disks * 2; i++) { blk_status_t status; struct bio *b =3D r1_bio->bios[i]; diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c index ec136e44aef7..889c4b7c1c35 100644 --- a/drivers/md/raid10.c +++ b/drivers/md/raid10.c @@ -2029,7 +2029,7 @@ static void sync_request_write(struct mddev *mddev, s= truct r10bio *r10_bio) fbio->bi_iter.bi_idx =3D 0; fpages =3D get_resync_pages(fbio)->pages; =20 - vcnt =3D (r10_bio->sectors + (PAGE_SIZE >> 9) - 1) >> (PAGE_SHIFT - 9); + vcnt =3D (r10_bio->sectors + (PAGE_SIZE >> 9) - 1) >> PAGE_SECTORS_SHIFT; /* now find blocks with errors */ for (i=3D0 ; i < conf->copies ; i++) { int j, d; diff --git a/drivers/md/raid5-cache.c b/drivers/md/raid5-cache.c index 9b6da759dca2..ad5470c12bb3 100644 --- a/drivers/md/raid5-cache.c +++ b/drivers/md/raid5-cache.c @@ -833,7 +833,7 @@ static void r5l_append_payload_meta(struct r5l_log *log= , u16 type, payload->header.type =3D cpu_to_le16(type); payload->header.flags =3D cpu_to_le16(0); payload->size =3D cpu_to_le32((1 + !!checksum2_valid) << - (PAGE_SHIFT - 9)); + PAGE_SECTORS_SHIFT); payload->location =3D cpu_to_le64(location); payload->checksum[0] =3D cpu_to_le32(checksum1); if (checksum2_valid) @@ -1042,7 +1042,7 @@ int r5l_write_stripe(struct r5l_log *log, struct stri= pe_head *sh) =20 mutex_lock(&log->io_mutex); /* meta + data */ - reserve =3D (1 + write_disks) << (PAGE_SHIFT - 9); + reserve =3D (1 + write_disks) << PAGE_SECTORS_SHIFT; =20 if (log->r5c_journal_mode =3D=3D R5C_JOURNAL_MODE_WRITE_THROUGH) { if (!r5l_has_free_space(log, reserve)) { @@ -2053,7 +2053,7 @@ r5l_recovery_verify_data_checksum_for_mb(struct r5l_l= og *log, le32_to_cpu(payload->size)); mb_offset +=3D sizeof(struct r5l_payload_data_parity) + sizeof(__le32) * - (le32_to_cpu(payload->size) >> (PAGE_SHIFT - 9)); + (le32_to_cpu(payload->size) >> PAGE_SECTORS_SHIFT); } =20 } @@ -2199,7 +2199,7 @@ r5c_recovery_analyze_meta_block(struct r5l_log *log, =20 mb_offset +=3D sizeof(struct r5l_payload_data_parity) + sizeof(__le32) * - (le32_to_cpu(payload->size) >> (PAGE_SHIFT - 9)); + (le32_to_cpu(payload->size) >> PAGE_SECTORS_SHIFT); } =20 return 0; @@ -2916,7 +2916,7 @@ int r5c_cache_data(struct r5l_log *log, struct stripe= _head *sh) =20 mutex_lock(&log->io_mutex); /* meta + data */ - reserve =3D (1 + pages) << (PAGE_SHIFT - 9); + reserve =3D (1 + pages) << PAGE_SECTORS_SHIFT; =20 if (test_bit(R5C_LOG_CRITICAL, &conf->cache_state) && sh->log_start =3D=3D MaxSector) diff --git a/drivers/md/raid5.h b/drivers/md/raid5.h index f90e0704bed9..49dfd8a9544a 100644 --- a/drivers/md/raid5.h +++ b/drivers/md/raid5.h @@ -473,7 +473,7 @@ struct disk_info { =20 #define NR_STRIPES 256 #define STRIPE_SIZE PAGE_SIZE -#define STRIPE_SHIFT (PAGE_SHIFT - 9) +#define STRIPE_SHIFT PAGE_SECTORS_SHIFT #define STRIPE_SECTORS (STRIPE_SIZE>>9) #define IO_THRESHOLD 1 #define BYPASS_THRESHOLD 1 diff --git a/drivers/mmc/core/host.c b/drivers/mmc/core/host.c index c8768726d925..4a23fb9d5642 100644 --- a/drivers/mmc/core/host.c +++ b/drivers/mmc/core/host.c @@ -18,6 +18,7 @@ #include #include #include +#include =20 #include #include @@ -427,7 +428,7 @@ struct mmc_host *mmc_alloc_host(int extra, struct devic= e *dev) =20 host->max_req_size =3D PAGE_SIZE; host->max_blk_size =3D 512; - host->max_blk_count =3D PAGE_SIZE / 512; + host->max_blk_count =3D PAGE_SECTORS; =20 host->fixed_drv_type =3D -EINVAL; host->ios.power_delay_ms =3D 10; diff --git a/drivers/nvme/host/fc.c b/drivers/nvme/host/fc.c index 5a70ac395d53..3401e82ea323 100644 --- a/drivers/nvme/host/fc.c +++ b/drivers/nvme/host/fc.c @@ -2674,7 +2674,7 @@ nvme_fc_create_association(struct nvme_fc_ctrl *ctrl) goto out_disconnect_admin_queue; =20 ctrl->ctrl.max_hw_sectors =3D - (ctrl->lport->ops->max_sgl_segments - 1) << (PAGE_SHIFT - 9); + (ctrl->lport->ops->max_sgl_segments - 1) << PAGE_SECTORS_SHIFT; =20 blk_mq_unquiesce_queue(ctrl->ctrl.admin_q); =20 diff --git a/drivers/nvme/target/loop.c b/drivers/nvme/target/loop.c index 4df4ebde208a..d43ac4b2ebef 100644 --- a/drivers/nvme/target/loop.c +++ b/drivers/nvme/target/loop.c @@ -382,7 +382,7 @@ static int nvme_loop_configure_admin_queue(struct nvme_= loop_ctrl *ctrl) goto out_cleanup_queue; =20 ctrl->ctrl.max_hw_sectors =3D - (NVME_LOOP_MAX_SEGMENTS - 1) << (PAGE_SHIFT - 9); + (NVME_LOOP_MAX_SEGMENTS - 1) << PAGE_SECTORS_SHIFT; =20 blk_mq_unquiesce_queue(ctrl->ctrl.admin_q); =20 diff --git a/drivers/scsi/xen-scsifront.c b/drivers/scsi/xen-scsifront.c index f0068e96a177..e6b29e54d07a 100644 --- a/drivers/scsi/xen-scsifront.c +++ b/drivers/scsi/xen-scsifront.c @@ -852,7 +852,7 @@ static int scsifront_probe(struct xenbus_device *dev, host->max_id =3D VSCSIIF_MAX_TARGET; host->max_channel =3D 0; host->max_lun =3D VSCSIIF_MAX_LUN; - host->max_sectors =3D (host->sg_tablesize - 1) * PAGE_SIZE / 512; + host->max_sectors =3D (host->sg_tablesize - 1) * PAGE_SECTORS; host->max_cmd_len =3D VSCSIIF_MAX_COMMAND_SIZE; =20 err =3D scsi_add_host(host, &dev->dev); @@ -1073,7 +1073,7 @@ static void scsifront_read_backend_params(struct xenb= us_device *dev, host->sg_tablesize, nr_segs); =20 host->sg_tablesize =3D nr_segs; - host->max_sectors =3D (nr_segs - 1) * PAGE_SIZE / 512; + host->max_sectors =3D (nr_segs - 1) * PAGE_SECTORS; } =20 static void scsifront_backend_changed(struct xenbus_device *dev, diff --git a/fs/erofs/internal.h b/fs/erofs/internal.h index c4c6dcdc89ad..210ca148cc0c 100644 --- a/fs/erofs/internal.h +++ b/fs/erofs/internal.h @@ -189,7 +189,7 @@ static inline int erofs_wait_on_workgroup_freezed(struc= t erofs_workgroup *grp) #define LOG_BLOCK_SIZE PAGE_SHIFT =20 #undef LOG_SECTORS_PER_BLOCK -#define LOG_SECTORS_PER_BLOCK (PAGE_SHIFT - 9) +#define LOG_SECTORS_PER_BLOCK PAGE_SECTORS_SHIFT =20 #undef SECTORS_PER_BLOCK #define SECTORS_PER_BLOCK (1 << SECTORS_PER_BLOCK) diff --git a/fs/ext2/dir.c b/fs/ext2/dir.c index 13318e255ebf..5014a6fc5c6e 100644 --- a/fs/ext2/dir.c +++ b/fs/ext2/dir.c @@ -385,7 +385,7 @@ struct ext2_dir_entry_2 *ext2_find_entry (struct inode = *dir, if (++n >=3D npages) n =3D 0; /* next page is past the blocks we've got */ - if (unlikely(n > (dir->i_blocks >> (PAGE_SHIFT - 9)))) { + if (unlikely(n > (dir->i_blocks >> PAGE_SECTORS_SHIFT))) { ext2_error(dir->i_sb, __func__, "dir %lu size %lld exceeds block count %llu", dir->i_ino, dir->i_size, diff --git a/fs/iomap/buffered-io.c b/fs/iomap/buffered-io.c index 7c84c4c027c4..60505fc156c5 100644 --- a/fs/iomap/buffered-io.c +++ b/fs/iomap/buffered-io.c @@ -29,7 +29,7 @@ struct iomap_page { atomic_t read_count; atomic_t write_count; spinlock_t uptodate_lock; - DECLARE_BITMAP(uptodate, PAGE_SIZE / 512); + DECLARE_BITMAP(uptodate, PAGE_SECTORS); }; =20 static inline struct iomap_page *to_iomap_page(struct page *page) diff --git a/fs/libfs.c b/fs/libfs.c index c686bd9caac6..00ca96752fcb 100644 --- a/fs/libfs.c +++ b/fs/libfs.c @@ -30,7 +30,7 @@ int simple_getattr(const struct path *path, struct kstat = *stat, { struct inode *inode =3D d_inode(path->dentry); generic_fillattr(inode, stat); - stat->blocks =3D inode->i_mapping->nrpages << (PAGE_SHIFT - 9); + stat->blocks =3D inode->i_mapping->nrpages << PAGE_SECTORS_SHIFT; return 0; } EXPORT_SYMBOL(simple_getattr); diff --git a/fs/nfs/blocklayout/blocklayout.h b/fs/nfs/blocklayout/blocklay= out.h index 716bc75e9ed2..22407751e0fd 100644 --- a/fs/nfs/blocklayout/blocklayout.h +++ b/fs/nfs/blocklayout/blocklayout.h @@ -40,8 +40,6 @@ #include "../pnfs.h" #include "../netns.h" =20 -#define PAGE_CACHE_SECTORS (PAGE_SIZE >> SECTOR_SHIFT) -#define PAGE_CACHE_SECTOR_SHIFT (PAGE_SHIFT - SECTOR_SHIFT) #define SECTOR_SIZE (1 << SECTOR_SHIFT) =20 struct pnfs_block_dev; diff --git a/fs/nilfs2/dir.c b/fs/nilfs2/dir.c index 81394e22d0a0..77c06e999828 100644 --- a/fs/nilfs2/dir.c +++ b/fs/nilfs2/dir.c @@ -368,7 +368,7 @@ nilfs_find_entry(struct inode *dir, const struct qstr *= qstr, if (++n >=3D npages) n =3D 0; /* next page is past the blocks we've got */ - if (unlikely(n > (dir->i_blocks >> (PAGE_SHIFT - 9)))) { + if (unlikely(n > (dir->i_blocks >> PAGE_SECTORS_SHIFT))) { nilfs_error(dir->i_sb, "dir %lu size %lld exceeds block count %llu", dir->i_ino, dir->i_size, diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index 053ea4b51988..98b2cbd0299a 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h @@ -910,6 +910,10 @@ static inline struct request_queue *bdev_get_queue(str= uct block_device *bdev) #define SECTOR_SIZE (1 << SECTOR_SHIFT) #endif =20 +#define PAGE_SECTORS_SHIFT (PAGE_SHIFT - SECTOR_SHIFT) +#define PAGE_SECTORS (1 << PAGE_SECTORS_SHIFT) +#define PAGE_SECTORS_MASK (~(PAGE_SECTORS - 1)) + /* * blk_rq_pos() : the current sector * blk_rq_bytes() : bytes left in the entire request diff --git a/include/linux/device-mapper.h b/include/linux/device-mapper.h index 475668c69dbc..c98a533f8ffa 100644 --- a/include/linux/device-mapper.h +++ b/include/linux/device-mapper.h @@ -141,7 +141,6 @@ typedef long (*dm_dax_direct_access_fn) (struct dm_targ= et *ti, pgoff_t pgoff, long nr_pages, void **kaddr, pfn_t *pfn); typedef size_t (*dm_dax_copy_iter_fn)(struct dm_target *ti, pgoff_t pgoff, void *addr, size_t bytes, struct iov_iter *i); -#define PAGE_SECTORS (PAGE_SIZE / 512) =20 void dm_error(const char *message); =20 diff --git a/mm/page_io.c b/mm/page_io.c index 76965be1d40e..9e3376ad60f5 100644 --- a/mm/page_io.c +++ b/mm/page_io.c @@ -38,7 +38,7 @@ static struct bio *get_swap_bio(gfp_t gfp_flags, =20 bio->bi_iter.bi_sector =3D map_swap_page(page, &bdev); bio_set_dev(bio, bdev); - bio->bi_iter.bi_sector <<=3D PAGE_SHIFT - 9; + bio->bi_iter.bi_sector <<=3D PAGE_SECTORS_SHIFT; bio->bi_end_io =3D end_io; =20 bio_add_page(bio, page, PAGE_SIZE * hpage_nr_pages(page), 0); @@ -266,7 +266,7 @@ int swap_writepage(struct page *page, struct writeback_= control *wbc) =20 static sector_t swap_page_sector(struct page *page) { - return (sector_t)__page_file_index(page) << (PAGE_SHIFT - 9); + return (sector_t)__page_file_index(page) << PAGE_SECTORS_SHIFT; } =20 static inline void count_swpout_vm_event(struct page *page) diff --git a/mm/swapfile.c b/mm/swapfile.c index b2a2e45c9a36..1cf3c80295d2 100644 --- a/mm/swapfile.c +++ b/mm/swapfile.c @@ -177,8 +177,8 @@ static int discard_swap(struct swap_info_struct *si) =20 /* Do not discard the swap header page! */ se =3D first_se(si); - start_block =3D (se->start_block + 1) << (PAGE_SHIFT - 9); - nr_blocks =3D ((sector_t)se->nr_pages - 1) << (PAGE_SHIFT - 9); + start_block =3D (se->start_block + 1) << PAGE_SECTORS_SHIFT; + nr_blocks =3D ((sector_t)se->nr_pages - 1) << PAGE_SECTORS_SHIFT; if (nr_blocks) { err =3D blkdev_issue_discard(si->bdev, start_block, nr_blocks, GFP_KERNEL, 0); @@ -188,8 +188,8 @@ static int discard_swap(struct swap_info_struct *si) } =20 for (se =3D next_se(se); se; se =3D next_se(se)) { - start_block =3D se->start_block << (PAGE_SHIFT - 9); - nr_blocks =3D (sector_t)se->nr_pages << (PAGE_SHIFT - 9); + start_block =3D se->start_block << PAGE_SECTORS_SHIFT; + nr_blocks =3D (sector_t)se->nr_pages << PAGE_SECTORS_SHIFT; =20 err =3D blkdev_issue_discard(si->bdev, start_block, nr_blocks, GFP_KERNEL, 0); @@ -240,8 +240,8 @@ static void discard_swap_cluster(struct swap_info_struc= t *si, start_page +=3D nr_blocks; nr_pages -=3D nr_blocks; =20 - start_block <<=3D PAGE_SHIFT - 9; - nr_blocks <<=3D PAGE_SHIFT - 9; + start_block <<=3D PAGE_SECTORS_SHIFT; + nr_blocks <<=3D PAGE_SECTORS_SHIFT; if (blkdev_issue_discard(si->bdev, start_block, nr_blocks, GFP_NOIO, 0)) break; --=20 2.24.1 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel