From: Sarthak Kukreti <sarthakkukreti@chromium.org> Hi, This patch series is an RFC of a mechanism to pass through provision requests on stacked thinly provisioned storage devices/filesystems. The linux kernel provides several mechanisms to set up thinly provisioned block storage abstractions (eg. dm-thin, loop devices over sparse files), either directly as block devices or backing storage for filesystems. Currently, short of writing data to either the device or filesystem, there is no way for users to pre-allocate space for use in such storage setups. Consider the following use-cases: 1) Suspend-to-disk and resume from a dm-thin device: In order to ensure that the underlying thinpool metadata is not modified during the suspend mechanism, the dm-thin device needs to be fully provisioned. 2) If a filesystem uses a loop device over a sparse file, fallocate() on the filesystem will allocate blocks for files but the underlying sparse file will remain intact. 3) Another example is virtual machine using a sparse file/dm-thin as a storage device; by default, allocations within the VM boundaries will not affect the host. 4) Several storage standards support mechanisms for thin provisioning on real hardware devices. For example: a. The NVMe spec 1.0b section 2.1.1 loosely talks about thin provisioning: "When the THINP bit in the NSFEAT field of the Identify Namespace data structure is set to ‘1’, the controller ... shall track the number of allocated blocks in the Namespace Utilization field" b. The SCSi Block Commands reference - 4 section references "Thin provisioned logical units", c. UFS 3.0 spec section 13.3.3 references "Thin provisioning". In all of the above situations, currently the only way for pre-allocating space is to issue writes (or use WRITE_ZEROES/WRITE_SAME). However, that does not scale well with larger pre-allocation sizes. This patchset introduces primitives to support block-level provisioning (note: the term 'provisioning' is used to prevent overloading the term 'allocations/pre-allocations') requests across filesystems and block devices. This allows fallocate() and file creation requests to reserve space across stacked layers of block devices and filesystems. Currently, the patchset covers a prototype on the device-mapper targets, loop device and ext4, but the same mechanism can be extended to other filesystems/block devices as well as extended for use with devices in 4 a-c. Patch 1 introduces REQ_OP_PROVISION as a new request type. The provision request acts like the inverse of a discard request; instead of notifying lower layers that the block range will no longer be used, provision acts as a request to lower layers to provision disk space for the given block range. Real hardware storage devices will currently disable the provisioing capability but for the standards listed in 4a.-c., REQ_OP_PROVISION can be overloaded for use as the provisioing primitive for future devices. Patch 2 implements REQ_OP_PROVISION handling for some of the device-mapper targets. This additionally adds support for pre-allocating space for thinly provisioned logical volumes via fallocate() Patch 3 implements the handling for virtio-blk. Patch 4 introduces an fallocate() mode (FALLOC_FL_PROVISION) that sends a provision request to the underlying block device (and beyond). This acts as the primary mechanism for file-level provisioing. Patch 5 wires up the loop device handling of REQ_OP_PROVISION. Patches 6-8 cover a prototype implementation for ext4, which includes wiring up the fallocate() implementation, introducing a filesystem level option (called 'provision') to control the default allocation behaviour and finally a file level override to retain current handling, even on filesystems mounted with 'provision' Testing: -------- - A backport of this patch series was tested on ChromiumOS using a 5.10 kernel. - File on ext4 on a thin logical volume: fallocate(FALLOC_FL_PROVISION) : 4.6s, dd if=/dev/zero of=...: 6 mins. TODOs: ------ 1) The stacked block devices (dm-*, loop etc.) currently unconditionally pass through provision requests. Add support for provision, similar to how discard handling is set up (with options to disable, passdown or passthrough requests). 2) Blktests and Xfstests for validating provisioning.
On Thu, Sep 15, 2022 at 09:48:18AM -0700, Sarthak Kukreti wrote: > From: Sarthak Kukreti <sarthakkukreti@chromium.org> > > Hi, > > This patch series is an RFC of a mechanism to pass through provision requests on stacked thinly provisioned storage devices/filesystems. > > The linux kernel provides several mechanisms to set up thinly provisioned block storage abstractions (eg. dm-thin, loop devices over sparse files), either directly as block devices or backing storage for filesystems. Currently, short of writing data to either the device or filesystem, there is no way for users to pre-allocate space for use in such storage setups. Consider the following use-cases: > > 1) Suspend-to-disk and resume from a dm-thin device: In order to ensure that the underlying thinpool metadata is not modified during the suspend mechanism, the dm-thin device needs to be fully provisioned. > 2) If a filesystem uses a loop device over a sparse file, fallocate() on the filesystem will allocate blocks for files but the underlying sparse file will remain intact. > 3) Another example is virtual machine using a sparse file/dm-thin as a storage device; by default, allocations within the VM boundaries will not affect the host. > 4) Several storage standards support mechanisms for thin provisioning on real hardware devices. For example: > a. The NVMe spec 1.0b section 2.1.1 loosely talks about thin provisioning: "When the THINP bit in the NSFEAT field of the Identify Namespace data structure is set to ‘1’, the controller ... shall track the number of allocated blocks in the Namespace Utilization field" > b. The SCSi Block Commands reference - 4 section references "Thin provisioned logical units", > c. UFS 3.0 spec section 13.3.3 references "Thin provisioning". When REQ_OP_PROVISION is sent on an already-allocated range of blocks, are those blocks zeroed? NVMe Write Zeroes with Deallocate=0 works this way, for example. That behavior is counterintuitive since the operation name suggests it just affects the logical block's provisioning state, not the contents of the blocks. > In all of the above situations, currently the only way for pre-allocating space is to issue writes (or use WRITE_ZEROES/WRITE_SAME). However, that does not scale well with larger pre-allocation sizes. What exactly is the issue with WRITE_ZEROES scalability? Are you referring to cases where the device doesn't support an efficient WRITE_ZEROES command and actually writes blocks filled with zeroes instead of updating internal allocation metadata cheaply? Stefan
On Thu, Sep 15, 2022 at 11:10 PM Stefan Hajnoczi <stefanha@redhat.com> wrote: > > On Thu, Sep 15, 2022 at 09:48:18AM -0700, Sarthak Kukreti wrote: > > From: Sarthak Kukreti <sarthakkukreti@chromium.org> > > > > Hi, > > > > This patch series is an RFC of a mechanism to pass through provision requests on stacked thinly provisioned storage devices/filesystems. > > > > The linux kernel provides several mechanisms to set up thinly provisioned block storage abstractions (eg. dm-thin, loop devices over sparse files), either directly as block devices or backing storage for filesystems. Currently, short of writing data to either the device or filesystem, there is no way for users to pre-allocate space for use in such storage setups. Consider the following use-cases: > > > > 1) Suspend-to-disk and resume from a dm-thin device: In order to ensure that the underlying thinpool metadata is not modified during the suspend mechanism, the dm-thin device needs to be fully provisioned. > > 2) If a filesystem uses a loop device over a sparse file, fallocate() on the filesystem will allocate blocks for files but the underlying sparse file will remain intact. > > 3) Another example is virtual machine using a sparse file/dm-thin as a storage device; by default, allocations within the VM boundaries will not affect the host. > > 4) Several storage standards support mechanisms for thin provisioning on real hardware devices. For example: > > a. The NVMe spec 1.0b section 2.1.1 loosely talks about thin provisioning: "When the THINP bit in the NSFEAT field of the Identify Namespace data structure is set to ‘1’, the controller ... shall track the number of allocated blocks in the Namespace Utilization field" > > b. The SCSi Block Commands reference - 4 section references "Thin provisioned logical units", > > c. UFS 3.0 spec section 13.3.3 references "Thin provisioning". > > When REQ_OP_PROVISION is sent on an already-allocated range of blocks, > are those blocks zeroed? NVMe Write Zeroes with Deallocate=0 works this > way, for example. That behavior is counterintuitive since the operation > name suggests it just affects the logical block's provisioning state, > not the contents of the blocks. > No, the blocks are not zeroed. The current implementation (in the dm patch) is to indeed look at the provisioned state of the logical block and provision if it is unmapped. if the block is already allocated, REQ_OP_PROVISION should have no effect on the contents of the block. Similarly, in the file semantics, sending an FALLOC_FL_PROVISION requests for extents already mapped should not affect the contents in the extents. > > In all of the above situations, currently the only way for pre-allocating space is to issue writes (or use WRITE_ZEROES/WRITE_SAME). However, that does not scale well with larger pre-allocation sizes. > > What exactly is the issue with WRITE_ZEROES scalability? Are you > referring to cases where the device doesn't support an efficient > WRITE_ZEROES command and actually writes blocks filled with zeroes > instead of updating internal allocation metadata cheaply? > Yes. On ChromiumOS, we regularly deal with storage devices that don't support WRITE_ZEROES or that need to have it disabled, via a quirk, due to a bug in the vendor's implementation. Using WRITE_ZEROES for allocation makes the allocation path quite slow for such devices (not to mention the effect on storage lifetime), so having a separate provisioning construct is very appealing. Even for devices that do support an efficient WRITE_ZEROES implementation but don't support logical provisioning per-se, I suppose that the allocation path might be a bit faster (the device driver's request queue would report 'max_provision_sectors'=0 and the request would be short circuited there) although I haven't benchmarked the difference. Sarthak > Stefan
On Fri, Sep 16, 2022 at 11:48:34AM -0700, Sarthak Kukreti wrote: > Yes. On ChromiumOS, we regularly deal with storage devices that don't > support WRITE_ZEROES or that need to have it disabled, via a quirk, > due to a bug in the vendor's implementation. So bloody punich the vendors for it. Unlike most of the Linux community your actually have purchasing power and you'd help everyone by making use of that instead adding hacks to upstream.
On 9/16/22 11:48, Sarthak Kukreti wrote: > Yes. On ChromiumOS, we regularly deal with storage devices that don't > support WRITE_ZEROES or that need to have it disabled, via a quirk, > due to a bug in the vendor's implementation. Using WRITE_ZEROES for > allocation makes the allocation path quite slow for such devices (not > to mention the effect on storage lifetime), so having a separate > provisioning construct is very appealing. Even for devices that do > support an efficient WRITE_ZEROES implementation but don't support > logical provisioning per-se, I suppose that the allocation path might > be a bit faster (the device driver's request queue would report > 'max_provision_sectors'=0 and the request would be short circuited > there) although I haven't benchmarked the difference. Some background information about why ChromiumOS uses thin provisioning instead of a single filesystem across the entire storage device would be welcome. Although UFS devices support thin provisioning I am not aware of any use cases in Android that would benefit from UFS thin provisioning support. Thanks, Bart.
On Fri, Sep 16, 2022 at 1:01 PM Bart Van Assche <bvanassche@acm.org> wrote: > > On 9/16/22 11:48, Sarthak Kukreti wrote: > > Yes. On ChromiumOS, we regularly deal with storage devices that don't > > support WRITE_ZEROES or that need to have it disabled, via a quirk, > > due to a bug in the vendor's implementation. Using WRITE_ZEROES for > > allocation makes the allocation path quite slow for such devices (not > > to mention the effect on storage lifetime), so having a separate > > provisioning construct is very appealing. Even for devices that do > > support an efficient WRITE_ZEROES implementation but don't support > > logical provisioning per-se, I suppose that the allocation path might > > be a bit faster (the device driver's request queue would report > > 'max_provision_sectors'=0 and the request would be short circuited > > there) although I haven't benchmarked the difference. > > Some background information about why ChromiumOS uses thin provisioning > instead of a single filesystem across the entire storage device would be > welcome. Although UFS devices support thin provisioning I am not aware > of any use cases in Android that would benefit from UFS thin > provisioning support. > Sure (and I'd be happy to put this in the cover letter, if you prefer; I didn't include it initially, since it seemed orthogonal to the discussion of the patchset)! On ChromiumOS, the primary driving force for using thin provisioning is to have flexible, segmented block storage, both per-user and for applications/virtual machines with several useful properties, for example: block-level encrypted user storage, snapshot based A-B updates for verified content, on-demand partitioning for short-lived use cases. Several of the other planned use-cases (like verified content retention over powerwash) require flexible on-demand block storage that is decoupled from the primary filesystem(s) so that we can have cryptographic erase for the user partitions and keep the on-demand, dm-verity backed executables intact. Best Sarthak > Thanks, > > Bart.
On Thu, Sep 15, 2022 at 09:48:18AM -0700, Sarthak Kukreti wrote: > From: Sarthak Kukreti <sarthakkukreti@chromium.org> > > Hi, > > This patch series is an RFC of a mechanism to pass through provision > requests on stacked thinly provisioned storage devices/filesystems. [Reflowed text] > The linux kernel provides several mechanisms to set up thinly > provisioned block storage abstractions (eg. dm-thin, loop devices over > sparse files), either directly as block devices or backing storage for > filesystems. Currently, short of writing data to either the device or > filesystem, there is no way for users to pre-allocate space for use in > such storage setups. Consider the following use-cases: > > 1) Suspend-to-disk and resume from a dm-thin device: In order to > ensure that the underlying thinpool metadata is not modified during > the suspend mechanism, the dm-thin device needs to be fully > provisioned. > 2) If a filesystem uses a loop device over a sparse file, fallocate() > on the filesystem will allocate blocks for files but the underlying > sparse file will remain intact. > 3) Another example is virtual machine using a sparse file/dm-thin as a > storage device; by default, allocations within the VM boundaries will > not affect the host. > 4) Several storage standards support mechanisms for thin provisioning > on real hardware devices. For example: > a. The NVMe spec 1.0b section 2.1.1 loosely talks about thin > provisioning: "When the THINP bit in the NSFEAT field of the > Identify Namespace data structure is set to ‘1’, the controller ... > shall track the number of allocated blocks in the Namespace > Utilization field" > b. The SCSi Block Commands reference - 4 section references "Thin > provisioned logical units", > c. UFS 3.0 spec section 13.3.3 references "Thin provisioning". > > In all of the above situations, currently the only way for > pre-allocating space is to issue writes (or use > WRITE_ZEROES/WRITE_SAME). However, that does not scale well with > larger pre-allocation sizes. > > This patchset introduces primitives to support block-level > provisioning (note: the term 'provisioning' is used to prevent > overloading the term 'allocations/pre-allocations') requests across > filesystems and block devices. This allows fallocate() and file > creation requests to reserve space across stacked layers of block > devices and filesystems. Currently, the patchset covers a prototype on > the device-mapper targets, loop device and ext4, but the same > mechanism can be extended to other filesystems/block devices as well > as extended for use with devices in 4 a-c. If you call REQ_OP_PROVISION on an unmapped LBA range of a block device and then try to read the provisioned blocks, what do you get? Zeroes? Random stale disk contents? I think I saw elsewhere in the thread that any mapped LBAs within the provisioning range are left alone (i.e. not zeroed) so I'll proceed on that basis. > Patch 1 introduces REQ_OP_PROVISION as a new request type. The > provision request acts like the inverse of a discard request; instead > of notifying lower layers that the block range will no longer be used, > provision acts as a request to lower layers to provision disk space > for the given block range. Real hardware storage devices will > currently disable the provisioing capability but for the standards > listed in 4a.-c., REQ_OP_PROVISION can be overloaded for use as the > provisioing primitive for future devices. > > Patch 2 implements REQ_OP_PROVISION handling for some of the > device-mapper targets. This additionally adds support for > pre-allocating space for thinly provisioned logical volumes via > fallocate() > > Patch 3 implements the handling for virtio-blk. > > Patch 4 introduces an fallocate() mode (FALLOC_FL_PROVISION) that > sends a provision request to the underlying block device (and beyond). > This acts as the primary mechanism for file-level provisioing. Personally, I think it's well within the definition of fallocate mode==0 (aka preallocate) for XFS to call REQ_OP_PROVISION on the blocks that it preallocates? XFS always sets the unwritten flag on the file mapping, so it doesn't matter if the device provisions space without zeroing the contents. That said, if devices are really allowed to expose stale disk blocks then for blkdev fallocate I think you could get away with reusin FALLOC_FL_NO_HIDE_STALE instead of introducing a new fallocate flag. > Patch 5 wires up the loop device handling of REQ_OP_PROVISION. > > Patches 6-8 cover a prototype implementation for ext4, which includes > wiring up the fallocate() implementation, introducing a filesystem > level option (called 'provision') to control the default allocation > behaviour and finally a file level override to retain current > handling, even on filesystems mounted with 'provision' Hmm, I'll have a look. > Testing: > -------- > - A backport of this patch series was tested on ChromiumOS using a > 5.10 kernel. > - File on ext4 on a thin logical volume: > fallocate(FALLOC_FL_PROVISION) : 4.6s, dd if=/dev/zero of=...: 6 mins. > > TODOs: > ------ > 1) The stacked block devices (dm-*, loop etc.) currently > unconditionally pass through provision requests. Add support for > provision, similar to how discard handling is set up (with options to > disable, passdown or passthrough requests). > 2) Blktests and Xfstests for validating provisioning. Yes.... --D > -- > dm-devel mailing list > dm-devel@redhat.com > https://listman.redhat.com/mailman/listinfo/dm-devel
On Fri, Sep 16, 2022 at 8:03 PM Darrick J. Wong <djwong@kernel.org> wrote: > > On Thu, Sep 15, 2022 at 09:48:18AM -0700, Sarthak Kukreti wrote: > > From: Sarthak Kukreti <sarthakkukreti@chromium.org> > > > > Hi, > > > > This patch series is an RFC of a mechanism to pass through provision > > requests on stacked thinly provisioned storage devices/filesystems. > > [Reflowed text] > > > The linux kernel provides several mechanisms to set up thinly > > provisioned block storage abstractions (eg. dm-thin, loop devices over > > sparse files), either directly as block devices or backing storage for > > filesystems. Currently, short of writing data to either the device or > > filesystem, there is no way for users to pre-allocate space for use in > > such storage setups. Consider the following use-cases: > > > > 1) Suspend-to-disk and resume from a dm-thin device: In order to > > ensure that the underlying thinpool metadata is not modified during > > the suspend mechanism, the dm-thin device needs to be fully > > provisioned. > > 2) If a filesystem uses a loop device over a sparse file, fallocate() > > on the filesystem will allocate blocks for files but the underlying > > sparse file will remain intact. > > 3) Another example is virtual machine using a sparse file/dm-thin as a > > storage device; by default, allocations within the VM boundaries will > > not affect the host. > > 4) Several storage standards support mechanisms for thin provisioning > > on real hardware devices. For example: > > a. The NVMe spec 1.0b section 2.1.1 loosely talks about thin > > provisioning: "When the THINP bit in the NSFEAT field of the > > Identify Namespace data structure is set to ‘1’, the controller ... > > shall track the number of allocated blocks in the Namespace > > Utilization field" > > b. The SCSi Block Commands reference - 4 section references "Thin > > provisioned logical units", > > c. UFS 3.0 spec section 13.3.3 references "Thin provisioning". > > > > In all of the above situations, currently the only way for > > pre-allocating space is to issue writes (or use > > WRITE_ZEROES/WRITE_SAME). However, that does not scale well with > > larger pre-allocation sizes. > > > > This patchset introduces primitives to support block-level > > provisioning (note: the term 'provisioning' is used to prevent > > overloading the term 'allocations/pre-allocations') requests across > > filesystems and block devices. This allows fallocate() and file > > creation requests to reserve space across stacked layers of block > > devices and filesystems. Currently, the patchset covers a prototype on > > the device-mapper targets, loop device and ext4, but the same > > mechanism can be extended to other filesystems/block devices as well > > as extended for use with devices in 4 a-c. > > If you call REQ_OP_PROVISION on an unmapped LBA range of a block device > and then try to read the provisioned blocks, what do you get? Zeroes? > Random stale disk contents? > > I think I saw elsewhere in the thread that any mapped LBAs within the > provisioning range are left alone (i.e. not zeroed) so I'll proceed on > that basis. > For block devices, I'd say it's definitely possible to get stale data, depending on the implementation of the allocation layer; for example, with dm-thinpool, the default setting via using LVM2 tools is to zero out blocks on allocation. But that's configurable and can be turned off to improve performance. Similarly, for actual devices that end up supporting thin provisioning, unless the specification absolutely mandates that an LBA contains zeroes post allocation, some implementations will definitely miss out on that (probably similar to the semantics of discard_zeroes_data today). I'm operating under the assumption that it's possible to get stale data from LBAs allocated using provision requests at the block layer and trying to see if we can create a safe default operating model from that. > > Patch 1 introduces REQ_OP_PROVISION as a new request type. The > > provision request acts like the inverse of a discard request; instead > > of notifying lower layers that the block range will no longer be used, > > provision acts as a request to lower layers to provision disk space > > for the given block range. Real hardware storage devices will > > currently disable the provisioing capability but for the standards > > listed in 4a.-c., REQ_OP_PROVISION can be overloaded for use as the > > provisioing primitive for future devices. > > > > Patch 2 implements REQ_OP_PROVISION handling for some of the > > device-mapper targets. This additionally adds support for > > pre-allocating space for thinly provisioned logical volumes via > > fallocate() > > > > Patch 3 implements the handling for virtio-blk. > > > > Patch 4 introduces an fallocate() mode (FALLOC_FL_PROVISION) that > > sends a provision request to the underlying block device (and beyond). > > This acts as the primary mechanism for file-level provisioing. > > Personally, I think it's well within the definition of fallocate mode==0 > (aka preallocate) for XFS to call REQ_OP_PROVISION on the blocks that it > preallocates? XFS always sets the unwritten flag on the file mapping, > so it doesn't matter if the device provisions space without zeroing the > contents. > > That said, if devices are really allowed to expose stale disk blocks > then for blkdev fallocate I think you could get away with reusin > FALLOC_FL_NO_HIDE_STALE instead of introducing a new fallocate flag. > For filesystems, I think it's reasonable to support the mode if and only if the filesystem can guarantee that unwritten extents return zero. For instance, in the current ext4 implementation, the provisioned extents are still marked as unwritten, which means a read from the file would still show all zeroes (which I think differs from the original FALLOC_FL_NO_HIDE implementation). That might be one more reason to keep the mode separate from the regular modes though; to drive home the point that it is only acceptable under the above conditions and that there's more to it than just adding blkdev_issue_provision(..) at the end of fs_fallocate(). Best Sarthak > > Patch 5 wires up the loop device handling of REQ_OP_PROVISION. > > > > Patches 6-8 cover a prototype implementation for ext4, which includes > > wiring up the fallocate() implementation, introducing a filesystem > > level option (called 'provision') to control the default allocation > > behaviour and finally a file level override to retain current > > handling, even on filesystems mounted with 'provision' > > Hmm, I'll have a look. > > > Testing: > > -------- > > - A backport of this patch series was tested on ChromiumOS using a > > 5.10 kernel. > > - File on ext4 on a thin logical volume: > > fallocate(FALLOC_FL_PROVISION) : 4.6s, dd if=/dev/zero of=...: 6 mins. > > > > TODOs: > > ------ > > 1) The stacked block devices (dm-*, loop etc.) currently > > unconditionally pass through provision requests. Add support for > > provision, similar to how discard handling is set up (with options to > > disable, passdown or passthrough requests). > > 2) Blktests and Xfstests for validating provisioning. > > Yes.... > > --D > > > -- > > dm-devel mailing list > > dm-devel@redhat.com > > https://listman.redhat.com/mailman/listinfo/dm-devel
On Sat, Sep 17, 2022 at 12:46:33PM -0700, Sarthak Kukreti wrote: > On Fri, Sep 16, 2022 at 8:03 PM Darrick J. Wong <djwong@kernel.org> wrote: > > > > On Thu, Sep 15, 2022 at 09:48:18AM -0700, Sarthak Kukreti wrote: > > > From: Sarthak Kukreti <sarthakkukreti@chromium.org> > > > > > > Hi, > > > > > > This patch series is an RFC of a mechanism to pass through provision > > > requests on stacked thinly provisioned storage devices/filesystems. > > > > [Reflowed text] > > > > > The linux kernel provides several mechanisms to set up thinly > > > provisioned block storage abstractions (eg. dm-thin, loop devices over > > > sparse files), either directly as block devices or backing storage for > > > filesystems. Currently, short of writing data to either the device or > > > filesystem, there is no way for users to pre-allocate space for use in > > > such storage setups. Consider the following use-cases: > > > > > > 1) Suspend-to-disk and resume from a dm-thin device: In order to > > > ensure that the underlying thinpool metadata is not modified during > > > the suspend mechanism, the dm-thin device needs to be fully > > > provisioned. > > > 2) If a filesystem uses a loop device over a sparse file, fallocate() > > > on the filesystem will allocate blocks for files but the underlying > > > sparse file will remain intact. > > > 3) Another example is virtual machine using a sparse file/dm-thin as a > > > storage device; by default, allocations within the VM boundaries will > > > not affect the host. > > > 4) Several storage standards support mechanisms for thin provisioning > > > on real hardware devices. For example: > > > a. The NVMe spec 1.0b section 2.1.1 loosely talks about thin > > > provisioning: "When the THINP bit in the NSFEAT field of the > > > Identify Namespace data structure is set to ‘1’, the controller ... > > > shall track the number of allocated blocks in the Namespace > > > Utilization field" > > > b. The SCSi Block Commands reference - 4 section references "Thin > > > provisioned logical units", > > > c. UFS 3.0 spec section 13.3.3 references "Thin provisioning". > > > > > > In all of the above situations, currently the only way for > > > pre-allocating space is to issue writes (or use > > > WRITE_ZEROES/WRITE_SAME). However, that does not scale well with > > > larger pre-allocation sizes. > > > > > > This patchset introduces primitives to support block-level > > > provisioning (note: the term 'provisioning' is used to prevent > > > overloading the term 'allocations/pre-allocations') requests across > > > filesystems and block devices. This allows fallocate() and file > > > creation requests to reserve space across stacked layers of block > > > devices and filesystems. Currently, the patchset covers a prototype on > > > the device-mapper targets, loop device and ext4, but the same > > > mechanism can be extended to other filesystems/block devices as well > > > as extended for use with devices in 4 a-c. > > > > If you call REQ_OP_PROVISION on an unmapped LBA range of a block device > > and then try to read the provisioned blocks, what do you get? Zeroes? > > Random stale disk contents? > > > > I think I saw elsewhere in the thread that any mapped LBAs within the > > provisioning range are left alone (i.e. not zeroed) so I'll proceed on > > that basis. > > > For block devices, I'd say it's definitely possible to get stale data, depending > on the implementation of the allocation layer; for example, with dm-thinpool, > the default setting via using LVM2 tools is to zero out blocks on allocation. > But that's configurable and can be turned off to improve performance. > > Similarly, for actual devices that end up supporting thin provisioning, unless > the specification absolutely mandates that an LBA contains zeroes post > allocation, some implementations will definitely miss out on that (probably > similar to the semantics of discard_zeroes_data today). I'm operating under > the assumption that it's possible to get stale data from LBAs allocated using > provision requests at the block layer and trying to see if we can create a > safe default operating model from that. Please explain the semantics of REQ_OP_PROVISION in the code/documentation in the next revision. Thanks, Stefan
© 2016 - 2026 Red Hat, Inc.