From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604706423; cv=none; d=zohomail.com; s=zohoarc; b=BrsYCRt8hHaqgs9f5CF8KqMS1MJxOGsKQFDQrxtpHWVCcnF51oM4BMo0MDtPpea/DZ146KNgybtcuV1t1JUsGDiSUs4KeeODvJh127UQdxBDXxQDkLGLl/6lOuNfoP2BlBKx8zXtj0HQXb+eaWuV3z2ErYJPsNWDz1IVyCTnfPI= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604706423; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=+zJJoBqXtzuKEsML0oG7xLAFvXmi4kzz0656zbJk0NY=; b=Ud+JmmblvtW1z1ZCO7r3TFY3Hj3G4dfNU2JKV/k/cgs8QpPUuarV0Md4HhhjOm8xtxBBK5BTF1wVVGAKlQTj5a/JN+BqrFBTWzlBkuDsDvoOmxA1V6BbS3IraxVdppnFKZgtf3XxeyE09nfdaPqxWZQSOx+BQnCciC/QjvA97Pw= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604706423291124.24559820579054; Fri, 6 Nov 2020 15:47:03 -0800 (PST) Received: from localhost ([::1]:38604 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBRd-00025H-VY for importer@patchew.org; Fri, 06 Nov 2020 18:47:02 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34144) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBO6-0005HG-6e; Fri, 06 Nov 2020 18:43:22 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57349) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBO0-0003Sw-Na; Fri, 06 Nov 2020 18:43:20 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:15 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:06 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:12 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706197; x=1636242197; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=QT37Zg41FIq+qHt+V884jfsn9aB/8VuGenGA9udWlZc=; b=l7KDs29AeTNNXts8csgrD6uDlz/JrhYcnoWeBKc5f23gLcsHxKqrXD6o gN/tG+ZiS8U1MZKtn+ySoOYXP/pLTPpv3G2JT/sb+UGx7eiZiKH3aRT+a Yf1XqgdbpjN/XEV09YwjUAIYl0d7ejNdPrLd+CQvuBKq0chLVwhj68ur+ uod7/ZL6O8P2eYwxocHkzvHtRrtl00lL4zKQpxxq/yW1vWG+JApOofL/5 vlTcuEcykCDrn/Ze2VRXQnt18zbytTAMH9cveV1jjZykePvLTuBTK7dc6 3bbKwoGrbyJvlmfLd8lneJ457nTBbXfG8jEhTWIyBn3+4urK+jnkqLgzj A==; IronPort-SDR: 81+kl1r7m6LDIR3axyDMjVeIizd0jZrbbvGCKuGpcRzh83DA6c2PaXRU+i9zthHaPcGAOnq6Hd IPNGN5Dl3btr4bsdNciRRKt1sP+6GjsRSPLJfer9D0e7yunXVsteLJcCi/aGf1FP/Kcw/4SkJG Ii78bbYURxFWP/apuuVX7npTRuh4+TYNkmbfYgPtacZ4dg85/YaIGVRWjq4I7l6vs57emTEsXu gGX+9eaxdY6UPJ7Mh4XOVJ2NRXTyaoKyL8w1EiXbPXcfW+a4+1yUnFP6HylEXzzop573EI422K TTI= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267054" IronPort-SDR: n0vIwh34sROd3Hncr+haq3jvc/K/iqogn2oLdu4s6mtPNNT8ZnhPzqyNlKdylrLVtTHSQhEhtb rJF8xwy1vDxxcTkmjcToXf3XcMLu+rrevou6Xt/BlnznEHu8F7TiiYfiZlFTQ4tKS5EQ9uZvHq M92UFlwexroA85kiK081ZElkKcjAxpB/nhuRPQPhriGPflzETdAIhfnMzl/Fq0YDtlXVFmDPkK QLQB7iI+EOrvWpw+jEUPs3qR6ptOXqc5dsEQtNPwcySYV7h0xTSUrnXgrd1vZBviKXemSG8tIB LrhhQ7EPrSC3xxTAFIsvFdRx IronPort-SDR: G07voAA+T9YumCnFk5rA/byfV//uGuP0ilQOqUpgloPFg1bwjSpPocLHg/LcuJFDn/gWqcPKn2 +bG2rOwMS3XoY/cApAC+/xXFqP2vFxzFTGGr4a6gLx66et+V4mW8KIAVVFy5yNg4n+LX3pP9GR I8GGJu87ck0rq7+ylFpEGbzzxZhsAHbdpY3U52bKQ2Fmh3crcBA7K6DXtWPjcuwxhVBXDfyu9L cVGdsgrwMOCJY0Oi0nyo0Gy/siyCe8xDYh00Cp/bmTVNg3/9Vwv+ZdaTRBx/fhEAFJ7bBWKskY Zns= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 01/12] hw/block/nvme: Add Commands Supported and Effects log Date: Sat, 7 Nov 2020 08:42:54 +0900 Message-Id: <20201106234305.21339-2-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" This log page becomes necessary to implement to allow checking for Zone Append command support in Zoned Namespace Command Set. This commit adds the code to report this log page for NVM Command Set only. The parts that are specific to zoned operation will be added later in the series. All incoming admin and i/o commands are now only processed if their corresponding support bits are set in this log. This provides an easy way to control what commands to support and what not to depending on set CC.CSS. Signed-off-by: Dmitry Fomichev Reviewed-by: Niklas Cassel --- hw/block/nvme-ns.h | 1 + include/block/nvme.h | 19 +++++++++ hw/block/nvme.c | 96 +++++++++++++++++++++++++++++++++++++++---- hw/block/trace-events | 1 + 4 files changed, 108 insertions(+), 9 deletions(-) diff --git a/hw/block/nvme-ns.h b/hw/block/nvme-ns.h index 83734f4606..ea8c2f785d 100644 --- a/hw/block/nvme-ns.h +++ b/hw/block/nvme-ns.h @@ -29,6 +29,7 @@ typedef struct NvmeNamespace { int32_t bootindex; int64_t size; NvmeIdNs id_ns; + const uint32_t *iocs; =20 NvmeNamespaceParams params; } NvmeNamespace; diff --git a/include/block/nvme.h b/include/block/nvme.h index 8a46d9cf01..f62cc90d49 100644 --- a/include/block/nvme.h +++ b/include/block/nvme.h @@ -745,10 +745,27 @@ enum NvmeSmartWarn { NVME_SMART_FAILED_VOLATILE_MEDIA =3D 1 << 4, }; =20 +typedef struct NvmeEffectsLog { + uint32_t acs[256]; + uint32_t iocs[256]; + uint8_t resv[2048]; +} NvmeEffectsLog; + +enum { + NVME_CMD_EFF_CSUPP =3D 1 << 0, + NVME_CMD_EFF_LBCC =3D 1 << 1, + NVME_CMD_EFF_NCC =3D 1 << 2, + NVME_CMD_EFF_NIC =3D 1 << 3, + NVME_CMD_EFF_CCC =3D 1 << 4, + NVME_CMD_EFF_CSE_MASK =3D 3 << 16, + NVME_CMD_EFF_UUID_SEL =3D 1 << 19, +}; + enum NvmeLogIdentifier { NVME_LOG_ERROR_INFO =3D 0x01, NVME_LOG_SMART_INFO =3D 0x02, NVME_LOG_FW_SLOT_INFO =3D 0x03, + NVME_LOG_CMD_EFFECTS =3D 0x05, }; =20 typedef struct QEMU_PACKED NvmePSD { @@ -861,6 +878,7 @@ enum NvmeIdCtrlFrmw { =20 enum NvmeIdCtrlLpa { NVME_LPA_NS_SMART =3D 1 << 0, + NVME_LPA_CSE =3D 1 << 1, NVME_LPA_EXTENDED =3D 1 << 2, }; =20 @@ -1060,6 +1078,7 @@ static inline void _nvme_check_size(void) QEMU_BUILD_BUG_ON(sizeof(NvmeErrorLog) !=3D 64); QEMU_BUILD_BUG_ON(sizeof(NvmeFwSlotInfoLog) !=3D 512); QEMU_BUILD_BUG_ON(sizeof(NvmeSmartLog) !=3D 512); + QEMU_BUILD_BUG_ON(sizeof(NvmeEffectsLog) !=3D 4096); QEMU_BUILD_BUG_ON(sizeof(NvmeIdCtrl) !=3D 4096); QEMU_BUILD_BUG_ON(sizeof(NvmeIdNs) !=3D 4096); QEMU_BUILD_BUG_ON(sizeof(NvmeSglDescriptor) !=3D 16); diff --git a/hw/block/nvme.c b/hw/block/nvme.c index 065e763e4f..702f7cc2e3 100644 --- a/hw/block/nvme.c +++ b/hw/block/nvme.c @@ -111,6 +111,28 @@ static const uint32_t nvme_feature_cap[NVME_FID_MAX] = =3D { [NVME_TIMESTAMP] =3D NVME_FEAT_CAP_CHANGE, }; =20 +static const uint32_t nvme_cse_acs[256] =3D { + [NVME_ADM_CMD_DELETE_SQ] =3D NVME_CMD_EFF_CSUPP, + [NVME_ADM_CMD_CREATE_SQ] =3D NVME_CMD_EFF_CSUPP, + [NVME_ADM_CMD_GET_LOG_PAGE] =3D NVME_CMD_EFF_CSUPP, + [NVME_ADM_CMD_DELETE_CQ] =3D NVME_CMD_EFF_CSUPP, + [NVME_ADM_CMD_CREATE_CQ] =3D NVME_CMD_EFF_CSUPP, + [NVME_ADM_CMD_IDENTIFY] =3D NVME_CMD_EFF_CSUPP, + [NVME_ADM_CMD_ABORT] =3D NVME_CMD_EFF_CSUPP, + [NVME_ADM_CMD_SET_FEATURES] =3D NVME_CMD_EFF_CSUPP, + [NVME_ADM_CMD_GET_FEATURES] =3D NVME_CMD_EFF_CSUPP, + [NVME_ADM_CMD_ASYNC_EV_REQ] =3D NVME_CMD_EFF_CSUPP, +}; + +static const uint32_t nvme_cse_iocs_none[256]; + +static const uint32_t nvme_cse_iocs_nvm[256] =3D { + [NVME_CMD_FLUSH] =3D NVME_CMD_EFF_CSUPP | NVME_CMD_EFF_= LBCC, + [NVME_CMD_WRITE_ZEROES] =3D NVME_CMD_EFF_CSUPP | NVME_CMD_EFF_= LBCC, + [NVME_CMD_WRITE] =3D NVME_CMD_EFF_CSUPP | NVME_CMD_EFF_= LBCC, + [NVME_CMD_READ] =3D NVME_CMD_EFF_CSUPP, +}; + static void nvme_process_sq(void *opaque); =20 static uint16_t nvme_cid(NvmeRequest *req) @@ -1022,10 +1044,6 @@ static uint16_t nvme_io_cmd(NvmeCtrl *n, NvmeRequest= *req) trace_pci_nvme_io_cmd(nvme_cid(req), nsid, nvme_sqid(req), req->cmd.opcode, nvme_io_opc_str(req->cmd.opcode= )); =20 - if (NVME_CC_CSS(n->bar.cc) =3D=3D NVME_CC_CSS_ADMIN_ONLY) { - return NVME_INVALID_OPCODE | NVME_DNR; - } - if (!nvme_nsid_valid(n, nsid)) { return NVME_INVALID_NSID | NVME_DNR; } @@ -1035,6 +1053,11 @@ static uint16_t nvme_io_cmd(NvmeCtrl *n, NvmeRequest= *req) return NVME_INVALID_FIELD | NVME_DNR; } =20 + if (!(req->ns->iocs[req->cmd.opcode] & NVME_CMD_EFF_CSUPP)) { + trace_pci_nvme_err_invalid_opc(req->cmd.opcode); + return NVME_INVALID_OPCODE | NVME_DNR; + } + switch (req->cmd.opcode) { case NVME_CMD_FLUSH: return nvme_flush(n, req); @@ -1044,8 +1067,7 @@ static uint16_t nvme_io_cmd(NvmeCtrl *n, NvmeRequest = *req) case NVME_CMD_READ: return nvme_rw(n, req); default: - trace_pci_nvme_err_invalid_opc(req->cmd.opcode); - return NVME_INVALID_OPCODE | NVME_DNR; + assert(false); } } =20 @@ -1282,6 +1304,37 @@ static uint16_t nvme_error_info(NvmeCtrl *n, uint8_t= rae, uint32_t buf_len, DMA_DIRECTION_FROM_DEVICE, req); } =20 +static uint16_t nvme_cmd_effects(NvmeCtrl *n, uint32_t buf_len, + uint64_t off, NvmeRequest *req) +{ + NvmeEffectsLog log =3D {}; + const uint32_t *src_iocs =3D NULL; + uint32_t trans_len; + + if (off >=3D sizeof(log)) { + trace_pci_nvme_err_invalid_log_page_offset(off, sizeof(log)); + return NVME_INVALID_FIELD | NVME_DNR; + } + + switch (NVME_CC_CSS(n->bar.cc)) { + case NVME_CC_CSS_NVM: + src_iocs =3D nvme_cse_iocs_nvm; + case NVME_CC_CSS_ADMIN_ONLY: + break; + } + + memcpy(log.acs, nvme_cse_acs, sizeof(nvme_cse_acs)); + + if (src_iocs) { + memcpy(log.iocs, src_iocs, sizeof(log.iocs)); + } + + trans_len =3D MIN(sizeof(log) - off, buf_len); + + return nvme_dma(n, ((uint8_t *)&log) + off, trans_len, + DMA_DIRECTION_FROM_DEVICE, req); +} + static uint16_t nvme_get_log(NvmeCtrl *n, NvmeRequest *req) { NvmeCmd *cmd =3D &req->cmd; @@ -1325,6 +1378,8 @@ static uint16_t nvme_get_log(NvmeCtrl *n, NvmeRequest= *req) return nvme_smart_info(n, rae, len, off, req); case NVME_LOG_FW_SLOT_INFO: return nvme_fw_log_info(n, len, off, req); + case NVME_LOG_CMD_EFFECTS: + return nvme_cmd_effects(n, len, off, req); default: trace_pci_nvme_err_invalid_log_page(nvme_cid(req), lid); return NVME_INVALID_FIELD | NVME_DNR; @@ -1912,6 +1967,11 @@ static uint16_t nvme_admin_cmd(NvmeCtrl *n, NvmeRequ= est *req) trace_pci_nvme_admin_cmd(nvme_cid(req), nvme_sqid(req), req->cmd.opcod= e, nvme_adm_opc_str(req->cmd.opcode)); =20 + if (!(nvme_cse_acs[req->cmd.opcode] & NVME_CMD_EFF_CSUPP)) { + trace_pci_nvme_err_invalid_admin_opc(req->cmd.opcode); + return NVME_INVALID_OPCODE | NVME_DNR; + } + switch (req->cmd.opcode) { case NVME_ADM_CMD_DELETE_SQ: return nvme_del_sq(n, req); @@ -1934,8 +1994,7 @@ static uint16_t nvme_admin_cmd(NvmeCtrl *n, NvmeReque= st *req) case NVME_ADM_CMD_ASYNC_EV_REQ: return nvme_aer(n, req); default: - trace_pci_nvme_err_invalid_admin_opc(req->cmd.opcode); - return NVME_INVALID_OPCODE | NVME_DNR; + assert(false); } } =20 @@ -2023,6 +2082,23 @@ static void nvme_clear_ctrl(NvmeCtrl *n) n->bar.cc =3D 0; } =20 +static void nvme_select_ns_iocs(NvmeCtrl *n) +{ + NvmeNamespace *ns; + int i; + + for (i =3D 1; i <=3D n->num_namespaces; i++) { + ns =3D nvme_ns(n, i); + if (!ns) { + continue; + } + ns->iocs =3D nvme_cse_iocs_none; + if (NVME_CC_CSS(n->bar.cc) !=3D NVME_CC_CSS_ADMIN_ONLY) { + ns->iocs =3D nvme_cse_iocs_nvm; + } + } +} + static int nvme_start_ctrl(NvmeCtrl *n) { uint32_t page_bits =3D NVME_CC_MPS(n->bar.cc) + 12; @@ -2121,6 +2197,8 @@ static int nvme_start_ctrl(NvmeCtrl *n) =20 QTAILQ_INIT(&n->aer_queue); =20 + nvme_select_ns_iocs(n); + return 0; } =20 @@ -2728,7 +2806,7 @@ static void nvme_init_ctrl(NvmeCtrl *n, PCIDevice *pc= i_dev) id->acl =3D 3; id->aerl =3D n->params.aerl; id->frmw =3D (NVME_NUM_FW_SLOTS << 1) | NVME_FRMW_SLOT1_RO; - id->lpa =3D NVME_LPA_NS_SMART | NVME_LPA_EXTENDED; + id->lpa =3D NVME_LPA_NS_SMART | NVME_LPA_CSE | NVME_LPA_EXTENDED; =20 /* recommended default value (~70 C) */ id->wctemp =3D cpu_to_le16(NVME_TEMPERATURE_WARNING); diff --git a/hw/block/trace-events b/hw/block/trace-events index 72abbbc2b4..33832a2021 100644 --- a/hw/block/trace-events +++ b/hw/block/trace-events @@ -102,6 +102,7 @@ pci_nvme_err_invalid_prp2_align(uint64_t prp2) "PRP2 is= not page aligned: 0x%"PR pci_nvme_err_invalid_opc(uint8_t opc) "invalid opcode 0x%"PRIx8"" pci_nvme_err_invalid_admin_opc(uint8_t opc) "invalid admin opcode 0x%"PRIx= 8"" pci_nvme_err_invalid_lba_range(uint64_t start, uint64_t len, uint64_t limi= t) "Invalid LBA start=3D%"PRIu64" len=3D%"PRIu64" limit=3D%"PRIu64"" +pci_nvme_err_invalid_log_page_offset(uint64_t ofs, uint64_t size) "must be= <=3D %"PRIu64", got %"PRIu64"" pci_nvme_err_invalid_del_sq(uint16_t qid) "invalid submission queue deleti= on, sid=3D%"PRIu16"" pci_nvme_err_invalid_create_sq_cqid(uint16_t cqid) "failed creating submis= sion queue, invalid cqid=3D%"PRIu16"" pci_nvme_err_invalid_create_sq_sqid(uint16_t sqid) "failed creating submis= sion queue, invalid sqid=3D%"PRIu16"" --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604706513; cv=none; d=zohomail.com; s=zohoarc; b=d0zQYpviHoNVdH/a09+KZIZwaRUV1IV6BeNStbM9xAsjqjEL//f5Jz4c3+0W8q4n8Ck9G9h3WJ0mHOcMKp6Y6OJBYNxX/sJGXbedU108YQWqbdOsZYiPPuLaEgVoNdV8Tbih/87yYsadAw7RdeG6OGQpzkiapOKm4Gtlf4XLU8A= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604706513; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=bC+HVd3k/maKfHDC6w5heiXuy13z0d+cBmmO0M2W8Vw=; b=AOho2u/0xmJ0tZ7jJShiP/RRQufHYdjaDMBbp0k/JkW1PauHALBzSHT90Nzf1yH87CtqNBVL+z61C/6ENaaalLakuaRqnSin6u5mZwGkjLupPKTYl3x8P93++oiNHVGAgVWpbohRInDdr3gose2UAxFGqaHPgwLsWtKjp/jsGyM= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604706513810826.3676254975597; Fri, 6 Nov 2020 15:48:33 -0800 (PST) Received: from localhost ([::1]:42842 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBT6-0003s2-Ql for importer@patchew.org; Fri, 06 Nov 2020 18:48:32 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34146) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBO6-0005HH-6p; Fri, 06 Nov 2020 18:43:22 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57357) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBO3-0003TB-AT; Fri, 06 Nov 2020 18:43:21 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:18 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:09 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:15 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706200; x=1636242200; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=jR2w/WGNNbVyESF4tPkSNOFiXsfAZlFecXFYaBR3/5E=; b=KlURHEdMC7vBfblCU9HaZkaBfoMETetXuHkxFBpwhTCrSo5FFMPigeHp qe8OekMbWLBBHa6Oszp7EdG7mAVh8DU+iNHBYm69KAUFmPct+BvZPx+L2 QlDvMVa1SqgPsY3ks5cWSXXL5vU7tyXLEiRuuKAJ3UCrEmOJApccDOcCx mhNJzOpYwtYyPzf+hkP6V+OxvgcUgA6SMLO3pPgeoYRRqGNWPybch8/G3 /kWUdoTjb7Ks24RUp7tU2kw5KTbDE1ZjcDoBwLiQXXe7igqvQvxn30lKF hYQRtFuml7sqp9DSicm19FniNVOtDS3OFGdD57dCGAKa7fretIIcVap8R Q==; IronPort-SDR: WvTBEgGIDuYW5b3DJrcLCw+WBjK0iIsDVxZFt7oXh5G6tEbTSw2Xf7ZSHuJxz47WPud/f0miTG vqUbOo5frkCeunxh6qIWIeB7PUsZUxNtwU+sGAUD1qZgKstsGVGcdYX+Z9sg7DHXBv2rSpjJVU gmsnxOTf7ep1YAcWTu7P7YjhUhhjOjHfrlDph/GuRHWAuJq9eMHbzE2HmLKHqkTaflBYsLSVbC oDp846OnXgJmhsOejhjBODnwleJI6VxPexjcF4tjcwa8AMP8Gq6p5eE91HPf9oJo1sMfEQVBG7 bdk= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267057" IronPort-SDR: nRuJr+I4+9aeoG1VnOAAhMIucvwtvXZzpkmlc2Okau80BJQU1j2v3+mIg7QsZASgDMXAMYb09R AQFnavTkOQZa16SogeSiq3namsJNlKc+FgcYpiFtxH1GgFXIpZbjka6SSW5wU8boClHGIIzMi0 OQkSVXMFWWE7bgFzsRnXN+uo/plFTa1wQbExWHOxAJv6m4zJa8S8O3aPt8vDNoYxu33ZaPL8PP ZAZBEbWxkXb5/TaKkDukZOA2K3eoBH7BzucGM6np+ZjqXln5rEtl1JWFliWgLGO8oWcSFHdcCF 860Jz7EbQEPpHJiDzWqyWB1N IronPort-SDR: 7uhjf2g0mRW9Z3k5x5yEi5T1QU1+sTugr0zBuh6pRMwz3kaM9TJe6MTex6EVwqohUpiKRUtFYX DvZQRvDZghRSSW4blsZL02pLpk4b9CtLaIKjaoTXU0GQf/wQiwcuJCn72D+vPwFa5VbHurg41P 3tEAw1ODB8hORoKY2Va9ql1JMTo/77gLNRNBg4qyN6q6MEHkza1kB7mNupxdceNvGv7N2oixfI AbgXC/3uAQKCIOc7CPNKlP/jAfNA+kLcACkSGgzlGroEH1NkHFMw1ihvY1NnARw7zJafTzJip4 pvw= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 02/12] hw/block/nvme: Generate namespace UUIDs Date: Sat, 7 Nov 2020 08:42:55 +0900 Message-Id: <20201106234305.21339-3-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" In NVMe 1.4, a namespace must report an ID descriptor of UUID type if it doesn't support EUI64 or NGUID. Add a new namespace property, "uuid", that provides the user the option to either specify the UUID explicitly or have a UUID generated automatically every time a namespace is initialized. Suggested-by: Klaus Jensen Signed-off-by: Dmitry Fomichev Reviewed-by: Klaus Jensen Reviewed-by: Keith Busch Reviewed-by: Niklas Cassel --- hw/block/nvme-ns.h | 1 + hw/block/nvme-ns.c | 1 + hw/block/nvme.c | 9 +++++---- 3 files changed, 7 insertions(+), 4 deletions(-) diff --git a/hw/block/nvme-ns.h b/hw/block/nvme-ns.h index ea8c2f785d..a38071884a 100644 --- a/hw/block/nvme-ns.h +++ b/hw/block/nvme-ns.h @@ -21,6 +21,7 @@ =20 typedef struct NvmeNamespaceParams { uint32_t nsid; + QemuUUID uuid; } NvmeNamespaceParams; =20 typedef struct NvmeNamespace { diff --git a/hw/block/nvme-ns.c b/hw/block/nvme-ns.c index b69cdaf27e..de735eb9f3 100644 --- a/hw/block/nvme-ns.c +++ b/hw/block/nvme-ns.c @@ -129,6 +129,7 @@ static void nvme_ns_realize(DeviceState *dev, Error **e= rrp) static Property nvme_ns_props[] =3D { DEFINE_BLOCK_PROPERTIES(NvmeNamespace, blkconf), DEFINE_PROP_UINT32("nsid", NvmeNamespace, params.nsid, 0), + DEFINE_PROP_UUID("uuid", NvmeNamespace, params.uuid), DEFINE_PROP_END_OF_LIST(), }; =20 diff --git a/hw/block/nvme.c b/hw/block/nvme.c index 702f7cc2e3..ed3f38f01d 100644 --- a/hw/block/nvme.c +++ b/hw/block/nvme.c @@ -1564,6 +1564,7 @@ static uint16_t nvme_identify_nslist(NvmeCtrl *n, Nvm= eRequest *req) =20 static uint16_t nvme_identify_ns_descr_list(NvmeCtrl *n, NvmeRequest *req) { + NvmeNamespace *ns; NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; uint32_t nsid =3D le32_to_cpu(c->nsid); uint8_t list[NVME_IDENTIFY_DATA_SIZE]; @@ -1583,7 +1584,8 @@ static uint16_t nvme_identify_ns_descr_list(NvmeCtrl = *n, NvmeRequest *req) return NVME_INVALID_NSID | NVME_DNR; } =20 - if (unlikely(!nvme_ns(n, nsid))) { + ns =3D nvme_ns(n, nsid); + if (unlikely(!ns)) { return NVME_INVALID_FIELD | NVME_DNR; } =20 @@ -1592,12 +1594,11 @@ static uint16_t nvme_identify_ns_descr_list(NvmeCtr= l *n, NvmeRequest *req) /* * Because the NGUID and EUI64 fields are 0 in the Identify Namespace = data * structure, a Namespace UUID (nidt =3D 0x3) must be reported in the - * Namespace Identification Descriptor. Add a very basic Namespace UUID - * here. + * Namespace Identification Descriptor. Add the namespace UUID here. */ ns_descrs->uuid.hdr.nidt =3D NVME_NIDT_UUID; ns_descrs->uuid.hdr.nidl =3D NVME_NIDT_UUID_LEN; - stl_be_p(&ns_descrs->uuid.v, nsid); + memcpy(&ns_descrs->uuid.v, ns->params.uuid.data, NVME_NIDT_UUID_LEN); =20 return nvme_dma(n, list, NVME_IDENTIFY_DATA_SIZE, DMA_DIRECTION_FROM_DEVICE, req); --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604706299; cv=none; d=zohomail.com; s=zohoarc; b=AIUmmCG8/MLgP7qjpB7GtuheW6rOSD7r4xKx/Hpa1IRO4gaq8gMWCptXHJz4GI+Jd5wxiONtxjEW916ts7Rvi8x9ElmQBiNf+HTeYZt0TeMv40OM7lvo+HIl8KbK9O6FIdXWoHQSkrzMEBs8HL9G89EUAg7QiK0qo4ixFw0gEVY= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604706299; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=r16gUjl3G5Ua8bOFSpQIEyBat5mVP5UppA+JLesKPOs=; b=RW4WIGgjFjv3hly9RxNoUPKNvs++AaLLiRf0MJGqiqbkcV9tNAzEQSt4R/ch3OlUPJJcGx3HpfE43lIjbweqRZSGGIVrEBqdmV7YAFLFZJ8FIF4zHC3lvU4pcEce5S0hwqYlHJ6JVGgh33gy9ZqQcEDSfuzWfCrC0dapr4ojG2M= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 160470629914148.42251519880176; Fri, 6 Nov 2020 15:44:59 -0800 (PST) Received: from localhost ([::1]:58902 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBPd-0007BN-Go for importer@patchew.org; Fri, 06 Nov 2020 18:44:57 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34212) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOD-0005K7-MM; Fri, 06 Nov 2020 18:43:31 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57347) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBO6-0003Sk-Fr; Fri, 06 Nov 2020 18:43:27 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:20 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:11 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:17 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706203; x=1636242203; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=6G60QSjZ5krlPLtstuo7O4+ni+lcVZttvDGIQYGvErY=; b=cG42IeA1rmUDUD3gBVRdLEUIroEPdNGAR15ddGgjpwLhjH1cTEEUaIim 6HhcyDSm7u8S+Oa2cmbNIfTsybd+6YE3wcV1OpfCGxL5b9fTRwDH4DMbL 3r5JGIQInGbFiIXVGUiyN+yk056jjpjadg/aVDGd6naumX+rubE7EFqVN 1TAlnbfIlLeAac0g6JBwybr63hLlxpoUrNY93LT9VxCJfy9gsL9x07KTD ZP2JNdno0jRglnO8Qq1DgHRqqLuQQLchJtNujSyIe33BWHjWOclyYSWFb kSz0yW81rY4EYF0cJjtvFIfnWcRCESHot8ImVeKlKQz+90gjNDfm87hJc Q==; IronPort-SDR: eVRk/l04tD7WHMybTbrYI5IrnyBFtqPqGh3biCRtDeuTreDZXsna9tk5MM8bHTcbnhpFeFXVNU sClJoGDgf8Dg2jULTJKzL3H3HKxr9vfrxRg6YyIaBxkSXgpe+/huvxAIf44JVZPXZC+7HR6VPt Tf0+S7yUPAJopj8c+wbdqI19UesYoQEfFTswgWTvo4fNl1MI+URZQiBusO2EhsFm8ET69dSSXK 00A6D8dgglg7GfA45AmQR/OnYQhShaMHW2dbDYhK3n/jihXaW0nFSaAKghzynI7jrq0S9JtEoo enE= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267058" IronPort-SDR: lPsAgmjEEX/sBbM6gJ4ThCQKCGwphh5CBKGobB7LSWSwSJswjOaVlkqIuTQZWeanpgc99VawWD kMiZmIcK2r4LeXJq3Z69PA3grh8KcE3uvDhsgmmS7bLUOgtntTK12PFNPNhXsRha/eEV6TA+Gi TbuZUWbsQyyRphiKXcyHbfsD571VdARODI7JpGJQhMHRH43akHvHT2jJEy/Mcn/B4T2A6KdokI YWqdu6XpGDNENIodF/0LWXXdX9i8x7IN+y2Q0S7IfTIarRZwzsPuW+2K+9KU6veDtUNBxT+Y9x Vs7H/9k+kNmKV8648wKmg8gt IronPort-SDR: jvqQwdkAZ804dQyqbFI+ZB9JkyifRd4C+4Ajm3NfSanxXjjw4LGo13/fDelabpFNG3cAnHz5Cx XPOIHzeZwee28OtYY0MNaeBopwOe6/3C7Pu/i59uw6G9uKwxutXprug2quaPwRaH61ONGlkLit TLCNUFT6yKVE8S7qf6vsb6smEaseBUhwM8cnq8b6PB/azIwyway5XEVcUmdpABR8z1RJpgOHvr HQYO00JtUXmLbCQDMoexEns8vUw/iZNK2FLFwqmgoiJWWKQzCqBcGxgxKYdZuoEu+mP9hovLHh Vvg= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 03/12] hw/block/nvme: Separate read and write handlers Date: Sat, 7 Nov 2020 08:42:56 +0900 Message-Id: <20201106234305.21339-4-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" With ZNS support in place, the majority of code in nvme_rw() has become read- or write-specific. Move these parts to two separate handlers, nvme_read() and nvme_write() to make the code more readable and to remove multiple is_write checks that so far existed in the i/o path. This is a refactoring patch, no change in functionality. Signed-off-by: Dmitry Fomichev Reviewed-by: Niklas Cassel Acked-by: Klaus Jensen --- hw/block/nvme.c | 91 ++++++++++++++++++++++++++++++------------- hw/block/trace-events | 3 +- 2 files changed, 67 insertions(+), 27 deletions(-) diff --git a/hw/block/nvme.c b/hw/block/nvme.c index ed3f38f01d..770e42a066 100644 --- a/hw/block/nvme.c +++ b/hw/block/nvme.c @@ -953,6 +953,54 @@ static uint16_t nvme_flush(NvmeCtrl *n, NvmeRequest *r= eq) return NVME_NO_COMPLETE; } =20 +static uint16_t nvme_read(NvmeCtrl *n, NvmeRequest *req) +{ + NvmeRwCmd *rw =3D (NvmeRwCmd *)&req->cmd; + NvmeNamespace *ns =3D req->ns; + uint64_t slba =3D le64_to_cpu(rw->slba); + uint32_t nlb =3D (uint32_t)le16_to_cpu(rw->nlb) + 1; + uint64_t data_size =3D nvme_l2b(ns, nlb); + uint64_t data_offset; + BlockBackend *blk =3D ns->blkconf.blk; + uint16_t status; + + trace_pci_nvme_read(nvme_cid(req), nvme_nsid(ns), nlb, data_size, slba= ); + + status =3D nvme_check_mdts(n, data_size); + if (status) { + trace_pci_nvme_err_mdts(nvme_cid(req), data_size); + goto invalid; + } + + status =3D nvme_check_bounds(n, ns, slba, nlb); + if (status) { + trace_pci_nvme_err_invalid_lba_range(slba, nlb, ns->id_ns.nsze); + goto invalid; + } + + status =3D nvme_map_dptr(n, data_size, req); + if (status) { + goto invalid; + } + + data_offset =3D nvme_l2b(ns, slba); + + block_acct_start(blk_get_stats(blk), &req->acct, data_size, + BLOCK_ACCT_READ); + if (req->qsg.sg) { + req->aiocb =3D dma_blk_read(blk, &req->qsg, data_offset, + BDRV_SECTOR_SIZE, nvme_rw_cb, req); + } else { + req->aiocb =3D blk_aio_preadv(blk, data_offset, &req->iov, 0, + nvme_rw_cb, req); + } + return NVME_NO_COMPLETE; + +invalid: + block_acct_invalid(blk_get_stats(blk), BLOCK_ACCT_READ); + return status | NVME_DNR; +} + static uint16_t nvme_write_zeroes(NvmeCtrl *n, NvmeRequest *req) { NvmeRwCmd *rw =3D (NvmeRwCmd *)&req->cmd; @@ -978,22 +1026,19 @@ static uint16_t nvme_write_zeroes(NvmeCtrl *n, NvmeR= equest *req) return NVME_NO_COMPLETE; } =20 -static uint16_t nvme_rw(NvmeCtrl *n, NvmeRequest *req) +static uint16_t nvme_write(NvmeCtrl *n, NvmeRequest *req) { NvmeRwCmd *rw =3D (NvmeRwCmd *)&req->cmd; NvmeNamespace *ns =3D req->ns; - uint32_t nlb =3D (uint32_t)le16_to_cpu(rw->nlb) + 1; uint64_t slba =3D le64_to_cpu(rw->slba); - + uint32_t nlb =3D (uint32_t)le16_to_cpu(rw->nlb) + 1; uint64_t data_size =3D nvme_l2b(ns, nlb); - uint64_t data_offset =3D nvme_l2b(ns, slba); - enum BlockAcctType acct =3D req->cmd.opcode =3D=3D NVME_CMD_WRITE ? - BLOCK_ACCT_WRITE : BLOCK_ACCT_READ; + uint64_t data_offset; BlockBackend *blk =3D ns->blkconf.blk; uint16_t status; =20 - trace_pci_nvme_rw(nvme_cid(req), nvme_io_opc_str(rw->opcode), - nvme_nsid(ns), nlb, data_size, slba); + trace_pci_nvme_write(nvme_cid(req), nvme_io_opc_str(rw->opcode), + nvme_nsid(ns), nlb, data_size, slba); =20 status =3D nvme_check_mdts(n, data_size); if (status) { @@ -1012,29 +1057,22 @@ static uint16_t nvme_rw(NvmeCtrl *n, NvmeRequest *r= eq) goto invalid; } =20 - block_acct_start(blk_get_stats(blk), &req->acct, data_size, acct); + data_offset =3D nvme_l2b(ns, slba); + + block_acct_start(blk_get_stats(blk), &req->acct, data_size, + BLOCK_ACCT_WRITE); if (req->qsg.sg) { - if (acct =3D=3D BLOCK_ACCT_WRITE) { - req->aiocb =3D dma_blk_write(blk, &req->qsg, data_offset, - BDRV_SECTOR_SIZE, nvme_rw_cb, req); - } else { - req->aiocb =3D dma_blk_read(blk, &req->qsg, data_offset, - BDRV_SECTOR_SIZE, nvme_rw_cb, req); - } + req->aiocb =3D dma_blk_write(blk, &req->qsg, data_offset, + BDRV_SECTOR_SIZE, nvme_rw_cb, req); } else { - if (acct =3D=3D BLOCK_ACCT_WRITE) { - req->aiocb =3D blk_aio_pwritev(blk, data_offset, &req->iov, 0, - nvme_rw_cb, req); - } else { - req->aiocb =3D blk_aio_preadv(blk, data_offset, &req->iov, 0, - nvme_rw_cb, req); - } + req->aiocb =3D blk_aio_pwritev(blk, data_offset, &req->iov, 0, + nvme_rw_cb, req); } return NVME_NO_COMPLETE; =20 invalid: - block_acct_invalid(blk_get_stats(ns->blkconf.blk), acct); - return status; + block_acct_invalid(blk_get_stats(blk), BLOCK_ACCT_WRITE); + return status | NVME_DNR; } =20 static uint16_t nvme_io_cmd(NvmeCtrl *n, NvmeRequest *req) @@ -1064,8 +1102,9 @@ static uint16_t nvme_io_cmd(NvmeCtrl *n, NvmeRequest = *req) case NVME_CMD_WRITE_ZEROES: return nvme_write_zeroes(n, req); case NVME_CMD_WRITE: + return nvme_write(n, req); case NVME_CMD_READ: - return nvme_rw(n, req); + return nvme_read(n, req); default: assert(false); } diff --git a/hw/block/trace-events b/hw/block/trace-events index 33832a2021..540c600931 100644 --- a/hw/block/trace-events +++ b/hw/block/trace-events @@ -40,7 +40,8 @@ pci_nvme_map_prp(uint64_t trans_len, uint32_t len, uint64= _t prp1, uint64_t prp2, pci_nvme_map_sgl(uint16_t cid, uint8_t typ, uint64_t len) "cid %"PRIu16" t= ype 0x%"PRIx8" len %"PRIu64"" pci_nvme_io_cmd(uint16_t cid, uint32_t nsid, uint16_t sqid, uint8_t opcode= , const char *opname) "cid %"PRIu16" nsid %"PRIu32" sqid %"PRIu16" opc 0x%"= PRIx8" opname '%s'" pci_nvme_admin_cmd(uint16_t cid, uint16_t sqid, uint8_t opcode, const char= *opname) "cid %"PRIu16" sqid %"PRIu16" opc 0x%"PRIx8" opname '%s'" -pci_nvme_rw(uint16_t cid, const char *verb, uint32_t nsid, uint32_t nlb, u= int64_t count, uint64_t lba) "cid %"PRIu16" opname '%s' nsid %"PRIu32" nlb = %"PRIu32" count %"PRIu64" lba 0x%"PRIx64"" +pci_nvme_read(uint16_t cid, uint32_t nsid, uint32_t nlb, uint64_t count, u= int64_t lba) "cid %"PRIu16" nsid %"PRIu32" nlb %"PRIu32" count %"PRIu64" lb= a 0x%"PRIx64"" +pci_nvme_write(uint16_t cid, const char *verb, uint32_t nsid, uint32_t nlb= , uint64_t count, uint64_t lba) "cid %"PRIu16" opname '%s' nsid %"PRIu32" n= lb %"PRIu32" count %"PRIu64" lba 0x%"PRIx64"" pci_nvme_rw_cb(uint16_t cid, const char *blkname) "cid %"PRIu16" blk '%s'" pci_nvme_write_zeroes(uint16_t cid, uint32_t nsid, uint64_t slba, uint32_t= nlb) "cid %"PRIu16" nsid %"PRIu32" slba %"PRIu64" nlb %"PRIu32"" pci_nvme_create_sq(uint64_t addr, uint16_t sqid, uint16_t cqid, uint16_t q= size, uint16_t qflags) "create submission queue, addr=3D0x%"PRIx64", sqid= =3D%"PRIu16", cqid=3D%"PRIu16", qsize=3D%"PRIu16", qflags=3D%"PRIu16"" --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604706627; cv=none; d=zohomail.com; s=zohoarc; b=fOMb6n9Ao7AnWj98ly+wucjLyNu5M3wsSTSloY1rn6Lj6pSY1kwu7PuuHjaKvfbX/BF/UFhNFluCwmTUZ+CxOjSG1aVoWKTaO6xx1PrDoF7Uh8PRwTjTWbVdC8CdCLCjHP52zzbCXM35VoyqpsmwrFtbfr2ph/n81Ht8fOPDyDI= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604706627; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=zuEXCu/aL1EmxBz1wOFGnYmhqbSR1tlG9wuPkeFzWNY=; b=RVP4ASBAdGh932AhWTr8lQwUgOwqhDx6U03rXDx8yQEqrh5as3WsiG8MZaKGv64cGjx0qhhdpXOdvtwYoQ5Rozhd/o3+f1lyDHZ+U7lx1Cj3FiJ21QIbHiTX2LQ1ryhB6Ii2pWAk7xXXS22uoI5EAwVOwwz4Y1r6mn0SXFTFZHQ= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604706627364433.68788832249095; Fri, 6 Nov 2020 15:50:27 -0800 (PST) Received: from localhost ([::1]:46598 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBUv-0005Tk-Nm for importer@patchew.org; Fri, 06 Nov 2020 18:50:25 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34234) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOF-0005Ll-Mq; Fri, 06 Nov 2020 18:43:31 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57360) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOB-0003Tq-Uq; Fri, 06 Nov 2020 18:43:31 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:23 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:14 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:20 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706208; x=1636242208; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=G3kD64FZFQBNwCAKQ6g+YVeLzgd0PXwIYGnld+o7ZrQ=; b=hDXKIku1ytFXPVMEpKNccP0xxJQCd8rzV5Jcb3F+Zf6/SrZmTPT3nDbD 5N8HemvSPOSAcq/GhXGnvjXFvWoU1DykCDvHLRRhYzIful7NmxWKPBx4j kbFaX2FPpdT5dNamcfzjHZlpaY4FV8Bz09YjfRx/GHacK9I/YuuUDUOME foIrym1UIwzZY8Ni1dQ3pVTxubFt3PhKfVVSHv2R8ostmi9yPSkeQOStI /AiAWC/vJ1YT7AZKMgDPVJZKGt4AeyI17ZahKonMOtrxMMUec+DHXppfn jzQZkHYRBYcJt4L0CekrthZkeJtFCA4AF+Qo0sDzqA9NjuWfPkgFNJQbO A==; IronPort-SDR: MIbK2XfzX3Aj9jr3cKRczTgNhxYNgzjUEin1UxmU4ok8qnSqfV4Y9EWEcf9QGFJEaw4s+uI/hA 2iDIrYXBmSdz6q2MiJ6kCyT1wfU98F3b9Anmj2cMIBd1oVAuDYJ1SumB0m2toCmLJadT4TAit4 I+thKvI19AVcFF+2++DnoO6uJie64PEe6SO6jA2QwSVnOLUUiCh8N84X62LyT8k+EPQjwzAM+I ubTjWlwCZpafiuOSNl4xqWqSRozOiKOf/lNgMbY5cKOIAVZAOI39EnoJhK0agShonBmB6j3Jvm wIc= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267061" IronPort-SDR: gxAxOcDNIxokSlWuSZ6a2m/QAIdrDTOIlLXYiyfER92JDrcD9nXxyiJMwA9jfFO365Chp88IOy KvRmEzD+4tEqgfNeYaQ1Ss3midfs8iK80sAxQOiZxkn+Vd8Fx7r00edunO+gz8W/rwx0fqFgVN EHCdAUye2+sHnw/+5x9gGS6yI3TSQelb/XKb3FcKgMyUSwuo6ZZ9QQ+FxlY6MEgiYYgo2HyNUC 3x0FPMlCT+0G+kbUzdWPIBa8sNKQzQUFPIkQQPh9ul7ZCmywTnYcLcHWcCUlxE5fpgqPcMl11e wHB2B1R9KNMyRr1rXORV/bv7 IronPort-SDR: 3bfwp0CI24LlCyo+hLjKBDRQ9mJhajMPgm+NF5ZDmO1Atutzmgo5eZ2qFzBFSOLH+mbtZXct5y PiDqJWrJRPho08x0H8Qzxb+OR6YPL2bQ455+JWqtyiu/LHThJtKnRTYw2cYaKxbBSUvsT9GOeJ BlopNpWnOfBMVoLJR326JoXQ919aq3pO+d+6kLoVNuJCLpWz1O+/LNn+f3vQiQfbEQDIDzKgU4 +PrFJIuhHVlcrCspaIOJALKw++SHyjSdsU4EMB7nPTuTGfekcxT5FmoM1KsSrW1apkgRhxVLFJ 53U= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 04/12] hw/block/nvme: Merge nvme_write_zeroes() with nvme_write() Date: Sat, 7 Nov 2020 08:42:57 +0900 Message-Id: <20201106234305.21339-5-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" nvme_write() now handles WRITE, WRITE ZEROES and ZONE_APPEND. Signed-off-by: Dmitry Fomichev Reviewed-by: Niklas Cassel Acked-by: Klaus Jensen --- hw/block/nvme.c | 72 +++++++++++++++++-------------------------- hw/block/trace-events | 1 - 2 files changed, 28 insertions(+), 45 deletions(-) diff --git a/hw/block/nvme.c b/hw/block/nvme.c index 770e42a066..48adbe84f5 100644 --- a/hw/block/nvme.c +++ b/hw/block/nvme.c @@ -1001,32 +1001,7 @@ invalid: return status | NVME_DNR; } =20 -static uint16_t nvme_write_zeroes(NvmeCtrl *n, NvmeRequest *req) -{ - NvmeRwCmd *rw =3D (NvmeRwCmd *)&req->cmd; - NvmeNamespace *ns =3D req->ns; - uint64_t slba =3D le64_to_cpu(rw->slba); - uint32_t nlb =3D (uint32_t)le16_to_cpu(rw->nlb) + 1; - uint64_t offset =3D nvme_l2b(ns, slba); - uint32_t count =3D nvme_l2b(ns, nlb); - uint16_t status; - - trace_pci_nvme_write_zeroes(nvme_cid(req), nvme_nsid(ns), slba, nlb); - - status =3D nvme_check_bounds(n, ns, slba, nlb); - if (status) { - trace_pci_nvme_err_invalid_lba_range(slba, nlb, ns->id_ns.nsze); - return status; - } - - block_acct_start(blk_get_stats(req->ns->blkconf.blk), &req->acct, 0, - BLOCK_ACCT_WRITE); - req->aiocb =3D blk_aio_pwrite_zeroes(req->ns->blkconf.blk, offset, cou= nt, - BDRV_REQ_MAY_UNMAP, nvme_rw_cb, req= ); - return NVME_NO_COMPLETE; -} - -static uint16_t nvme_write(NvmeCtrl *n, NvmeRequest *req) +static uint16_t nvme_write(NvmeCtrl *n, NvmeRequest *req, bool wrz) { NvmeRwCmd *rw =3D (NvmeRwCmd *)&req->cmd; NvmeNamespace *ns =3D req->ns; @@ -1040,10 +1015,12 @@ static uint16_t nvme_write(NvmeCtrl *n, NvmeRequest= *req) trace_pci_nvme_write(nvme_cid(req), nvme_io_opc_str(rw->opcode), nvme_nsid(ns), nlb, data_size, slba); =20 - status =3D nvme_check_mdts(n, data_size); - if (status) { - trace_pci_nvme_err_mdts(nvme_cid(req), data_size); - goto invalid; + if (!wrz) { + status =3D nvme_check_mdts(n, data_size); + if (status) { + trace_pci_nvme_err_mdts(nvme_cid(req), data_size); + goto invalid; + } } =20 status =3D nvme_check_bounds(n, ns, slba, nlb); @@ -1052,21 +1029,28 @@ static uint16_t nvme_write(NvmeCtrl *n, NvmeRequest= *req) goto invalid; } =20 - status =3D nvme_map_dptr(n, data_size, req); - if (status) { - goto invalid; - } - data_offset =3D nvme_l2b(ns, slba); =20 - block_acct_start(blk_get_stats(blk), &req->acct, data_size, - BLOCK_ACCT_WRITE); - if (req->qsg.sg) { - req->aiocb =3D dma_blk_write(blk, &req->qsg, data_offset, - BDRV_SECTOR_SIZE, nvme_rw_cb, req); + if (!wrz) { + status =3D nvme_map_dptr(n, data_size, req); + if (status) { + goto invalid; + } + + block_acct_start(blk_get_stats(blk), &req->acct, data_size, + BLOCK_ACCT_WRITE); + if (req->qsg.sg) { + req->aiocb =3D dma_blk_write(blk, &req->qsg, data_offset, + BDRV_SECTOR_SIZE, nvme_rw_cb, req); + } else { + req->aiocb =3D blk_aio_pwritev(blk, data_offset, &req->iov, 0, + nvme_rw_cb, req); + } } else { - req->aiocb =3D blk_aio_pwritev(blk, data_offset, &req->iov, 0, - nvme_rw_cb, req); + block_acct_start(blk_get_stats(blk), &req->acct, 0, BLOCK_ACCT_WRI= TE); + req->aiocb =3D blk_aio_pwrite_zeroes(blk, data_offset, data_size, + BDRV_REQ_MAY_UNMAP, nvme_rw_cb, + req); } return NVME_NO_COMPLETE; =20 @@ -1100,9 +1084,9 @@ static uint16_t nvme_io_cmd(NvmeCtrl *n, NvmeRequest = *req) case NVME_CMD_FLUSH: return nvme_flush(n, req); case NVME_CMD_WRITE_ZEROES: - return nvme_write_zeroes(n, req); + return nvme_write(n, req, true); case NVME_CMD_WRITE: - return nvme_write(n, req); + return nvme_write(n, req, false); case NVME_CMD_READ: return nvme_read(n, req); default: diff --git a/hw/block/trace-events b/hw/block/trace-events index 540c600931..e67e96c2b5 100644 --- a/hw/block/trace-events +++ b/hw/block/trace-events @@ -43,7 +43,6 @@ pci_nvme_admin_cmd(uint16_t cid, uint16_t sqid, uint8_t o= pcode, const char *opna pci_nvme_read(uint16_t cid, uint32_t nsid, uint32_t nlb, uint64_t count, u= int64_t lba) "cid %"PRIu16" nsid %"PRIu32" nlb %"PRIu32" count %"PRIu64" lb= a 0x%"PRIx64"" pci_nvme_write(uint16_t cid, const char *verb, uint32_t nsid, uint32_t nlb= , uint64_t count, uint64_t lba) "cid %"PRIu16" opname '%s' nsid %"PRIu32" n= lb %"PRIu32" count %"PRIu64" lba 0x%"PRIx64"" pci_nvme_rw_cb(uint16_t cid, const char *blkname) "cid %"PRIu16" blk '%s'" -pci_nvme_write_zeroes(uint16_t cid, uint32_t nsid, uint64_t slba, uint32_t= nlb) "cid %"PRIu16" nsid %"PRIu32" slba %"PRIu64" nlb %"PRIu32"" pci_nvme_create_sq(uint64_t addr, uint16_t sqid, uint16_t cqid, uint16_t q= size, uint16_t qflags) "create submission queue, addr=3D0x%"PRIx64", sqid= =3D%"PRIu16", cqid=3D%"PRIu16", qsize=3D%"PRIu16", qflags=3D%"PRIu16"" pci_nvme_create_cq(uint64_t addr, uint16_t cqid, uint16_t vector, uint16_t= size, uint16_t qflags, int ien) "create completion queue, addr=3D0x%"PRIx6= 4", cqid=3D%"PRIu16", vector=3D%"PRIu16", qsize=3D%"PRIu16", qflags=3D%"PRI= u16", ien=3D%d" pci_nvme_del_sq(uint16_t qid) "deleting submission queue sqid=3D%"PRIu16"" --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604706771; cv=none; d=zohomail.com; s=zohoarc; b=EBtkvCK1MGsvNbV81V4ic8RjTwKfgKQJLaw3QgqDLbzZnQH6ibO1wecPWdkMyQUevGHil3RWw5jck4AellL563SyqcugDdXDz6g47APSOr3hxG6t487iAt94lDNU8kvwesYRd75oWxywV8xg0o+M14Di4G5EYNmtO0wOFElxYNk= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604706771; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=1XlE6oycZFzolID0P636Dbodn4SCBkmcuo/r3mko31I=; b=hDUvvr0RdYpy5462C/CxpqlWAvbXrF/9fYlwdIT9iNA3NnaC5tpfv4Zw9nd9zqbOja/69fpvUFdL0FFEAKaIXNo/vKwNJzlC3CyX8gMqzRt9+HlCKrBOPCcqdlQzkgrLrB4t60gczlV457ITg3V5jvwt7BPlIle+W7Lzz0UYkbY= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604706771658248.94049611758214; Fri, 6 Nov 2020 15:52:51 -0800 (PST) Received: from localhost ([::1]:50164 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBXG-00076z-6H for importer@patchew.org; Fri, 06 Nov 2020 18:52:50 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34260) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOG-0005O4-OG; Fri, 06 Nov 2020 18:43:32 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57362) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOB-0003U4-V2; Fri, 06 Nov 2020 18:43:32 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:26 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:17 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:23 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706208; x=1636242208; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=+ppyMgq050tOprIgbi9IDz8nw1YexhJOX71OXLYK6SQ=; b=BIncTqFcX3lZk7xHSMkg0dE6hDQqKx5t3Op63bXrp1fOKX4cWWvBQrX8 76rKPbzSXog++hi/wdBi8dC3w72yMTlzIpTRdn1cWbWhtA4g5KxPxGMXO DPR9PPmveQHlJGQh920ke0aiSj6ksaLnJegBQvOZKd1fUwEF7DB7jQnDq /+AzOv52LeNB6c71EVPHcKIz/sS+p46WvD3wLxO+/Ct0F0xHglE7Po6T+ 0dFKrA1h59fPa7e9fzlVTZ4v2cZmiw3ZXfiIrcEWpc2B5uqmjWfROCOLL X3QRDg0xCr5ZA2S7x6qSIc4XVdYgfQUg9E8aQW9WGRCabemNgS0ENZlia g==; IronPort-SDR: lZCqfqKstMXDwA2NmkOzMCMCytSCIvSCPt87g+gKvNq5w6Jw7IOfGWB8+Q41c1V2oH+TLfyUu/ mr4rzg2AJPk2BgOnqFRDfTC8FjQyu06NDdomxAHwwgybJNYZx7BTnEAGNj3QQBHzGXOyAG/IhF hwQTx1uTVosTqPFL38JMgnnBPzXH42SNQ4/dupLkPei0ilK14PKsMJZpw34pWPeiieLpr4sg/w QVSAUGLWIEpoHGkikpvtx9YoKeNZ22QF6TWLywnmpAIlYWGIUaYF4OQPBsK80Po5utPUg4Q5pD k54= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267064" IronPort-SDR: +4nN9fHzYR3aGYLzrNs07s79ufMEceq9rqItM2VejjvwuksvppWjLWC6uKIylYxQ4RdI6vbWWb QWm0gP+6yeWTr4N7gVz+SouX5z2JKQ0SGFvHFNt6qKyxSe3GFGw+rlkfELOMxLWFxivX1sOIJd z8GDgfseDzuXOULmTruE2+zY5qiLqa0NeDfX/EgGflIVfwqZUdc7yVAZay0HMlDvLwIL89B8bT +igTmv/JQvNDAgOwo9jl3t4inGgSji84pYLKmzOJK+YviYOVVedWLJEGCSSb6322Kna0cEDXLN rUnaTwgdNHI6G+JNlLz56Yqk IronPort-SDR: g9p9q635i2P3LcCKcADhWKVfZoBbOQp8a+ifChwMouHdKVRarbFX8rHnn7qfZu3Vtg4ijyd709 cie69jYmjAQDUdaT1oUMZta6YXZ+yfYGEPDvM8cRkOcx59UxexS2WKpT0A1v6Gm+5a84uZGjut cjoM5ELsDdUCXZjook0WgokShiYsQcOTsXwDg0yOoAu1mWFqAY0llPfbXBEgqzrEOhICAjWM7o oq1iFhGYJhN4Qjd8EgHn14InQadq7JJBtjErzTV41YBTELhXZ+kxjueRWdJZ4SexOvtzVBF+MJ IZA= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 05/12] hw/block/nvme: Add support for Namespace Types Date: Sat, 7 Nov 2020 08:42:58 +0900 Message-Id: <20201106234305.21339-6-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" From: Niklas Cassel Define the structures and constants required to implement Namespace Types support. Namespace Types introduce a new command set, "I/O Command Sets", that allows the host to retrieve the command sets associated with a namespace. Introduce support for the command set and enable detection for the NVM Command Set. The new workflows for identify commands rely heavily on zero-filled identify structs. E.g., certain CNS commands are defined to return a zero-filled identify struct when an inactive namespace NSID is supplied. Add a helper function in order to avoid code duplication when reporting zero-filled identify structures. Signed-off-by: Niklas Cassel Signed-off-by: Dmitry Fomichev Reviewed-by: Keith Busch --- hw/block/nvme-ns.h | 1 + include/block/nvme.h | 66 +++++++++++---- hw/block/nvme-ns.c | 2 + hw/block/nvme.c | 188 +++++++++++++++++++++++++++++++++++------- hw/block/trace-events | 7 ++ 5 files changed, 219 insertions(+), 45 deletions(-) diff --git a/hw/block/nvme-ns.h b/hw/block/nvme-ns.h index a38071884a..d795e44bab 100644 --- a/hw/block/nvme-ns.h +++ b/hw/block/nvme-ns.h @@ -31,6 +31,7 @@ typedef struct NvmeNamespace { int64_t size; NvmeIdNs id_ns; const uint32_t *iocs; + uint8_t csi; =20 NvmeNamespaceParams params; } NvmeNamespace; diff --git a/include/block/nvme.h b/include/block/nvme.h index f62cc90d49..af23514713 100644 --- a/include/block/nvme.h +++ b/include/block/nvme.h @@ -84,6 +84,7 @@ enum NvmeCapMask { =20 enum NvmeCapCss { NVME_CAP_CSS_NVM =3D 1 << 0, + NVME_CAP_CSS_CSI_SUPP =3D 1 << 6, NVME_CAP_CSS_ADMIN_ONLY =3D 1 << 7, }; =20 @@ -117,9 +118,25 @@ enum NvmeCcMask { =20 enum NvmeCcCss { NVME_CC_CSS_NVM =3D 0x0, + NVME_CC_CSS_CSI =3D 0x6, NVME_CC_CSS_ADMIN_ONLY =3D 0x7, }; =20 +#define NVME_SET_CC_EN(cc, val) \ + (cc |=3D (uint32_t)((val) & CC_EN_MASK) << CC_EN_SHIFT) +#define NVME_SET_CC_CSS(cc, val) \ + (cc |=3D (uint32_t)((val) & CC_CSS_MASK) << CC_CSS_SHIFT) +#define NVME_SET_CC_MPS(cc, val) \ + (cc |=3D (uint32_t)((val) & CC_MPS_MASK) << CC_MPS_SHIFT) +#define NVME_SET_CC_AMS(cc, val) \ + (cc |=3D (uint32_t)((val) & CC_AMS_MASK) << CC_AMS_SHIFT) +#define NVME_SET_CC_SHN(cc, val) \ + (cc |=3D (uint32_t)((val) & CC_SHN_MASK) << CC_SHN_SHIFT) +#define NVME_SET_CC_IOSQES(cc, val) \ + (cc |=3D (uint32_t)((val) & CC_IOSQES_MASK) << CC_IOSQES_SHIFT) +#define NVME_SET_CC_IOCQES(cc, val) \ + (cc |=3D (uint32_t)((val) & CC_IOCQES_MASK) << CC_IOCQES_SHIFT) + enum NvmeCstsShift { CSTS_RDY_SHIFT =3D 0, CSTS_CFS_SHIFT =3D 1, @@ -534,8 +551,13 @@ typedef struct QEMU_PACKED NvmeIdentify { uint64_t rsvd2[2]; uint64_t prp1; uint64_t prp2; - uint32_t cns; - uint32_t rsvd11[5]; + uint8_t cns; + uint8_t rsvd10; + uint16_t ctrlid; + uint16_t nvmsetid; + uint8_t rsvd11; + uint8_t csi; + uint32_t rsvd12[4]; } NvmeIdentify; =20 typedef struct QEMU_PACKED NvmeRwCmd { @@ -656,6 +678,7 @@ enum NvmeStatusCodes { NVME_SGL_DESCR_TYPE_INVALID =3D 0x0011, NVME_INVALID_USE_OF_CMB =3D 0x0012, NVME_INVALID_PRP_OFFSET =3D 0x0013, + NVME_CMD_SET_CMB_REJECTED =3D 0x002b, NVME_LBA_RANGE =3D 0x0080, NVME_CAP_EXCEEDED =3D 0x0081, NVME_NS_NOT_READY =3D 0x0082, @@ -782,11 +805,15 @@ typedef struct QEMU_PACKED NvmePSD { =20 #define NVME_IDENTIFY_DATA_SIZE 4096 =20 -enum { - NVME_ID_CNS_NS =3D 0x0, - NVME_ID_CNS_CTRL =3D 0x1, - NVME_ID_CNS_NS_ACTIVE_LIST =3D 0x2, - NVME_ID_CNS_NS_DESCR_LIST =3D 0x3, +enum NvmeIdCns { + NVME_ID_CNS_NS =3D 0x00, + NVME_ID_CNS_CTRL =3D 0x01, + NVME_ID_CNS_NS_ACTIVE_LIST =3D 0x02, + NVME_ID_CNS_NS_DESCR_LIST =3D 0x03, + NVME_ID_CNS_CS_NS =3D 0x05, + NVME_ID_CNS_CS_CTRL =3D 0x06, + NVME_ID_CNS_CS_NS_ACTIVE_LIST =3D 0x07, + NVME_ID_CNS_IO_COMMAND_SET =3D 0x1c, }; =20 typedef struct QEMU_PACKED NvmeIdCtrl { @@ -934,6 +961,7 @@ enum NvmeFeatureIds { NVME_WRITE_ATOMICITY =3D 0xa, NVME_ASYNCHRONOUS_EVENT_CONF =3D 0xb, NVME_TIMESTAMP =3D 0xe, + NVME_COMMAND_SET_PROFILE =3D 0x19, NVME_SOFTWARE_PROGRESS_MARKER =3D 0x80, NVME_FID_MAX =3D 0x100, }; @@ -1018,18 +1046,26 @@ typedef struct QEMU_PACKED NvmeIdNsDescr { uint8_t rsvd2[2]; } NvmeIdNsDescr; =20 -enum { - NVME_NIDT_EUI64_LEN =3D 8, - NVME_NIDT_NGUID_LEN =3D 16, - NVME_NIDT_UUID_LEN =3D 16, +enum NvmeNsIdentifierLength { + NVME_NIDL_EUI64 =3D 8, + NVME_NIDL_NGUID =3D 16, + NVME_NIDL_UUID =3D 16, + NVME_NIDL_CSI =3D 1, }; =20 enum NvmeNsIdentifierType { - NVME_NIDT_EUI64 =3D 0x1, - NVME_NIDT_NGUID =3D 0x2, - NVME_NIDT_UUID =3D 0x3, + NVME_NIDT_EUI64 =3D 0x01, + NVME_NIDT_NGUID =3D 0x02, + NVME_NIDT_UUID =3D 0x03, + NVME_NIDT_CSI =3D 0x04, }; =20 +enum NvmeCsi { + NVME_CSI_NVM =3D 0x00, +}; + +#define NVME_SET_CSI(vec, csi) (vec |=3D (uint8_t)(1 << (csi))) + /*Deallocate Logical Block Features*/ #define NVME_ID_NS_DLFEAT_GUARD_CRC(dlfeat) ((dlfeat) & 0x10) #define NVME_ID_NS_DLFEAT_WRITE_ZEROES(dlfeat) ((dlfeat) & 0x08) @@ -1080,8 +1116,8 @@ static inline void _nvme_check_size(void) QEMU_BUILD_BUG_ON(sizeof(NvmeSmartLog) !=3D 512); QEMU_BUILD_BUG_ON(sizeof(NvmeEffectsLog) !=3D 4096); QEMU_BUILD_BUG_ON(sizeof(NvmeIdCtrl) !=3D 4096); + QEMU_BUILD_BUG_ON(sizeof(NvmeIdNsDescr) !=3D 4); QEMU_BUILD_BUG_ON(sizeof(NvmeIdNs) !=3D 4096); QEMU_BUILD_BUG_ON(sizeof(NvmeSglDescriptor) !=3D 16); - QEMU_BUILD_BUG_ON(sizeof(NvmeIdNsDescr) !=3D 4); } #endif diff --git a/hw/block/nvme-ns.c b/hw/block/nvme-ns.c index de735eb9f3..c0362426cc 100644 --- a/hw/block/nvme-ns.c +++ b/hw/block/nvme-ns.c @@ -41,6 +41,8 @@ static void nvme_ns_init(NvmeNamespace *ns) =20 id_ns->nsze =3D cpu_to_le64(nvme_ns_nlbas(ns)); =20 + ns->csi =3D NVME_CSI_NVM; + /* no thin provisioning */ id_ns->ncap =3D id_ns->nsze; id_ns->nuse =3D id_ns->ncap; diff --git a/hw/block/nvme.c b/hw/block/nvme.c index 48adbe84f5..bb82dd9975 100644 --- a/hw/block/nvme.c +++ b/hw/block/nvme.c @@ -1327,7 +1327,7 @@ static uint16_t nvme_error_info(NvmeCtrl *n, uint8_t = rae, uint32_t buf_len, DMA_DIRECTION_FROM_DEVICE, req); } =20 -static uint16_t nvme_cmd_effects(NvmeCtrl *n, uint32_t buf_len, +static uint16_t nvme_cmd_effects(NvmeCtrl *n, uint8_t csi, uint32_t buf_le= n, uint64_t off, NvmeRequest *req) { NvmeEffectsLog log =3D {}; @@ -1342,8 +1342,15 @@ static uint16_t nvme_cmd_effects(NvmeCtrl *n, uint32= _t buf_len, switch (NVME_CC_CSS(n->bar.cc)) { case NVME_CC_CSS_NVM: src_iocs =3D nvme_cse_iocs_nvm; + /* fall through */ case NVME_CC_CSS_ADMIN_ONLY: break; + case NVME_CC_CSS_CSI: + switch (csi) { + case NVME_CSI_NVM: + src_iocs =3D nvme_cse_iocs_nvm; + break; + } } =20 memcpy(log.acs, nvme_cse_acs, sizeof(nvme_cse_acs)); @@ -1369,6 +1376,7 @@ static uint16_t nvme_get_log(NvmeCtrl *n, NvmeRequest= *req) uint8_t lid =3D dw10 & 0xff; uint8_t lsp =3D (dw10 >> 8) & 0xf; uint8_t rae =3D (dw10 >> 15) & 0x1; + uint8_t csi =3D le32_to_cpu(cmd->cdw14) >> 24; uint32_t numdl, numdu; uint64_t off, lpol, lpou; size_t len; @@ -1402,7 +1410,7 @@ static uint16_t nvme_get_log(NvmeCtrl *n, NvmeRequest= *req) case NVME_LOG_FW_SLOT_INFO: return nvme_fw_log_info(n, len, off, req); case NVME_LOG_CMD_EFFECTS: - return nvme_cmd_effects(n, len, off, req); + return nvme_cmd_effects(n, csi, len, off, req); default: trace_pci_nvme_err_invalid_log_page(nvme_cid(req), lid); return NVME_INVALID_FIELD | NVME_DNR; @@ -1516,6 +1524,13 @@ static uint16_t nvme_create_cq(NvmeCtrl *n, NvmeRequ= est *req) return NVME_SUCCESS; } =20 +static uint16_t nvme_rpt_empty_id_struct(NvmeCtrl *n, NvmeRequest *req) +{ + uint8_t id[NVME_IDENTIFY_DATA_SIZE] =3D {}; + + return nvme_dma(n, id, sizeof(id), DMA_DIRECTION_FROM_DEVICE, req); +} + static uint16_t nvme_identify_ctrl(NvmeCtrl *n, NvmeRequest *req) { trace_pci_nvme_identify_ctrl(); @@ -1524,11 +1539,23 @@ static uint16_t nvme_identify_ctrl(NvmeCtrl *n, Nvm= eRequest *req) DMA_DIRECTION_FROM_DEVICE, req); } =20 +static uint16_t nvme_identify_ctrl_csi(NvmeCtrl *n, NvmeRequest *req) +{ + NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; + + trace_pci_nvme_identify_ctrl_csi(c->csi); + + if (c->csi =3D=3D NVME_CSI_NVM) { + return nvme_rpt_empty_id_struct(n, req); + } + + return NVME_INVALID_FIELD | NVME_DNR; +} + static uint16_t nvme_identify_ns(NvmeCtrl *n, NvmeRequest *req) { NvmeNamespace *ns; NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; - NvmeIdNs *id_ns, inactive =3D { 0 }; uint32_t nsid =3D le32_to_cpu(c->nsid); =20 trace_pci_nvme_identify_ns(nsid); @@ -1539,23 +1566,46 @@ static uint16_t nvme_identify_ns(NvmeCtrl *n, NvmeR= equest *req) =20 ns =3D nvme_ns(n, nsid); if (unlikely(!ns)) { - id_ns =3D &inactive; - } else { - id_ns =3D &ns->id_ns; + return nvme_rpt_empty_id_struct(n, req); } =20 - return nvme_dma(n, (uint8_t *)id_ns, sizeof(NvmeIdNs), + return nvme_dma(n, (uint8_t *)&ns->id_ns, sizeof(NvmeIdNs), DMA_DIRECTION_FROM_DEVICE, req); } =20 +static uint16_t nvme_identify_ns_csi(NvmeCtrl *n, NvmeRequest *req) +{ + NvmeNamespace *ns; + NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; + uint32_t nsid =3D le32_to_cpu(c->nsid); + + trace_pci_nvme_identify_ns_csi(nsid, c->csi); + + if (!nvme_nsid_valid(n, nsid) || nsid =3D=3D NVME_NSID_BROADCAST) { + return NVME_INVALID_NSID | NVME_DNR; + } + + ns =3D nvme_ns(n, nsid); + if (unlikely(!ns)) { + return nvme_rpt_empty_id_struct(n, req); + } + + if (c->csi =3D=3D NVME_CSI_NVM) { + return nvme_rpt_empty_id_struct(n, req); + } + + return NVME_INVALID_FIELD | NVME_DNR; +} + static uint16_t nvme_identify_nslist(NvmeCtrl *n, NvmeRequest *req) { + NvmeNamespace *ns; NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; - static const int data_len =3D NVME_IDENTIFY_DATA_SIZE; uint32_t min_nsid =3D le32_to_cpu(c->nsid); - uint32_t *list; - uint16_t ret; - int j =3D 0; + uint8_t list[NVME_IDENTIFY_DATA_SIZE] =3D {}; + static const int data_len =3D sizeof(list); + uint32_t *list_ptr =3D (uint32_t *)list; + int i, j =3D 0; =20 trace_pci_nvme_identify_nslist(min_nsid); =20 @@ -1569,20 +1619,61 @@ static uint16_t nvme_identify_nslist(NvmeCtrl *n, N= vmeRequest *req) return NVME_INVALID_NSID | NVME_DNR; } =20 - list =3D g_malloc0(data_len); - for (int i =3D 1; i <=3D n->num_namespaces; i++) { - if (i <=3D min_nsid || !nvme_ns(n, i)) { + for (i =3D 1; i <=3D n->num_namespaces; i++) { + ns =3D nvme_ns(n, i); + if (!ns) { continue; } - list[j++] =3D cpu_to_le32(i); + if (ns->params.nsid <=3D min_nsid) { + continue; + } + list_ptr[j++] =3D cpu_to_le32(ns->params.nsid); if (j =3D=3D data_len / sizeof(uint32_t)) { break; } } - ret =3D nvme_dma(n, (uint8_t *)list, data_len, DMA_DIRECTION_FROM_DEVI= CE, - req); - g_free(list); - return ret; + + return nvme_dma(n, list, data_len, DMA_DIRECTION_FROM_DEVICE, req); +} + +static uint16_t nvme_identify_nslist_csi(NvmeCtrl *n, NvmeRequest *req) +{ + NvmeNamespace *ns; + NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; + uint32_t min_nsid =3D le32_to_cpu(c->nsid); + uint8_t list[NVME_IDENTIFY_DATA_SIZE] =3D {}; + static const int data_len =3D sizeof(list); + uint32_t *list_ptr =3D (uint32_t *)list; + int i, j =3D 0; + + trace_pci_nvme_identify_nslist_csi(min_nsid, c->csi); + + /* + * Same as in nvme_identify_nslist(), 0xffffffff/0xfffffffe are invali= d. + */ + if (min_nsid >=3D NVME_NSID_BROADCAST - 1) { + return NVME_INVALID_NSID | NVME_DNR; + } + + if (c->csi !=3D NVME_CSI_NVM) { + return NVME_INVALID_FIELD | NVME_DNR; + } + + for (i =3D 1; i <=3D n->num_namespaces; i++) { + ns =3D nvme_ns(n, i); + if (!ns) { + continue; + } + if (ns->params.nsid <=3D min_nsid) { + continue; + } + list_ptr[j++] =3D cpu_to_le32(ns->params.nsid); + if (j =3D=3D data_len / sizeof(uint32_t)) { + break; + } + } + + return nvme_dma(n, list, data_len, DMA_DIRECTION_FROM_DEVICE, req); } =20 static uint16_t nvme_identify_ns_descr_list(NvmeCtrl *n, NvmeRequest *req) @@ -1590,13 +1681,17 @@ static uint16_t nvme_identify_ns_descr_list(NvmeCtr= l *n, NvmeRequest *req) NvmeNamespace *ns; NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; uint32_t nsid =3D le32_to_cpu(c->nsid); - uint8_t list[NVME_IDENTIFY_DATA_SIZE]; + uint8_t list[NVME_IDENTIFY_DATA_SIZE] =3D {}; =20 struct data { struct { NvmeIdNsDescr hdr; - uint8_t v[16]; + uint8_t v[NVME_NIDL_UUID]; } uuid; + struct { + NvmeIdNsDescr hdr; + uint8_t v; + } csi; }; =20 struct data *ns_descrs =3D (struct data *)list; @@ -1612,19 +1707,31 @@ static uint16_t nvme_identify_ns_descr_list(NvmeCtr= l *n, NvmeRequest *req) return NVME_INVALID_FIELD | NVME_DNR; } =20 - memset(list, 0x0, sizeof(list)); - /* * Because the NGUID and EUI64 fields are 0 in the Identify Namespace = data * structure, a Namespace UUID (nidt =3D 0x3) must be reported in the * Namespace Identification Descriptor. Add the namespace UUID here. */ ns_descrs->uuid.hdr.nidt =3D NVME_NIDT_UUID; - ns_descrs->uuid.hdr.nidl =3D NVME_NIDT_UUID_LEN; - memcpy(&ns_descrs->uuid.v, ns->params.uuid.data, NVME_NIDT_UUID_LEN); + ns_descrs->uuid.hdr.nidl =3D NVME_NIDL_UUID; + memcpy(&ns_descrs->uuid.v, ns->params.uuid.data, NVME_NIDL_UUID); =20 - return nvme_dma(n, list, NVME_IDENTIFY_DATA_SIZE, - DMA_DIRECTION_FROM_DEVICE, req); + ns_descrs->csi.hdr.nidt =3D NVME_NIDT_CSI; + ns_descrs->csi.hdr.nidl =3D NVME_NIDL_CSI; + ns_descrs->csi.v =3D ns->csi; + + return nvme_dma(n, list, sizeof(list), DMA_DIRECTION_FROM_DEVICE, req); +} + +static uint16_t nvme_identify_cmd_set(NvmeCtrl *n, NvmeRequest *req) +{ + uint8_t list[NVME_IDENTIFY_DATA_SIZE] =3D {}; + static const int data_len =3D sizeof(list); + + trace_pci_nvme_identify_cmd_set(); + + NVME_SET_CSI(*list, NVME_CSI_NVM); + return nvme_dma(n, list, data_len, DMA_DIRECTION_FROM_DEVICE, req); } =20 static uint16_t nvme_identify(NvmeCtrl *n, NvmeRequest *req) @@ -1634,12 +1741,20 @@ static uint16_t nvme_identify(NvmeCtrl *n, NvmeRequ= est *req) switch (le32_to_cpu(c->cns)) { case NVME_ID_CNS_NS: return nvme_identify_ns(n, req); + case NVME_ID_CNS_CS_NS: + return nvme_identify_ns_csi(n, req); case NVME_ID_CNS_CTRL: return nvme_identify_ctrl(n, req); + case NVME_ID_CNS_CS_CTRL: + return nvme_identify_ctrl_csi(n, req); case NVME_ID_CNS_NS_ACTIVE_LIST: return nvme_identify_nslist(n, req); + case NVME_ID_CNS_CS_NS_ACTIVE_LIST: + return nvme_identify_nslist_csi(n, req); case NVME_ID_CNS_NS_DESCR_LIST: return nvme_identify_ns_descr_list(n, req); + case NVME_ID_CNS_IO_COMMAND_SET: + return nvme_identify_cmd_set(n, req); default: trace_pci_nvme_err_invalid_identify_cns(le32_to_cpu(c->cns)); return NVME_INVALID_FIELD | NVME_DNR; @@ -1820,7 +1935,9 @@ defaults: if (iv =3D=3D n->admin_cq.vector) { result |=3D NVME_INTVC_NOCOALESCING; } - + break; + case NVME_COMMAND_SET_PROFILE: + result =3D 0; break; default: result =3D nvme_feature_default[fid]; @@ -1961,6 +2078,12 @@ static uint16_t nvme_set_feature(NvmeCtrl *n, NvmeRe= quest *req) break; case NVME_TIMESTAMP: return nvme_set_feature_timestamp(n, req); + case NVME_COMMAND_SET_PROFILE: + if (dw11 & 0x1ff) { + trace_pci_nvme_err_invalid_iocsci(dw11 & 0x1ff); + return NVME_CMD_SET_CMB_REJECTED | NVME_DNR; + } + break; default: return NVME_FEAT_NOT_CHANGEABLE | NVME_DNR; } @@ -2117,8 +2240,12 @@ static void nvme_select_ns_iocs(NvmeCtrl *n) continue; } ns->iocs =3D nvme_cse_iocs_none; - if (NVME_CC_CSS(n->bar.cc) !=3D NVME_CC_CSS_ADMIN_ONLY) { - ns->iocs =3D nvme_cse_iocs_nvm; + switch (ns->csi) { + case NVME_CSI_NVM: + if (NVME_CC_CSS(n->bar.cc) !=3D NVME_CC_CSS_ADMIN_ONLY) { + ns->iocs =3D nvme_cse_iocs_nvm; + } + break; } } } @@ -2859,6 +2986,7 @@ static void nvme_init_ctrl(NvmeCtrl *n, PCIDevice *pc= i_dev) NVME_CAP_SET_CQR(n->bar.cap, 1); NVME_CAP_SET_TO(n->bar.cap, 0xf); NVME_CAP_SET_CSS(n->bar.cap, NVME_CAP_CSS_NVM); + NVME_CAP_SET_CSS(n->bar.cap, NVME_CAP_CSS_CSI_SUPP); NVME_CAP_SET_CSS(n->bar.cap, NVME_CAP_CSS_ADMIN_ONLY); NVME_CAP_SET_MPSMAX(n->bar.cap, 4); =20 diff --git a/hw/block/trace-events b/hw/block/trace-events index e67e96c2b5..8b29423132 100644 --- a/hw/block/trace-events +++ b/hw/block/trace-events @@ -48,8 +48,12 @@ pci_nvme_create_cq(uint64_t addr, uint16_t cqid, uint16_= t vector, uint16_t size, pci_nvme_del_sq(uint16_t qid) "deleting submission queue sqid=3D%"PRIu16"" pci_nvme_del_cq(uint16_t cqid) "deleted completion queue, cqid=3D%"PRIu16"" pci_nvme_identify_ctrl(void) "identify controller" +pci_nvme_identify_ctrl_csi(uint8_t csi) "identify controller, csi=3D0x%"PR= Ix8"" pci_nvme_identify_ns(uint32_t ns) "nsid %"PRIu32"" +pci_nvme_identify_ns_csi(uint32_t ns, uint8_t csi) "nsid=3D%"PRIu32", csi= =3D0x%"PRIx8"" pci_nvme_identify_nslist(uint32_t ns) "nsid %"PRIu32"" +pci_nvme_identify_nslist_csi(uint16_t ns, uint8_t csi) "nsid=3D%"PRIu16", = csi=3D0x%"PRIx8"" +pci_nvme_identify_cmd_set(void) "identify i/o command set" pci_nvme_identify_ns_descr_list(uint32_t ns) "nsid %"PRIu32"" pci_nvme_get_log(uint16_t cid, uint8_t lid, uint8_t lsp, uint8_t rae, uint= 32_t len, uint64_t off) "cid %"PRIu16" lid 0x%"PRIx8" lsp 0x%"PRIx8" rae 0x= %"PRIx8" len %"PRIu32" off %"PRIu64"" pci_nvme_getfeat(uint16_t cid, uint32_t nsid, uint8_t fid, uint8_t sel, ui= nt32_t cdw11) "cid %"PRIu16" nsid 0x%"PRIx32" fid 0x%"PRIx8" sel 0x%"PRIx8"= cdw11 0x%"PRIx32"" @@ -103,6 +107,8 @@ pci_nvme_err_invalid_opc(uint8_t opc) "invalid opcode 0= x%"PRIx8"" pci_nvme_err_invalid_admin_opc(uint8_t opc) "invalid admin opcode 0x%"PRIx= 8"" pci_nvme_err_invalid_lba_range(uint64_t start, uint64_t len, uint64_t limi= t) "Invalid LBA start=3D%"PRIu64" len=3D%"PRIu64" limit=3D%"PRIu64"" pci_nvme_err_invalid_log_page_offset(uint64_t ofs, uint64_t size) "must be= <=3D %"PRIu64", got %"PRIu64"" +pci_nvme_err_only_nvm_cmd_set_avail(void) "setting 110b CC.CSS, but only N= VM command set is enabled" +pci_nvme_err_invalid_iocsci(uint32_t idx) "unsupported command set combina= tion index %"PRIu32"" pci_nvme_err_invalid_del_sq(uint16_t qid) "invalid submission queue deleti= on, sid=3D%"PRIu16"" pci_nvme_err_invalid_create_sq_cqid(uint16_t cqid) "failed creating submis= sion queue, invalid cqid=3D%"PRIu16"" pci_nvme_err_invalid_create_sq_sqid(uint16_t sqid) "failed creating submis= sion queue, invalid sqid=3D%"PRIu16"" @@ -159,6 +165,7 @@ pci_nvme_ub_db_wr_invalid_cq(uint32_t qid) "completion = queue doorbell write for pci_nvme_ub_db_wr_invalid_cqhead(uint32_t qid, uint16_t new_head) "complet= ion queue doorbell write value beyond queue size, cqid=3D%"PRIu32", new_hea= d=3D%"PRIu16", ignoring" pci_nvme_ub_db_wr_invalid_sq(uint32_t qid) "submission queue doorbell writ= e for nonexistent queue, sqid=3D%"PRIu32", ignoring" pci_nvme_ub_db_wr_invalid_sqtail(uint32_t qid, uint16_t new_tail) "submiss= ion queue doorbell write value beyond queue size, sqid=3D%"PRIu32", new_hea= d=3D%"PRIu16", ignoring" +pci_nvme_ub_unknown_css_value(void) "unknown value in cc.css field" =20 # xen-block.c xen_block_realize(const char *type, uint32_t disk, uint32_t partition) "%s= d%up%u" --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604706300; cv=none; d=zohomail.com; s=zohoarc; b=AM11TSSxnujdrAlKb0kOQKu5AuARh6PryAzecAiyB3iWm2RLymm2hIvQyw94VPmaWeTMshoDMNoPE8FnvIMzKfKCBfGJzevIyKtzWEyvr/jItd6JxbdV3jv0Y0TufauE9T0nnTGzp1yVaigKyxSwOqLumJtg0CgHlu0GOXtKbqw= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604706300; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=Vwpru4/cAGkKQrzNlEgBXd8hdnLQ6k5cjHj3gIIujmw=; b=IGloEPT0wdc9EaBaWUCVyAl7N6DD/K4zNntNiRv7YzMWjzvdL/TLOaKXRodYYzLzYFhMpNEfXksn3k4du9Wi7P4V1V7VBx2hcXAGBWfEVMhFv4EBna4Swwe1rlt3OJmg8HP4EfI4KQYXqIROTdck767E4tzsK88WVIDXYbEDcg8= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604706300925926.5305948377464; Fri, 6 Nov 2020 15:45:00 -0800 (PST) Received: from localhost ([::1]:59042 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBPf-0007En-9M for importer@patchew.org; Fri, 06 Nov 2020 18:44:59 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34270) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOH-0005Q4-Dw; Fri, 06 Nov 2020 18:43:33 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57347) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOE-0003Sk-Bs; Fri, 06 Nov 2020 18:43:33 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:29 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:20 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:25 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706211; x=1636242211; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=ioG8CtacGXPm8N0F0AXEBY1dx1hdRzuXtXfPGQ3KeOw=; b=gkXD4a7DRB/of4YkkwVYfZ7NZKailH+LF0XJ4WT2TwbXDG0eyjClXv2c QmKUx3oSmSXB1J+syKD5N7M9GdEzmQnWTra8zsH1oTCSNm7zmggJzbx7x +hWNH4wf6omdw0v6y8nLF5HPV8WbtO2JK7FVs5lM+af6YG59xx6iDw3Qq g5sDc58Ux9gPeFmGm0jldHa/DgkUlvB3v1c7l0ZE5P7myhVeIzdfzbw79 j8zbxwMKIN6MtHVi1gsGAUBAhAydCRJDq9UcBcF/7n8yUHmLdxoukDPmi 4ujAB1DySSGJqIhpHGkPRRjzSuzWsyfDgGbiH5jAeNxj4dWnh6ZbwvXJV Q==; IronPort-SDR: u6QFlePsS1HYcpY3gbs2kHZ2NhedEfaY/ZjIskY5quZuJcPJvRk/HibqYLWnZKjGbJQKCHLgYB 1V8vpifQ5x6kMNMPAwJqzkLhAnGDUXR6TRCNco7wI/tUqf269TKsCkhjGQj7tEPx3fJf/VSe8h glNZmVTmTXeeVmqyni/KC1y4Qpx0Psygz/2zVHLISjAAG4O5f/7Rh/scUhVBH+oaC9Q/dquraq IP+Cvzm3ZT5GiStod0tBktzuCmlV7wV8l0Le33cFv3o0ryAgAF9Ls8lNC8AQIcXqDeI3/jUbvn x8I= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267067" IronPort-SDR: 37pMl1OvNSBMOm/mn8fcUDGhBYb9Rz6AIDmSqR6ItMvCyJStdcFyf0J/c52YWKnOc2bZYvsfYm bvZ8dOh/g+nn2/K0INqdhj0jfdOxmJHsOoZj1RD7x9TqJa+/bPD6akioxnYaJ6sZVTerdq8KIY YsJya7/TI7+WbSV5zm6NjLpOjoemusaPkRu2Pde46Uw/AB06nRxNjwA4UsWT8wuNy1FAdKL8vp be3EWLWiIjDhZiTDGcFhSSEW26F+LRdlq17PGIAIwxJzpCCMxjj26kbXx8ydZko1xIg17xJn3t wxu8M6DVYTY7GVyrVeeLseH4 IronPort-SDR: o+e0cQrqoKkmr/p3NqPK4aMCseyJo4zcUdCzJ6MQus6K22rZa9tBMf8egN9GbxtfE83ZyoC9QG eWSp41xvpj4DrfumZj3qU8JVNW7xyJpJvhE7jQJkhLUCVrHn/0w0vn3IUx04O7lWYRA0uKnUUZ dkLm8CmjcfMgACde3P7FBl+8hFcwclAa+EZ+4GHeGlQg9Ety3pZYtE7jV0Umu0wsHcjFo/Kobo vMlkk7DLJ3/zyqtwzDxkC+S3kpSTqOUWyD+zdl5oKWGKY9vjElSGDIWErqzP6NOepZ+sEHG8Wx euk= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 06/12] hw/block/nvme: Support allocated CNS command variants Date: Sat, 7 Nov 2020 08:42:59 +0900 Message-Id: <20201106234305.21339-7-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" From: Niklas Cassel Many CNS commands have "allocated" command variants. These include a namespace as long as it is allocated, that is a namespace is included regardless if it is active (attached) or not. While these commands are optional (they are mandatory for controllers supporting the namespace attachment command), our QEMU implementation is more complete by actually providing support for these CNS values. However, since our QEMU model currently does not support the namespace attachment command, these new allocated CNS commands will return the same result as the active CNS command variants. In NVMe, a namespace is active if it exists and is attached to the controller. Add a new Boolean namespace flag, "attached", to provide the most basic namespace attachment support. The default value for this new flag is true. Also, implement the logic in the new CNS values to include/exclude namespaces based on this new property. The only thing missing is hooking up the actual Namespace Attachment command opcode, which will allow a user to toggle the "attached" flag per namespace. The reason for not hooking up this command completely is because the NVMe specification requires the namespace management command to be supported if the namespace attachment command is supported. Signed-off-by: Niklas Cassel Signed-off-by: Dmitry Fomichev Reviewed-by: Keith Busch --- hw/block/nvme-ns.h | 1 + include/block/nvme.h | 20 +++++++++++-------- hw/block/nvme-ns.c | 1 + hw/block/nvme.c | 46 +++++++++++++++++++++++++++++++++----------- 4 files changed, 49 insertions(+), 19 deletions(-) diff --git a/hw/block/nvme-ns.h b/hw/block/nvme-ns.h index d795e44bab..2d9cd29d07 100644 --- a/hw/block/nvme-ns.h +++ b/hw/block/nvme-ns.h @@ -31,6 +31,7 @@ typedef struct NvmeNamespace { int64_t size; NvmeIdNs id_ns; const uint32_t *iocs; + bool attached; uint8_t csi; =20 NvmeNamespaceParams params; diff --git a/include/block/nvme.h b/include/block/nvme.h index af23514713..394db19022 100644 --- a/include/block/nvme.h +++ b/include/block/nvme.h @@ -806,14 +806,18 @@ typedef struct QEMU_PACKED NvmePSD { #define NVME_IDENTIFY_DATA_SIZE 4096 =20 enum NvmeIdCns { - NVME_ID_CNS_NS =3D 0x00, - NVME_ID_CNS_CTRL =3D 0x01, - NVME_ID_CNS_NS_ACTIVE_LIST =3D 0x02, - NVME_ID_CNS_NS_DESCR_LIST =3D 0x03, - NVME_ID_CNS_CS_NS =3D 0x05, - NVME_ID_CNS_CS_CTRL =3D 0x06, - NVME_ID_CNS_CS_NS_ACTIVE_LIST =3D 0x07, - NVME_ID_CNS_IO_COMMAND_SET =3D 0x1c, + NVME_ID_CNS_NS =3D 0x00, + NVME_ID_CNS_CTRL =3D 0x01, + NVME_ID_CNS_NS_ACTIVE_LIST =3D 0x02, + NVME_ID_CNS_NS_DESCR_LIST =3D 0x03, + NVME_ID_CNS_CS_NS =3D 0x05, + NVME_ID_CNS_CS_CTRL =3D 0x06, + NVME_ID_CNS_CS_NS_ACTIVE_LIST =3D 0x07, + NVME_ID_CNS_NS_PRESENT_LIST =3D 0x10, + NVME_ID_CNS_NS_PRESENT =3D 0x11, + NVME_ID_CNS_CS_NS_PRESENT_LIST =3D 0x1a, + NVME_ID_CNS_CS_NS_PRESENT =3D 0x1b, + NVME_ID_CNS_IO_COMMAND_SET =3D 0x1c, }; =20 typedef struct QEMU_PACKED NvmeIdCtrl { diff --git a/hw/block/nvme-ns.c b/hw/block/nvme-ns.c index c0362426cc..e191ef9be0 100644 --- a/hw/block/nvme-ns.c +++ b/hw/block/nvme-ns.c @@ -42,6 +42,7 @@ static void nvme_ns_init(NvmeNamespace *ns) id_ns->nsze =3D cpu_to_le64(nvme_ns_nlbas(ns)); =20 ns->csi =3D NVME_CSI_NVM; + ns->attached =3D true; =20 /* no thin provisioning */ id_ns->ncap =3D id_ns->nsze; diff --git a/hw/block/nvme.c b/hw/block/nvme.c index bb82dd9975..7495cdb5ef 100644 --- a/hw/block/nvme.c +++ b/hw/block/nvme.c @@ -1236,6 +1236,7 @@ static uint16_t nvme_smart_info(NvmeCtrl *n, uint8_t = rae, uint32_t buf_len, uint32_t trans_len; NvmeNamespace *ns; time_t current_ms; + int i; =20 if (off >=3D sizeof(smart)) { return NVME_INVALID_FIELD | NVME_DNR; @@ -1246,10 +1247,7 @@ static uint16_t nvme_smart_info(NvmeCtrl *n, uint8_t= rae, uint32_t buf_len, if (!ns) { return NVME_INVALID_NSID | NVME_DNR; } - nvme_set_blk_stats(ns, &stats); } else { - int i; - for (i =3D 1; i <=3D n->num_namespaces; i++) { ns =3D nvme_ns(n, i); if (!ns) { @@ -1552,7 +1550,8 @@ static uint16_t nvme_identify_ctrl_csi(NvmeCtrl *n, N= vmeRequest *req) return NVME_INVALID_FIELD | NVME_DNR; } =20 -static uint16_t nvme_identify_ns(NvmeCtrl *n, NvmeRequest *req) +static uint16_t nvme_identify_ns(NvmeCtrl *n, NvmeRequest *req, + bool only_active) { NvmeNamespace *ns; NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; @@ -1569,11 +1568,16 @@ static uint16_t nvme_identify_ns(NvmeCtrl *n, NvmeR= equest *req) return nvme_rpt_empty_id_struct(n, req); } =20 + if (only_active && !ns->attached) { + return nvme_rpt_empty_id_struct(n, req); + } + return nvme_dma(n, (uint8_t *)&ns->id_ns, sizeof(NvmeIdNs), DMA_DIRECTION_FROM_DEVICE, req); } =20 -static uint16_t nvme_identify_ns_csi(NvmeCtrl *n, NvmeRequest *req) +static uint16_t nvme_identify_ns_csi(NvmeCtrl *n, NvmeRequest *req, + bool only_active) { NvmeNamespace *ns; NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; @@ -1590,6 +1594,10 @@ static uint16_t nvme_identify_ns_csi(NvmeCtrl *n, Nv= meRequest *req) return nvme_rpt_empty_id_struct(n, req); } =20 + if (only_active && !ns->attached) { + return nvme_rpt_empty_id_struct(n, req); + } + if (c->csi =3D=3D NVME_CSI_NVM) { return nvme_rpt_empty_id_struct(n, req); } @@ -1597,7 +1605,8 @@ static uint16_t nvme_identify_ns_csi(NvmeCtrl *n, Nvm= eRequest *req) return NVME_INVALID_FIELD | NVME_DNR; } =20 -static uint16_t nvme_identify_nslist(NvmeCtrl *n, NvmeRequest *req) +static uint16_t nvme_identify_nslist(NvmeCtrl *n, NvmeRequest *req, + bool only_active) { NvmeNamespace *ns; NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; @@ -1627,6 +1636,9 @@ static uint16_t nvme_identify_nslist(NvmeCtrl *n, Nvm= eRequest *req) if (ns->params.nsid <=3D min_nsid) { continue; } + if (only_active && !ns->attached) { + continue; + } list_ptr[j++] =3D cpu_to_le32(ns->params.nsid); if (j =3D=3D data_len / sizeof(uint32_t)) { break; @@ -1636,7 +1648,8 @@ static uint16_t nvme_identify_nslist(NvmeCtrl *n, Nvm= eRequest *req) return nvme_dma(n, list, data_len, DMA_DIRECTION_FROM_DEVICE, req); } =20 -static uint16_t nvme_identify_nslist_csi(NvmeCtrl *n, NvmeRequest *req) +static uint16_t nvme_identify_nslist_csi(NvmeCtrl *n, NvmeRequest *req, + bool only_active) { NvmeNamespace *ns; NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; @@ -1667,6 +1680,9 @@ static uint16_t nvme_identify_nslist_csi(NvmeCtrl *n,= NvmeRequest *req) if (ns->params.nsid <=3D min_nsid) { continue; } + if (only_active && !ns->attached) { + continue; + } list_ptr[j++] =3D cpu_to_le32(ns->params.nsid); if (j =3D=3D data_len / sizeof(uint32_t)) { break; @@ -1740,17 +1756,25 @@ static uint16_t nvme_identify(NvmeCtrl *n, NvmeRequ= est *req) =20 switch (le32_to_cpu(c->cns)) { case NVME_ID_CNS_NS: - return nvme_identify_ns(n, req); + /* fall through */ + case NVME_ID_CNS_NS_PRESENT: + return nvme_identify_ns(n, req, true); case NVME_ID_CNS_CS_NS: - return nvme_identify_ns_csi(n, req); + /* fall through */ + case NVME_ID_CNS_CS_NS_PRESENT: + return nvme_identify_ns_csi(n, req, true); case NVME_ID_CNS_CTRL: return nvme_identify_ctrl(n, req); case NVME_ID_CNS_CS_CTRL: return nvme_identify_ctrl_csi(n, req); case NVME_ID_CNS_NS_ACTIVE_LIST: - return nvme_identify_nslist(n, req); + /* fall through */ + case NVME_ID_CNS_NS_PRESENT_LIST: + return nvme_identify_nslist(n, req, true); case NVME_ID_CNS_CS_NS_ACTIVE_LIST: - return nvme_identify_nslist_csi(n, req); + /* fall through */ + case NVME_ID_CNS_CS_NS_PRESENT_LIST: + return nvme_identify_nslist_csi(n, req, true); case NVME_ID_CNS_NS_DESCR_LIST: return nvme_identify_ns_descr_list(n, req); case NVME_ID_CNS_IO_COMMAND_SET: --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604706900; cv=none; d=zohomail.com; s=zohoarc; b=PSwQCBxcOHuWbJlh8QHMVteh6f3NYwifsSC6dbLQ5yyjCI0ueinsZqiQTUaQZwWZ62ejx9FzBuQx3garIUgIBxovD2PL/2br7GHk+3bZJ4HgoPkKXG6etoqgjWTo2dDzQa0N+wPCMb3xv5KbIXvVt8aOzqGjk3/hR4AhGO7f2Vc= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604706900; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=Ac4zpZMjP+mnpF2qxYP0XMJ3wxLP1CdPdveCj1aUwW4=; b=mZ8G+96pAVOs1uzwWN5aB56mNflLvOhT8bN172fc52hXwH3l1+xc4MZPovdiUCe07BRDVnTd3s0xeUdLzibLNHdUbSQcHB5wXrEmNxiu8LO0qHJCTBlI3O/v0aljlB78V2/q1Erf8yE4EbxB2bvCC6fUB1QFsB2QIfAnyxIpKic= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604706900024624.5223361236979; Fri, 6 Nov 2020 15:55:00 -0800 (PST) Received: from localhost ([::1]:56556 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBZK-0001M4-Vs for importer@patchew.org; Fri, 06 Nov 2020 18:54:59 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34282) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOI-0005SA-CI; Fri, 06 Nov 2020 18:43:34 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57360) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOG-0003Tq-BK; Fri, 06 Nov 2020 18:43:34 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:31 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:22 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:28 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706213; x=1636242213; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=g063222WlgUGzbpsQdE8ESMz45RVE42CPVFav5X88Us=; b=RiAy3J0c0dC7az2FYMAGsZTX/OKGaV0IrvONIGAsMtfJTbNFtvOd0uPR 8VRNl2DKzi0ND3cGxVZvStcwRANMGoPV2DmZeNPOb0LSkMtOdGMKsEWVj HI3QJXBQtlJ7krddekixobnCHWv25A18KPIMN9y70nMw6XJZN9ZpWNcCN 6lxNohEeP+xp+84b1W5LVt2zsSji+uEp+C2F09MJJ/VErBju5E/J+Ploy ohIz5Z7E+qRBlyRyXmXB4p7RpgCcWAsEtfb1MaiRuT98sml9kqn1qvMWf 1fsQeodnVsRblf2ziqgPde2VUfzLuPlsUd/hsF8FZzimMQhGxzP7MjzXg A==; IronPort-SDR: ewtCoAUuunRnAMMQ2l9Ozj8HLFi3ZuScEqUWOXZcQO3Mh/pddoae8w00SDad3UC6j/sTbwL+gJ 4jtyUOsyaMXw1rk+9d79h3uO2bezqSi7FI2p3OUClZO338NiMJdmE9bOfahInF1Ij4A77Jmgpk NjLD2M0ACe1T82F4B9c5qS1LqhN4lgLHUUfTNjHdMoRQsQW67YPIp8EwaqbAlHt7etmuFH1mNi JvfxRVVdlE3E3dfUKYRDJymumnRzX5Pkdf/L616Gky9jMV4feSx6wZ+6Hsy+Y0Q+WFTZ7o1qqC aQY= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267071" IronPort-SDR: 55o6HdOnW4Vv/pQrCMmRc0LSt0bLVUfySDVvDgOtT9UjCjNaBvIZDdA0Ij/1QbCRuvxsWJI+Et Yj7Z4vPFRUXwq39iY4tNQVqvetlEM/y+v6fqvhu9oZFUCZMULQEm4DkvnFyi6OXPN7fkN1D+fp K2O4sG/vcHrSPybIowgxR+ONfvulnDc8v/ZNMDpHbXXRRKd9dkZcO5843vbLw3U1TmdZI6qspg f1Aiobw9+3Mupsw5mFKsj+ErDdebDiis3XvmnPfuddP47CBG+qsCYrN69/sZpO7m+sS8+aB+3R Ymg4EqwTM9dN7rwZWR+jt1gK IronPort-SDR: wxDKU7jQOtvn0suT9acN7pJP29WOvGxiLY7MrNbCLrmM1WzN72iJAlL/qs2IpuAIIbKxOiyULo TB5PSdpM/4JkcQvt4sbLn3tUaYimpVZPkp8rmPSJEuv25yTS3V/NrC2goEpXr9XtjrkT6PuNWo l0szem1Dl8kSNR9ECIIcRFq932ce6DtEdWarIuv1zIdOVBj4pPV1wmnsyrMdyN6n2srs3jWgTU VUovQpzraIsvvRSYXFnkOpZ8iqwG648MYsAlUkIpmwoPq8eGO91uERpKWtI1hpbciBYFH1vFAz cEQ= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 07/12] block/nvme: Make ZNS-related definitions Date: Sat, 7 Nov 2020 08:43:00 +0900 Message-Id: <20201106234305.21339-8-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" Define values and structures that are needed to support Zoned Namespace Command Set (NVMe TP 4053). Signed-off-by: Dmitry Fomichev --- include/block/nvme.h | 114 ++++++++++++++++++++++++++++++++++++++++++- 1 file changed, 113 insertions(+), 1 deletion(-) diff --git a/include/block/nvme.h b/include/block/nvme.h index 394db19022..752623b4f9 100644 --- a/include/block/nvme.h +++ b/include/block/nvme.h @@ -489,6 +489,9 @@ enum NvmeIoCommands { NVME_CMD_COMPARE =3D 0x05, NVME_CMD_WRITE_ZEROES =3D 0x08, NVME_CMD_DSM =3D 0x09, + NVME_CMD_ZONE_MGMT_SEND =3D 0x79, + NVME_CMD_ZONE_MGMT_RECV =3D 0x7a, + NVME_CMD_ZONE_APPEND =3D 0x7d, }; =20 typedef struct QEMU_PACKED NvmeDeleteQ { @@ -648,9 +651,13 @@ typedef struct QEMU_PACKED NvmeAerResult { uint8_t resv; } NvmeAerResult; =20 +typedef struct QEMU_PACKED NvmeZonedResult { + uint64_t slba; +} NvmeZonedResult; + typedef struct QEMU_PACKED NvmeCqe { uint32_t result; - uint32_t rsvd; + uint32_t dw1; uint16_t sq_head; uint16_t sq_id; uint16_t cid; @@ -679,6 +686,7 @@ enum NvmeStatusCodes { NVME_INVALID_USE_OF_CMB =3D 0x0012, NVME_INVALID_PRP_OFFSET =3D 0x0013, NVME_CMD_SET_CMB_REJECTED =3D 0x002b, + NVME_INVALID_CMD_SET =3D 0x002c, NVME_LBA_RANGE =3D 0x0080, NVME_CAP_EXCEEDED =3D 0x0081, NVME_NS_NOT_READY =3D 0x0082, @@ -703,6 +711,14 @@ enum NvmeStatusCodes { NVME_CONFLICTING_ATTRS =3D 0x0180, NVME_INVALID_PROT_INFO =3D 0x0181, NVME_WRITE_TO_RO =3D 0x0182, + NVME_ZONE_BOUNDARY_ERROR =3D 0x01b8, + NVME_ZONE_FULL =3D 0x01b9, + NVME_ZONE_READ_ONLY =3D 0x01ba, + NVME_ZONE_OFFLINE =3D 0x01bb, + NVME_ZONE_INVALID_WRITE =3D 0x01bc, + NVME_ZONE_TOO_MANY_ACTIVE =3D 0x01bd, + NVME_ZONE_TOO_MANY_OPEN =3D 0x01be, + NVME_ZONE_INVAL_TRANSITION =3D 0x01bf, NVME_WRITE_FAULT =3D 0x0280, NVME_UNRECOVERED_READ =3D 0x0281, NVME_E2E_GUARD_ERROR =3D 0x0282, @@ -887,6 +903,11 @@ typedef struct QEMU_PACKED NvmeIdCtrl { uint8_t vs[1024]; } NvmeIdCtrl; =20 +typedef struct NvmeIdCtrlZoned { + uint8_t zasl; + uint8_t rsvd1[4095]; +} NvmeIdCtrlZoned; + enum NvmeIdCtrlOacs { NVME_OACS_SECURITY =3D 1 << 0, NVME_OACS_FORMAT =3D 1 << 1, @@ -1012,6 +1033,12 @@ typedef struct QEMU_PACKED NvmeLBAF { uint8_t rp; } NvmeLBAF; =20 +typedef struct QEMU_PACKED NvmeLBAFE { + uint64_t zsze; + uint8_t zdes; + uint8_t rsvd9[7]; +} NvmeLBAFE; + #define NVME_NSID_BROADCAST 0xffffffff =20 typedef struct QEMU_PACKED NvmeIdNs { @@ -1066,10 +1093,24 @@ enum NvmeNsIdentifierType { =20 enum NvmeCsi { NVME_CSI_NVM =3D 0x00, + NVME_CSI_ZONED =3D 0x02, }; =20 #define NVME_SET_CSI(vec, csi) (vec |=3D (uint8_t)(1 << (csi))) =20 +typedef struct QEMU_PACKED NvmeIdNsZoned { + uint16_t zoc; + uint16_t ozcs; + uint32_t mar; + uint32_t mor; + uint32_t rrl; + uint32_t frl; + uint8_t rsvd20[2796]; + NvmeLBAFE lbafe[16]; + uint8_t rsvd3072[768]; + uint8_t vs[256]; +} NvmeIdNsZoned; + /*Deallocate Logical Block Features*/ #define NVME_ID_NS_DLFEAT_GUARD_CRC(dlfeat) ((dlfeat) & 0x10) #define NVME_ID_NS_DLFEAT_WRITE_ZEROES(dlfeat) ((dlfeat) & 0x08) @@ -1101,10 +1142,76 @@ enum NvmeIdNsDps { DPS_FIRST_EIGHT =3D 8, }; =20 +enum NvmeZoneAttr { + NVME_ZA_FINISHED_BY_CTLR =3D 1 << 0, + NVME_ZA_FINISH_RECOMMENDED =3D 1 << 1, + NVME_ZA_RESET_RECOMMENDED =3D 1 << 2, + NVME_ZA_ZD_EXT_VALID =3D 1 << 7, +}; + +typedef struct QEMU_PACKED NvmeZoneReportHeader { + uint64_t nr_zones; + uint8_t rsvd[56]; +} NvmeZoneReportHeader; + +enum NvmeZoneReceiveAction { + NVME_ZONE_REPORT =3D 0, + NVME_ZONE_REPORT_EXTENDED =3D 1, +}; + +enum NvmeZoneReportType { + NVME_ZONE_REPORT_ALL =3D 0, + NVME_ZONE_REPORT_EMPTY =3D 1, + NVME_ZONE_REPORT_IMPLICITLY_OPEN =3D 2, + NVME_ZONE_REPORT_EXPLICITLY_OPEN =3D 3, + NVME_ZONE_REPORT_CLOSED =3D 4, + NVME_ZONE_REPORT_FULL =3D 5, + NVME_ZONE_REPORT_READ_ONLY =3D 6, + NVME_ZONE_REPORT_OFFLINE =3D 7, +}; + +enum NvmeZoneType { + NVME_ZONE_TYPE_RESERVED =3D 0x00, + NVME_ZONE_TYPE_SEQ_WRITE =3D 0x02, +}; + +enum NvmeZoneSendAction { + NVME_ZONE_ACTION_RSD =3D 0x00, + NVME_ZONE_ACTION_CLOSE =3D 0x01, + NVME_ZONE_ACTION_FINISH =3D 0x02, + NVME_ZONE_ACTION_OPEN =3D 0x03, + NVME_ZONE_ACTION_RESET =3D 0x04, + NVME_ZONE_ACTION_OFFLINE =3D 0x05, + NVME_ZONE_ACTION_SET_ZD_EXT =3D 0x10, +}; + +typedef struct QEMU_PACKED NvmeZoneDescr { + uint8_t zt; + uint8_t zs; + uint8_t za; + uint8_t rsvd3[5]; + uint64_t zcap; + uint64_t zslba; + uint64_t wp; + uint8_t rsvd32[32]; +} NvmeZoneDescr; + +enum NvmeZoneState { + NVME_ZONE_STATE_RESERVED =3D 0x00, + NVME_ZONE_STATE_EMPTY =3D 0x01, + NVME_ZONE_STATE_IMPLICITLY_OPEN =3D 0x02, + NVME_ZONE_STATE_EXPLICITLY_OPEN =3D 0x03, + NVME_ZONE_STATE_CLOSED =3D 0x04, + NVME_ZONE_STATE_READ_ONLY =3D 0x0D, + NVME_ZONE_STATE_FULL =3D 0x0E, + NVME_ZONE_STATE_OFFLINE =3D 0x0F, +}; + static inline void _nvme_check_size(void) { QEMU_BUILD_BUG_ON(sizeof(NvmeBar) !=3D 4096); QEMU_BUILD_BUG_ON(sizeof(NvmeAerResult) !=3D 4); + QEMU_BUILD_BUG_ON(sizeof(NvmeZonedResult) !=3D 8); QEMU_BUILD_BUG_ON(sizeof(NvmeCqe) !=3D 16); QEMU_BUILD_BUG_ON(sizeof(NvmeDsmRange) !=3D 16); QEMU_BUILD_BUG_ON(sizeof(NvmeCmd) !=3D 64); @@ -1120,8 +1227,13 @@ static inline void _nvme_check_size(void) QEMU_BUILD_BUG_ON(sizeof(NvmeSmartLog) !=3D 512); QEMU_BUILD_BUG_ON(sizeof(NvmeEffectsLog) !=3D 4096); QEMU_BUILD_BUG_ON(sizeof(NvmeIdCtrl) !=3D 4096); + QEMU_BUILD_BUG_ON(sizeof(NvmeIdCtrlZoned) !=3D 4096); QEMU_BUILD_BUG_ON(sizeof(NvmeIdNsDescr) !=3D 4); + QEMU_BUILD_BUG_ON(sizeof(NvmeLBAF) !=3D 4); + QEMU_BUILD_BUG_ON(sizeof(NvmeLBAFE) !=3D 16); QEMU_BUILD_BUG_ON(sizeof(NvmeIdNs) !=3D 4096); + QEMU_BUILD_BUG_ON(sizeof(NvmeIdNsZoned) !=3D 4096); QEMU_BUILD_BUG_ON(sizeof(NvmeSglDescriptor) !=3D 16); + QEMU_BUILD_BUG_ON(sizeof(NvmeZoneDescr) !=3D 64); } #endif --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604707022; cv=none; d=zohomail.com; s=zohoarc; b=Rj0JM6srMFzOqGv2/G7J8IAaP5jCu8WeeVYpBiG2ONhdarr9rpta3BTwZMlahPU6SCqCJ6OFv3r2plkllC8UhxvSU8Xr+piybfFvfHNan/WD+yk7TomCN4YqeA28k0Jv9F9nkQTjaoK/L2Z8RZiq0rM4pVfL/rGTH5nhDgYR4dw= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604707022; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=j42b9q/dpLN39XziMJhqvnHjKrJt8aCYi3MNALBZEds=; b=TaArNEAeIFsR1eC71h47hsw3MnAdMA6o+Y7qc58iLKifetKkkXX5lZ60v1nwFCNAMLzNvX0p3qK26PvreLZ5iggY0Eh1STFhALMwLau2i+66RlFTqMlPP2NHyeVFbnnxmgEstBDDe3XQaehOolru6iE5cVsUksq2gA+QvSQlBgY= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604707022176864.6197193840482; Fri, 6 Nov 2020 15:57:02 -0800 (PST) Received: from localhost ([::1]:60166 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBbI-00030Z-Sz for importer@patchew.org; Fri, 06 Nov 2020 18:57:00 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34320) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBON-0005hX-Ku; Fri, 06 Nov 2020 18:43:39 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57360) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOJ-0003Tq-1M; Fri, 06 Nov 2020 18:43:39 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:34 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:25 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:31 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706215; x=1636242215; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=hXh5KueJNF5HWuYGtdYQ9Q1DbHpFoGs6JEU/YA2VpWA=; b=o73CmhcmxpAjP1mLpKik42P1jM9VM0+V7yU+KPMwL4JMqNlFNeE7Fah8 OOYQeXC1oljGY3qi31VLTyBH/74etqkArQ2z7pPPUIQS4MrKIXIOPp6h5 l34pIwCEItbCwHJ02XWr1XSWaFbiXPkjLmfJufc+d6QolY/iLvkdM+8ef b4zz4JZqvuxlCkCRFyHTOI3bVZfyQDvziXMPVZ33mluqPPwbirZkAISw7 nAxlr2jb3KYHHfdFQ+uo0FAeuKmOQ+afmySKqLcVO35b4xGjTWC69a+T7 QTFIZINiGPuutLU2m9PPtbzZJcTLzQL+Bg8OX9TXG6F/HNnIARaiVLDbh w==; IronPort-SDR: zy5FxdO+H+XnbUd+1fLMemHwGF8ULWO8DHHlAXVXPRfuscbxBraTDR0Dq/9rcDnUt0LWgj0pAq d9RQ94GIJJ26oBlpu0WnQKLSrV0p507rVPbm6bUZ0FySXWSN4sSEIjJBAK+OD9gRsrjqXDgoZ4 yHdbU8zMXe+elf5BIMcrpnsu6BabEZ1ddQ21yLa5Y+FasqCBHjZvIFo2XTBjSjWFsvdJxBYy3O D4tFxct4GNWKIg+ms0zf5UdfoZSXbeQ0uaASZexuGw3NTIxJJqRu+wvDpgCyIYTOp3mAsE5nJR xFM= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267072" IronPort-SDR: aqaozhyn+1/wqTGYHx3SewvILlznwswBoM8hg6rKBCGEvkESor5ltOyzuYDesKxO1GgYbCMsXG vvUeG/1n3GTr5FEgY7y8JRX6l19bRQXEFeZM+aHa7rTPyg6vqykMH8yfg5JRpnDNVUDeH7NKuY E44V6pg5ncZiIBquPnIFivevNsA8jDoMVAFYUdpA3XbgnAQiWDGARRZiqwW+3aP1DTRbnuoIGE ++v2ced/ffoZCOgltoK2QPdQVal+KRUXLU0jP1AQmQEYmsrC8WXzH6TTYB4N71oSVL7Xv6Bq27 s3G8Pd6pmv3DbyliFRSWCZSn IronPort-SDR: JjaGHv/ASmDRjvL1kowI9f060/+AJ7j5Y+QQqplKvWahclXq9Aw+ojhjHw6uma86WBXoRt+gxt uTPTXAM1sfK3TErocBKDbUQG4e+cCif85Z6nQkw9hjSsSeRv6plW9c6A7kJpINHal7vFxEd2Kn GhnNrdkQPJpSxhT1D/qgCBU1UbD92qz4mZJCsoIFhgH9H+Slmco9ZFhKnWvwBW1jWmubwODdW3 e3bkRA7orEK8IFPfwGBrvykxBrmDu0IYIT5cASgRb98EP0ZMgwOJqP8+HqVg+zfIQxazHAo99I /Ls= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 08/12] hw/block/nvme: Support Zoned Namespace Command Set Date: Sat, 7 Nov 2020 08:43:01 +0900 Message-Id: <20201106234305.21339-9-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" The emulation code has been changed to advertise NVM Command Set when "zoned" device property is not set (default) and Zoned Namespace Command Set otherwise. Define values and structures that are needed to support Zoned Namespace Command Set (NVMe TP 4053) in PCI NVMe controller emulator. Define trace events where needed in newly introduced code. In order to improve scalability, all open, closed and full zones are organized in separate linked lists. Consequently, almost all zone operations don't require scanning of the entire zone array (which potentially can be quite large) - it is only necessary to enumerate one or more zone lists. Handlers for three new NVMe commands introduced in Zoned Namespace Command Set specification are added, namely for Zone Management Receive, Zone Management Send and Zone Append. Device initialization code has been extended to create a proper configuration for zoned operation using device properties. Read/Write command handler is modified to only allow writes at the write pointer if the namespace is zoned. For Zone Append command, writes implicitly happen at the write pointer and the starting write pointer value is returned as the result of the command. Write Zeroes handler is modified to add zoned checks that are identical to those done as a part of Write flow. Subsequent commits in this series add ZDE support and checks for active and open zone limits. Signed-off-by: Niklas Cassel Signed-off-by: Hans Holmberg Signed-off-by: Ajay Joshi Signed-off-by: Chaitanya Kulkarni Signed-off-by: Matias Bjorling Signed-off-by: Aravind Ramesh Signed-off-by: Shin'ichiro Kawasaki Signed-off-by: Adam Manzanares Signed-off-by: Dmitry Fomichev Reviewed-by: Niklas Cassel --- hw/block/nvme-ns.h | 54 +++ hw/block/nvme.h | 8 + hw/block/nvme-ns.c | 173 ++++++++ hw/block/nvme.c | 972 +++++++++++++++++++++++++++++++++++++++++- hw/block/trace-events | 18 +- 5 files changed, 1210 insertions(+), 15 deletions(-) diff --git a/hw/block/nvme-ns.h b/hw/block/nvme-ns.h index 2d9cd29d07..d2631ff5a3 100644 --- a/hw/block/nvme-ns.h +++ b/hw/block/nvme-ns.h @@ -19,9 +19,20 @@ #define NVME_NS(obj) \ OBJECT_CHECK(NvmeNamespace, (obj), TYPE_NVME_NS) =20 +typedef struct NvmeZone { + NvmeZoneDescr d; + uint64_t w_ptr; + QTAILQ_ENTRY(NvmeZone) entry; +} NvmeZone; + typedef struct NvmeNamespaceParams { uint32_t nsid; QemuUUID uuid; + + bool zoned; + bool cross_zone_read; + uint64_t zone_size_bs; + uint64_t zone_cap_bs; } NvmeNamespaceParams; =20 typedef struct NvmeNamespace { @@ -34,6 +45,18 @@ typedef struct NvmeNamespace { bool attached; uint8_t csi; =20 + NvmeIdNsZoned *id_ns_zoned; + NvmeZone *zone_array; + QTAILQ_HEAD(, NvmeZone) exp_open_zones; + QTAILQ_HEAD(, NvmeZone) imp_open_zones; + QTAILQ_HEAD(, NvmeZone) closed_zones; + QTAILQ_HEAD(, NvmeZone) full_zones; + uint32_t num_zones; + uint64_t zone_size; + uint64_t zone_capacity; + uint64_t zone_array_size; + uint32_t zone_size_log2; + NvmeNamespaceParams params; } NvmeNamespace; =20 @@ -71,8 +94,39 @@ static inline size_t nvme_l2b(NvmeNamespace *ns, uint64_= t lba) =20 typedef struct NvmeCtrl NvmeCtrl; =20 +static inline uint8_t nvme_get_zone_state(NvmeZone *zone) +{ + return zone->d.zs >> 4; +} + +static inline void nvme_set_zone_state(NvmeZone *zone, enum NvmeZoneState = state) +{ + zone->d.zs =3D state << 4; +} + +static inline uint64_t nvme_zone_rd_boundary(NvmeNamespace *ns, NvmeZone *= zone) +{ + return zone->d.zslba + ns->zone_size; +} + +static inline uint64_t nvme_zone_wr_boundary(NvmeZone *zone) +{ + return zone->d.zslba + zone->d.zcap; +} + +static inline bool nvme_wp_is_valid(NvmeZone *zone) +{ + uint8_t st =3D nvme_get_zone_state(zone); + + return st !=3D NVME_ZONE_STATE_FULL && + st !=3D NVME_ZONE_STATE_READ_ONLY && + st !=3D NVME_ZONE_STATE_OFFLINE; +} + int nvme_ns_setup(NvmeCtrl *n, NvmeNamespace *ns, Error **errp); void nvme_ns_drain(NvmeNamespace *ns); void nvme_ns_flush(NvmeNamespace *ns); +void nvme_ns_shutdown(NvmeNamespace *ns); +void nvme_ns_cleanup(NvmeNamespace *ns); =20 #endif /* NVME_NS_H */ diff --git a/hw/block/nvme.h b/hw/block/nvme.h index e080a2318a..4cb0615128 100644 --- a/hw/block/nvme.h +++ b/hw/block/nvme.h @@ -6,6 +6,9 @@ =20 #define NVME_MAX_NAMESPACES 256 =20 +#define NVME_DEFAULT_ZONE_SIZE (128 * MiB) +#define NVME_DEFAULT_MAX_ZA_SIZE (128 * KiB) + typedef struct NvmeParams { char *serial; uint32_t num_queues; /* deprecated since 5.1 */ @@ -16,6 +19,7 @@ typedef struct NvmeParams { uint32_t aer_max_queued; uint8_t mdts; bool use_intel_id; + uint32_t zasl_bs; } NvmeParams; =20 typedef struct NvmeAsyncEvent { @@ -28,6 +32,8 @@ typedef struct NvmeRequest { struct NvmeNamespace *ns; BlockAIOCB *aiocb; uint16_t status; + uint64_t fill_off; + uint32_t fill_len; NvmeCqe cqe; NvmeCmd cmd; BlockAcctCookie acct; @@ -147,6 +153,8 @@ typedef struct NvmeCtrl { QTAILQ_HEAD(, NvmeAsyncEvent) aer_queue; int aer_queued; =20 + uint8_t zasl; + NvmeNamespace namespace; NvmeNamespace *namespaces[NVME_MAX_NAMESPACES]; NvmeSQueue **sq; diff --git a/hw/block/nvme-ns.c b/hw/block/nvme-ns.c index e191ef9be0..e6db7f7d3b 100644 --- a/hw/block/nvme-ns.c +++ b/hw/block/nvme-ns.c @@ -25,6 +25,7 @@ #include "hw/qdev-properties.h" #include "hw/qdev-core.h" =20 +#include "trace.h" #include "nvme.h" #include "nvme-ns.h" =20 @@ -77,6 +78,151 @@ static int nvme_ns_init_blk(NvmeCtrl *n, NvmeNamespace = *ns, Error **errp) return 0; } =20 +static int nvme_calc_zone_geometry(NvmeNamespace *ns, Error **errp) +{ + uint64_t zone_size, zone_cap; + uint32_t nz, lbasz =3D ns->blkconf.logical_block_size; + + if (ns->params.zone_size_bs) { + zone_size =3D ns->params.zone_size_bs; + } else { + zone_size =3D NVME_DEFAULT_ZONE_SIZE; + } + if (ns->params.zone_cap_bs) { + zone_cap =3D ns->params.zone_cap_bs; + } else { + zone_cap =3D zone_size; + } + if (zone_cap > zone_size) { + error_setg(errp, "zone capacity %luB exceeds zone size %luB", + zone_cap, zone_size); + return -1; + } + if (zone_size < lbasz) { + error_setg(errp, "zone size %luB too small, must be at least %uB", + zone_size, lbasz); + return -1; + } + if (zone_cap < lbasz) { + error_setg(errp, "zone capacity %luB too small, must be at least %= uB", + zone_cap, lbasz); + return -1; + } + ns->zone_size =3D zone_size / lbasz; + ns->zone_capacity =3D zone_cap / lbasz; + + nz =3D DIV_ROUND_UP(ns->size / lbasz, ns->zone_size); + ns->num_zones =3D nz; + ns->zone_array_size =3D sizeof(NvmeZone) * nz; + ns->zone_size_log2 =3D 0; + if (is_power_of_2(ns->zone_size)) { + ns->zone_size_log2 =3D 63 - clz64(ns->zone_size); + } + + return 0; +} + +static void nvme_init_zone_state(NvmeNamespace *ns) +{ + uint64_t start =3D 0, zone_size =3D ns->zone_size; + uint64_t capacity =3D ns->num_zones * zone_size; + NvmeZone *zone; + int i; + + ns->zone_array =3D g_malloc0(ns->zone_array_size); + + QTAILQ_INIT(&ns->exp_open_zones); + QTAILQ_INIT(&ns->imp_open_zones); + QTAILQ_INIT(&ns->closed_zones); + QTAILQ_INIT(&ns->full_zones); + + zone =3D ns->zone_array; + for (i =3D 0; i < ns->num_zones; i++, zone++) { + if (start + zone_size > capacity) { + zone_size =3D capacity - start; + } + zone->d.zt =3D NVME_ZONE_TYPE_SEQ_WRITE; + nvme_set_zone_state(zone, NVME_ZONE_STATE_EMPTY); + zone->d.za =3D 0; + zone->d.zcap =3D ns->zone_capacity; + zone->d.zslba =3D start; + zone->d.wp =3D start; + zone->w_ptr =3D start; + start +=3D zone_size; + } +} + +static int nvme_zoned_init_ns(NvmeCtrl *n, NvmeNamespace *ns, int lba_inde= x, + Error **errp) +{ + NvmeIdNsZoned *id_ns_z; + + if (nvme_calc_zone_geometry(ns, errp) !=3D 0) { + return -1; + } + + nvme_init_zone_state(ns); + + id_ns_z =3D g_malloc0(sizeof(NvmeIdNsZoned)); + + /* MAR/MOR are zeroes-based, 0xffffffff means no limit */ + id_ns_z->mar =3D 0xffffffff; + id_ns_z->mor =3D 0xffffffff; + id_ns_z->zoc =3D 0; + id_ns_z->ozcs =3D ns->params.cross_zone_read ? 0x01 : 0x00; + + id_ns_z->lbafe[lba_index].zsze =3D cpu_to_le64(ns->zone_size); + id_ns_z->lbafe[lba_index].zdes =3D 0; + + ns->csi =3D NVME_CSI_ZONED; + ns->id_ns.nsze =3D cpu_to_le64(ns->zone_size * ns->num_zones); + ns->id_ns.ncap =3D cpu_to_le64(ns->zone_capacity * ns->num_zones); + ns->id_ns.nuse =3D ns->id_ns.ncap; + + ns->id_ns_zoned =3D id_ns_z; + + return 0; +} + +static void nvme_clear_zone(NvmeNamespace *ns, NvmeZone *zone) +{ + uint8_t state; + + zone->w_ptr =3D zone->d.wp; + state =3D nvme_get_zone_state(zone); + if (zone->d.wp !=3D zone->d.zslba) { + if (state !=3D NVME_ZONE_STATE_CLOSED) { + trace_pci_nvme_clear_ns_close(state, zone->d.zslba); + nvme_set_zone_state(zone, NVME_ZONE_STATE_CLOSED); + } + QTAILQ_INSERT_HEAD(&ns->closed_zones, zone, entry); + } else { + trace_pci_nvme_clear_ns_reset(state, zone->d.zslba); + nvme_set_zone_state(zone, NVME_ZONE_STATE_EMPTY); + } +} + +/* + * Close all the zones that are currently open. + */ +static void nvme_zoned_ns_shutdown(NvmeNamespace *ns) +{ + NvmeZone *zone, *next; + + QTAILQ_FOREACH_SAFE(zone, &ns->closed_zones, entry, next) { + QTAILQ_REMOVE(&ns->closed_zones, zone, entry); + nvme_clear_zone(ns, zone); + } + QTAILQ_FOREACH_SAFE(zone, &ns->imp_open_zones, entry, next) { + QTAILQ_REMOVE(&ns->imp_open_zones, zone, entry); + nvme_clear_zone(ns, zone); + } + QTAILQ_FOREACH_SAFE(zone, &ns->exp_open_zones, entry, next) { + QTAILQ_REMOVE(&ns->exp_open_zones, zone, entry); + nvme_clear_zone(ns, zone); + } +} + static int nvme_ns_check_constraints(NvmeNamespace *ns, Error **errp) { if (!ns->blkconf.blk) { @@ -98,6 +244,12 @@ int nvme_ns_setup(NvmeCtrl *n, NvmeNamespace *ns, Error= **errp) } =20 nvme_ns_init(ns); + if (ns->params.zoned) { + if (nvme_zoned_init_ns(n, ns, 0, errp) !=3D 0) { + return -1; + } + } + if (nvme_register_namespace(n, ns, errp)) { return -1; } @@ -115,6 +267,21 @@ void nvme_ns_flush(NvmeNamespace *ns) blk_flush(ns->blkconf.blk); } =20 +void nvme_ns_shutdown(NvmeNamespace *ns) +{ + if (ns->params.zoned) { + nvme_zoned_ns_shutdown(ns); + } +} + +void nvme_ns_cleanup(NvmeNamespace *ns) +{ + if (ns->params.zoned) { + g_free(ns->id_ns_zoned); + g_free(ns->zone_array); + } +} + static void nvme_ns_realize(DeviceState *dev, Error **errp) { NvmeNamespace *ns =3D NVME_NS(dev); @@ -133,6 +300,12 @@ static Property nvme_ns_props[] =3D { DEFINE_BLOCK_PROPERTIES(NvmeNamespace, blkconf), DEFINE_PROP_UINT32("nsid", NvmeNamespace, params.nsid, 0), DEFINE_PROP_UUID("uuid", NvmeNamespace, params.uuid), + DEFINE_PROP_BOOL("zoned", NvmeNamespace, params.zoned, false), + DEFINE_PROP_SIZE("zoned.zsze", NvmeNamespace, params.zone_size_bs, + NVME_DEFAULT_ZONE_SIZE), + DEFINE_PROP_SIZE("zoned.zcap", NvmeNamespace, params.zone_cap_bs, 0), + DEFINE_PROP_BOOL("zoned.cross_read", NvmeNamespace, + params.cross_zone_read, false), DEFINE_PROP_END_OF_LIST(), }; =20 diff --git a/hw/block/nvme.c b/hw/block/nvme.c index 7495cdb5ef..f5390e6863 100644 --- a/hw/block/nvme.c +++ b/hw/block/nvme.c @@ -133,6 +133,16 @@ static const uint32_t nvme_cse_iocs_nvm[256] =3D { [NVME_CMD_READ] =3D NVME_CMD_EFF_CSUPP, }; =20 +static const uint32_t nvme_cse_iocs_zoned[256] =3D { + [NVME_CMD_FLUSH] =3D NVME_CMD_EFF_CSUPP | NVME_CMD_EFF_= LBCC, + [NVME_CMD_WRITE_ZEROES] =3D NVME_CMD_EFF_CSUPP | NVME_CMD_EFF_= LBCC, + [NVME_CMD_WRITE] =3D NVME_CMD_EFF_CSUPP | NVME_CMD_EFF_= LBCC, + [NVME_CMD_READ] =3D NVME_CMD_EFF_CSUPP, + [NVME_CMD_ZONE_APPEND] =3D NVME_CMD_EFF_CSUPP | NVME_CMD_EFF_= LBCC, + [NVME_CMD_ZONE_MGMT_SEND] =3D NVME_CMD_EFF_CSUPP, + [NVME_CMD_ZONE_MGMT_RECV] =3D NVME_CMD_EFF_CSUPP, +}; + static void nvme_process_sq(void *opaque); =20 static uint16_t nvme_cid(NvmeRequest *req) @@ -149,6 +159,46 @@ static uint16_t nvme_sqid(NvmeRequest *req) return le16_to_cpu(req->sq->sqid); } =20 +static void nvme_assign_zone_state(NvmeNamespace *ns, NvmeZone *zone, + uint8_t state) +{ + if (QTAILQ_IN_USE(zone, entry)) { + switch (nvme_get_zone_state(zone)) { + case NVME_ZONE_STATE_EXPLICITLY_OPEN: + QTAILQ_REMOVE(&ns->exp_open_zones, zone, entry); + break; + case NVME_ZONE_STATE_IMPLICITLY_OPEN: + QTAILQ_REMOVE(&ns->imp_open_zones, zone, entry); + break; + case NVME_ZONE_STATE_CLOSED: + QTAILQ_REMOVE(&ns->closed_zones, zone, entry); + break; + case NVME_ZONE_STATE_FULL: + QTAILQ_REMOVE(&ns->full_zones, zone, entry); + } + } + + nvme_set_zone_state(zone, state); + + switch (state) { + case NVME_ZONE_STATE_EXPLICITLY_OPEN: + QTAILQ_INSERT_TAIL(&ns->exp_open_zones, zone, entry); + break; + case NVME_ZONE_STATE_IMPLICITLY_OPEN: + QTAILQ_INSERT_TAIL(&ns->imp_open_zones, zone, entry); + break; + case NVME_ZONE_STATE_CLOSED: + QTAILQ_INSERT_TAIL(&ns->closed_zones, zone, entry); + break; + case NVME_ZONE_STATE_FULL: + QTAILQ_INSERT_TAIL(&ns->full_zones, zone, entry); + case NVME_ZONE_STATE_READ_ONLY: + break; + default: + zone->d.za =3D 0; + } +} + static bool nvme_addr_is_cmb(NvmeCtrl *n, hwaddr addr) { hwaddr low =3D n->ctrl_mem.addr; @@ -900,6 +950,319 @@ static inline uint16_t nvme_check_bounds(NvmeCtrl *n,= NvmeNamespace *ns, return NVME_SUCCESS; } =20 +static void nvme_fill_read_data(NvmeRequest *req, uint64_t offset, + uint32_t max_len) +{ + QEMUSGList *qsg =3D &req->qsg; + QEMUIOVector *iov =3D &req->iov; + ScatterGatherEntry *entry; + uint32_t len, ent_len; + + if (qsg->nsg > 0) { + entry =3D qsg->sg; + len =3D qsg->size; + if (max_len) { + len =3D MIN(len, max_len); + } + for (; len > 0; len -=3D ent_len) { + ent_len =3D MIN(len, entry->len); + if (offset > ent_len) { + offset -=3D ent_len; + } else if (offset !=3D 0) { + dma_memory_set(qsg->as, entry->base + offset, + 0, ent_len - offset); + offset =3D 0; + } else { + dma_memory_set(qsg->as, entry->base, 0, ent_len); + } + entry++; + } + } else if (iov->iov) { + len =3D iov_size(iov->iov, iov->niov); + if (max_len) { + len =3D MIN(len, max_len); + } + qemu_iovec_memset(iov, offset, 0, len - offset); + } +} + +static inline uint32_t nvme_zone_idx(NvmeNamespace *ns, uint64_t slba) +{ + return ns->zone_size_log2 > 0 ? slba >> ns->zone_size_log2 : + slba / ns->zone_size; +} + +static inline NvmeZone *nvme_get_zone_by_slba(NvmeNamespace *ns, uint64_t = slba) +{ + uint32_t zone_idx =3D nvme_zone_idx(ns, slba); + + assert(zone_idx < ns->num_zones); + return &ns->zone_array[zone_idx]; +} + +static uint16_t nvme_zone_state_ok_to_write(NvmeZone *zone) +{ + uint16_t status; + + switch (nvme_get_zone_state(zone)) { + case NVME_ZONE_STATE_EMPTY: + case NVME_ZONE_STATE_IMPLICITLY_OPEN: + case NVME_ZONE_STATE_EXPLICITLY_OPEN: + case NVME_ZONE_STATE_CLOSED: + status =3D NVME_SUCCESS; + break; + case NVME_ZONE_STATE_FULL: + status =3D NVME_ZONE_FULL; + break; + case NVME_ZONE_STATE_OFFLINE: + status =3D NVME_ZONE_OFFLINE; + break; + case NVME_ZONE_STATE_READ_ONLY: + status =3D NVME_ZONE_READ_ONLY; + break; + default: + assert(false); + } + + return status; +} + +static uint16_t nvme_check_zone_write(NvmeCtrl *n, NvmeNamespace *ns, + NvmeZone *zone, uint64_t slba, + uint32_t nlb, bool append) +{ + uint16_t status; + + if (unlikely((slba + nlb) > nvme_zone_wr_boundary(zone))) { + status =3D NVME_ZONE_BOUNDARY_ERROR; + } else { + status =3D nvme_zone_state_ok_to_write(zone); + } + + if (status !=3D NVME_SUCCESS) { + trace_pci_nvme_err_zone_write_not_ok(slba, nlb, status); + } else { + assert(nvme_wp_is_valid(zone)); + if (append) { + if (unlikely(slba !=3D zone->d.zslba)) { + trace_pci_nvme_err_append_not_at_start(slba, zone->d.zslba= ); + status =3D NVME_ZONE_INVALID_WRITE; + } + if (nvme_l2b(ns, nlb) > (n->page_size << n->zasl)) { + trace_pci_nvme_err_append_too_large(slba, nlb, n->zasl); + status =3D NVME_INVALID_FIELD; + } + } else if (unlikely(slba !=3D zone->w_ptr)) { + trace_pci_nvme_err_write_not_at_wp(slba, zone->d.zslba, + zone->w_ptr); + status =3D NVME_ZONE_INVALID_WRITE; + } + } + + return status; +} + +static uint16_t nvme_zone_state_ok_to_read(NvmeZone *zone) +{ + uint16_t status; + + switch (nvme_get_zone_state(zone)) { + case NVME_ZONE_STATE_EMPTY: + case NVME_ZONE_STATE_IMPLICITLY_OPEN: + case NVME_ZONE_STATE_EXPLICITLY_OPEN: + case NVME_ZONE_STATE_FULL: + case NVME_ZONE_STATE_CLOSED: + case NVME_ZONE_STATE_READ_ONLY: + status =3D NVME_SUCCESS; + break; + case NVME_ZONE_STATE_OFFLINE: + status =3D NVME_ZONE_OFFLINE | NVME_DNR; + break; + default: + assert(false); + } + + return status; +} + +typedef struct NvmeReadFillCtx { + uint64_t pre_rd_fill_slba; + uint64_t read_slba; + uint64_t post_rd_fill_slba; + + uint32_t pre_rd_fill_nlb; + uint32_t read_nlb; + uint32_t post_rd_fill_nlb; +} NvmeReadFillCtx; + +static uint16_t nvme_check_zone_read(NvmeNamespace *ns, uint64_t slba, + uint32_t nlb, NvmeReadFillCtx *rfc) +{ + NvmeZone *zone =3D nvme_get_zone_by_slba(ns, slba); + NvmeZone *next_zone; + uint64_t bndry =3D nvme_zone_rd_boundary(ns, zone); + uint64_t end =3D slba + nlb, wp1, wp2; + uint16_t status; + + rfc->read_slba =3D slba; + rfc->read_nlb =3D nlb; + + status =3D nvme_zone_state_ok_to_read(zone); + if (status !=3D NVME_SUCCESS) { + ; + } else if (likely(end <=3D bndry)) { + if (end > zone->w_ptr) { + wp1 =3D zone->w_ptr; + if (slba >=3D wp1) { + /* No i/o necessary, just fill */ + rfc->pre_rd_fill_slba =3D slba; + rfc->pre_rd_fill_nlb =3D nlb; + rfc->read_nlb =3D 0; + } else { + rfc->read_nlb =3D wp1 - slba; + rfc->post_rd_fill_slba =3D wp1; + rfc->post_rd_fill_nlb =3D nlb - rfc->read_nlb; + } + } + } else if (!ns->params.cross_zone_read) { + status =3D NVME_ZONE_BOUNDARY_ERROR; + } else { + /* + * Read across zone boundary - look at the next zone. + * Earlier bounds checks ensure that the current zone + * is not the last one. + */ + next_zone =3D zone + 1; + status =3D nvme_zone_state_ok_to_read(next_zone); + if (status !=3D NVME_SUCCESS) { + ; + } else if (end > nvme_zone_rd_boundary(ns, next_zone)) { + /* + * As zone size is much larger than a typical maximum + * i/o size in real hardware, only allow the i/o range + * to span no more than one pair of zones. + */ + status =3D NVME_ZONE_BOUNDARY_ERROR; + } else { + wp1 =3D zone->w_ptr; + wp2 =3D next_zone->w_ptr; + if (wp2 =3D=3D bndry) { + if (slba >=3D wp1) { + /* Again, no i/o necessary, just fill */ + rfc->pre_rd_fill_slba =3D slba; + rfc->pre_rd_fill_nlb =3D nlb; + rfc->read_nlb =3D 0; + } else { + rfc->read_nlb =3D wp1 - slba; + rfc->post_rd_fill_slba =3D wp1; + rfc->post_rd_fill_nlb =3D nlb - rfc->read_nlb; + } + } else if (slba < wp1) { + if (end > wp2) { + if (wp1 =3D=3D bndry) { + rfc->post_rd_fill_slba =3D wp2; + rfc->post_rd_fill_nlb =3D end - wp2; + rfc->read_nlb =3D wp2 - slba; + } else { + rfc->pre_rd_fill_slba =3D wp2; + rfc->pre_rd_fill_nlb =3D end - wp2; + rfc->read_nlb =3D wp2 - slba; + rfc->post_rd_fill_slba =3D wp1; + rfc->post_rd_fill_nlb =3D bndry - wp1; + } + } else { + rfc->post_rd_fill_slba =3D wp1; + rfc->post_rd_fill_nlb =3D bndry - wp1; + } + } else { + if (end > wp2) { + rfc->pre_rd_fill_slba =3D slba; + rfc->pre_rd_fill_nlb =3D end - slba; + rfc->read_slba =3D bndry; + rfc->read_nlb =3D wp2 - bndry; + } else { + rfc->read_slba =3D bndry; + rfc->read_nlb =3D end - bndry; + rfc->post_rd_fill_slba =3D slba; + rfc->post_rd_fill_nlb =3D bndry - slba; + } + } + } + } + + return status; +} + +static bool nvme_finalize_zoned_write(NvmeNamespace *ns, NvmeRequest *req, + bool failed) +{ + NvmeRwCmd *rw =3D (NvmeRwCmd *)&req->cmd; + NvmeZone *zone; + NvmeZonedResult *res =3D (NvmeZonedResult *)&req->cqe; + uint64_t slba, start_wp =3D res->slba; + uint32_t nlb; + + if (rw->opcode !=3D NVME_CMD_WRITE && + rw->opcode !=3D NVME_CMD_ZONE_APPEND && + rw->opcode !=3D NVME_CMD_WRITE_ZEROES) { + return false; + } + + slba =3D le64_to_cpu(rw->slba); + nlb =3D le16_to_cpu(rw->nlb) + 1; + zone =3D nvme_get_zone_by_slba(ns, slba); + + if (!failed && zone->w_ptr < start_wp + nlb) { + /* + * A preceding queued write to the zone has failed, + * now this write is not at the WP, fail it too. + */ + failed =3D true; + } + + if (failed) { + res->slba =3D 0; + } else if (zone->w_ptr =3D=3D nvme_zone_wr_boundary(zone)) { + switch (nvme_get_zone_state(zone)) { + case NVME_ZONE_STATE_IMPLICITLY_OPEN: + case NVME_ZONE_STATE_EXPLICITLY_OPEN: + case NVME_ZONE_STATE_CLOSED: + case NVME_ZONE_STATE_EMPTY: + nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_FULL); + /* fall through */ + case NVME_ZONE_STATE_FULL: + break; + default: + assert(false); + } + zone->d.wp =3D zone->w_ptr; + } else { + zone->d.wp +=3D nlb; + } + + return failed; +} + +static uint64_t nvme_advance_zone_wp(NvmeNamespace *ns, NvmeZone *zone, + uint32_t nlb) +{ + uint64_t result =3D zone->w_ptr; + uint8_t zs; + + zone->w_ptr +=3D nlb; + + if (zone->w_ptr < nvme_zone_wr_boundary(zone)) { + zs =3D nvme_get_zone_state(zone); + switch (zs) { + case NVME_ZONE_STATE_EMPTY: + case NVME_ZONE_STATE_CLOSED: + nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_IMPLICITLY_OP= EN); + } + } + + return result; +} + static void nvme_rw_cb(void *opaque, int ret) { NvmeRequest *req =3D opaque; @@ -914,10 +1277,26 @@ static void nvme_rw_cb(void *opaque, int ret) trace_pci_nvme_rw_cb(nvme_cid(req), blk_name(blk)); =20 if (!ret) { - block_acct_done(stats, acct); + if (ns->params.zoned) { + if (nvme_finalize_zoned_write(ns, req, false)) { + ret =3D EIO; + block_acct_failed(stats, acct); + req->status =3D NVME_ZONE_INVALID_WRITE; + } else if (req->fill_len) { + nvme_fill_read_data(req, req->fill_off, req->fill_len); + req->fill_len =3D 0; + } + } + if (!ret) { + block_acct_done(stats, acct); + } } else { uint16_t status; =20 + if (ns->params.zoned) { + nvme_finalize_zoned_write(ns, req, true); + } + block_acct_failed(stats, acct); =20 switch (req->cmd.opcode) { @@ -960,7 +1339,9 @@ static uint16_t nvme_read(NvmeCtrl *n, NvmeRequest *re= q) uint64_t slba =3D le64_to_cpu(rw->slba); uint32_t nlb =3D (uint32_t)le16_to_cpu(rw->nlb) + 1; uint64_t data_size =3D nvme_l2b(ns, nlb); - uint64_t data_offset; + uint64_t data_offset, fill_off; + uint32_t fill_len; + NvmeReadFillCtx rfc =3D {}; BlockBackend *blk =3D ns->blkconf.blk; uint16_t status; =20 @@ -978,11 +1359,40 @@ static uint16_t nvme_read(NvmeCtrl *n, NvmeRequest *= req) goto invalid; } =20 + if (ns->params.zoned) { + status =3D nvme_check_zone_read(ns, slba, nlb, &rfc); + if (status !=3D NVME_SUCCESS) { + trace_pci_nvme_err_zone_read_not_ok(slba, nlb, status); + goto invalid; + } + } + status =3D nvme_map_dptr(n, data_size, req); if (status) { goto invalid; } =20 + if (ns->params.zoned) { + if (rfc.pre_rd_fill_nlb) { + fill_off =3D nvme_l2b(ns, rfc.pre_rd_fill_slba - slba); + fill_len =3D nvme_l2b(ns, rfc.pre_rd_fill_nlb); + nvme_fill_read_data(req, fill_off, fill_len); + } + if (!rfc.read_nlb) { + /* No backend I/O necessary, only needed to fill the buffer */ + req->status =3D NVME_SUCCESS; + return NVME_SUCCESS; + } + if (rfc.post_rd_fill_nlb) { + req->fill_off =3D nvme_l2b(ns, rfc.post_rd_fill_slba - slba); + req->fill_len =3D nvme_l2b(ns, rfc.post_rd_fill_nlb); + } else { + req->fill_len =3D 0; + } + slba =3D rfc.read_slba; + data_size =3D nvme_l2b(ns, rfc.read_nlb); + } + data_offset =3D nvme_l2b(ns, slba); =20 block_acct_start(blk_get_stats(blk), &req->acct, data_size, @@ -1001,7 +1411,7 @@ invalid: return status | NVME_DNR; } =20 -static uint16_t nvme_write(NvmeCtrl *n, NvmeRequest *req, bool wrz) +static uint16_t nvme_write(NvmeCtrl *n, NvmeRequest *req, bool append, boo= l wrz) { NvmeRwCmd *rw =3D (NvmeRwCmd *)&req->cmd; NvmeNamespace *ns =3D req->ns; @@ -1009,6 +1419,8 @@ static uint16_t nvme_write(NvmeCtrl *n, NvmeRequest *= req, bool wrz) uint32_t nlb =3D (uint32_t)le16_to_cpu(rw->nlb) + 1; uint64_t data_size =3D nvme_l2b(ns, nlb); uint64_t data_offset; + NvmeZone *zone; + NvmeZonedResult *res =3D (NvmeZonedResult *)&req->cqe; BlockBackend *blk =3D ns->blkconf.blk; uint16_t status; =20 @@ -1029,6 +1441,25 @@ static uint16_t nvme_write(NvmeCtrl *n, NvmeRequest = *req, bool wrz) goto invalid; } =20 + if (ns->params.zoned) { + zone =3D nvme_get_zone_by_slba(ns, slba); + + status =3D nvme_check_zone_write(n, ns, zone, slba, nlb, append); + if (status !=3D NVME_SUCCESS) { + goto invalid; + } + + if (append) { + slba =3D zone->w_ptr; + } + + res->slba =3D nvme_advance_zone_wp(ns, zone, nlb); + } else if (append) { + trace_pci_nvme_err_invalid_opc(rw->opcode); + status =3D NVME_INVALID_OPCODE; + goto invalid; + } + data_offset =3D nvme_l2b(ns, slba); =20 if (!wrz) { @@ -1059,6 +1490,436 @@ invalid: return status | NVME_DNR; } =20 +static uint16_t nvme_get_mgmt_zone_slba_idx(NvmeNamespace *ns, NvmeCmd *c, + uint64_t *slba, uint32_t *zone= _idx) +{ + uint32_t dw10 =3D le32_to_cpu(c->cdw10); + uint32_t dw11 =3D le32_to_cpu(c->cdw11); + + if (!ns->params.zoned) { + trace_pci_nvme_err_invalid_opc(c->opcode); + return NVME_INVALID_OPCODE | NVME_DNR; + } + + *slba =3D ((uint64_t)dw11) << 32 | dw10; + if (unlikely(*slba >=3D ns->id_ns.nsze)) { + trace_pci_nvme_err_invalid_lba_range(*slba, 0, ns->id_ns.nsze); + *slba =3D 0; + return NVME_LBA_RANGE | NVME_DNR; + } + + *zone_idx =3D nvme_zone_idx(ns, *slba); + assert(*zone_idx < ns->num_zones); + + return NVME_SUCCESS; +} + +typedef uint16_t (*op_handler_t)(NvmeNamespace *, NvmeZone *, + uint8_t); + +enum NvmeZoneProcessingMask { + NVME_PROC_CURRENT_ZONE =3D 0, + NVME_PROC_IMP_OPEN_ZONES =3D 1 << 0, + NVME_PROC_EXP_OPEN_ZONES =3D 1 << 1, + NVME_PROC_CLOSED_ZONES =3D 1 << 2, + NVME_PROC_READ_ONLY_ZONES =3D 1 << 3, + NVME_PROC_FULL_ZONES =3D 1 << 4, +}; + +static uint16_t nvme_open_zone(NvmeNamespace *ns, NvmeZone *zone, + uint8_t state) +{ + switch (state) { + case NVME_ZONE_STATE_EMPTY: + case NVME_ZONE_STATE_CLOSED: + case NVME_ZONE_STATE_IMPLICITLY_OPEN: + nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_EXPLICITLY_OPEN); + /* fall through */ + case NVME_ZONE_STATE_EXPLICITLY_OPEN: + return NVME_SUCCESS; + } + + return NVME_ZONE_INVAL_TRANSITION; +} + +static uint16_t nvme_close_zone(NvmeNamespace *ns, NvmeZone *zone, + uint8_t state) +{ + switch (state) { + case NVME_ZONE_STATE_EXPLICITLY_OPEN: + case NVME_ZONE_STATE_IMPLICITLY_OPEN: + nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_CLOSED); + /* fall through */ + case NVME_ZONE_STATE_CLOSED: + return NVME_SUCCESS; + } + + return NVME_ZONE_INVAL_TRANSITION; +} + +static uint16_t nvme_finish_zone(NvmeNamespace *ns, NvmeZone *zone, + uint8_t state) +{ + switch (state) { + case NVME_ZONE_STATE_EXPLICITLY_OPEN: + case NVME_ZONE_STATE_IMPLICITLY_OPEN: + case NVME_ZONE_STATE_CLOSED: + case NVME_ZONE_STATE_EMPTY: + zone->w_ptr =3D nvme_zone_wr_boundary(zone); + zone->d.wp =3D zone->w_ptr; + nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_FULL); + /* fall through */ + case NVME_ZONE_STATE_FULL: + return NVME_SUCCESS; + } + + return NVME_ZONE_INVAL_TRANSITION; +} + +static uint16_t nvme_reset_zone(NvmeNamespace *ns, NvmeZone *zone, + uint8_t state) +{ + switch (state) { + case NVME_ZONE_STATE_EXPLICITLY_OPEN: + case NVME_ZONE_STATE_IMPLICITLY_OPEN: + case NVME_ZONE_STATE_CLOSED: + case NVME_ZONE_STATE_FULL: + zone->w_ptr =3D zone->d.zslba; + zone->d.wp =3D zone->w_ptr; + nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_EMPTY); + /* fall through */ + case NVME_ZONE_STATE_EMPTY: + return NVME_SUCCESS; + } + + return NVME_ZONE_INVAL_TRANSITION; +} + +static uint16_t nvme_offline_zone(NvmeNamespace *ns, NvmeZone *zone, + uint8_t state) +{ + switch (state) { + case NVME_ZONE_STATE_READ_ONLY: + nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_OFFLINE); + /* fall through */ + case NVME_ZONE_STATE_OFFLINE: + return NVME_SUCCESS; + } + + return NVME_ZONE_INVAL_TRANSITION; +} + +static uint16_t nvme_bulk_proc_zone(NvmeNamespace *ns, NvmeZone *zone, + enum NvmeZoneProcessingMask proc_mask, + op_handler_t op_hndlr) +{ + uint16_t status =3D NVME_SUCCESS; + uint8_t zs =3D nvme_get_zone_state(zone); + bool proc_zone =3D false; + + switch (zs) { + case NVME_ZONE_STATE_IMPLICITLY_OPEN: + proc_zone =3D proc_mask & NVME_PROC_IMP_OPEN_ZONES; + break; + case NVME_ZONE_STATE_EXPLICITLY_OPEN: + proc_zone =3D proc_mask & NVME_PROC_EXP_OPEN_ZONES; + break; + case NVME_ZONE_STATE_CLOSED: + proc_zone =3D proc_mask & NVME_PROC_CLOSED_ZONES; + break; + case NVME_ZONE_STATE_READ_ONLY: + proc_zone =3D proc_mask & NVME_PROC_READ_ONLY_ZONES; + break; + case NVME_ZONE_STATE_FULL: + proc_zone =3D proc_mask & NVME_PROC_FULL_ZONES; + } + + if (proc_zone) { + status =3D op_hndlr(ns, zone, zs); + } + + return status; +} + +static uint16_t nvme_do_zone_op(NvmeNamespace *ns, NvmeZone *zone, + enum NvmeZoneProcessingMask proc_mask, + op_handler_t op_hndlr) +{ + NvmeZone *next; + uint16_t status =3D NVME_SUCCESS; + int i; + + if (!proc_mask) { + status =3D op_hndlr(ns, zone, nvme_get_zone_state(zone)); + } else { + if (proc_mask & NVME_PROC_CLOSED_ZONES) { + QTAILQ_FOREACH_SAFE(zone, &ns->closed_zones, entry, next) { + status =3D nvme_bulk_proc_zone(ns, zone, proc_mask, op_hnd= lr); + if (status !=3D NVME_SUCCESS) { + goto out; + } + } + } + if (proc_mask & NVME_PROC_IMP_OPEN_ZONES) { + QTAILQ_FOREACH_SAFE(zone, &ns->imp_open_zones, entry, next) { + status =3D nvme_bulk_proc_zone(ns, zone, proc_mask, op_hnd= lr); + if (status !=3D NVME_SUCCESS) { + goto out; + } + } + } + if (proc_mask & NVME_PROC_EXP_OPEN_ZONES) { + QTAILQ_FOREACH_SAFE(zone, &ns->exp_open_zones, entry, next) { + status =3D nvme_bulk_proc_zone(ns, zone, proc_mask, op_hnd= lr); + if (status !=3D NVME_SUCCESS) { + goto out; + } + } + } + if (proc_mask & NVME_PROC_FULL_ZONES) { + QTAILQ_FOREACH_SAFE(zone, &ns->full_zones, entry, next) { + status =3D nvme_bulk_proc_zone(ns, zone, proc_mask, op_hnd= lr); + if (status !=3D NVME_SUCCESS) { + goto out; + } + } + } + + if (proc_mask & NVME_PROC_READ_ONLY_ZONES) { + for (i =3D 0; i < ns->num_zones; i++, zone++) { + status =3D nvme_bulk_proc_zone(ns, zone, proc_mask, op_hnd= lr); + if (status !=3D NVME_SUCCESS) { + goto out; + } + } + } + } + +out: + return status; +} + +static uint16_t nvme_zone_mgmt_send(NvmeCtrl *n, NvmeRequest *req) +{ + NvmeCmd *cmd =3D (NvmeCmd *)&req->cmd; + NvmeNamespace *ns =3D req->ns; + NvmeZone *zone; + uint32_t dw13 =3D le32_to_cpu(cmd->cdw13); + uint64_t slba =3D 0; + uint32_t zone_idx =3D 0; + uint16_t status; + uint8_t action; + bool all; + enum NvmeZoneProcessingMask proc_mask =3D NVME_PROC_CURRENT_ZONE; + + action =3D dw13 & 0xff; + all =3D dw13 & 0x100; + + req->status =3D NVME_SUCCESS; + + if (!all) { + status =3D nvme_get_mgmt_zone_slba_idx(ns, cmd, &slba, &zone_idx); + if (status) { + return status; + } + } + + zone =3D &ns->zone_array[zone_idx]; + if (slba !=3D zone->d.zslba) { + trace_pci_nvme_err_unaligned_zone_cmd(action, slba, zone->d.zslba); + return NVME_INVALID_FIELD | NVME_DNR; + } + + switch (action) { + + case NVME_ZONE_ACTION_OPEN: + if (all) { + proc_mask =3D NVME_PROC_CLOSED_ZONES; + } + trace_pci_nvme_open_zone(slba, zone_idx, all); + status =3D nvme_do_zone_op(ns, zone, proc_mask, nvme_open_zone); + break; + + case NVME_ZONE_ACTION_CLOSE: + if (all) { + proc_mask =3D NVME_PROC_IMP_OPEN_ZONES | NVME_PROC_EXP_OPEN_ZO= NES; + } + trace_pci_nvme_close_zone(slba, zone_idx, all); + status =3D nvme_do_zone_op(ns, zone, proc_mask, nvme_close_zone); + break; + + case NVME_ZONE_ACTION_FINISH: + if (all) { + proc_mask =3D NVME_PROC_IMP_OPEN_ZONES | NVME_PROC_EXP_OPEN_ZO= NES | + NVME_PROC_CLOSED_ZONES; + } + trace_pci_nvme_finish_zone(slba, zone_idx, all); + status =3D nvme_do_zone_op(ns, zone, proc_mask, nvme_finish_zone); + break; + + case NVME_ZONE_ACTION_RESET: + if (all) { + proc_mask =3D NVME_PROC_IMP_OPEN_ZONES | NVME_PROC_EXP_OPEN_ZO= NES | + NVME_PROC_CLOSED_ZONES | NVME_PROC_FULL_ZONES; + } + trace_pci_nvme_reset_zone(slba, zone_idx, all); + status =3D nvme_do_zone_op(ns, zone, proc_mask, nvme_reset_zone); + break; + + case NVME_ZONE_ACTION_OFFLINE: + if (all) { + proc_mask =3D NVME_PROC_READ_ONLY_ZONES; + } + trace_pci_nvme_offline_zone(slba, zone_idx, all); + status =3D nvme_do_zone_op(ns, zone, proc_mask, nvme_offline_zone); + break; + + case NVME_ZONE_ACTION_SET_ZD_EXT: + trace_pci_nvme_set_descriptor_extension(slba, zone_idx); + return NVME_INVALID_FIELD | NVME_DNR; + break; + + default: + trace_pci_nvme_err_invalid_mgmt_action(action); + status =3D NVME_INVALID_FIELD; + } + + if (status =3D=3D NVME_ZONE_INVAL_TRANSITION) { + trace_pci_nvme_err_invalid_zone_state_transition(action, slba, + zone->d.za); + } + if (status) { + status |=3D NVME_DNR; + } + + return status; +} + +static bool nvme_zone_matches_filter(uint32_t zafs, NvmeZone *zl) +{ + int zs =3D nvme_get_zone_state(zl); + + switch (zafs) { + case NVME_ZONE_REPORT_ALL: + return true; + case NVME_ZONE_REPORT_EMPTY: + return zs =3D=3D NVME_ZONE_STATE_EMPTY; + case NVME_ZONE_REPORT_IMPLICITLY_OPEN: + return zs =3D=3D NVME_ZONE_STATE_IMPLICITLY_OPEN; + case NVME_ZONE_REPORT_EXPLICITLY_OPEN: + return zs =3D=3D NVME_ZONE_STATE_EXPLICITLY_OPEN; + case NVME_ZONE_REPORT_CLOSED: + return zs =3D=3D NVME_ZONE_STATE_CLOSED; + case NVME_ZONE_REPORT_FULL: + return zs =3D=3D NVME_ZONE_STATE_FULL; + case NVME_ZONE_REPORT_READ_ONLY: + return zs =3D=3D NVME_ZONE_STATE_READ_ONLY; + case NVME_ZONE_REPORT_OFFLINE: + return zs =3D=3D NVME_ZONE_STATE_OFFLINE; + default: + return false; + } +} + +static uint16_t nvme_zone_mgmt_recv(NvmeCtrl *n, NvmeRequest *req) +{ + NvmeCmd *cmd =3D (NvmeCmd *)&req->cmd; + NvmeNamespace *ns =3D req->ns; + /* cdw12 is zero-based number of dwords to return. Convert to bytes */ + uint32_t data_size =3D (le32_to_cpu(cmd->cdw12) + 1) << 2; + uint32_t dw13 =3D le32_to_cpu(cmd->cdw13); + uint32_t zone_idx, zra, zrasf, partial; + uint64_t max_zones, nr_zones =3D 0; + uint16_t status; + uint64_t slba; + NvmeZoneDescr *z; + NvmeZone *zs; + NvmeZoneReportHeader *header; + void *buf, *buf_p; + size_t zone_entry_sz; + + req->status =3D NVME_SUCCESS; + + status =3D nvme_get_mgmt_zone_slba_idx(ns, cmd, &slba, &zone_idx); + if (status) { + return status; + } + + zra =3D dw13 & 0xff; + if (zra !=3D NVME_ZONE_REPORT) { + return NVME_INVALID_FIELD | NVME_DNR; + } + + zrasf =3D (dw13 >> 8) & 0xff; + if (zrasf > NVME_ZONE_REPORT_OFFLINE) { + return NVME_INVALID_FIELD | NVME_DNR; + } + + if (data_size < sizeof(NvmeZoneReportHeader)) { + return NVME_INVALID_FIELD | NVME_DNR; + } + + status =3D nvme_check_mdts(n, data_size); + if (status) { + trace_pci_nvme_err_mdts(nvme_cid(req), data_size); + return status; + } + + partial =3D (dw13 >> 16) & 0x01; + + zone_entry_sz =3D sizeof(NvmeZoneDescr); + + max_zones =3D (data_size - sizeof(NvmeZoneReportHeader)) / zone_entry_= sz; + buf =3D g_malloc0(data_size); + + header =3D (NvmeZoneReportHeader *)buf; + buf_p =3D buf + sizeof(NvmeZoneReportHeader); + + while (zone_idx < ns->num_zones && nr_zones < max_zones) { + zs =3D &ns->zone_array[zone_idx]; + + if (!nvme_zone_matches_filter(zrasf, zs)) { + zone_idx++; + continue; + } + + z =3D (NvmeZoneDescr *)buf_p; + buf_p +=3D sizeof(NvmeZoneDescr); + nr_zones++; + + z->zt =3D zs->d.zt; + z->zs =3D zs->d.zs; + z->zcap =3D cpu_to_le64(zs->d.zcap); + z->zslba =3D cpu_to_le64(zs->d.zslba); + z->za =3D zs->d.za; + + if (nvme_wp_is_valid(zs)) { + z->wp =3D cpu_to_le64(zs->d.wp); + } else { + z->wp =3D cpu_to_le64(~0ULL); + } + + zone_idx++; + } + + if (!partial) { + for (; zone_idx < ns->num_zones; zone_idx++) { + zs =3D &ns->zone_array[zone_idx]; + if (nvme_zone_matches_filter(zrasf, zs)) { + nr_zones++; + } + } + } + header->nr_zones =3D cpu_to_le64(nr_zones); + + status =3D nvme_dma(n, (uint8_t *)buf, data_size, + DMA_DIRECTION_FROM_DEVICE, req); + + g_free(buf); + + return status; +} + static uint16_t nvme_io_cmd(NvmeCtrl *n, NvmeRequest *req) { uint32_t nsid =3D le32_to_cpu(req->cmd.nsid); @@ -1084,11 +1945,17 @@ static uint16_t nvme_io_cmd(NvmeCtrl *n, NvmeReques= t *req) case NVME_CMD_FLUSH: return nvme_flush(n, req); case NVME_CMD_WRITE_ZEROES: - return nvme_write(n, req, true); + return nvme_write(n, req, false, true); + case NVME_CMD_ZONE_APPEND: + return nvme_write(n, req, true, false); case NVME_CMD_WRITE: - return nvme_write(n, req, false); + return nvme_write(n, req, false, false); case NVME_CMD_READ: return nvme_read(n, req); + case NVME_CMD_ZONE_MGMT_SEND: + return nvme_zone_mgmt_send(n, req); + case NVME_CMD_ZONE_MGMT_RECV: + return nvme_zone_mgmt_recv(n, req); default: assert(false); } @@ -1348,6 +2215,9 @@ static uint16_t nvme_cmd_effects(NvmeCtrl *n, uint8_t= csi, uint32_t buf_len, case NVME_CSI_NVM: src_iocs =3D nvme_cse_iocs_nvm; break; + case NVME_CSI_ZONED: + src_iocs =3D nvme_cse_iocs_zoned; + break; } } =20 @@ -1529,6 +2399,16 @@ static uint16_t nvme_rpt_empty_id_struct(NvmeCtrl *n= , NvmeRequest *req) return nvme_dma(n, id, sizeof(id), DMA_DIRECTION_FROM_DEVICE, req); } =20 +static inline bool nvme_csi_has_nvm_support(NvmeNamespace *ns) +{ + switch (ns->csi) { + case NVME_CSI_NVM: + case NVME_CSI_ZONED: + return true; + } + return false; +} + static uint16_t nvme_identify_ctrl(NvmeCtrl *n, NvmeRequest *req) { trace_pci_nvme_identify_ctrl(); @@ -1540,11 +2420,16 @@ static uint16_t nvme_identify_ctrl(NvmeCtrl *n, Nvm= eRequest *req) static uint16_t nvme_identify_ctrl_csi(NvmeCtrl *n, NvmeRequest *req) { NvmeIdentify *c =3D (NvmeIdentify *)&req->cmd; + NvmeIdCtrlZoned id =3D {}; =20 trace_pci_nvme_identify_ctrl_csi(c->csi); =20 if (c->csi =3D=3D NVME_CSI_NVM) { return nvme_rpt_empty_id_struct(n, req); + } else if (c->csi =3D=3D NVME_CSI_ZONED) { + id.zasl =3D n->zasl; + return nvme_dma(n, (uint8_t *)&id, sizeof(id), + DMA_DIRECTION_FROM_DEVICE, req); } =20 return NVME_INVALID_FIELD | NVME_DNR; @@ -1572,8 +2457,12 @@ static uint16_t nvme_identify_ns(NvmeCtrl *n, NvmeRe= quest *req, return nvme_rpt_empty_id_struct(n, req); } =20 - return nvme_dma(n, (uint8_t *)&ns->id_ns, sizeof(NvmeIdNs), - DMA_DIRECTION_FROM_DEVICE, req); + if (c->csi =3D=3D NVME_CSI_NVM && nvme_csi_has_nvm_support(ns)) { + return nvme_dma(n, (uint8_t *)&ns->id_ns, sizeof(NvmeIdNs), + DMA_DIRECTION_FROM_DEVICE, req); + } + + return NVME_INVALID_CMD_SET | NVME_DNR; } =20 static uint16_t nvme_identify_ns_csi(NvmeCtrl *n, NvmeRequest *req, @@ -1598,8 +2487,11 @@ static uint16_t nvme_identify_ns_csi(NvmeCtrl *n, Nv= meRequest *req, return nvme_rpt_empty_id_struct(n, req); } =20 - if (c->csi =3D=3D NVME_CSI_NVM) { + if (c->csi =3D=3D NVME_CSI_NVM && nvme_csi_has_nvm_support(ns)) { return nvme_rpt_empty_id_struct(n, req); + } else if (c->csi =3D=3D NVME_CSI_ZONED && ns->csi =3D=3D NVME_CSI_ZON= ED) { + return nvme_dma(n, (uint8_t *)ns->id_ns_zoned, sizeof(NvmeIdNsZone= d), + DMA_DIRECTION_FROM_DEVICE, req); } =20 return NVME_INVALID_FIELD | NVME_DNR; @@ -1668,7 +2560,7 @@ static uint16_t nvme_identify_nslist_csi(NvmeCtrl *n,= NvmeRequest *req, return NVME_INVALID_NSID | NVME_DNR; } =20 - if (c->csi !=3D NVME_CSI_NVM) { + if (c->csi !=3D NVME_CSI_NVM && c->csi !=3D NVME_CSI_ZONED) { return NVME_INVALID_FIELD | NVME_DNR; } =20 @@ -1677,7 +2569,7 @@ static uint16_t nvme_identify_nslist_csi(NvmeCtrl *n,= NvmeRequest *req, if (!ns) { continue; } - if (ns->params.nsid <=3D min_nsid) { + if (ns->params.nsid <=3D min_nsid || c->csi !=3D ns->csi) { continue; } if (only_active && !ns->attached) { @@ -1747,6 +2639,8 @@ static uint16_t nvme_identify_cmd_set(NvmeCtrl *n, Nv= meRequest *req) trace_pci_nvme_identify_cmd_set(); =20 NVME_SET_CSI(*list, NVME_CSI_NVM); + NVME_SET_CSI(*list, NVME_CSI_ZONED); + return nvme_dma(n, list, data_len, DMA_DIRECTION_FROM_DEVICE, req); } =20 @@ -2206,7 +3100,7 @@ static void nvme_process_sq(void *opaque) } } =20 -static void nvme_clear_ctrl(NvmeCtrl *n) +static void nvme_clear_ctrl(NvmeCtrl *n, bool shutdown) { NvmeNamespace *ns; int i; @@ -2250,6 +3144,17 @@ static void nvme_clear_ctrl(NvmeCtrl *n) nvme_ns_flush(ns); } =20 + if (shutdown) { + for (i =3D 1; i <=3D n->num_namespaces; i++) { + ns =3D nvme_ns(n, i); + if (!ns) { + continue; + } + + nvme_ns_shutdown(ns); + } + } + n->bar.cc =3D 0; } =20 @@ -2270,6 +3175,13 @@ static void nvme_select_ns_iocs(NvmeCtrl *n) ns->iocs =3D nvme_cse_iocs_nvm; } break; + case NVME_CSI_ZONED: + if (NVME_CC_CSS(n->bar.cc) =3D=3D NVME_CC_CSS_CSI) { + ns->iocs =3D nvme_cse_iocs_zoned; + } else if (NVME_CC_CSS(n->bar.cc) =3D=3D NVME_CC_CSS_NVM) { + ns->iocs =3D nvme_cse_iocs_nvm; + } + break; } } } @@ -2368,6 +3280,17 @@ static int nvme_start_ctrl(NvmeCtrl *n) nvme_init_sq(&n->admin_sq, n, n->bar.asq, 0, 0, NVME_AQA_ASQS(n->bar.aqa) + 1); =20 + if (!n->params.zasl_bs) { + n->zasl =3D n->params.mdts; + } else { + if (n->params.zasl_bs < n->page_size) { + trace_pci_nvme_err_startfail_zasl_too_small(n->params.zasl_bs, + n->page_size); + return -1; + } + n->zasl =3D 31 - clz32(n->params.zasl_bs / n->page_size); + } + nvme_set_timestamp(n, 0ULL); =20 QTAILQ_INIT(&n->aer_queue); @@ -2440,12 +3363,12 @@ static void nvme_write_bar(NvmeCtrl *n, hwaddr offs= et, uint64_t data, } } else if (!NVME_CC_EN(data) && NVME_CC_EN(n->bar.cc)) { trace_pci_nvme_mmio_stopped(); - nvme_clear_ctrl(n); + nvme_clear_ctrl(n, false); n->bar.csts &=3D ~NVME_CSTS_READY; } if (NVME_CC_SHN(data) && !(NVME_CC_SHN(n->bar.cc))) { trace_pci_nvme_mmio_shutdown_set(); - nvme_clear_ctrl(n); + nvme_clear_ctrl(n, true); n->bar.cc =3D data; n->bar.csts |=3D NVME_CSTS_SHST_COMPLETE; } else if (!NVME_CC_SHN(data) && NVME_CC_SHN(n->bar.cc)) { @@ -2792,6 +3715,13 @@ static void nvme_check_constraints(NvmeCtrl *n, Erro= r **errp) =20 host_memory_backend_set_mapped(n->pmrdev, true); } + + if (n->params.zasl_bs) { + if (!is_power_of_2(n->params.zasl_bs)) { + error_setg(errp, "zone append size limit has to be a power of = 2"); + return; + } + } } =20 static void nvme_init_state(NvmeCtrl *n) @@ -3056,8 +3986,20 @@ static void nvme_realize(PCIDevice *pci_dev, Error *= *errp) static void nvme_exit(PCIDevice *pci_dev) { NvmeCtrl *n =3D NVME(pci_dev); + NvmeNamespace *ns; + int i; + + nvme_clear_ctrl(n, true); + + for (i =3D 1; i <=3D n->num_namespaces; i++) { + ns =3D nvme_ns(n, i); + if (!ns) { + continue; + } + + nvme_ns_cleanup(ns); + } =20 - nvme_clear_ctrl(n); g_free(n->cq); g_free(n->sq); g_free(n->aer_reqs); @@ -3085,6 +4027,8 @@ static Property nvme_props[] =3D { DEFINE_PROP_UINT32("aer_max_queued", NvmeCtrl, params.aer_max_queued, = 64), DEFINE_PROP_UINT8("mdts", NvmeCtrl, params.mdts, 7), DEFINE_PROP_BOOL("use-intel-id", NvmeCtrl, params.use_intel_id, false), + DEFINE_PROP_SIZE32("zoned.append_size_limit", NvmeCtrl, params.zasl_bs, + NVME_DEFAULT_MAX_ZA_SIZE), DEFINE_PROP_END_OF_LIST(), }; =20 diff --git a/hw/block/trace-events b/hw/block/trace-events index 8b29423132..4d910bb942 100644 --- a/hw/block/trace-events +++ b/hw/block/trace-events @@ -89,6 +89,14 @@ pci_nvme_mmio_start_success(void) "setting controller en= able bit succeeded" pci_nvme_mmio_stopped(void) "cleared controller enable bit" pci_nvme_mmio_shutdown_set(void) "shutdown bit set" pci_nvme_mmio_shutdown_cleared(void) "shutdown bit cleared" +pci_nvme_open_zone(uint64_t slba, uint32_t zone_idx, int all) "open zone, = slba=3D%"PRIu64", idx=3D%"PRIu32", all=3D%"PRIi32"" +pci_nvme_close_zone(uint64_t slba, uint32_t zone_idx, int all) "close zone= , slba=3D%"PRIu64", idx=3D%"PRIu32", all=3D%"PRIi32"" +pci_nvme_finish_zone(uint64_t slba, uint32_t zone_idx, int all) "finish zo= ne, slba=3D%"PRIu64", idx=3D%"PRIu32", all=3D%"PRIi32"" +pci_nvme_reset_zone(uint64_t slba, uint32_t zone_idx, int all) "reset zone= , slba=3D%"PRIu64", idx=3D%"PRIu32", all=3D%"PRIi32"" +pci_nvme_offline_zone(uint64_t slba, uint32_t zone_idx, int all) "offline = zone, slba=3D%"PRIu64", idx=3D%"PRIu32", all=3D%"PRIi32"" +pci_nvme_set_descriptor_extension(uint64_t slba, uint32_t zone_idx) "set z= one descriptor extension, slba=3D%"PRIu64", idx=3D%"PRIu32"" +pci_nvme_clear_ns_close(uint32_t state, uint64_t slba) "zone state=3D%"PRI= u32", slba=3D%"PRIu64" transitioned to Closed state" +pci_nvme_clear_ns_reset(uint32_t state, uint64_t slba) "zone state=3D%"PRI= u32", slba=3D%"PRIu64" transitioned to Empty state" =20 # nvme traces for error conditions pci_nvme_err_mdts(uint16_t cid, size_t len) "cid %"PRIu16" len %zu" @@ -107,7 +115,13 @@ pci_nvme_err_invalid_opc(uint8_t opc) "invalid opcode = 0x%"PRIx8"" pci_nvme_err_invalid_admin_opc(uint8_t opc) "invalid admin opcode 0x%"PRIx= 8"" pci_nvme_err_invalid_lba_range(uint64_t start, uint64_t len, uint64_t limi= t) "Invalid LBA start=3D%"PRIu64" len=3D%"PRIu64" limit=3D%"PRIu64"" pci_nvme_err_invalid_log_page_offset(uint64_t ofs, uint64_t size) "must be= <=3D %"PRIu64", got %"PRIu64"" -pci_nvme_err_only_nvm_cmd_set_avail(void) "setting 110b CC.CSS, but only N= VM command set is enabled" +pci_nvme_err_unaligned_zone_cmd(uint8_t action, uint64_t slba, uint64_t zs= lba) "unaligned zone op 0x%"PRIx32", got slba=3D%"PRIu64", zslba=3D%"PRIu64= "" +pci_nvme_err_invalid_zone_state_transition(uint8_t action, uint64_t slba, = uint8_t attrs) "action=3D0x%"PRIx8", slba=3D%"PRIu64", attrs=3D0x%"PRIx32"" +pci_nvme_err_write_not_at_wp(uint64_t slba, uint64_t zone, uint64_t wp) "w= riting at slba=3D%"PRIu64", zone=3D%"PRIu64", but wp=3D%"PRIu64"" +pci_nvme_err_append_not_at_start(uint64_t slba, uint64_t zone) "appending = at slba=3D%"PRIu64", but zone=3D%"PRIu64"" +pci_nvme_err_zone_write_not_ok(uint64_t slba, uint32_t nlb, uint32_t statu= s) "slba=3D%"PRIu64", nlb=3D%"PRIu32", status=3D0x%"PRIx16"" +pci_nvme_err_zone_read_not_ok(uint64_t slba, uint32_t nlb, uint32_t status= ) "slba=3D%"PRIu64", nlb=3D%"PRIu32", status=3D0x%"PRIx16"" +pci_nvme_err_append_too_large(uint64_t slba, uint32_t nlb, uint8_t zasl) "= slba=3D%"PRIu64", nlb=3D%"PRIu32", zasl=3D%"PRIu8"" pci_nvme_err_invalid_iocsci(uint32_t idx) "unsupported command set combina= tion index %"PRIu32"" pci_nvme_err_invalid_del_sq(uint16_t qid) "invalid submission queue deleti= on, sid=3D%"PRIu16"" pci_nvme_err_invalid_create_sq_cqid(uint16_t cqid) "failed creating submis= sion queue, invalid cqid=3D%"PRIu16"" @@ -141,7 +155,9 @@ pci_nvme_err_startfail_sqent_too_large(uint8_t log2ps, = uint8_t maxlog2ps) "nvme_ pci_nvme_err_startfail_css(uint8_t css) "nvme_start_ctrl failed because in= valid command set selected:%u" pci_nvme_err_startfail_asqent_sz_zero(void) "nvme_start_ctrl failed becaus= e the admin submission queue size is zero" pci_nvme_err_startfail_acqent_sz_zero(void) "nvme_start_ctrl failed becaus= e the admin completion queue size is zero" +pci_nvme_err_startfail_zasl_too_small(uint32_t zasl, uint32_t pagesz) "nvm= e_start_ctrl failed because zone append size limit %"PRIu32" is too small, = needs to be >=3D %"PRIu32"" pci_nvme_err_startfail(void) "setting controller enable bit failed" +pci_nvme_err_invalid_mgmt_action(int action) "action=3D0x%"PRIx8"" =20 # Traces for undefined behavior pci_nvme_ub_mmiowr_misaligned32(uint64_t offset) "MMIO write not 32-bit al= igned, offset=3D0x%"PRIx64"" --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604707098; cv=none; d=zohomail.com; s=zohoarc; b=ICEzTMnfsM1dVbtunG7TwEKrCwyF3Qo7qZS0Jye8UoeJpP+H4UwbDDzwxdEAwIYvFERZtq3AFjeiwbkT02VtZ0Rr2Iel8f0zhqoaj/Xkt9sjpqYmSlZf+DZb0ajci7uzoBCyJJTR2v3x0Ub8IhFTwCcjAvu5yRrENxu1PL7Bnq0= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604707098; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=XkOC00bNxULnrWaUsMvMmkB++xvTtBZLCBDtdggfpHk=; b=XNKKpVwYZ7UiNonDes7WnFRK+Lt8EnP83Q4Pm32w25EFBOdZsBDNj4JMP8PNYBEFmwV+6u8uAXhIn0CRTWc4sZnQwnN6ijDbquraracpdMCrXe1m1ulCy/yI+p2/kuH0lCv+rROGhhY1w2l5er0cj471bHHPIt5fOsub+O4D3J4= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604707098171801.221980079677; Fri, 6 Nov 2020 15:58:18 -0800 (PST) Received: from localhost ([::1]:34086 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBcW-0003yX-Rc for importer@patchew.org; Fri, 06 Nov 2020 18:58:16 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34326) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOP-0005mY-7U; Fri, 06 Nov 2020 18:43:41 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57376) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOL-0003V2-78; Fri, 06 Nov 2020 18:43:40 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:36 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:27 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:34 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706217; x=1636242217; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=uhMYU7fhme0K9tPS7WLxkmrjtyjoUnongLOGA3/tMno=; b=KrtbrNa82EBOijhvt+wCIy9epe0rPDVb7APRodRV4wl6cTkvE4oBypzf SEUZf81jehEQ5aHGxI8a/Bk1Bklqw4rNzuf4jpSGcBlXfBs77qiERo/qi QOLMCELC+aoSQkSGUHlwLA2sbVlrJ+piWnpzRzkZcVoKTKLYsEng1CNUg pNC9QViU2hc8L8tl8vcAb8peVeALjN4EgNYHSesgMgETjhZs6o6tXPzFD WbhJfP5I7GPaHjXgTK4i9/6CTiqT/pYNz38IF4R4twm/+Y/g2z3gOat4e nPD0mfY40L+xU+of+fy2UQA/SB008k1swCWITDASRKq6cgBcJrFZcZjFL A==; IronPort-SDR: Qi0JAKBwSGy8yB7yMEo4dnIdDqBlw1VhVn34amsbhPOIFuE71LBVhQzLHx3bhSvIQ55NFbGfXb ifdxqfuh4NkC27o+jrXWAsaKps8gIiUYxr0p2hoqTQyZ9zwvibteXeEOcvBnRMOac0j+5+Kc13 M+c8/lg2Jyi0VyxHz4NNFXHi5HtC5leHCxCXQeNwVXtY3bZiSzzRbL63Koan6cQtoa2BUJZiBV AuhIbUywslz98yP9+jWcwPfxqNc2E7QGtu9DIQeTL8zqP6HDCZQ3nRQFPbUOiHeD0uc7HA3ppk 6Z0= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267074" IronPort-SDR: pQGULjxK0aG9AAk+Y7+kxngzhgk0lIP0y5azSmfdV1SJhn3n3cSg4Bgzc6gQqsbArfLnI6DllR zIaV5HC2XQhX9b0AnV9pbYQld8p626rwgyjUdlpAljI4/+vVSMgBRBUuPjNmhESFBtvvj2T9zH 3AHEVWUyP+cfgI9b9+SP83B3eTCuVKHnI+lV+4NpOlywWauXear+2EbAppylbler5ZEAcxbtOm 9iN1bz2Z/+8G0IY8pvQ25urqxenXLalChz+19YGXkTIZiqtr+6SmUv67N+UIr7MEWg34bwQeGh 2Prqc39k5+qo5aidcIv8g4ks IronPort-SDR: w2N7p88GIimh1uJSFmOerWwsINwbSFVECWt6cc13bHm4lY349suBki/rhD3ThsaslHki+ssP/V VMIp0OFMx4ogdfVixIWSu5+772ljpPt1SmtNCZx7CV8DmkC0Lgcz5tO9tek+AhjBLBh48XDzB3 CZP6gz5Av0rLK4UOvo7NzIiPJMICuUSKfdo8UW0PbTxs1G+WKIbEA3k+kFZkALikSx7KIaJc20 EpjoJU3F/aGY88FTn+afxaOyr1Gr9vw0V+pg7y5wUUahkeHvKS5X7o15c8Qa6/xSN6ljT04bmr 4rA= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 09/12] hw/block/nvme: Introduce max active and open zone limits Date: Sat, 7 Nov 2020 08:43:02 +0900 Message-Id: <20201106234305.21339-10-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" Add two module properties, "zoned.max_active" and "zoned.max_open" to control the maximum number of zones that can be active or open. Once these variables are set to non-default values, these limits are checked during I/O and Too Many Active or Too Many Open command status is returned if they are exceeded. Signed-off-by: Hans Holmberg Signed-off-by: Dmitry Fomichev Reviewed-by: Niklas Cassel --- hw/block/nvme-ns.h | 41 +++++++++++++++++++ hw/block/nvme-ns.c | 30 +++++++++++++- hw/block/nvme.c | 94 +++++++++++++++++++++++++++++++++++++++++++ hw/block/trace-events | 2 + 4 files changed, 165 insertions(+), 2 deletions(-) diff --git a/hw/block/nvme-ns.h b/hw/block/nvme-ns.h index d2631ff5a3..421bab0a57 100644 --- a/hw/block/nvme-ns.h +++ b/hw/block/nvme-ns.h @@ -33,6 +33,8 @@ typedef struct NvmeNamespaceParams { bool cross_zone_read; uint64_t zone_size_bs; uint64_t zone_cap_bs; + uint32_t max_active_zones; + uint32_t max_open_zones; } NvmeNamespaceParams; =20 typedef struct NvmeNamespace { @@ -56,6 +58,8 @@ typedef struct NvmeNamespace { uint64_t zone_capacity; uint64_t zone_array_size; uint32_t zone_size_log2; + int32_t nr_open_zones; + int32_t nr_active_zones; =20 NvmeNamespaceParams params; } NvmeNamespace; @@ -123,6 +127,43 @@ static inline bool nvme_wp_is_valid(NvmeZone *zone) st !=3D NVME_ZONE_STATE_OFFLINE; } =20 +static inline void nvme_aor_inc_open(NvmeNamespace *ns) +{ + assert(ns->nr_open_zones >=3D 0); + if (ns->params.max_open_zones) { + ns->nr_open_zones++; + assert(ns->nr_open_zones <=3D ns->params.max_open_zones); + } +} + +static inline void nvme_aor_dec_open(NvmeNamespace *ns) +{ + if (ns->params.max_open_zones) { + assert(ns->nr_open_zones > 0); + ns->nr_open_zones--; + } + assert(ns->nr_open_zones >=3D 0); +} + +static inline void nvme_aor_inc_active(NvmeNamespace *ns) +{ + assert(ns->nr_active_zones >=3D 0); + if (ns->params.max_active_zones) { + ns->nr_active_zones++; + assert(ns->nr_active_zones <=3D ns->params.max_active_zones); + } +} + +static inline void nvme_aor_dec_active(NvmeNamespace *ns) +{ + if (ns->params.max_active_zones) { + assert(ns->nr_active_zones > 0); + ns->nr_active_zones--; + assert(ns->nr_active_zones >=3D ns->nr_open_zones); + } + assert(ns->nr_active_zones >=3D 0); +} + int nvme_ns_setup(NvmeCtrl *n, NvmeNamespace *ns, Error **errp); void nvme_ns_drain(NvmeNamespace *ns); void nvme_ns_flush(NvmeNamespace *ns); diff --git a/hw/block/nvme-ns.c b/hw/block/nvme-ns.c index e6db7f7d3b..2e45838c15 100644 --- a/hw/block/nvme-ns.c +++ b/hw/block/nvme-ns.c @@ -119,6 +119,20 @@ static int nvme_calc_zone_geometry(NvmeNamespace *ns, = Error **errp) ns->zone_size_log2 =3D 63 - clz64(ns->zone_size); } =20 + /* Make sure that the values of all ZNS properties are sane */ + if (ns->params.max_open_zones > nz) { + error_setg(errp, + "max_open_zones value %u exceeds the number of zones %u= ", + ns->params.max_open_zones, nz); + return -1; + } + if (ns->params.max_active_zones > nz) { + error_setg(errp, + "max_active_zones value %u exceeds the number of zones = %u", + ns->params.max_active_zones, nz); + return -1; + } + return 0; } =20 @@ -166,8 +180,8 @@ static int nvme_zoned_init_ns(NvmeCtrl *n, NvmeNamespac= e *ns, int lba_index, id_ns_z =3D g_malloc0(sizeof(NvmeIdNsZoned)); =20 /* MAR/MOR are zeroes-based, 0xffffffff means no limit */ - id_ns_z->mar =3D 0xffffffff; - id_ns_z->mor =3D 0xffffffff; + id_ns_z->mar =3D cpu_to_le32(ns->params.max_active_zones - 1); + id_ns_z->mor =3D cpu_to_le32(ns->params.max_open_zones - 1); id_ns_z->zoc =3D 0; id_ns_z->ozcs =3D ns->params.cross_zone_read ? 0x01 : 0x00; =20 @@ -195,6 +209,7 @@ static void nvme_clear_zone(NvmeNamespace *ns, NvmeZone= *zone) trace_pci_nvme_clear_ns_close(state, zone->d.zslba); nvme_set_zone_state(zone, NVME_ZONE_STATE_CLOSED); } + nvme_aor_inc_active(ns); QTAILQ_INSERT_HEAD(&ns->closed_zones, zone, entry); } else { trace_pci_nvme_clear_ns_reset(state, zone->d.zslba); @@ -211,16 +226,23 @@ static void nvme_zoned_ns_shutdown(NvmeNamespace *ns) =20 QTAILQ_FOREACH_SAFE(zone, &ns->closed_zones, entry, next) { QTAILQ_REMOVE(&ns->closed_zones, zone, entry); + nvme_aor_dec_active(ns); nvme_clear_zone(ns, zone); } QTAILQ_FOREACH_SAFE(zone, &ns->imp_open_zones, entry, next) { QTAILQ_REMOVE(&ns->imp_open_zones, zone, entry); + nvme_aor_dec_open(ns); + nvme_aor_dec_active(ns); nvme_clear_zone(ns, zone); } QTAILQ_FOREACH_SAFE(zone, &ns->exp_open_zones, entry, next) { QTAILQ_REMOVE(&ns->exp_open_zones, zone, entry); + nvme_aor_dec_open(ns); + nvme_aor_dec_active(ns); nvme_clear_zone(ns, zone); } + + assert(ns->nr_open_zones =3D=3D 0); } =20 static int nvme_ns_check_constraints(NvmeNamespace *ns, Error **errp) @@ -306,6 +328,10 @@ static Property nvme_ns_props[] =3D { DEFINE_PROP_SIZE("zoned.zcap", NvmeNamespace, params.zone_cap_bs, 0), DEFINE_PROP_BOOL("zoned.cross_read", NvmeNamespace, params.cross_zone_read, false), + DEFINE_PROP_UINT32("zoned.max_active", NvmeNamespace, + params.max_active_zones, 0), + DEFINE_PROP_UINT32("zoned.max_open", NvmeNamespace, + params.max_open_zones, 0), DEFINE_PROP_END_OF_LIST(), }; =20 diff --git a/hw/block/nvme.c b/hw/block/nvme.c index f5390e6863..c6b3a5dcf7 100644 --- a/hw/block/nvme.c +++ b/hw/block/nvme.c @@ -199,6 +199,26 @@ static void nvme_assign_zone_state(NvmeNamespace *ns, = NvmeZone *zone, } } =20 +/* + * Check if we can open a zone without exceeding open/active limits. + * AOR stands for "Active and Open Resources" (see TP 4053 section 2.5). + */ +static int nvme_aor_check(NvmeNamespace *ns, uint32_t act, uint32_t opn) +{ + if (ns->params.max_active_zones !=3D 0 && + ns->nr_active_zones + act > ns->params.max_active_zones) { + trace_pci_nvme_err_insuff_active_res(ns->params.max_active_zones); + return NVME_ZONE_TOO_MANY_ACTIVE | NVME_DNR; + } + if (ns->params.max_open_zones !=3D 0 && + ns->nr_open_zones + opn > ns->params.max_open_zones) { + trace_pci_nvme_err_insuff_open_res(ns->params.max_open_zones); + return NVME_ZONE_TOO_MANY_OPEN | NVME_DNR; + } + + return NVME_SUCCESS; +} + static bool nvme_addr_is_cmb(NvmeCtrl *n, hwaddr addr) { hwaddr low =3D n->ctrl_mem.addr; @@ -1193,6 +1213,41 @@ static uint16_t nvme_check_zone_read(NvmeNamespace *= ns, uint64_t slba, return status; } =20 +static void nvme_auto_transition_zone(NvmeNamespace *ns, bool implicit, + bool adding_active) +{ + NvmeZone *zone; + + if (implicit && ns->params.max_open_zones && + ns->nr_open_zones =3D=3D ns->params.max_open_zones) { + zone =3D QTAILQ_FIRST(&ns->imp_open_zones); + if (zone) { + /* + * Automatically close this implicitly open zone. + */ + QTAILQ_REMOVE(&ns->imp_open_zones, zone, entry); + nvme_aor_dec_open(ns); + nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_CLOSED); + } + } +} + +static uint16_t nvme_auto_open_zone(NvmeNamespace *ns, NvmeZone *zone) +{ + uint16_t status =3D NVME_SUCCESS; + uint8_t zs =3D nvme_get_zone_state(zone); + + if (zs =3D=3D NVME_ZONE_STATE_EMPTY) { + nvme_auto_transition_zone(ns, true, true); + status =3D nvme_aor_check(ns, 1, 1); + } else if (zs =3D=3D NVME_ZONE_STATE_CLOSED) { + nvme_auto_transition_zone(ns, true, false); + status =3D nvme_aor_check(ns, 0, 1); + } + + return status; +} + static bool nvme_finalize_zoned_write(NvmeNamespace *ns, NvmeRequest *req, bool failed) { @@ -1226,7 +1281,11 @@ static bool nvme_finalize_zoned_write(NvmeNamespace = *ns, NvmeRequest *req, switch (nvme_get_zone_state(zone)) { case NVME_ZONE_STATE_IMPLICITLY_OPEN: case NVME_ZONE_STATE_EXPLICITLY_OPEN: + nvme_aor_dec_open(ns); + /* fall through */ case NVME_ZONE_STATE_CLOSED: + nvme_aor_dec_active(ns); + /* fall through */ case NVME_ZONE_STATE_EMPTY: nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_FULL); /* fall through */ @@ -1255,7 +1314,10 @@ static uint64_t nvme_advance_zone_wp(NvmeNamespace *= ns, NvmeZone *zone, zs =3D nvme_get_zone_state(zone); switch (zs) { case NVME_ZONE_STATE_EMPTY: + nvme_aor_inc_active(ns); + /* fall through */ case NVME_ZONE_STATE_CLOSED: + nvme_aor_inc_open(ns); nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_IMPLICITLY_OP= EN); } } @@ -1449,6 +1511,11 @@ static uint16_t nvme_write(NvmeCtrl *n, NvmeRequest = *req, bool append, bool wrz) goto invalid; } =20 + status =3D nvme_auto_open_zone(ns, zone); + if (status !=3D NVME_SUCCESS) { + goto invalid; + } + if (append) { slba =3D zone->w_ptr; } @@ -1529,9 +1596,27 @@ enum NvmeZoneProcessingMask { static uint16_t nvme_open_zone(NvmeNamespace *ns, NvmeZone *zone, uint8_t state) { + uint16_t status; + switch (state) { case NVME_ZONE_STATE_EMPTY: + nvme_auto_transition_zone(ns, false, true); + status =3D nvme_aor_check(ns, 1, 0); + if (status !=3D NVME_SUCCESS) { + return status; + } + nvme_aor_inc_active(ns); + /* fall through */ case NVME_ZONE_STATE_CLOSED: + status =3D nvme_aor_check(ns, 0, 1); + if (status !=3D NVME_SUCCESS) { + if (state =3D=3D NVME_ZONE_STATE_EMPTY) { + nvme_aor_dec_active(ns); + } + return status; + } + nvme_aor_inc_open(ns); + /* fall through */ case NVME_ZONE_STATE_IMPLICITLY_OPEN: nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_EXPLICITLY_OPEN); /* fall through */ @@ -1548,6 +1633,7 @@ static uint16_t nvme_close_zone(NvmeNamespace *ns, Nv= meZone *zone, switch (state) { case NVME_ZONE_STATE_EXPLICITLY_OPEN: case NVME_ZONE_STATE_IMPLICITLY_OPEN: + nvme_aor_dec_open(ns); nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_CLOSED); /* fall through */ case NVME_ZONE_STATE_CLOSED: @@ -1563,7 +1649,11 @@ static uint16_t nvme_finish_zone(NvmeNamespace *ns, = NvmeZone *zone, switch (state) { case NVME_ZONE_STATE_EXPLICITLY_OPEN: case NVME_ZONE_STATE_IMPLICITLY_OPEN: + nvme_aor_dec_open(ns); + /* fall through */ case NVME_ZONE_STATE_CLOSED: + nvme_aor_dec_active(ns); + /* fall through */ case NVME_ZONE_STATE_EMPTY: zone->w_ptr =3D nvme_zone_wr_boundary(zone); zone->d.wp =3D zone->w_ptr; @@ -1582,7 +1672,11 @@ static uint16_t nvme_reset_zone(NvmeNamespace *ns, N= vmeZone *zone, switch (state) { case NVME_ZONE_STATE_EXPLICITLY_OPEN: case NVME_ZONE_STATE_IMPLICITLY_OPEN: + nvme_aor_dec_open(ns); + /* fall through */ case NVME_ZONE_STATE_CLOSED: + nvme_aor_dec_active(ns); + /* fall through */ case NVME_ZONE_STATE_FULL: zone->w_ptr =3D zone->d.zslba; zone->d.wp =3D zone->w_ptr; diff --git a/hw/block/trace-events b/hw/block/trace-events index 4d910bb942..e674522883 100644 --- a/hw/block/trace-events +++ b/hw/block/trace-events @@ -122,6 +122,8 @@ pci_nvme_err_append_not_at_start(uint64_t slba, uint64_= t zone) "appending at slb pci_nvme_err_zone_write_not_ok(uint64_t slba, uint32_t nlb, uint32_t statu= s) "slba=3D%"PRIu64", nlb=3D%"PRIu32", status=3D0x%"PRIx16"" pci_nvme_err_zone_read_not_ok(uint64_t slba, uint32_t nlb, uint32_t status= ) "slba=3D%"PRIu64", nlb=3D%"PRIu32", status=3D0x%"PRIx16"" pci_nvme_err_append_too_large(uint64_t slba, uint32_t nlb, uint8_t zasl) "= slba=3D%"PRIu64", nlb=3D%"PRIu32", zasl=3D%"PRIu8"" +pci_nvme_err_insuff_active_res(uint32_t max_active) "max_active=3D%"PRIu32= " zone limit exceeded" +pci_nvme_err_insuff_open_res(uint32_t max_open) "max_open=3D%"PRIu32" zone= limit exceeded" pci_nvme_err_invalid_iocsci(uint32_t idx) "unsupported command set combina= tion index %"PRIu32"" pci_nvme_err_invalid_del_sq(uint16_t qid) "invalid submission queue deleti= on, sid=3D%"PRIu16"" pci_nvme_err_invalid_create_sq_cqid(uint16_t cqid) "failed creating submis= sion queue, invalid cqid=3D%"PRIu16"" --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604707161; cv=none; d=zohomail.com; s=zohoarc; b=HfGrL0frH18BJvoVNtGTZ7JOePjf8NLHCdXmNz0FoDpGq09JbAN3Ow4GarSmmCvqgTm+gpJAH6912oWq50mhZIN2aLOkbjILArHo7aZIUlaEkIS0dAm81koZq2SNKugni3rD0jjx9FPytSvOePP5C85bi2nW+RemoudthD5gslY= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604707161; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=6WxBzt3N08mcJXTGCW73Di5y+IWTbCXnd3tscUQPlAY=; b=UX1G2O/hO8IMWXHtObKrReyQhR0Inp8V52Ajod4NjCJI5lN1MZNb89hagS8ZcKsNiXfN+KR5dGfsLCVQ86uWXQLLeD94KSB8nl08kNoLqolCiu/k4CThHi0cahBwgcsqNYQVQ1dvkZXeWHCMU0PqQnHSaZXQPIjaqw+bAsJsbrg= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604707161869626.2188869326235; Fri, 6 Nov 2020 15:59:21 -0800 (PST) Received: from localhost ([::1]:36200 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBdY-0004tC-Ou for importer@patchew.org; Fri, 06 Nov 2020 18:59:20 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34336) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOQ-0005qX-L5; Fri, 06 Nov 2020 18:43:42 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57360) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOO-0003Tq-Bj; Fri, 06 Nov 2020 18:43:42 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:39 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:30 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:36 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706221; x=1636242221; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=qPsMt6Py9z6j9y71gIC3n0L3IIyZmGFlYik7fCYTXcw=; b=F6/6rAXSFjkfD5TyhoR9oyp7A9a99DVy1PZVGwhb495nhJgz2bGReNPE V5g6412xdCtsU8WgSmZ8y9pzRgamgY2LP2CyKnbtPh2Ra53vKrmBLW+8u 0ZIkf9IOBm5G2dRlgQDrVQYlg4UkGOqt1mQv0YTIb8+ymU5FN5W6IFx4w f1xqroMGZUDQ/k/IUmoVhRxdk4A12k2aC+fkjqJmuVmgKUs5Xbdjw4fmJ 2vQWYbfswgNcndZEWhIzapWKKaSMReHx1ECuvtCzTj6nzTDaM94Lcu+9z 1BlnC3fji1fQdveu80xXoRNwYpwGymjCJ/46GN8IWnbESaEiFbEE3lA7O A==; IronPort-SDR: cThB30qgDrYNtxCuLvLkcCJAix5F58kmjeiXV/WvQHWQd2vHDygFgKT0q7qkXxe9HjjK71oxeC F4O3Mo5VeBZuJGMbFj1jPt3jjaLHo3D6IvbvYkVPtyaAb4LDptpx8yLYDs3EWGWowgZFVgMfX1 edy7Dz2TUDiPKHSiADkGTyu9XXV1ehdt300S6gzRXte+Afe/t5UTSQ6cXMlzhjGmUUjkLkBVHl jpplMde75xD3OD9vHE0JdwhIFRzbuTnrqg4+V55/F19GltvGGKATPYYXlI4EGVKewVWRW48o5v BN4= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267076" IronPort-SDR: k2JnEs3xjvEeSWr5uPezZxRfSO3JgEfWNF5rlyDvOhCp4V0nqiJoPqzQ63TYarTN3GHUyZcySA PwfpHqRBpExltMBpJI/IPwD92EH42A77ENU1iqmu0J9SA9lyU0IZgRDuS/93aujyS2e/8ySaef 8f7veeCNEw3EuFgmE/sLRAJPfd2E/EZ/dWmTO+DvS0+gN9X/cB108TRlIVc5DA5FgQgGJWk97I NF2tq5lYDKuwy0GN1IWHpXOC132IZuvO4/ZVN/TwsNIIbtwyxW2MHwvStMg04CPqf4Qg/VnlO0 ziYaVlihhO/YtNYFzJcDRWE6 IronPort-SDR: UwHjA+46+wJCteUNTumATB4AKoh5RIXR7tvY+9pNQo1VYpSM0QjXm9gaMS4z6VBRm2f+D9B99C 9I8matuxoFHnwWRgnAL57bFErEXVeiKOkLXfrHyWOr93qV2NtVICVIgc8fZXp1NMiCkKoWiaNB OHyko58UxeMi63rTaKuZjZGzXvJFxNcOVhy4Ou8fFYVlg3uanrRZ8Gpio2HvMq5m4PeNTh/pOS S2+38aUepr8pboiezQ3FkxDyBvrFMlffCWhzuxCsiBAgTUKdLRtRQ15ITthBsrkdGYeqL+PYtP XkE= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 10/12] hw/block/nvme: Support Zone Descriptor Extensions Date: Sat, 7 Nov 2020 08:43:03 +0900 Message-Id: <20201106234305.21339-11-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" Zone Descriptor Extension is a label that can be assigned to a zone. It can be set to an Empty zone and it stays assigned until the zone is reset. This commit adds a new optional module property, "zoned.descr_ext_size". Its value must be a multiple of 64 bytes. If this value is non-zero, it becomes possible to assign extensions of that size to any Empty zones. The default value for this property is 0, therefore setting extensions is disabled by default. Signed-off-by: Hans Holmberg Signed-off-by: Dmitry Fomichev Reviewed-by: Klaus Jensen Reviewed-by: Niklas Cassel --- hw/block/nvme-ns.h | 8 +++++++ hw/block/nvme-ns.c | 25 ++++++++++++++++++-- hw/block/nvme.c | 54 +++++++++++++++++++++++++++++++++++++++++-- hw/block/trace-events | 2 ++ 4 files changed, 85 insertions(+), 4 deletions(-) diff --git a/hw/block/nvme-ns.h b/hw/block/nvme-ns.h index 421bab0a57..50a6a0e1ac 100644 --- a/hw/block/nvme-ns.h +++ b/hw/block/nvme-ns.h @@ -35,6 +35,7 @@ typedef struct NvmeNamespaceParams { uint64_t zone_cap_bs; uint32_t max_active_zones; uint32_t max_open_zones; + uint32_t zd_extension_size; } NvmeNamespaceParams; =20 typedef struct NvmeNamespace { @@ -58,6 +59,7 @@ typedef struct NvmeNamespace { uint64_t zone_capacity; uint64_t zone_array_size; uint32_t zone_size_log2; + uint8_t *zd_extensions; int32_t nr_open_zones; int32_t nr_active_zones; =20 @@ -127,6 +129,12 @@ static inline bool nvme_wp_is_valid(NvmeZone *zone) st !=3D NVME_ZONE_STATE_OFFLINE; } =20 +static inline uint8_t *nvme_get_zd_extension(NvmeNamespace *ns, + uint32_t zone_idx) +{ + return &ns->zd_extensions[zone_idx * ns->params.zd_extension_size]; +} + static inline void nvme_aor_inc_open(NvmeNamespace *ns) { assert(ns->nr_open_zones >=3D 0); diff --git a/hw/block/nvme-ns.c b/hw/block/nvme-ns.c index 2e45838c15..85dc73cf06 100644 --- a/hw/block/nvme-ns.c +++ b/hw/block/nvme-ns.c @@ -133,6 +133,18 @@ static int nvme_calc_zone_geometry(NvmeNamespace *ns, = Error **errp) return -1; } =20 + if (ns->params.zd_extension_size) { + if (ns->params.zd_extension_size & 0x3f) { + error_setg(errp, + "zone descriptor extension size must be a multiple of 64B"= ); + return -1; + } + if ((ns->params.zd_extension_size >> 6) > 0xff) { + error_setg(errp, "zone descriptor extension size is too large"= ); + return -1; + } + } + return 0; } =20 @@ -144,6 +156,10 @@ static void nvme_init_zone_state(NvmeNamespace *ns) int i; =20 ns->zone_array =3D g_malloc0(ns->zone_array_size); + if (ns->params.zd_extension_size) { + ns->zd_extensions =3D g_malloc0(ns->params.zd_extension_size * + ns->num_zones); + } =20 QTAILQ_INIT(&ns->exp_open_zones); QTAILQ_INIT(&ns->imp_open_zones); @@ -186,7 +202,8 @@ static int nvme_zoned_init_ns(NvmeCtrl *n, NvmeNamespac= e *ns, int lba_index, id_ns_z->ozcs =3D ns->params.cross_zone_read ? 0x01 : 0x00; =20 id_ns_z->lbafe[lba_index].zsze =3D cpu_to_le64(ns->zone_size); - id_ns_z->lbafe[lba_index].zdes =3D 0; + id_ns_z->lbafe[lba_index].zdes =3D + ns->params.zd_extension_size >> 6; /* Units of 64B */ =20 ns->csi =3D NVME_CSI_ZONED; ns->id_ns.nsze =3D cpu_to_le64(ns->zone_size * ns->num_zones); @@ -204,7 +221,8 @@ static void nvme_clear_zone(NvmeNamespace *ns, NvmeZone= *zone) =20 zone->w_ptr =3D zone->d.wp; state =3D nvme_get_zone_state(zone); - if (zone->d.wp !=3D zone->d.zslba) { + if (zone->d.wp !=3D zone->d.zslba || + (zone->d.za & NVME_ZA_ZD_EXT_VALID)) { if (state !=3D NVME_ZONE_STATE_CLOSED) { trace_pci_nvme_clear_ns_close(state, zone->d.zslba); nvme_set_zone_state(zone, NVME_ZONE_STATE_CLOSED); @@ -301,6 +319,7 @@ void nvme_ns_cleanup(NvmeNamespace *ns) if (ns->params.zoned) { g_free(ns->id_ns_zoned); g_free(ns->zone_array); + g_free(ns->zd_extensions); } } =20 @@ -332,6 +351,8 @@ static Property nvme_ns_props[] =3D { params.max_active_zones, 0), DEFINE_PROP_UINT32("zoned.max_open", NvmeNamespace, params.max_open_zones, 0), + DEFINE_PROP_UINT32("zoned.descr_ext_size", NvmeNamespace, + params.zd_extension_size, 0), DEFINE_PROP_END_OF_LIST(), }; =20 diff --git a/hw/block/nvme.c b/hw/block/nvme.c index c6b3a5dcf7..e82e3be821 100644 --- a/hw/block/nvme.c +++ b/hw/block/nvme.c @@ -1703,6 +1703,26 @@ static uint16_t nvme_offline_zone(NvmeNamespace *ns,= NvmeZone *zone, return NVME_ZONE_INVAL_TRANSITION; } =20 +static uint16_t nvme_set_zd_ext(NvmeNamespace *ns, NvmeZone *zone) +{ + uint16_t status; + uint8_t state =3D nvme_get_zone_state(zone); + + if (state =3D=3D NVME_ZONE_STATE_EMPTY) { + nvme_auto_transition_zone(ns, false, true); + status =3D nvme_aor_check(ns, 1, 0); + if (status !=3D NVME_SUCCESS) { + return status; + } + nvme_aor_inc_active(ns); + zone->d.za |=3D NVME_ZA_ZD_EXT_VALID; + nvme_assign_zone_state(ns, zone, NVME_ZONE_STATE_CLOSED); + return NVME_SUCCESS; + } + + return NVME_ZONE_INVAL_TRANSITION; +} + static uint16_t nvme_bulk_proc_zone(NvmeNamespace *ns, NvmeZone *zone, enum NvmeZoneProcessingMask proc_mask, op_handler_t op_hndlr) @@ -1798,6 +1818,7 @@ static uint16_t nvme_zone_mgmt_send(NvmeCtrl *n, Nvme= Request *req) NvmeCmd *cmd =3D (NvmeCmd *)&req->cmd; NvmeNamespace *ns =3D req->ns; NvmeZone *zone; + uint8_t *zd_ext; uint32_t dw13 =3D le32_to_cpu(cmd->cdw13); uint64_t slba =3D 0; uint32_t zone_idx =3D 0; @@ -1870,7 +1891,22 @@ static uint16_t nvme_zone_mgmt_send(NvmeCtrl *n, Nvm= eRequest *req) =20 case NVME_ZONE_ACTION_SET_ZD_EXT: trace_pci_nvme_set_descriptor_extension(slba, zone_idx); - return NVME_INVALID_FIELD | NVME_DNR; + if (all || !ns->params.zd_extension_size) { + return NVME_INVALID_FIELD | NVME_DNR; + } + zd_ext =3D nvme_get_zd_extension(ns, zone_idx); + status =3D nvme_dma(n, zd_ext, ns->params.zd_extension_size, + DMA_DIRECTION_TO_DEVICE, req); + if (status) { + trace_pci_nvme_err_zd_extension_map_error(zone_idx); + return status; + } + + status =3D nvme_set_zd_ext(ns, zone); + if (status =3D=3D NVME_SUCCESS) { + trace_pci_nvme_zd_extension_set(zone_idx); + return status; + } break; =20 default: @@ -1940,7 +1976,10 @@ static uint16_t nvme_zone_mgmt_recv(NvmeCtrl *n, Nvm= eRequest *req) } =20 zra =3D dw13 & 0xff; - if (zra !=3D NVME_ZONE_REPORT) { + if (zra !=3D NVME_ZONE_REPORT && zra !=3D NVME_ZONE_REPORT_EXTENDED) { + return NVME_INVALID_FIELD | NVME_DNR; + } + if (zra =3D=3D NVME_ZONE_REPORT_EXTENDED && !ns->params.zd_extension_s= ize) { return NVME_INVALID_FIELD | NVME_DNR; } =20 @@ -1962,6 +2001,9 @@ static uint16_t nvme_zone_mgmt_recv(NvmeCtrl *n, Nvme= Request *req) partial =3D (dw13 >> 16) & 0x01; =20 zone_entry_sz =3D sizeof(NvmeZoneDescr); + if (zra =3D=3D NVME_ZONE_REPORT_EXTENDED) { + zone_entry_sz +=3D ns->params.zd_extension_size; + } =20 max_zones =3D (data_size - sizeof(NvmeZoneReportHeader)) / zone_entry_= sz; buf =3D g_malloc0(data_size); @@ -1993,6 +2035,14 @@ static uint16_t nvme_zone_mgmt_recv(NvmeCtrl *n, Nvm= eRequest *req) z->wp =3D cpu_to_le64(~0ULL); } =20 + if (zra =3D=3D NVME_ZONE_REPORT_EXTENDED) { + if (zs->d.za & NVME_ZA_ZD_EXT_VALID) { + memcpy(buf_p, nvme_get_zd_extension(ns, zone_idx), + ns->params.zd_extension_size); + } + buf_p +=3D ns->params.zd_extension_size; + } + zone_idx++; } =20 diff --git a/hw/block/trace-events b/hw/block/trace-events index e674522883..d42d2c8d61 100644 --- a/hw/block/trace-events +++ b/hw/block/trace-events @@ -95,6 +95,7 @@ pci_nvme_finish_zone(uint64_t slba, uint32_t zone_idx, in= t all) "finish zone, sl pci_nvme_reset_zone(uint64_t slba, uint32_t zone_idx, int all) "reset zone= , slba=3D%"PRIu64", idx=3D%"PRIu32", all=3D%"PRIi32"" pci_nvme_offline_zone(uint64_t slba, uint32_t zone_idx, int all) "offline = zone, slba=3D%"PRIu64", idx=3D%"PRIu32", all=3D%"PRIi32"" pci_nvme_set_descriptor_extension(uint64_t slba, uint32_t zone_idx) "set z= one descriptor extension, slba=3D%"PRIu64", idx=3D%"PRIu32"" +pci_nvme_zd_extension_set(uint32_t zone_idx) "set descriptor extension for= zone_idx=3D%"PRIu32"" pci_nvme_clear_ns_close(uint32_t state, uint64_t slba) "zone state=3D%"PRI= u32", slba=3D%"PRIu64" transitioned to Closed state" pci_nvme_clear_ns_reset(uint32_t state, uint64_t slba) "zone state=3D%"PRI= u32", slba=3D%"PRIu64" transitioned to Empty state" =20 @@ -124,6 +125,7 @@ pci_nvme_err_zone_read_not_ok(uint64_t slba, uint32_t n= lb, uint32_t status) "slb pci_nvme_err_append_too_large(uint64_t slba, uint32_t nlb, uint8_t zasl) "= slba=3D%"PRIu64", nlb=3D%"PRIu32", zasl=3D%"PRIu8"" pci_nvme_err_insuff_active_res(uint32_t max_active) "max_active=3D%"PRIu32= " zone limit exceeded" pci_nvme_err_insuff_open_res(uint32_t max_open) "max_open=3D%"PRIu32" zone= limit exceeded" +pci_nvme_err_zd_extension_map_error(uint32_t zone_idx) "can't map descript= or extension for zone_idx=3D%"PRIu32"" pci_nvme_err_invalid_iocsci(uint32_t idx) "unsupported command set combina= tion index %"PRIu32"" pci_nvme_err_invalid_del_sq(uint16_t qid) "invalid submission queue deleti= on, sid=3D%"PRIu16"" pci_nvme_err_invalid_create_sq_cqid(uint16_t cqid) "failed creating submis= sion queue, invalid cqid=3D%"PRIu16"" --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604706480; cv=none; d=zohomail.com; s=zohoarc; b=Df0+zIQDP1OYd+oqGke+4zIzjt+RPo7Ph/qf2Eto82q7X6QfaIbqPDAHaBRhP7GbCC7Gz0CFafTsz7uwSnNTkdiukF9grqCzKbVEC6PHCr7gtvOpMhU2szgJXen7QFkzhleN0O+Jvkfbe3yyz2BzknJooiWyPUyHNymadDSK7o8= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604706480; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=MCI/Zf06lifLWhS6TByDUSCXVaz4G4EWrLvPBWtejYE=; b=Sf+akOSUAEHmQUW+PEvWaqQDBMO1ibDWzYXw4CVQgRv9OmNjHPA1sHuSDH5TowyICALXCiO+RbdfBO8CYHZNOu3QnKY3MU2V4MXUfSDBu5FlPS+isRhCXeYRrpldd6Jna0MxN8aC9qF6KGhHWPGtFNuDjLFrwNqdmaBTj5sgPEY= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604706480657341.2899405060276; Fri, 6 Nov 2020 15:48:00 -0800 (PST) Received: from localhost ([::1]:40606 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBSZ-0002uG-Kh for importer@patchew.org; Fri, 06 Nov 2020 18:47:59 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34360) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOT-0005xm-81; Fri, 06 Nov 2020 18:43:45 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57360) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOR-0003Tq-C9; Fri, 06 Nov 2020 18:43:44 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:43 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:33 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:39 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706224; x=1636242224; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=Y+4Hkf4JH02Acl3UmgoBUtRDqe6KVXNmEtKOCZ+YtyY=; b=RM/LIEx0Yqji8g14RRsY2j7+NVTLqwgLoa+Sf5ZBQ8i1bX+Xp/UmFYoB gMvkxZX+HG5hiwZjduuaqxDif+atu7HTMiZNHOVwquuKy/9ksoLLFM8yE JoSJHiCNEnIuD21AbAdYvNOckI/GdUnc/sqf3yr0uZGPIUBGTs3SBYeOn YYcDo40XIuwQ52Z/HH0ql0WmJB/Kvqi/hRzqScX1U1BGiT/TTZJEZMSkb 0fZdv9FlH/Dnfk6R560XBHPUxjDpw87AUaBSNjS/6JjUMf4+vb+x4wWhb cN+hAsYKo4dRf4+lTq33mHVY8lIZxIPbrnIXXU8w8XeGRK/uD23O1KqrZ A==; IronPort-SDR: x/32uDi2W330jtptJ0cubV2YquQntA3pKj1GEZN1Ziob7ujTGezMRCCnpPykg99Bl1jnwvGYiw jRgHPn6j4DwOFyP4NC+VD8+HTH9mqCucnAefzfVCIOh6h3FVa2ygxMwjQlaPtnCxA24ZB1hCOE yCi0BfJRw5FtUbcU1DTktxyNEIlmJWzJMjtXWcauINqT2/czIO6YFW4Pgoaa7exXGOnH5nWxtg ax68dY+JFx3TgeamLaISAL2Zquj+RlFvzJyY6FBOt5g0EzArysomwGBkvpuyiWwMCshp8w6JxI 3FE= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267078" IronPort-SDR: bZc8qEBcqsLLt8u96Gg8bJg1RiHf+Scgxe/PbGGWNMfd2e5R1xBILCq7we0iVlPEEWr+Ifkekd AVr9XxHst8QBGxNX8P9cvGTdkKDlhOpTHZ5ZHZN6bdzZuArEhKldhND3UfSB0uo1Cjm2XBOMVU WCBgMuveTTIOkjiNcINCvrBeeBdVzGQebpO2bhbfYM5kj0L1fIrGUwcg4znZ/FvMIgfiAYxdF9 x05m+Ctr2cKp72ry0ztaWF837zqmk4HmsITwXp6+28m8jVdQs+/Hin4poerYcGtSou1FJ77Tfh MJtx0YWOOlZiH3JglQSsXa44 IronPort-SDR: WRG7ra1xj1qi5gbCGA/GawMVoRZ4iI3YvcLNL1tR1xwi69cqbCG8dcaP+Uh1DjBXVLHr2pHdzE VTL4L1Suc8/W99zmYkInJdYbRkPJBT9vNkYHG1C+Gz88Pvo+imJipZN+ZKXL/3dZm+WKF534X/ 7/hfOnSuIM2j+xDNouI+Ii7VTXJjOoBNguwOWoo/VA5P2PnwIJZrV8SDhjL/54XXCaE4qtcuPa ZUTT3F8E1KWtA2lBngk+SnMxoUErmxIvRYMa4R5a0ZnjkOLhEFkyRvyWKzJ8k8+1LaMRZlhsNV 09c= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 11/12] hw/block/nvme: Add injection of Offline/Read-Only zones Date: Sat, 7 Nov 2020 08:43:04 +0900 Message-Id: <20201106234305.21339-12-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" ZNS specification defines two zone conditions for the zones that no longer can function properly, possibly because of flash wear or other internal fault. It is useful to be able to "inject" a small number of such zones for testing purposes. This commit defines two optional device properties, "offline_zones" and "rdonly_zones". Users can assign non-zero values to these variables to specify the number of zones to be initialized as Offline or Read-Only. The actual number of injected zones may be smaller than the requested amount - Read-Only and Offline counts are expected to be much smaller than the total number of zones on a drive. Signed-off-by: Dmitry Fomichev Reviewed-by: Niklas Cassel --- hw/block/nvme-ns.h | 2 ++ hw/block/nvme-ns.c | 52 ++++++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 54 insertions(+) diff --git a/hw/block/nvme-ns.h b/hw/block/nvme-ns.h index 50a6a0e1ac..b30478e5d7 100644 --- a/hw/block/nvme-ns.h +++ b/hw/block/nvme-ns.h @@ -36,6 +36,8 @@ typedef struct NvmeNamespaceParams { uint32_t max_active_zones; uint32_t max_open_zones; uint32_t zd_extension_size; + uint32_t nr_offline_zones; + uint32_t nr_rdonly_zones; } NvmeNamespaceParams; =20 typedef struct NvmeNamespace { diff --git a/hw/block/nvme-ns.c b/hw/block/nvme-ns.c index 85dc73cf06..5e4a6705cd 100644 --- a/hw/block/nvme-ns.c +++ b/hw/block/nvme-ns.c @@ -21,6 +21,7 @@ #include "sysemu/sysemu.h" #include "sysemu/block-backend.h" #include "qapi/error.h" +#include "crypto/random.h" =20 #include "hw/qdev-properties.h" #include "hw/qdev-core.h" @@ -145,6 +146,20 @@ static int nvme_calc_zone_geometry(NvmeNamespace *ns, = Error **errp) } } =20 + if (ns->params.max_open_zones < nz) { + if (ns->params.nr_offline_zones > nz - ns->params.max_open_zones) { + error_setg(errp, "offline_zones value %u is too large", + ns->params.nr_offline_zones); + return -1; + } + if (ns->params.nr_rdonly_zones > + nz - ns->params.max_open_zones - ns->params.nr_offline_zones) { + error_setg(errp, "rdonly_zones value %u is too large", + ns->params.nr_rdonly_zones); + return -1; + } + } + return 0; } =20 @@ -153,7 +168,9 @@ static void nvme_init_zone_state(NvmeNamespace *ns) uint64_t start =3D 0, zone_size =3D ns->zone_size; uint64_t capacity =3D ns->num_zones * zone_size; NvmeZone *zone; + uint32_t rnd; int i; + uint16_t zs; =20 ns->zone_array =3D g_malloc0(ns->zone_array_size); if (ns->params.zd_extension_size) { @@ -180,6 +197,37 @@ static void nvme_init_zone_state(NvmeNamespace *ns) zone->w_ptr =3D start; start +=3D zone_size; } + + /* If required, make some zones Offline or Read Only */ + + for (i =3D 0; i < ns->params.nr_offline_zones; i++) { + do { + qcrypto_random_bytes(&rnd, sizeof(rnd), NULL); + rnd %=3D ns->num_zones; + } while (rnd < ns->params.max_open_zones); + zone =3D &ns->zone_array[rnd]; + zs =3D nvme_get_zone_state(zone); + if (zs !=3D NVME_ZONE_STATE_OFFLINE) { + nvme_set_zone_state(zone, NVME_ZONE_STATE_OFFLINE); + } else { + i--; + } + } + + for (i =3D 0; i < ns->params.nr_rdonly_zones; i++) { + do { + qcrypto_random_bytes(&rnd, sizeof(rnd), NULL); + rnd %=3D ns->num_zones; + } while (rnd < ns->params.max_open_zones); + zone =3D &ns->zone_array[rnd]; + zs =3D nvme_get_zone_state(zone); + if (zs !=3D NVME_ZONE_STATE_OFFLINE && + zs !=3D NVME_ZONE_STATE_READ_ONLY) { + nvme_set_zone_state(zone, NVME_ZONE_STATE_READ_ONLY); + } else { + i--; + } + } } =20 static int nvme_zoned_init_ns(NvmeCtrl *n, NvmeNamespace *ns, int lba_inde= x, @@ -353,6 +401,10 @@ static Property nvme_ns_props[] =3D { params.max_open_zones, 0), DEFINE_PROP_UINT32("zoned.descr_ext_size", NvmeNamespace, params.zd_extension_size, 0), + DEFINE_PROP_UINT32("zoned.offline_zones", NvmeNamespace, + params.nr_offline_zones, 0), + DEFINE_PROP_UINT32("zoned.rdonly_zones", NvmeNamespace, + params.nr_rdonly_zones, 0), DEFINE_PROP_END_OF_LIST(), }; =20 --=20 2.21.0 From nobody Sun Apr 28 00:44:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail(p=none dis=none) header.from=wdc.com ARC-Seal: i=1; a=rsa-sha256; t=1604706340; cv=none; d=zohomail.com; s=zohoarc; b=mx7jR3UGOpfbhVCMZ6+gIC9CfMCqTvOqttG/CoIf0q5CvVO79gqSyzzaHJLbrKN4jMbLvYLcYh2jE48mVHhiCVY6yoVxeHFtxtnt53UZWH7o4YtNXIri3jqzBwwLYiXbn/dI01cRmRkXOyybLPjtAsAtKP3pOmxvpddZHcFv2+s= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1604706340; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=eRDaHqn3hXiwX84CGz4BsT3N+zxQC0JHDTy79vwu2lw=; b=TlllI9DCyCZWNB6HhAbuFwePsBi68unv65vI9dSIYSMDINuDx2H/nIPLnZ8GTw0PvCTFX8m4MmGxT/Aot7y9JL4cweYrOxkf5FIT1b8X6HhIoeXHAFfXcUKr6k+iruosLL1E++ky5Fj3mS43419CgKcTnbPiW3GFVPxbdQHAlfw= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=fail header.i=@wdc.com; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=fail header.from= (p=none dis=none) header.from= Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1604706340745778.1796481731028; Fri, 6 Nov 2020 15:45:40 -0800 (PST) Received: from localhost ([::1]:34336 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kbBQJ-0000FS-HT for importer@patchew.org; Fri, 06 Nov 2020 18:45:39 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34394) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOW-00065z-7w; Fri, 06 Nov 2020 18:43:48 -0500 Received: from esa6.hgst.iphmx.com ([216.71.154.45]:57360) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kbBOT-0003Tq-Sl; Fri, 06 Nov 2020 18:43:47 -0500 Received: from h199-255-45-15.hgst.com (HELO uls-op-cesaep02.wdc.com) ([199.255.45.15]) by ob1.hgst.iphmx.com with ESMTP; 07 Nov 2020 07:43:45 +0800 Received: from uls-op-cesaip01.wdc.com ([10.248.3.36]) by uls-op-cesaep02.wdc.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Nov 2020 15:28:36 -0800 Received: from unknown (HELO redsun50.ssa.fujisawa.hgst.com) ([10.149.66.24]) by uls-op-cesaip01.wdc.com with ESMTP; 06 Nov 2020 15:43:42 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=wdc.com; i=@wdc.com; q=dns/txt; s=dkim.wdc.com; t=1604706226; x=1636242226; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=fYRGLGEOoxEBnGfF2tyEuYZSUoEdxDjp8wNk3+vvDGE=; b=HCQpyZ27xbVRX4rzWs+g6Oe0loLfjrPILTkps/JusEkw4h8UXVuWIzsX wOyED/iejWUHuKyFxB3AEoCz0HTNhdhx69KcRWo6fWnYpTFN1TKoNEQGB eMUewM/8fKeCxr5Ib2lR0T7KQG/sHYEeIL1C1EPvcaXFPNZeGgAZEsdD9 W51e17cYDGORAN6gbBXcj6gtJqUUG/wGthQGb6PgYXIsm8IzPIgAfsRYb iOqycUZ1+Mt3qGAdfzkwbA3MYgzO3s9qYFmmuwlePkpiSk4fBIxCpNlv7 NksUJ1wEQ9J8i3ldrLRoWNh3KXSgWONr3JbCUinM+/F5sXi7TTaxPqz5G g==; IronPort-SDR: 2NW0z+4hkOzEFlzaCSQPo2kmAm6/pL/LqWnhbyu5Gx99iT+LKG+drUZPA4dW9OrNn5Ep4yc6GK LrdJQ0c3F16DD1wPmdVz/cTDSj9xuY9J38el7o3I0EnAORw6S88owo3OF1IRo1SaLFy/TPFIoZ Lh2n1vPZ6jCO6hisaXZe9XyzyAyf/O/7u1hcCLFHAm2Vf5RW7ubRai1t81PwAkj+MzLowWUHak ZzDvJTCir+RzwzxkNubVD0JoUC+eOypNjy3YtwU5hH7Oewbb8sLlLylPziOilvngeublKN60hr SIE= X-IronPort-AV: E=Sophos;i="5.77,457,1596470400"; d="scan'208";a="153267080" IronPort-SDR: F/6FZxJQr88nyu/0O6RWpt4/wSn1FPxf8yyxdfkqEqSa9ns96RjkF6ZYK8FJzr9J4y5TlujZnY 4+YFyfOTTfij5MAbr64Z44dnwwsZ7cybvNMgcmzFmBaAlJzNKcFy4+HrGr+Mc7xsV+OivKNcRr 6WJOhlnkINvA4X4cx5m97VwvDCeG2mAV4hRSo1XVWLVPajzJGUo4X7fMArOwnrAhYOfA9m52dr IIgZ/VWw7G/DsO+MVRBmKhPsNeaE4iGPhLRCZBBUL8coK6OGdPaRkQG00zI12PXjukXi+90jVU q7symgHTHBLB/czHwYt+7RPW IronPort-SDR: 1tqXDhHyaKLSVGTR1PEtg1K6N3iblJH1/I+w4I93OFJLIsq6wLzQSSwOlMZDmINfHpeL4aUUXc Tc3YFDzGBK6+IQpC0iE1eqNv0F3BpX9ijAu4qjxP9oeaKVe1KT4P4y69tWrB2wzgtCoYbB1KIL gwcnyLrYp506zpA2afOnYQTOp7AzA3NcSrDDCaJMi5TkFULUxjD6RIeqQqbdQDrCtmj89y9tV0 133z2/sXgp1lmF5uFKLBvQWk+R/qGx3N0TnfM/RtqbRTBbBIY7q82F+/94jD9qpJmKiiovZmQ7 XQM= WDCIronportException: Internal From: Dmitry Fomichev To: Keith Busch , Klaus Jensen , Kevin Wolf , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Max Reitz , Maxim Levitsky , Fam Zheng Subject: [PATCH v10 12/12] hw/block/nvme: Document zoned parameters in usage text Date: Sat, 7 Nov 2020 08:43:05 +0900 Message-Id: <20201106234305.21339-13-dmitry.fomichev@wdc.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20201106234305.21339-1-dmitry.fomichev@wdc.com> References: <20201106234305.21339-1-dmitry.fomichev@wdc.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=216.71.154.45; envelope-from=prvs=572b21b8d=dmitry.fomichev@wdc.com; helo=esa6.hgst.iphmx.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/11/06 18:43:12 X-ACL-Warn: Detected OS = FreeBSD 9.x or newer [fuzzy] X-Spam_score_int: -43 X-Spam_score: -4.4 X-Spam_bar: ---- X-Spam_report: (-4.4 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Niklas Cassel , Damien Le Moal , qemu-block@nongnu.org, Dmitry Fomichev , qemu-devel@nongnu.org, Alistair Francis , Matias Bjorling Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: "Qemu-devel" X-ZohoMail-DKIM: fail (Header signature does not verify) Content-Type: text/plain; charset="utf-8" Added brief descriptions of the new device properties that are now available to users to configure features of Zoned Namespace Command Set in the emulator. This patch is for documentation only, no functionality change. Signed-off-by: Dmitry Fomichev Reviewed-by: Niklas Cassel --- hw/block/nvme.c | 47 ++++++++++++++++++++++++++++++++++++++++++----- 1 file changed, 42 insertions(+), 5 deletions(-) diff --git a/hw/block/nvme.c b/hw/block/nvme.c index e82e3be821..6043f95116 100644 --- a/hw/block/nvme.c +++ b/hw/block/nvme.c @@ -9,7 +9,7 @@ */ =20 /** - * Reference Specs: http://www.nvmexpress.org, 1.2, 1.1, 1.0e + * Reference Specs: http://www.nvmexpress.org, 1.4, 1.3, 1.2, 1.1, 1.0e * * https://nvmexpress.org/developers/nvme-specification/ */ @@ -22,8 +22,9 @@ * [pmrdev=3D,] \ * max_ioqpairs=3D, \ * aerl=3D, aer_max_queued=3D, \ - * mdts=3D - * -device nvme-ns,drive=3D,bus=3Dbus_name,nsid=3D + * mdts=3D,zoned.append_size_limit=3D \ + * -device nvme-ns,drive=3D,bus=3D,nsid=3D,\ + * zoned=3D * * Note cmb_size_mb denotes size of CMB in MB. CMB is assumed to be at * offset 0 in BAR2 and supports only WDS, RDS and SQS for now. @@ -41,14 +42,50 @@ * ~~~~~~~~~~~~~~~~~~~~~~ * - `aerl` * The Asynchronous Event Request Limit (AERL). Indicates the maximum nu= mber - * of concurrently outstanding Asynchronous Event Request commands suppo= ert + * of concurrently outstanding Asynchronous Event Request commands suppo= rt * by the controller. This is a 0's based value. * * - `aer_max_queued` * This is the maximum number of events that the device will enqueue for - * completion when there are no oustanding AERs. When the maximum number= of + * completion when there are no outstanding AERs. When the maximum numbe= r of * enqueued events are reached, subsequent events will be dropped. * + * - `zoned.append_size_limit` + * The maximum I/O size in bytes that is allowed in Zone Append command. + * The default is 128KiB. Since internally this this value is maintained= as + * ZASL =3D log2( / ), some values assig= ned + * to this property may be rounded down and result in a lower maximum ZA + * data size being in effect. By setting this property to 0, users can m= ake + * ZASL to be equal to MDTS. This property only affects zoned namespaces. + * + * Setting `zoned` to true selects Zoned Command Set at the namespace. + * In this case, the following namespace properties are available to confi= gure + * zoned operation: + * zoned.zsze=3D + * The number may be followed by K, M, G as in kilo-, mega- or gig= a-. + * + * zoned.zcap=3D + * The value 0 (default) forces zone capacity to be the same as zo= ne + * size. The value of this property may not exceed zone size. + * + * zoned.descr_ext_size=3D + * This value needs to be specified in 64B units. If it is zero, + * namespace(s) will not support zone descriptor extensions. + * + * zoned.max_active=3D + * The default value means there is no limit to the number of + * concurrently active zones. + * + * zoned.max_open=3D + * The default value means there is no limit to the number of + * concurrently open zones. + * + * zoned.offline_zones=3D + * + * zoned.rdonly_zones=3D + * + * zoned.cross_zone_read=3D + * Setting this property to true enables Read Across Zone Boundari= es. */ =20 #include "qemu/osdep.h" --=20 2.21.0