1
The following changes since commit ac793156f650ae2d77834932d72224175ee69086:
1
The following changes since commit 813bac3d8d70d85cb7835f7945eb9eed84c2d8d0:
2
2
3
Merge remote-tracking branch 'remotes/pmaydell/tags/pull-target-arm-20201020-1' into staging (2020-10-20 21:11:35 +0100)
3
Merge tag '2023q3-bsd-user-pull-request' of https://gitlab.com/bsdimp/qemu into staging (2023-08-29 08:58:00 -0400)
4
4
5
are available in the Git repository at:
5
are available in the Git repository at:
6
6
7
https://gitlab.com/stefanha/qemu.git tags/block-pull-request
7
https://gitlab.com/stefanha/qemu.git tags/block-pull-request
8
8
9
for you to fetch changes up to 32a3fd65e7e3551337fd26bfc0e2f899d70c028c:
9
for you to fetch changes up to 3f5f2285bfcdd855508a55da7875fb92de1a6ed0:
10
10
11
iotests: add commit top->base cases to 274 (2020-10-22 09:55:39 +0100)
11
tests/qemu-iotests/197: add testcase for CoR with subclusters (2023-08-29 13:19:56 -0400)
12
12
13
----------------------------------------------------------------
13
----------------------------------------------------------------
14
Pull request
14
Pull request
15
15
16
v2:
16
v2:
17
* Fix format string issues on 32-bit hosts [Peter]
17
- Fix authorship information lost by the mailing list for Andrey Drobyshev
18
* Fix qemu-nbd.c CONFIG_POSIX ifdef issue [Eric]
19
* Fix missing eventfd.h header on macOS [Peter]
20
* Drop unreliable vhost-user-blk test (will send a new patch when ready) [Peter]
21
22
This pull request contains the vhost-user-blk server by Coiby Xu along with my
23
additions, block/nvme.c alignment and hardware error statistics by Philippe
24
Mathieu-Daudé, and bdrv_co_block_status_above() fixes by Vladimir
25
Sementsov-Ogievskiy.
26
18
27
----------------------------------------------------------------
19
----------------------------------------------------------------
28
20
29
Coiby Xu (6):
21
Andrey Drobyshev (3):
30
libvhost-user: Allow vu_message_read to be replaced
22
block: add subcluster_size field to BlockDriverInfo
31
libvhost-user: remove watch for kick_fd when de-initialize vu-dev
23
block/io: align requests to subcluster_size
32
util/vhost-user-server: generic vhost user server
24
tests/qemu-iotests/197: add testcase for CoR with subclusters
33
block: move logical block size check function to a common utility
34
function
35
block/export: vhost-user block device backend server
36
MAINTAINERS: Add vhost-user block device backend server maintainer
37
25
38
Philippe Mathieu-Daudé (1):
26
Fabiano Rosas (1):
39
block/nvme: Add driver statistics for access alignment and hw errors
27
block-migration: Ensure we don't crash during migration cleanup
40
28
41
Stefan Hajnoczi (16):
29
Jeuk Kim (4):
42
util/vhost-user-server: s/fileds/fields/ typo fix
30
hw/ufs: Initial commit for emulated Universal-Flash-Storage
43
util/vhost-user-server: drop unnecessary QOM cast
31
hw/ufs: Support for Query Transfer Requests
44
util/vhost-user-server: drop unnecessary watch deletion
32
hw/ufs: Support for UFS logical unit
45
block/export: consolidate request structs into VuBlockReq
33
tests/qtest: Introduce tests for UFS
46
util/vhost-user-server: drop unused DevicePanicNotifier
47
util/vhost-user-server: fix memory leak in vu_message_read()
48
util/vhost-user-server: check EOF when reading payload
49
util/vhost-user-server: rework vu_client_trip() coroutine lifecycle
50
block/export: report flush errors
51
block/export: convert vhost-user-blk server to block export API
52
util/vhost-user-server: move header to include/
53
util/vhost-user-server: use static library in meson.build
54
qemu-storage-daemon: avoid compiling blockdev_ss twice
55
block: move block exports to libblockdev
56
block/export: add iothread and fixed-iothread options
57
block/export: add vhost-user-blk multi-queue support
58
34
59
Vladimir Sementsov-Ogievskiy (5):
35
MAINTAINERS | 7 +
60
block/io: fix bdrv_co_block_status_above
36
docs/specs/pci-ids.rst | 2 +
61
block/io: bdrv_common_block_status_above: support include_base
37
meson.build | 1 +
62
block/io: bdrv_common_block_status_above: support bs == base
38
hw/ufs/trace.h | 1 +
63
block/io: fix bdrv_is_allocated_above
39
hw/ufs/ufs.h | 131 +++
64
iotests: add commit top->base cases to 274
40
include/block/block-common.h | 5 +
65
41
include/block/block-io.h | 8 +-
66
MAINTAINERS | 9 +
42
include/block/ufs.h | 1090 +++++++++++++++++++++++++
67
qapi/block-core.json | 24 +-
43
include/hw/pci/pci.h | 1 +
68
qapi/block-export.json | 36 +-
44
include/hw/pci/pci_ids.h | 1 +
69
block/coroutines.h | 2 +
45
include/scsi/constants.h | 1 +
70
block/export/vhost-user-blk-server.h | 19 +
46
block.c | 7 +
71
contrib/libvhost-user/libvhost-user.h | 21 +
47
block/io.c | 50 +-
72
include/qemu/vhost-user-server.h | 65 +++
48
block/mirror.c | 8 +-
73
util/block-helpers.h | 19 +
49
block/qcow2.c | 1 +
74
block/export/export.c | 37 +-
50
hw/ufs/lu.c | 1445 ++++++++++++++++++++++++++++++++
75
block/export/vhost-user-blk-server.c | 431 ++++++++++++++++++++
51
hw/ufs/ufs.c | 1494 ++++++++++++++++++++++++++++++++++
76
block/io.c | 132 +++---
52
migration/block.c | 11 +-
77
block/nvme.c | 27 ++
53
tests/qtest/ufs-test.c | 584 +++++++++++++
78
block/qcow2.c | 16 +-
54
hw/Kconfig | 1 +
79
contrib/libvhost-user/libvhost-user-glib.c | 2 +-
55
hw/meson.build | 1 +
80
contrib/libvhost-user/libvhost-user.c | 15 +-
56
hw/ufs/Kconfig | 4 +
81
hw/core/qdev-properties-system.c | 31 +-
57
hw/ufs/meson.build | 1 +
82
nbd/server.c | 2 -
58
hw/ufs/trace-events | 58 ++
83
qemu-nbd.c | 21 +-
59
tests/qemu-iotests/197 | 29 +
84
softmmu/vl.c | 4 +
60
tests/qemu-iotests/197.out | 24 +
85
stubs/blk-exp-close-all.c | 7 +
61
tests/qtest/meson.build | 1 +
86
tests/vhost-user-bridge.c | 2 +
62
27 files changed, 4932 insertions(+), 35 deletions(-)
87
tools/virtiofsd/fuse_virtio.c | 4 +-
63
create mode 100644 hw/ufs/trace.h
88
util/block-helpers.c | 46 +++
64
create mode 100644 hw/ufs/ufs.h
89
util/vhost-user-server.c | 446 +++++++++++++++++++++
65
create mode 100644 include/block/ufs.h
90
block/export/meson.build | 3 +-
66
create mode 100644 hw/ufs/lu.c
91
contrib/libvhost-user/meson.build | 1 +
67
create mode 100644 hw/ufs/ufs.c
92
meson.build | 22 +-
68
create mode 100644 tests/qtest/ufs-test.c
93
nbd/meson.build | 2 +
69
create mode 100644 hw/ufs/Kconfig
94
storage-daemon/meson.build | 3 +-
70
create mode 100644 hw/ufs/meson.build
95
stubs/meson.build | 1 +
71
create mode 100644 hw/ufs/trace-events
96
tests/qemu-iotests/274 | 20 +
97
tests/qemu-iotests/274.out | 68 ++++
98
util/meson.build | 4 +
99
33 files changed, 1420 insertions(+), 122 deletions(-)
100
create mode 100644 block/export/vhost-user-blk-server.h
101
create mode 100644 include/qemu/vhost-user-server.h
102
create mode 100644 util/block-helpers.h
103
create mode 100644 block/export/vhost-user-blk-server.c
104
create mode 100644 stubs/blk-exp-close-all.c
105
create mode 100644 util/block-helpers.c
106
create mode 100644 util/vhost-user-server.c
107
72
108
--
73
--
109
2.26.2
74
2.41.0
110
diff view generated by jsdifflib
Deleted patch
1
From: Philippe Mathieu-Daudé <philmd@redhat.com>
2
1
3
Keep statistics of some hardware errors, and number of
4
aligned/unaligned I/O accesses.
5
6
QMP example booting a full RHEL 8.3 aarch64 guest:
7
8
{ "execute": "query-blockstats" }
9
{
10
"return": [
11
{
12
"device": "",
13
"node-name": "drive0",
14
"stats": {
15
"flush_total_time_ns": 6026948,
16
"wr_highest_offset": 3383991230464,
17
"wr_total_time_ns": 807450995,
18
"failed_wr_operations": 0,
19
"failed_rd_operations": 0,
20
"wr_merged": 3,
21
"wr_bytes": 50133504,
22
"failed_unmap_operations": 0,
23
"failed_flush_operations": 0,
24
"account_invalid": false,
25
"rd_total_time_ns": 1846979900,
26
"flush_operations": 130,
27
"wr_operations": 659,
28
"rd_merged": 1192,
29
"rd_bytes": 218244096,
30
"account_failed": false,
31
"idle_time_ns": 2678641497,
32
"rd_operations": 7406,
33
},
34
"driver-specific": {
35
"driver": "nvme",
36
"completion-errors": 0,
37
"unaligned-accesses": 2959,
38
"aligned-accesses": 4477
39
},
40
"qdev": "/machine/peripheral-anon/device[0]/virtio-backend"
41
}
42
]
43
}
44
45
Suggested-by: Stefan Hajnoczi <stefanha@gmail.com>
46
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
47
Acked-by: Markus Armbruster <armbru@redhat.com>
48
Message-id: 20201001162939.1567915-1-philmd@redhat.com
49
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
50
---
51
qapi/block-core.json | 24 +++++++++++++++++++++++-
52
block/nvme.c | 27 +++++++++++++++++++++++++++
53
2 files changed, 50 insertions(+), 1 deletion(-)
54
55
diff --git a/qapi/block-core.json b/qapi/block-core.json
56
index XXXXXXX..XXXXXXX 100644
57
--- a/qapi/block-core.json
58
+++ b/qapi/block-core.json
59
@@ -XXX,XX +XXX,XX @@
60
'discard-nb-failed': 'uint64',
61
'discard-bytes-ok': 'uint64' } }
62
63
+##
64
+# @BlockStatsSpecificNvme:
65
+#
66
+# NVMe driver statistics
67
+#
68
+# @completion-errors: The number of completion errors.
69
+#
70
+# @aligned-accesses: The number of aligned accesses performed by
71
+# the driver.
72
+#
73
+# @unaligned-accesses: The number of unaligned accesses performed by
74
+# the driver.
75
+#
76
+# Since: 5.2
77
+##
78
+{ 'struct': 'BlockStatsSpecificNvme',
79
+ 'data': {
80
+ 'completion-errors': 'uint64',
81
+ 'aligned-accesses': 'uint64',
82
+ 'unaligned-accesses': 'uint64' } }
83
+
84
##
85
# @BlockStatsSpecific:
86
#
87
@@ -XXX,XX +XXX,XX @@
88
'discriminator': 'driver',
89
'data': {
90
'file': 'BlockStatsSpecificFile',
91
- 'host_device': 'BlockStatsSpecificFile' } }
92
+ 'host_device': 'BlockStatsSpecificFile',
93
+ 'nvme': 'BlockStatsSpecificNvme' } }
94
95
##
96
# @BlockStats:
97
diff --git a/block/nvme.c b/block/nvme.c
98
index XXXXXXX..XXXXXXX 100644
99
--- a/block/nvme.c
100
+++ b/block/nvme.c
101
@@ -XXX,XX +XXX,XX @@ struct BDRVNVMeState {
102
103
/* PCI address (required for nvme_refresh_filename()) */
104
char *device;
105
+
106
+ struct {
107
+ uint64_t completion_errors;
108
+ uint64_t aligned_accesses;
109
+ uint64_t unaligned_accesses;
110
+ } stats;
111
};
112
113
#define NVME_BLOCK_OPT_DEVICE "device"
114
@@ -XXX,XX +XXX,XX @@ static bool nvme_process_completion(NVMeQueuePair *q)
115
break;
116
}
117
ret = nvme_translate_error(c);
118
+ if (ret) {
119
+ s->stats.completion_errors++;
120
+ }
121
q->cq.head = (q->cq.head + 1) % NVME_QUEUE_SIZE;
122
if (!q->cq.head) {
123
q->cq_phase = !q->cq_phase;
124
@@ -XXX,XX +XXX,XX @@ static int nvme_co_prw(BlockDriverState *bs, uint64_t offset, uint64_t bytes,
125
assert(QEMU_IS_ALIGNED(bytes, s->page_size));
126
assert(bytes <= s->max_transfer);
127
if (nvme_qiov_aligned(bs, qiov)) {
128
+ s->stats.aligned_accesses++;
129
return nvme_co_prw_aligned(bs, offset, bytes, qiov, is_write, flags);
130
}
131
+ s->stats.unaligned_accesses++;
132
trace_nvme_prw_buffered(s, offset, bytes, qiov->niov, is_write);
133
buf = qemu_try_memalign(s->page_size, bytes);
134
135
@@ -XXX,XX +XXX,XX @@ static void nvme_unregister_buf(BlockDriverState *bs, void *host)
136
qemu_vfio_dma_unmap(s->vfio, host);
137
}
138
139
+static BlockStatsSpecific *nvme_get_specific_stats(BlockDriverState *bs)
140
+{
141
+ BlockStatsSpecific *stats = g_new(BlockStatsSpecific, 1);
142
+ BDRVNVMeState *s = bs->opaque;
143
+
144
+ stats->driver = BLOCKDEV_DRIVER_NVME;
145
+ stats->u.nvme = (BlockStatsSpecificNvme) {
146
+ .completion_errors = s->stats.completion_errors,
147
+ .aligned_accesses = s->stats.aligned_accesses,
148
+ .unaligned_accesses = s->stats.unaligned_accesses,
149
+ };
150
+
151
+ return stats;
152
+}
153
+
154
static const char *const nvme_strong_runtime_opts[] = {
155
NVME_BLOCK_OPT_DEVICE,
156
NVME_BLOCK_OPT_NAMESPACE,
157
@@ -XXX,XX +XXX,XX @@ static BlockDriver bdrv_nvme = {
158
.bdrv_refresh_filename = nvme_refresh_filename,
159
.bdrv_refresh_limits = nvme_refresh_limits,
160
.strong_runtime_opts = nvme_strong_runtime_opts,
161
+ .bdrv_get_specific_stats = nvme_get_specific_stats,
162
163
.bdrv_detach_aio_context = nvme_detach_aio_context,
164
.bdrv_attach_aio_context = nvme_attach_aio_context,
165
--
166
2.26.2
167
diff view generated by jsdifflib
Deleted patch
1
From: Coiby Xu <coiby.xu@gmail.com>
2
1
3
Allow vu_message_read to be replaced by one which will make use of the
4
QIOChannel functions. Thus reading vhost-user message won't stall the
5
guest. For slave channel, we still use the default vu_message_read.
6
7
Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com>
8
Signed-off-by: Coiby Xu <coiby.xu@gmail.com>
9
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
10
Message-id: 20200918080912.321299-2-coiby.xu@gmail.com
11
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
12
---
13
contrib/libvhost-user/libvhost-user.h | 21 +++++++++++++++++++++
14
contrib/libvhost-user/libvhost-user-glib.c | 2 +-
15
contrib/libvhost-user/libvhost-user.c | 14 +++++++-------
16
tests/vhost-user-bridge.c | 2 ++
17
tools/virtiofsd/fuse_virtio.c | 4 ++--
18
5 files changed, 33 insertions(+), 10 deletions(-)
19
20
diff --git a/contrib/libvhost-user/libvhost-user.h b/contrib/libvhost-user/libvhost-user.h
21
index XXXXXXX..XXXXXXX 100644
22
--- a/contrib/libvhost-user/libvhost-user.h
23
+++ b/contrib/libvhost-user/libvhost-user.h
24
@@ -XXX,XX +XXX,XX @@
25
*/
26
#define VHOST_USER_MAX_RAM_SLOTS 32
27
28
+#define VHOST_USER_HDR_SIZE offsetof(VhostUserMsg, payload.u64)
29
+
30
typedef enum VhostSetConfigType {
31
VHOST_SET_CONFIG_TYPE_MASTER = 0,
32
VHOST_SET_CONFIG_TYPE_MIGRATION = 1,
33
@@ -XXX,XX +XXX,XX @@ typedef uint64_t (*vu_get_features_cb) (VuDev *dev);
34
typedef void (*vu_set_features_cb) (VuDev *dev, uint64_t features);
35
typedef int (*vu_process_msg_cb) (VuDev *dev, VhostUserMsg *vmsg,
36
int *do_reply);
37
+typedef bool (*vu_read_msg_cb) (VuDev *dev, int sock, VhostUserMsg *vmsg);
38
typedef void (*vu_queue_set_started_cb) (VuDev *dev, int qidx, bool started);
39
typedef bool (*vu_queue_is_processed_in_order_cb) (VuDev *dev, int qidx);
40
typedef int (*vu_get_config_cb) (VuDev *dev, uint8_t *config, uint32_t len);
41
@@ -XXX,XX +XXX,XX @@ struct VuDev {
42
bool broken;
43
uint16_t max_queues;
44
45
+ /* @read_msg: custom method to read vhost-user message
46
+ *
47
+ * Read data from vhost_user socket fd and fill up
48
+ * the passed VhostUserMsg *vmsg struct.
49
+ *
50
+ * If reading fails, it should close the received set of file
51
+ * descriptors as socket message's auxiliary data.
52
+ *
53
+ * For the details, please refer to vu_message_read in libvhost-user.c
54
+ * which will be used by default if not custom method is provided when
55
+ * calling vu_init
56
+ *
57
+ * Returns: true if vhost-user message successfully received,
58
+ * otherwise return false.
59
+ *
60
+ */
61
+ vu_read_msg_cb read_msg;
62
/* @set_watch: add or update the given fd to the watch set,
63
* call cb when condition is met */
64
vu_set_watch_cb set_watch;
65
@@ -XXX,XX +XXX,XX @@ bool vu_init(VuDev *dev,
66
uint16_t max_queues,
67
int socket,
68
vu_panic_cb panic,
69
+ vu_read_msg_cb read_msg,
70
vu_set_watch_cb set_watch,
71
vu_remove_watch_cb remove_watch,
72
const VuDevIface *iface);
73
diff --git a/contrib/libvhost-user/libvhost-user-glib.c b/contrib/libvhost-user/libvhost-user-glib.c
74
index XXXXXXX..XXXXXXX 100644
75
--- a/contrib/libvhost-user/libvhost-user-glib.c
76
+++ b/contrib/libvhost-user/libvhost-user-glib.c
77
@@ -XXX,XX +XXX,XX @@ vug_init(VugDev *dev, uint16_t max_queues, int socket,
78
g_assert(dev);
79
g_assert(iface);
80
81
- if (!vu_init(&dev->parent, max_queues, socket, panic, set_watch,
82
+ if (!vu_init(&dev->parent, max_queues, socket, panic, NULL, set_watch,
83
remove_watch, iface)) {
84
return false;
85
}
86
diff --git a/contrib/libvhost-user/libvhost-user.c b/contrib/libvhost-user/libvhost-user.c
87
index XXXXXXX..XXXXXXX 100644
88
--- a/contrib/libvhost-user/libvhost-user.c
89
+++ b/contrib/libvhost-user/libvhost-user.c
90
@@ -XXX,XX +XXX,XX @@
91
/* The version of inflight buffer */
92
#define INFLIGHT_VERSION 1
93
94
-#define VHOST_USER_HDR_SIZE offsetof(VhostUserMsg, payload.u64)
95
-
96
/* The version of the protocol we support */
97
#define VHOST_USER_VERSION 1
98
#define LIBVHOST_USER_DEBUG 0
99
@@ -XXX,XX +XXX,XX @@ have_userfault(void)
100
}
101
102
static bool
103
-vu_message_read(VuDev *dev, int conn_fd, VhostUserMsg *vmsg)
104
+vu_message_read_default(VuDev *dev, int conn_fd, VhostUserMsg *vmsg)
105
{
106
char control[CMSG_SPACE(VHOST_MEMORY_BASELINE_NREGIONS * sizeof(int))] = {};
107
struct iovec iov = {
108
@@ -XXX,XX +XXX,XX @@ vu_process_message_reply(VuDev *dev, const VhostUserMsg *vmsg)
109
goto out;
110
}
111
112
- if (!vu_message_read(dev, dev->slave_fd, &msg_reply)) {
113
+ if (!vu_message_read_default(dev, dev->slave_fd, &msg_reply)) {
114
goto out;
115
}
116
117
@@ -XXX,XX +XXX,XX @@ vu_set_mem_table_exec_postcopy(VuDev *dev, VhostUserMsg *vmsg)
118
/* Wait for QEMU to confirm that it's registered the handler for the
119
* faults.
120
*/
121
- if (!vu_message_read(dev, dev->sock, vmsg) ||
122
+ if (!dev->read_msg(dev, dev->sock, vmsg) ||
123
vmsg->size != sizeof(vmsg->payload.u64) ||
124
vmsg->payload.u64 != 0) {
125
vu_panic(dev, "failed to receive valid ack for postcopy set-mem-table");
126
@@ -XXX,XX +XXX,XX @@ vu_dispatch(VuDev *dev)
127
int reply_requested;
128
bool need_reply, success = false;
129
130
- if (!vu_message_read(dev, dev->sock, &vmsg)) {
131
+ if (!dev->read_msg(dev, dev->sock, &vmsg)) {
132
goto end;
133
}
134
135
@@ -XXX,XX +XXX,XX @@ vu_init(VuDev *dev,
136
uint16_t max_queues,
137
int socket,
138
vu_panic_cb panic,
139
+ vu_read_msg_cb read_msg,
140
vu_set_watch_cb set_watch,
141
vu_remove_watch_cb remove_watch,
142
const VuDevIface *iface)
143
@@ -XXX,XX +XXX,XX @@ vu_init(VuDev *dev,
144
145
dev->sock = socket;
146
dev->panic = panic;
147
+ dev->read_msg = read_msg ? read_msg : vu_message_read_default;
148
dev->set_watch = set_watch;
149
dev->remove_watch = remove_watch;
150
dev->iface = iface;
151
@@ -XXX,XX +XXX,XX @@ static void _vu_queue_notify(VuDev *dev, VuVirtq *vq, bool sync)
152
153
vu_message_write(dev, dev->slave_fd, &vmsg);
154
if (ack) {
155
- vu_message_read(dev, dev->slave_fd, &vmsg);
156
+ vu_message_read_default(dev, dev->slave_fd, &vmsg);
157
}
158
return;
159
}
160
diff --git a/tests/vhost-user-bridge.c b/tests/vhost-user-bridge.c
161
index XXXXXXX..XXXXXXX 100644
162
--- a/tests/vhost-user-bridge.c
163
+++ b/tests/vhost-user-bridge.c
164
@@ -XXX,XX +XXX,XX @@ vubr_accept_cb(int sock, void *ctx)
165
VHOST_USER_BRIDGE_MAX_QUEUES,
166
conn_fd,
167
vubr_panic,
168
+ NULL,
169
vubr_set_watch,
170
vubr_remove_watch,
171
&vuiface)) {
172
@@ -XXX,XX +XXX,XX @@ vubr_new(const char *path, bool client)
173
VHOST_USER_BRIDGE_MAX_QUEUES,
174
dev->sock,
175
vubr_panic,
176
+ NULL,
177
vubr_set_watch,
178
vubr_remove_watch,
179
&vuiface)) {
180
diff --git a/tools/virtiofsd/fuse_virtio.c b/tools/virtiofsd/fuse_virtio.c
181
index XXXXXXX..XXXXXXX 100644
182
--- a/tools/virtiofsd/fuse_virtio.c
183
+++ b/tools/virtiofsd/fuse_virtio.c
184
@@ -XXX,XX +XXX,XX @@ int virtio_session_mount(struct fuse_session *se)
185
se->vu_socketfd = data_sock;
186
se->virtio_dev->se = se;
187
pthread_rwlock_init(&se->virtio_dev->vu_dispatch_rwlock, NULL);
188
- vu_init(&se->virtio_dev->dev, 2, se->vu_socketfd, fv_panic, fv_set_watch,
189
- fv_remove_watch, &fv_iface);
190
+ vu_init(&se->virtio_dev->dev, 2, se->vu_socketfd, fv_panic, NULL,
191
+ fv_set_watch, fv_remove_watch, &fv_iface);
192
193
return 0;
194
}
195
--
196
2.26.2
197
diff view generated by jsdifflib
Deleted patch
1
From: Coiby Xu <coiby.xu@gmail.com>
2
1
3
When the client is running in gdb and quit command is run in gdb,
4
QEMU will still dispatch the event which will cause segment fault in
5
the callback function.
6
7
Signed-off-by: Coiby Xu <coiby.xu@gmail.com>
8
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
9
Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com>
10
Message-id: 20200918080912.321299-3-coiby.xu@gmail.com
11
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
12
---
13
contrib/libvhost-user/libvhost-user.c | 1 +
14
1 file changed, 1 insertion(+)
15
16
diff --git a/contrib/libvhost-user/libvhost-user.c b/contrib/libvhost-user/libvhost-user.c
17
index XXXXXXX..XXXXXXX 100644
18
--- a/contrib/libvhost-user/libvhost-user.c
19
+++ b/contrib/libvhost-user/libvhost-user.c
20
@@ -XXX,XX +XXX,XX @@ vu_deinit(VuDev *dev)
21
}
22
23
if (vq->kick_fd != -1) {
24
+ dev->remove_watch(dev, vq->kick_fd);
25
close(vq->kick_fd);
26
vq->kick_fd = -1;
27
}
28
--
29
2.26.2
30
diff view generated by jsdifflib
1
From: Coiby Xu <coiby.xu@gmail.com>
1
From: Jeuk Kim <jeuk20.kim@gmail.com>
2
2
3
By making use of libvhost-user, block device drive can be shared to
3
Universal Flash Storage (UFS) is a high-performance mass storage device
4
the connected vhost-user client. Only one client can connect to the
4
with a serial interface. It is primarily used as a high-performance
5
server one time.
5
data storage device for embedded applications.
6
6
7
Since vhost-user-server needs a block drive to be created first, delay
7
This commit contains code for UFS device to be recognized
8
the creation of this object.
8
as a UFS PCI device.
9
Patches to handle UFS logical unit and Transfer Request will follow.
9
10
10
Suggested-by: Kevin Wolf <kwolf@redhat.com>
11
Signed-off-by: Jeuk Kim <jeuk20.kim@samsung.com>
11
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
12
Signed-off-by: Coiby Xu <coiby.xu@gmail.com>
13
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
12
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
14
Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com>
13
Message-id: 9f3db32fe1c708090a6bb764d456973b5abef55f.1691062912.git.jeuk20.kim@samsung.com
15
Message-id: 20200918080912.321299-6-coiby.xu@gmail.com
16
[Shorten "vhost_user_blk_server" string to "vhost_user_blk" to avoid the
17
following compiler warning:
18
../block/export/vhost-user-blk-server.c:178:50: error: ‘%s’ directive output truncated writing 21 bytes into a region of size 20 [-Werror=format-truncation=]
19
and fix "Invalid size %ld ..." ssize_t format string arguments for
20
32-bit hosts.
21
--Stefan]
22
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
14
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
23
---
15
---
24
block/export/vhost-user-blk-server.h | 36 ++
16
MAINTAINERS | 6 +
25
block/export/vhost-user-blk-server.c | 661 +++++++++++++++++++++++++++
17
docs/specs/pci-ids.rst | 2 +
26
softmmu/vl.c | 4 +
18
meson.build | 1 +
27
block/meson.build | 1 +
19
hw/ufs/trace.h | 1 +
28
4 files changed, 702 insertions(+)
20
hw/ufs/ufs.h | 42 ++
29
create mode 100644 block/export/vhost-user-blk-server.h
21
include/block/ufs.h | 1090 ++++++++++++++++++++++++++++++++++++++
30
create mode 100644 block/export/vhost-user-blk-server.c
22
include/hw/pci/pci.h | 1 +
23
include/hw/pci/pci_ids.h | 1 +
24
hw/ufs/ufs.c | 278 ++++++++++
25
hw/Kconfig | 1 +
26
hw/meson.build | 1 +
27
hw/ufs/Kconfig | 4 +
28
hw/ufs/meson.build | 1 +
29
hw/ufs/trace-events | 32 ++
30
14 files changed, 1461 insertions(+)
31
create mode 100644 hw/ufs/trace.h
32
create mode 100644 hw/ufs/ufs.h
33
create mode 100644 include/block/ufs.h
34
create mode 100644 hw/ufs/ufs.c
35
create mode 100644 hw/ufs/Kconfig
36
create mode 100644 hw/ufs/meson.build
37
create mode 100644 hw/ufs/trace-events
31
38
32
diff --git a/block/export/vhost-user-blk-server.h b/block/export/vhost-user-blk-server.h
39
diff --git a/MAINTAINERS b/MAINTAINERS
40
index XXXXXXX..XXXXXXX 100644
41
--- a/MAINTAINERS
42
+++ b/MAINTAINERS
43
@@ -XXX,XX +XXX,XX @@ F: tests/qtest/nvme-test.c
44
F: docs/system/devices/nvme.rst
45
T: git git://git.infradead.org/qemu-nvme.git nvme-next
46
47
+ufs
48
+M: Jeuk Kim <jeuk20.kim@samsung.com>
49
+S: Supported
50
+F: hw/ufs/*
51
+F: include/block/ufs.h
52
+
53
megasas
54
M: Hannes Reinecke <hare@suse.com>
55
L: qemu-block@nongnu.org
56
diff --git a/docs/specs/pci-ids.rst b/docs/specs/pci-ids.rst
57
index XXXXXXX..XXXXXXX 100644
58
--- a/docs/specs/pci-ids.rst
59
+++ b/docs/specs/pci-ids.rst
60
@@ -XXX,XX +XXX,XX @@ PCI devices (other than virtio):
61
PCI PVPanic device (``-device pvpanic-pci``)
62
1b36:0012
63
PCI ACPI ERST device (``-device acpi-erst``)
64
+1b36:0013
65
+ PCI UFS device (``-device ufs``)
66
67
All these devices are documented in :doc:`index`.
68
69
diff --git a/meson.build b/meson.build
70
index XXXXXXX..XXXXXXX 100644
71
--- a/meson.build
72
+++ b/meson.build
73
@@ -XXX,XX +XXX,XX @@ if have_system
74
'hw/ssi',
75
'hw/timer',
76
'hw/tpm',
77
+ 'hw/ufs',
78
'hw/usb',
79
'hw/vfio',
80
'hw/virtio',
81
diff --git a/hw/ufs/trace.h b/hw/ufs/trace.h
33
new file mode 100644
82
new file mode 100644
34
index XXXXXXX..XXXXXXX
83
index XXXXXXX..XXXXXXX
35
--- /dev/null
84
--- /dev/null
36
+++ b/block/export/vhost-user-blk-server.h
85
+++ b/hw/ufs/trace.h
37
@@ -XXX,XX +XXX,XX @@
86
@@ -0,0 +1 @@
38
+/*
87
+#include "trace/trace-hw_ufs.h"
39
+ * Sharing QEMU block devices via vhost-user protocal
88
diff --git a/hw/ufs/ufs.h b/hw/ufs/ufs.h
40
+ *
41
+ * Copyright (c) Coiby Xu <coiby.xu@gmail.com>.
42
+ * Copyright (c) 2020 Red Hat, Inc.
43
+ *
44
+ * This work is licensed under the terms of the GNU GPL, version 2 or
45
+ * later. See the COPYING file in the top-level directory.
46
+ */
47
+
48
+#ifndef VHOST_USER_BLK_SERVER_H
49
+#define VHOST_USER_BLK_SERVER_H
50
+#include "util/vhost-user-server.h"
51
+
52
+typedef struct VuBlockDev VuBlockDev;
53
+#define TYPE_VHOST_USER_BLK_SERVER "vhost-user-blk-server"
54
+#define VHOST_USER_BLK_SERVER(obj) \
55
+ OBJECT_CHECK(VuBlockDev, obj, TYPE_VHOST_USER_BLK_SERVER)
56
+
57
+/* vhost user block device */
58
+struct VuBlockDev {
59
+ Object parent_obj;
60
+ char *node_name;
61
+ SocketAddress *addr;
62
+ AioContext *ctx;
63
+ VuServer vu_server;
64
+ bool running;
65
+ uint32_t blk_size;
66
+ BlockBackend *backend;
67
+ QIOChannelSocket *sioc;
68
+ QTAILQ_ENTRY(VuBlockDev) next;
69
+ struct virtio_blk_config blkcfg;
70
+ bool writable;
71
+};
72
+
73
+#endif /* VHOST_USER_BLK_SERVER_H */
74
diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c
75
new file mode 100644
89
new file mode 100644
76
index XXXXXXX..XXXXXXX
90
index XXXXXXX..XXXXXXX
77
--- /dev/null
91
--- /dev/null
78
+++ b/block/export/vhost-user-blk-server.c
92
+++ b/hw/ufs/ufs.h
79
@@ -XXX,XX +XXX,XX @@
93
@@ -XXX,XX +XXX,XX @@
80
+/*
94
+/*
81
+ * Sharing QEMU block devices via vhost-user protocal
95
+ * QEMU UFS
82
+ *
96
+ *
83
+ * Parts of the code based on nbd/server.c.
97
+ * Copyright (c) 2023 Samsung Electronics Co., Ltd. All rights reserved.
84
+ *
98
+ *
85
+ * Copyright (c) Coiby Xu <coiby.xu@gmail.com>.
99
+ * Written by Jeuk Kim <jeuk20.kim@samsung.com>
86
+ * Copyright (c) 2020 Red Hat, Inc.
87
+ *
100
+ *
88
+ * This work is licensed under the terms of the GNU GPL, version 2 or
101
+ * SPDX-License-Identifier: GPL-2.0-or-later
89
+ * later. See the COPYING file in the top-level directory.
102
+ */
90
+ */
103
+
104
+#ifndef HW_UFS_UFS_H
105
+#define HW_UFS_UFS_H
106
+
107
+#include "hw/pci/pci_device.h"
108
+#include "hw/scsi/scsi.h"
109
+#include "block/ufs.h"
110
+
111
+#define UFS_MAX_LUS 32
112
+#define UFS_BLOCK_SIZE 4096
113
+
114
+typedef struct UfsParams {
115
+ char *serial;
116
+ uint8_t nutrs; /* Number of UTP Transfer Request Slots */
117
+ uint8_t nutmrs; /* Number of UTP Task Management Request Slots */
118
+} UfsParams;
119
+
120
+typedef struct UfsHc {
121
+ PCIDevice parent_obj;
122
+ MemoryRegion iomem;
123
+ UfsReg reg;
124
+ UfsParams params;
125
+ uint32_t reg_size;
126
+
127
+ qemu_irq irq;
128
+ QEMUBH *doorbell_bh;
129
+ QEMUBH *complete_bh;
130
+} UfsHc;
131
+
132
+#define TYPE_UFS "ufs"
133
+#define UFS(obj) OBJECT_CHECK(UfsHc, (obj), TYPE_UFS)
134
+
135
+#endif /* HW_UFS_UFS_H */
136
diff --git a/include/block/ufs.h b/include/block/ufs.h
137
new file mode 100644
138
index XXXXXXX..XXXXXXX
139
--- /dev/null
140
+++ b/include/block/ufs.h
141
@@ -XXX,XX +XXX,XX @@
142
+/* SPDX-License-Identifier: GPL-2.0-or-later */
143
+
144
+#ifndef BLOCK_UFS_H
145
+#define BLOCK_UFS_H
146
+
147
+#include "hw/registerfields.h"
148
+
149
+typedef struct QEMU_PACKED UfsReg {
150
+ uint32_t cap;
151
+ uint32_t rsvd0;
152
+ uint32_t ver;
153
+ uint32_t rsvd1;
154
+ uint32_t hcpid;
155
+ uint32_t hcmid;
156
+ uint32_t ahit;
157
+ uint32_t rsvd2;
158
+ uint32_t is;
159
+ uint32_t ie;
160
+ uint32_t rsvd3[2];
161
+ uint32_t hcs;
162
+ uint32_t hce;
163
+ uint32_t uecpa;
164
+ uint32_t uecdl;
165
+ uint32_t uecn;
166
+ uint32_t uect;
167
+ uint32_t uecdme;
168
+ uint32_t utriacr;
169
+ uint32_t utrlba;
170
+ uint32_t utrlbau;
171
+ uint32_t utrldbr;
172
+ uint32_t utrlclr;
173
+ uint32_t utrlrsr;
174
+ uint32_t utrlcnr;
175
+ uint32_t rsvd4[2];
176
+ uint32_t utmrlba;
177
+ uint32_t utmrlbau;
178
+ uint32_t utmrldbr;
179
+ uint32_t utmrlclr;
180
+ uint32_t utmrlrsr;
181
+ uint32_t rsvd5[3];
182
+ uint32_t uiccmd;
183
+ uint32_t ucmdarg1;
184
+ uint32_t ucmdarg2;
185
+ uint32_t ucmdarg3;
186
+ uint32_t rsvd6[4];
187
+ uint32_t rsvd7[4];
188
+ uint32_t rsvd8[16];
189
+ uint32_t ccap;
190
+} UfsReg;
191
+
192
+REG32(CAP, offsetof(UfsReg, cap))
193
+ FIELD(CAP, NUTRS, 0, 5)
194
+ FIELD(CAP, RTT, 8, 8)
195
+ FIELD(CAP, NUTMRS, 16, 3)
196
+ FIELD(CAP, AUTOH8, 23, 1)
197
+ FIELD(CAP, 64AS, 24, 1)
198
+ FIELD(CAP, OODDS, 25, 1)
199
+ FIELD(CAP, UICDMETMS, 26, 1)
200
+ FIELD(CAP, CS, 28, 1)
201
+REG32(VER, offsetof(UfsReg, ver))
202
+REG32(HCPID, offsetof(UfsReg, hcpid))
203
+REG32(HCMID, offsetof(UfsReg, hcmid))
204
+REG32(AHIT, offsetof(UfsReg, ahit))
205
+REG32(IS, offsetof(UfsReg, is))
206
+ FIELD(IS, UTRCS, 0, 1)
207
+ FIELD(IS, UDEPRI, 1, 1)
208
+ FIELD(IS, UE, 2, 1)
209
+ FIELD(IS, UTMS, 3, 1)
210
+ FIELD(IS, UPMS, 4, 1)
211
+ FIELD(IS, UHXS, 5, 1)
212
+ FIELD(IS, UHES, 6, 1)
213
+ FIELD(IS, ULLS, 7, 1)
214
+ FIELD(IS, ULSS, 8, 1)
215
+ FIELD(IS, UTMRCS, 9, 1)
216
+ FIELD(IS, UCCS, 10, 1)
217
+ FIELD(IS, DFES, 11, 1)
218
+ FIELD(IS, UTPES, 12, 1)
219
+ FIELD(IS, HCFES, 16, 1)
220
+ FIELD(IS, SBFES, 17, 1)
221
+ FIELD(IS, CEFES, 18, 1)
222
+REG32(IE, offsetof(UfsReg, ie))
223
+ FIELD(IE, UTRCE, 0, 1)
224
+ FIELD(IE, UDEPRIE, 1, 1)
225
+ FIELD(IE, UEE, 2, 1)
226
+ FIELD(IE, UTMSE, 3, 1)
227
+ FIELD(IE, UPMSE, 4, 1)
228
+ FIELD(IE, UHXSE, 5, 1)
229
+ FIELD(IE, UHESE, 6, 1)
230
+ FIELD(IE, ULLSE, 7, 1)
231
+ FIELD(IE, ULSSE, 8, 1)
232
+ FIELD(IE, UTMRCE, 9, 1)
233
+ FIELD(IE, UCCE, 10, 1)
234
+ FIELD(IE, DFEE, 11, 1)
235
+ FIELD(IE, UTPEE, 12, 1)
236
+ FIELD(IE, HCFEE, 16, 1)
237
+ FIELD(IE, SBFEE, 17, 1)
238
+ FIELD(IE, CEFEE, 18, 1)
239
+REG32(HCS, offsetof(UfsReg, hcs))
240
+ FIELD(HCS, DP, 0, 1)
241
+ FIELD(HCS, UTRLRDY, 1, 1)
242
+ FIELD(HCS, UTMRLRDY, 2, 1)
243
+ FIELD(HCS, UCRDY, 3, 1)
244
+ FIELD(HCS, UPMCRS, 8, 3)
245
+REG32(HCE, offsetof(UfsReg, hce))
246
+ FIELD(HCE, HCE, 0, 1)
247
+ FIELD(HCE, CGE, 1, 1)
248
+REG32(UECPA, offsetof(UfsReg, uecpa))
249
+REG32(UECDL, offsetof(UfsReg, uecdl))
250
+REG32(UECN, offsetof(UfsReg, uecn))
251
+REG32(UECT, offsetof(UfsReg, uect))
252
+REG32(UECDME, offsetof(UfsReg, uecdme))
253
+REG32(UTRIACR, offsetof(UfsReg, utriacr))
254
+REG32(UTRLBA, offsetof(UfsReg, utrlba))
255
+ FIELD(UTRLBA, UTRLBA, 9, 22)
256
+REG32(UTRLBAU, offsetof(UfsReg, utrlbau))
257
+REG32(UTRLDBR, offsetof(UfsReg, utrldbr))
258
+REG32(UTRLCLR, offsetof(UfsReg, utrlclr))
259
+REG32(UTRLRSR, offsetof(UfsReg, utrlrsr))
260
+REG32(UTRLCNR, offsetof(UfsReg, utrlcnr))
261
+REG32(UTMRLBA, offsetof(UfsReg, utmrlba))
262
+ FIELD(UTMRLBA, UTMRLBA, 9, 22)
263
+REG32(UTMRLBAU, offsetof(UfsReg, utmrlbau))
264
+REG32(UTMRLDBR, offsetof(UfsReg, utmrldbr))
265
+REG32(UTMRLCLR, offsetof(UfsReg, utmrlclr))
266
+REG32(UTMRLRSR, offsetof(UfsReg, utmrlrsr))
267
+REG32(UICCMD, offsetof(UfsReg, uiccmd))
268
+REG32(UCMDARG1, offsetof(UfsReg, ucmdarg1))
269
+REG32(UCMDARG2, offsetof(UfsReg, ucmdarg2))
270
+REG32(UCMDARG3, offsetof(UfsReg, ucmdarg3))
271
+REG32(CCAP, offsetof(UfsReg, ccap))
272
+
273
+#define UFS_INTR_MASK \
274
+ ((1 << R_IS_CEFES_SHIFT) | (1 << R_IS_SBFES_SHIFT) | \
275
+ (1 << R_IS_HCFES_SHIFT) | (1 << R_IS_UTPES_SHIFT) | \
276
+ (1 << R_IS_DFES_SHIFT) | (1 << R_IS_UCCS_SHIFT) | \
277
+ (1 << R_IS_UTMRCS_SHIFT) | (1 << R_IS_ULSS_SHIFT) | \
278
+ (1 << R_IS_ULLS_SHIFT) | (1 << R_IS_UHES_SHIFT) | \
279
+ (1 << R_IS_UHXS_SHIFT) | (1 << R_IS_UPMS_SHIFT) | \
280
+ (1 << R_IS_UTMS_SHIFT) | (1 << R_IS_UE_SHIFT) | \
281
+ (1 << R_IS_UDEPRI_SHIFT) | (1 << R_IS_UTRCS_SHIFT))
282
+
283
+#define UFS_UPIU_HEADER_TRANSACTION_TYPE_SHIFT 24
284
+#define UFS_UPIU_HEADER_TRANSACTION_TYPE_MASK 0xff
285
+#define UFS_UPIU_HEADER_TRANSACTION_TYPE(dword0) \
286
+ ((be32_to_cpu(dword0) >> UFS_UPIU_HEADER_TRANSACTION_TYPE_SHIFT) & \
287
+ UFS_UPIU_HEADER_TRANSACTION_TYPE_MASK)
288
+
289
+#define UFS_UPIU_HEADER_QUERY_FUNC_SHIFT 16
290
+#define UFS_UPIU_HEADER_QUERY_FUNC_MASK 0xff
291
+#define UFS_UPIU_HEADER_QUERY_FUNC(dword1) \
292
+ ((be32_to_cpu(dword1) >> UFS_UPIU_HEADER_QUERY_FUNC_SHIFT) & \
293
+ UFS_UPIU_HEADER_QUERY_FUNC_MASK)
294
+
295
+#define UFS_UPIU_HEADER_DATA_SEGMENT_LENGTH_SHIFT 0
296
+#define UFS_UPIU_HEADER_DATA_SEGMENT_LENGTH_MASK 0xffff
297
+#define UFS_UPIU_HEADER_DATA_SEGMENT_LENGTH(dword2) \
298
+ ((be32_to_cpu(dword2) >> UFS_UPIU_HEADER_DATA_SEGMENT_LENGTH_SHIFT) & \
299
+ UFS_UPIU_HEADER_DATA_SEGMENT_LENGTH_MASK)
300
+
301
+typedef struct QEMU_PACKED DeviceDescriptor {
302
+ uint8_t length;
303
+ uint8_t descriptor_idn;
304
+ uint8_t device;
305
+ uint8_t device_class;
306
+ uint8_t device_sub_class;
307
+ uint8_t protocol;
308
+ uint8_t number_lu;
309
+ uint8_t number_wlu;
310
+ uint8_t boot_enable;
311
+ uint8_t descr_access_en;
312
+ uint8_t init_power_mode;
313
+ uint8_t high_priority_lun;
314
+ uint8_t secure_removal_type;
315
+ uint8_t security_lu;
316
+ uint8_t background_ops_term_lat;
317
+ uint8_t init_active_icc_level;
318
+ uint16_t spec_version;
319
+ uint16_t manufacture_date;
320
+ uint8_t manufacturer_name;
321
+ uint8_t product_name;
322
+ uint8_t serial_number;
323
+ uint8_t oem_id;
324
+ uint16_t manufacturer_id;
325
+ uint8_t ud_0_base_offset;
326
+ uint8_t ud_config_p_length;
327
+ uint8_t device_rtt_cap;
328
+ uint16_t periodic_rtc_update;
329
+ uint8_t ufs_features_support;
330
+ uint8_t ffu_timeout;
331
+ uint8_t queue_depth;
332
+ uint16_t device_version;
333
+ uint8_t num_secure_wp_area;
334
+ uint32_t psa_max_data_size;
335
+ uint8_t psa_state_timeout;
336
+ uint8_t product_revision_level;
337
+ uint8_t reserved[36];
338
+ uint32_t extended_ufs_features_support;
339
+ uint8_t write_booster_buffer_preserve_user_space_en;
340
+ uint8_t write_booster_buffer_type;
341
+ uint32_t num_shared_write_booster_buffer_alloc_units;
342
+} DeviceDescriptor;
343
+
344
+typedef struct QEMU_PACKED GeometryDescriptor {
345
+ uint8_t length;
346
+ uint8_t descriptor_idn;
347
+ uint8_t media_technology;
348
+ uint8_t reserved;
349
+ uint64_t total_raw_device_capacity;
350
+ uint8_t max_number_lu;
351
+ uint32_t segment_size;
352
+ uint8_t allocation_unit_size;
353
+ uint8_t min_addr_block_size;
354
+ uint8_t optimal_read_block_size;
355
+ uint8_t optimal_write_block_size;
356
+ uint8_t max_in_buffer_size;
357
+ uint8_t max_out_buffer_size;
358
+ uint8_t rpmb_read_write_size;
359
+ uint8_t dynamic_capacity_resource_policy;
360
+ uint8_t data_ordering;
361
+ uint8_t max_context_id_number;
362
+ uint8_t sys_data_tag_unit_size;
363
+ uint8_t sys_data_tag_res_size;
364
+ uint8_t supported_sec_r_types;
365
+ uint16_t supported_memory_types;
366
+ uint32_t system_code_max_n_alloc_u;
367
+ uint16_t system_code_cap_adj_fac;
368
+ uint32_t non_persist_max_n_alloc_u;
369
+ uint16_t non_persist_cap_adj_fac;
370
+ uint32_t enhanced_1_max_n_alloc_u;
371
+ uint16_t enhanced_1_cap_adj_fac;
372
+ uint32_t enhanced_2_max_n_alloc_u;
373
+ uint16_t enhanced_2_cap_adj_fac;
374
+ uint32_t enhanced_3_max_n_alloc_u;
375
+ uint16_t enhanced_3_cap_adj_fac;
376
+ uint32_t enhanced_4_max_n_alloc_u;
377
+ uint16_t enhanced_4_cap_adj_fac;
378
+ uint32_t optimal_logical_block_size;
379
+ uint8_t reserved2[7];
380
+ uint32_t write_booster_buffer_max_n_alloc_units;
381
+ uint8_t device_max_write_booster_l_us;
382
+ uint8_t write_booster_buffer_cap_adj_fac;
383
+ uint8_t supported_write_booster_buffer_user_space_reduction_types;
384
+ uint8_t supported_write_booster_buffer_types;
385
+} GeometryDescriptor;
386
+
387
+#define UFS_GEOMETRY_CAPACITY_SHIFT 9
388
+
389
+typedef struct QEMU_PACKED UnitDescriptor {
390
+ uint8_t length;
391
+ uint8_t descriptor_idn;
392
+ uint8_t unit_index;
393
+ uint8_t lu_enable;
394
+ uint8_t boot_lun_id;
395
+ uint8_t lu_write_protect;
396
+ uint8_t lu_queue_depth;
397
+ uint8_t psa_sensitive;
398
+ uint8_t memory_type;
399
+ uint8_t data_reliability;
400
+ uint8_t logical_block_size;
401
+ uint64_t logical_block_count;
402
+ uint32_t erase_block_size;
403
+ uint8_t provisioning_type;
404
+ uint64_t phy_mem_resource_count;
405
+ uint16_t context_capabilities;
406
+ uint8_t large_unit_granularity_m1;
407
+ uint8_t reserved[6];
408
+ uint32_t lu_num_write_booster_buffer_alloc_units;
409
+} UnitDescriptor;
410
+
411
+typedef struct QEMU_PACKED RpmbUnitDescriptor {
412
+ uint8_t length;
413
+ uint8_t descriptor_idn;
414
+ uint8_t unit_index;
415
+ uint8_t lu_enable;
416
+ uint8_t boot_lun_id;
417
+ uint8_t lu_write_protect;
418
+ uint8_t lu_queue_depth;
419
+ uint8_t psa_sensitive;
420
+ uint8_t memory_type;
421
+ uint8_t reserved;
422
+ uint8_t logical_block_size;
423
+ uint64_t logical_block_count;
424
+ uint32_t erase_block_size;
425
+ uint8_t provisioning_type;
426
+ uint64_t phy_mem_resource_count;
427
+ uint8_t reserved2[3];
428
+} RpmbUnitDescriptor;
429
+
430
+typedef struct QEMU_PACKED PowerParametersDescriptor {
431
+ uint8_t length;
432
+ uint8_t descriptor_idn;
433
+ uint16_t active_icc_levels_vcc[16];
434
+ uint16_t active_icc_levels_vccq[16];
435
+ uint16_t active_icc_levels_vccq_2[16];
436
+} PowerParametersDescriptor;
437
+
438
+typedef struct QEMU_PACKED InterconnectDescriptor {
439
+ uint8_t length;
440
+ uint8_t descriptor_idn;
441
+ uint16_t bcd_unipro_version;
442
+ uint16_t bcd_mphy_version;
443
+} InterconnectDescriptor;
444
+
445
+typedef struct QEMU_PACKED StringDescriptor {
446
+ uint8_t length;
447
+ uint8_t descriptor_idn;
448
+ uint16_t UC[126];
449
+} StringDescriptor;
450
+
451
+typedef struct QEMU_PACKED DeviceHealthDescriptor {
452
+ uint8_t length;
453
+ uint8_t descriptor_idn;
454
+ uint8_t pre_eol_info;
455
+ uint8_t device_life_time_est_a;
456
+ uint8_t device_life_time_est_b;
457
+ uint8_t vendor_prop_info[32];
458
+ uint32_t refresh_total_count;
459
+ uint32_t refresh_progress;
460
+} DeviceHealthDescriptor;
461
+
462
+typedef struct QEMU_PACKED Flags {
463
+ uint8_t reserved;
464
+ uint8_t device_init;
465
+ uint8_t permanent_wp_en;
466
+ uint8_t power_on_wp_en;
467
+ uint8_t background_ops_en;
468
+ uint8_t device_life_span_mode_en;
469
+ uint8_t purge_enable;
470
+ uint8_t refresh_enable;
471
+ uint8_t phy_resource_removal;
472
+ uint8_t busy_rtc;
473
+ uint8_t reserved2;
474
+ uint8_t permanently_disable_fw_update;
475
+ uint8_t reserved3[2];
476
+ uint8_t wb_en;
477
+ uint8_t wb_buffer_flush_en;
478
+ uint8_t wb_buffer_flush_during_hibernate;
479
+ uint8_t reserved4[2];
480
+} Flags;
481
+
482
+typedef struct Attributes {
483
+ uint8_t boot_lun_en;
484
+ uint8_t reserved;
485
+ uint8_t current_power_mode;
486
+ uint8_t active_icc_level;
487
+ uint8_t out_of_order_data_en;
488
+ uint8_t background_op_status;
489
+ uint8_t purge_status;
490
+ uint8_t max_data_in_size;
491
+ uint8_t max_data_out_size;
492
+ uint32_t dyn_cap_needed;
493
+ uint8_t ref_clk_freq;
494
+ uint8_t config_descr_lock;
495
+ uint8_t max_num_of_rtt;
496
+ uint16_t exception_event_control;
497
+ uint16_t exception_event_status;
498
+ uint32_t seconds_passed;
499
+ uint16_t context_conf;
500
+ uint8_t device_ffu_status;
501
+ uint8_t psa_state;
502
+ uint32_t psa_data_size;
503
+ uint8_t ref_clk_gating_wait_time;
504
+ uint8_t device_case_rough_temperaure;
505
+ uint8_t device_too_high_temp_boundary;
506
+ uint8_t device_too_low_temp_boundary;
507
+ uint8_t throttling_status;
508
+ uint8_t wb_buffer_flush_status;
509
+ uint8_t available_wb_buffer_size;
510
+ uint8_t wb_buffer_life_time_est;
511
+ uint32_t current_wb_buffer_size;
512
+ uint8_t refresh_status;
513
+ uint8_t refresh_freq;
514
+ uint8_t refresh_unit;
515
+ uint8_t refresh_method;
516
+} Attributes;
517
+
518
+#define UFS_TRANSACTION_SPECIFIC_FIELD_SIZE 20
519
+#define UFS_MAX_QUERY_DATA_SIZE 256
520
+
521
+/* Command response result code */
522
+typedef enum CommandRespCode {
523
+ COMMAND_RESULT_SUCESS = 0x00,
524
+ COMMAND_RESULT_FAIL = 0x01,
525
+} CommandRespCode;
526
+
527
+enum {
528
+ UFS_UPIU_FLAG_UNDERFLOW = 0x20,
529
+ UFS_UPIU_FLAG_OVERFLOW = 0x40,
530
+};
531
+
532
+typedef struct QEMU_PACKED UtpUpiuHeader {
533
+ uint8_t trans_type;
534
+ uint8_t flags;
535
+ uint8_t lun;
536
+ uint8_t task_tag;
537
+ uint8_t iid_cmd_set_type;
538
+ uint8_t query_func;
539
+ uint8_t response;
540
+ uint8_t scsi_status;
541
+ uint8_t ehs_len;
542
+ uint8_t device_inf;
543
+ uint16_t data_segment_length;
544
+} UtpUpiuHeader;
545
+
546
+/*
547
+ * The code below is copied from the linux kernel
548
+ * ("include/uapi/scsi/scsi_bsg_ufs.h") and modified to fit the qemu style.
549
+ */
550
+
551
+typedef struct QEMU_PACKED UtpUpiuQuery {
552
+ uint8_t opcode;
553
+ uint8_t idn;
554
+ uint8_t index;
555
+ uint8_t selector;
556
+ uint16_t reserved_osf;
557
+ uint16_t length;
558
+ uint32_t value;
559
+ uint32_t reserved[2];
560
+ /* EHS length should be 0. We don't have to worry about EHS area. */
561
+ uint8_t data[UFS_MAX_QUERY_DATA_SIZE];
562
+} UtpUpiuQuery;
563
+
564
+#define UFS_CDB_SIZE 16
565
+
566
+/*
567
+ * struct UtpUpiuCmd - Command UPIU structure
568
+ * @data_transfer_len: Data Transfer Length DW-3
569
+ * @cdb: Command Descriptor Block CDB DW-4 to DW-7
570
+ */
571
+typedef struct QEMU_PACKED UtpUpiuCmd {
572
+ uint32_t exp_data_transfer_len;
573
+ uint8_t cdb[UFS_CDB_SIZE];
574
+} UtpUpiuCmd;
575
+
576
+/*
577
+ * struct UtpUpiuReq - general upiu request structure
578
+ * @header:UPIU header structure DW-0 to DW-2
579
+ * @sc: fields structure for scsi command DW-3 to DW-7
580
+ * @qr: fields structure for query request DW-3 to DW-7
581
+ * @uc: use utp_upiu_query to host the 4 dwords of uic command
582
+ */
583
+typedef struct QEMU_PACKED UtpUpiuReq {
584
+ UtpUpiuHeader header;
585
+ union {
586
+ UtpUpiuCmd sc;
587
+ UtpUpiuQuery qr;
588
+ };
589
+} UtpUpiuReq;
590
+
591
+/*
592
+ * The code below is copied from the linux kernel ("include/ufs/ufshci.h") and
593
+ * modified to fit the qemu style.
594
+ */
595
+
596
+enum {
597
+ PWR_OK = 0x0,
598
+ PWR_LOCAL = 0x01,
599
+ PWR_REMOTE = 0x02,
600
+ PWR_BUSY = 0x03,
601
+ PWR_ERROR_CAP = 0x04,
602
+ PWR_FATAL_ERROR = 0x05,
603
+};
604
+
605
+/* UIC Commands */
606
+enum uic_cmd_dme {
607
+ UIC_CMD_DME_GET = 0x01,
608
+ UIC_CMD_DME_SET = 0x02,
609
+ UIC_CMD_DME_PEER_GET = 0x03,
610
+ UIC_CMD_DME_PEER_SET = 0x04,
611
+ UIC_CMD_DME_POWERON = 0x10,
612
+ UIC_CMD_DME_POWEROFF = 0x11,
613
+ UIC_CMD_DME_ENABLE = 0x12,
614
+ UIC_CMD_DME_RESET = 0x14,
615
+ UIC_CMD_DME_END_PT_RST = 0x15,
616
+ UIC_CMD_DME_LINK_STARTUP = 0x16,
617
+ UIC_CMD_DME_HIBER_ENTER = 0x17,
618
+ UIC_CMD_DME_HIBER_EXIT = 0x18,
619
+ UIC_CMD_DME_TEST_MODE = 0x1A,
620
+};
621
+
622
+/* UIC Config result code / Generic error code */
623
+enum {
624
+ UIC_CMD_RESULT_SUCCESS = 0x00,
625
+ UIC_CMD_RESULT_INVALID_ATTR = 0x01,
626
+ UIC_CMD_RESULT_FAILURE = 0x01,
627
+ UIC_CMD_RESULT_INVALID_ATTR_VALUE = 0x02,
628
+ UIC_CMD_RESULT_READ_ONLY_ATTR = 0x03,
629
+ UIC_CMD_RESULT_WRITE_ONLY_ATTR = 0x04,
630
+ UIC_CMD_RESULT_BAD_INDEX = 0x05,
631
+ UIC_CMD_RESULT_LOCKED_ATTR = 0x06,
632
+ UIC_CMD_RESULT_BAD_TEST_FEATURE_INDEX = 0x07,
633
+ UIC_CMD_RESULT_PEER_COMM_FAILURE = 0x08,
634
+ UIC_CMD_RESULT_BUSY = 0x09,
635
+ UIC_CMD_RESULT_DME_FAILURE = 0x0A,
636
+};
637
+
638
+#define MASK_UIC_COMMAND_RESULT 0xFF
639
+
640
+/*
641
+ * Request Descriptor Definitions
642
+ */
643
+
644
+/* Transfer request command type */
645
+enum {
646
+ UTP_CMD_TYPE_SCSI = 0x0,
647
+ UTP_CMD_TYPE_UFS = 0x1,
648
+ UTP_CMD_TYPE_DEV_MANAGE = 0x2,
649
+};
650
+
651
+/* To accommodate UFS2.0 required Command type */
652
+enum {
653
+ UTP_CMD_TYPE_UFS_STORAGE = 0x1,
654
+};
655
+
656
+enum {
657
+ UTP_SCSI_COMMAND = 0x00000000,
658
+ UTP_NATIVE_UFS_COMMAND = 0x10000000,
659
+ UTP_DEVICE_MANAGEMENT_FUNCTION = 0x20000000,
660
+ UTP_REQ_DESC_INT_CMD = 0x01000000,
661
+ UTP_REQ_DESC_CRYPTO_ENABLE_CMD = 0x00800000,
662
+};
663
+
664
+/* UTP Transfer Request Data Direction (DD) */
665
+enum {
666
+ UTP_NO_DATA_TRANSFER = 0x00000000,
667
+ UTP_HOST_TO_DEVICE = 0x02000000,
668
+ UTP_DEVICE_TO_HOST = 0x04000000,
669
+};
670
+
671
+/* Overall command status values */
672
+enum UtpOcsCodes {
673
+ OCS_SUCCESS = 0x0,
674
+ OCS_INVALID_CMD_TABLE_ATTR = 0x1,
675
+ OCS_INVALID_PRDT_ATTR = 0x2,
676
+ OCS_MISMATCH_DATA_BUF_SIZE = 0x3,
677
+ OCS_MISMATCH_RESP_UPIU_SIZE = 0x4,
678
+ OCS_PEER_COMM_FAILURE = 0x5,
679
+ OCS_ABORTED = 0x6,
680
+ OCS_FATAL_ERROR = 0x7,
681
+ OCS_DEVICE_FATAL_ERROR = 0x8,
682
+ OCS_INVALID_CRYPTO_CONFIG = 0x9,
683
+ OCS_GENERAL_CRYPTO_ERROR = 0xa,
684
+ OCS_INVALID_COMMAND_STATUS = 0xf,
685
+};
686
+
687
+enum {
688
+ MASK_OCS = 0x0F,
689
+};
690
+
691
+/*
692
+ * struct UfshcdSgEntry - UFSHCI PRD Entry
693
+ * @addr: Physical address; DW-0 and DW-1.
694
+ * @reserved: Reserved for future use DW-2
695
+ * @size: size of physical segment DW-3
696
+ */
697
+typedef struct QEMU_PACKED UfshcdSgEntry {
698
+ uint64_t addr;
699
+ uint32_t reserved;
700
+ uint32_t size;
701
+ /*
702
+ * followed by variant-specific fields if
703
+ * CONFIG_SCSI_UFS_VARIABLE_SG_ENTRY_SIZE has been defined.
704
+ */
705
+} UfshcdSgEntry;
706
+
707
+/*
708
+ * struct RequestDescHeader - Descriptor Header common to both UTRD and UTMRD
709
+ * @dword0: Descriptor Header DW0
710
+ * @dword1: Descriptor Header DW1
711
+ * @dword2: Descriptor Header DW2
712
+ * @dword3: Descriptor Header DW3
713
+ */
714
+typedef struct QEMU_PACKED RequestDescHeader {
715
+ uint32_t dword_0;
716
+ uint32_t dword_1;
717
+ uint32_t dword_2;
718
+ uint32_t dword_3;
719
+} RequestDescHeader;
720
+
721
+/*
722
+ * struct UtpTransferReqDesc - UTP Transfer Request Descriptor (UTRD)
723
+ * @header: UTRD header DW-0 to DW-3
724
+ * @command_desc_base_addr_lo: UCD base address low DW-4
725
+ * @command_desc_base_addr_hi: UCD base address high DW-5
726
+ * @response_upiu_length: response UPIU length DW-6
727
+ * @response_upiu_offset: response UPIU offset DW-6
728
+ * @prd_table_length: Physical region descriptor length DW-7
729
+ * @prd_table_offset: Physical region descriptor offset DW-7
730
+ */
731
+typedef struct QEMU_PACKED UtpTransferReqDesc {
732
+ /* DW 0-3 */
733
+ RequestDescHeader header;
734
+
735
+ /* DW 4-5*/
736
+ uint32_t command_desc_base_addr_lo;
737
+ uint32_t command_desc_base_addr_hi;
738
+
739
+ /* DW 6 */
740
+ uint16_t response_upiu_length;
741
+ uint16_t response_upiu_offset;
742
+
743
+ /* DW 7 */
744
+ uint16_t prd_table_length;
745
+ uint16_t prd_table_offset;
746
+} UtpTransferReqDesc;
747
+
748
+/*
749
+ * UTMRD structure.
750
+ */
751
+typedef struct QEMU_PACKED UtpTaskReqDesc {
752
+ /* DW 0-3 */
753
+ RequestDescHeader header;
754
+
755
+ /* DW 4-11 - Task request UPIU structure */
756
+ struct {
757
+ UtpUpiuHeader req_header;
758
+ uint32_t input_param1;
759
+ uint32_t input_param2;
760
+ uint32_t input_param3;
761
+ uint32_t reserved1[2];
762
+ } upiu_req;
763
+
764
+ /* DW 12-19 - Task Management Response UPIU structure */
765
+ struct {
766
+ UtpUpiuHeader rsp_header;
767
+ uint32_t output_param1;
768
+ uint32_t output_param2;
769
+ uint32_t reserved2[3];
770
+ } upiu_rsp;
771
+} UtpTaskReqDesc;
772
+
773
+/*
774
+ * The code below is copied from the linux kernel ("include/ufs/ufs.h") and
775
+ * modified to fit the qemu style.
776
+ */
777
+
778
+#define GENERAL_UPIU_REQUEST_SIZE (sizeof(UtpUpiuReq))
779
+#define QUERY_DESC_MAX_SIZE 255
780
+#define QUERY_DESC_MIN_SIZE 2
781
+#define QUERY_DESC_HDR_SIZE 2
782
+#define QUERY_OSF_SIZE (GENERAL_UPIU_REQUEST_SIZE - (sizeof(UtpUpiuHeader)))
783
+#define UFS_SENSE_SIZE 18
784
+
785
+/*
786
+ * UFS device may have standard LUs and LUN id could be from 0x00 to
787
+ * 0x7F. Standard LUs use "Peripheral Device Addressing Format".
788
+ * UFS device may also have the Well Known LUs (also referred as W-LU)
789
+ * which again could be from 0x00 to 0x7F. For W-LUs, device only use
790
+ * the "Extended Addressing Format" which means the W-LUNs would be
791
+ * from 0xc100 (SCSI_W_LUN_BASE) onwards.
792
+ * This means max. LUN number reported from UFS device could be 0xC17F.
793
+ */
794
+#define UFS_UPIU_MAX_UNIT_NUM_ID 0x7F
795
+#define UFS_UPIU_WLUN_ID (1 << 7)
796
+
797
+/* WriteBooster buffer is available only for the logical unit from 0 to 7 */
798
+#define UFS_UPIU_MAX_WB_LUN_ID 8
799
+
800
+/*
801
+ * WriteBooster buffer lifetime has a limit setted by vendor.
802
+ * If it is over the limit, WriteBooster feature will be disabled.
803
+ */
804
+#define UFS_WB_EXCEED_LIFETIME 0x0B
805
+
806
+/*
807
+ * In UFS Spec, the Extra Header Segment (EHS) starts from byte 32 in UPIU
808
+ * request/response packet
809
+ */
810
+#define EHS_OFFSET_IN_RESPONSE 32
811
+
812
+/* Well known logical unit id in LUN field of UPIU */
813
+enum {
814
+ UFS_UPIU_REPORT_LUNS_WLUN = 0x81,
815
+ UFS_UPIU_UFS_DEVICE_WLUN = 0xD0,
816
+ UFS_UPIU_BOOT_WLUN = 0xB0,
817
+ UFS_UPIU_RPMB_WLUN = 0xC4,
818
+};
819
+
820
+/*
821
+ * UFS Protocol Information Unit related definitions
822
+ */
823
+
824
+/* Task management functions */
825
+enum {
826
+ UFS_ABORT_TASK = 0x01,
827
+ UFS_ABORT_TASK_SET = 0x02,
828
+ UFS_CLEAR_TASK_SET = 0x04,
829
+ UFS_LOGICAL_RESET = 0x08,
830
+ UFS_QUERY_TASK = 0x80,
831
+ UFS_QUERY_TASK_SET = 0x81,
832
+};
833
+
834
+/* UTP UPIU Transaction Codes Initiator to Target */
835
+enum {
836
+ UPIU_TRANSACTION_NOP_OUT = 0x00,
837
+ UPIU_TRANSACTION_COMMAND = 0x01,
838
+ UPIU_TRANSACTION_DATA_OUT = 0x02,
839
+ UPIU_TRANSACTION_TASK_REQ = 0x04,
840
+ UPIU_TRANSACTION_QUERY_REQ = 0x16,
841
+};
842
+
843
+/* UTP UPIU Transaction Codes Target to Initiator */
844
+enum {
845
+ UPIU_TRANSACTION_NOP_IN = 0x20,
846
+ UPIU_TRANSACTION_RESPONSE = 0x21,
847
+ UPIU_TRANSACTION_DATA_IN = 0x22,
848
+ UPIU_TRANSACTION_TASK_RSP = 0x24,
849
+ UPIU_TRANSACTION_READY_XFER = 0x31,
850
+ UPIU_TRANSACTION_QUERY_RSP = 0x36,
851
+ UPIU_TRANSACTION_REJECT_UPIU = 0x3F,
852
+};
853
+
854
+/* UPIU Read/Write flags */
855
+enum {
856
+ UPIU_CMD_FLAGS_NONE = 0x00,
857
+ UPIU_CMD_FLAGS_WRITE = 0x20,
858
+ UPIU_CMD_FLAGS_READ = 0x40,
859
+};
860
+
861
+/* UPIU Task Attributes */
862
+enum {
863
+ UPIU_TASK_ATTR_SIMPLE = 0x00,
864
+ UPIU_TASK_ATTR_ORDERED = 0x01,
865
+ UPIU_TASK_ATTR_HEADQ = 0x02,
866
+ UPIU_TASK_ATTR_ACA = 0x03,
867
+};
868
+
869
+/* UPIU Query request function */
870
+enum {
871
+ UPIU_QUERY_FUNC_STANDARD_READ_REQUEST = 0x01,
872
+ UPIU_QUERY_FUNC_STANDARD_WRITE_REQUEST = 0x81,
873
+};
874
+
875
+/* Flag idn for Query Requests*/
876
+enum flag_idn {
877
+ QUERY_FLAG_IDN_FDEVICEINIT = 0x01,
878
+ QUERY_FLAG_IDN_PERMANENT_WPE = 0x02,
879
+ QUERY_FLAG_IDN_PWR_ON_WPE = 0x03,
880
+ QUERY_FLAG_IDN_BKOPS_EN = 0x04,
881
+ QUERY_FLAG_IDN_LIFE_SPAN_MODE_ENABLE = 0x05,
882
+ QUERY_FLAG_IDN_PURGE_ENABLE = 0x06,
883
+ QUERY_FLAG_IDN_REFRESH_ENABLE = 0x07,
884
+ QUERY_FLAG_IDN_FPHYRESOURCEREMOVAL = 0x08,
885
+ QUERY_FLAG_IDN_BUSY_RTC = 0x09,
886
+ QUERY_FLAG_IDN_RESERVED3 = 0x0A,
887
+ QUERY_FLAG_IDN_PERMANENTLY_DISABLE_FW_UPDATE = 0x0B,
888
+ QUERY_FLAG_IDN_WB_EN = 0x0E,
889
+ QUERY_FLAG_IDN_WB_BUFF_FLUSH_EN = 0x0F,
890
+ QUERY_FLAG_IDN_WB_BUFF_FLUSH_DURING_HIBERN8 = 0x10,
891
+ QUERY_FLAG_IDN_HPB_RESET = 0x11,
892
+ QUERY_FLAG_IDN_HPB_EN = 0x12,
893
+ QUERY_FLAG_IDN_COUNT,
894
+};
895
+
896
+/* Attribute idn for Query requests */
897
+enum attr_idn {
898
+ QUERY_ATTR_IDN_BOOT_LU_EN = 0x00,
899
+ QUERY_ATTR_IDN_MAX_HPB_SINGLE_CMD = 0x01,
900
+ QUERY_ATTR_IDN_POWER_MODE = 0x02,
901
+ QUERY_ATTR_IDN_ACTIVE_ICC_LVL = 0x03,
902
+ QUERY_ATTR_IDN_OOO_DATA_EN = 0x04,
903
+ QUERY_ATTR_IDN_BKOPS_STATUS = 0x05,
904
+ QUERY_ATTR_IDN_PURGE_STATUS = 0x06,
905
+ QUERY_ATTR_IDN_MAX_DATA_IN = 0x07,
906
+ QUERY_ATTR_IDN_MAX_DATA_OUT = 0x08,
907
+ QUERY_ATTR_IDN_DYN_CAP_NEEDED = 0x09,
908
+ QUERY_ATTR_IDN_REF_CLK_FREQ = 0x0A,
909
+ QUERY_ATTR_IDN_CONF_DESC_LOCK = 0x0B,
910
+ QUERY_ATTR_IDN_MAX_NUM_OF_RTT = 0x0C,
911
+ QUERY_ATTR_IDN_EE_CONTROL = 0x0D,
912
+ QUERY_ATTR_IDN_EE_STATUS = 0x0E,
913
+ QUERY_ATTR_IDN_SECONDS_PASSED = 0x0F,
914
+ QUERY_ATTR_IDN_CNTX_CONF = 0x10,
915
+ QUERY_ATTR_IDN_CORR_PRG_BLK_NUM = 0x11,
916
+ QUERY_ATTR_IDN_RESERVED2 = 0x12,
917
+ QUERY_ATTR_IDN_RESERVED3 = 0x13,
918
+ QUERY_ATTR_IDN_FFU_STATUS = 0x14,
919
+ QUERY_ATTR_IDN_PSA_STATE = 0x15,
920
+ QUERY_ATTR_IDN_PSA_DATA_SIZE = 0x16,
921
+ QUERY_ATTR_IDN_REF_CLK_GATING_WAIT_TIME = 0x17,
922
+ QUERY_ATTR_IDN_CASE_ROUGH_TEMP = 0x18,
923
+ QUERY_ATTR_IDN_HIGH_TEMP_BOUND = 0x19,
924
+ QUERY_ATTR_IDN_LOW_TEMP_BOUND = 0x1A,
925
+ QUERY_ATTR_IDN_THROTTLING_STATUS = 0x1B,
926
+ QUERY_ATTR_IDN_WB_FLUSH_STATUS = 0x1C,
927
+ QUERY_ATTR_IDN_AVAIL_WB_BUFF_SIZE = 0x1D,
928
+ QUERY_ATTR_IDN_WB_BUFF_LIFE_TIME_EST = 0x1E,
929
+ QUERY_ATTR_IDN_CURR_WB_BUFF_SIZE = 0x1F,
930
+ QUERY_ATTR_IDN_REFRESH_STATUS = 0x2C,
931
+ QUERY_ATTR_IDN_REFRESH_FREQ = 0x2D,
932
+ QUERY_ATTR_IDN_REFRESH_UNIT = 0x2E,
933
+ QUERY_ATTR_IDN_COUNT,
934
+};
935
+
936
+/* Descriptor idn for Query requests */
937
+enum desc_idn {
938
+ QUERY_DESC_IDN_DEVICE = 0x0,
939
+ QUERY_DESC_IDN_CONFIGURATION = 0x1,
940
+ QUERY_DESC_IDN_UNIT = 0x2,
941
+ QUERY_DESC_IDN_RFU_0 = 0x3,
942
+ QUERY_DESC_IDN_INTERCONNECT = 0x4,
943
+ QUERY_DESC_IDN_STRING = 0x5,
944
+ QUERY_DESC_IDN_RFU_1 = 0x6,
945
+ QUERY_DESC_IDN_GEOMETRY = 0x7,
946
+ QUERY_DESC_IDN_POWER = 0x8,
947
+ QUERY_DESC_IDN_HEALTH = 0x9,
948
+ QUERY_DESC_IDN_MAX,
949
+};
950
+
951
+enum desc_header_offset {
952
+ QUERY_DESC_LENGTH_OFFSET = 0x00,
953
+ QUERY_DESC_DESC_TYPE_OFFSET = 0x01,
954
+};
955
+
956
+/* Unit descriptor parameters offsets in bytes*/
957
+enum unit_desc_param {
958
+ UNIT_DESC_PARAM_LEN = 0x0,
959
+ UNIT_DESC_PARAM_TYPE = 0x1,
960
+ UNIT_DESC_PARAM_UNIT_INDEX = 0x2,
961
+ UNIT_DESC_PARAM_LU_ENABLE = 0x3,
962
+ UNIT_DESC_PARAM_BOOT_LUN_ID = 0x4,
963
+ UNIT_DESC_PARAM_LU_WR_PROTECT = 0x5,
964
+ UNIT_DESC_PARAM_LU_Q_DEPTH = 0x6,
965
+ UNIT_DESC_PARAM_PSA_SENSITIVE = 0x7,
966
+ UNIT_DESC_PARAM_MEM_TYPE = 0x8,
967
+ UNIT_DESC_PARAM_DATA_RELIABILITY = 0x9,
968
+ UNIT_DESC_PARAM_LOGICAL_BLK_SIZE = 0xA,
969
+ UNIT_DESC_PARAM_LOGICAL_BLK_COUNT = 0xB,
970
+ UNIT_DESC_PARAM_ERASE_BLK_SIZE = 0x13,
971
+ UNIT_DESC_PARAM_PROVISIONING_TYPE = 0x17,
972
+ UNIT_DESC_PARAM_PHY_MEM_RSRC_CNT = 0x18,
973
+ UNIT_DESC_PARAM_CTX_CAPABILITIES = 0x20,
974
+ UNIT_DESC_PARAM_LARGE_UNIT_SIZE_M1 = 0x22,
975
+ UNIT_DESC_PARAM_HPB_LU_MAX_ACTIVE_RGNS = 0x23,
976
+ UNIT_DESC_PARAM_HPB_PIN_RGN_START_OFF = 0x25,
977
+ UNIT_DESC_PARAM_HPB_NUM_PIN_RGNS = 0x27,
978
+ UNIT_DESC_PARAM_WB_BUF_ALLOC_UNITS = 0x29,
979
+};
980
+
981
+/* RPMB Unit descriptor parameters offsets in bytes*/
982
+enum rpmb_unit_desc_param {
983
+ RPMB_UNIT_DESC_PARAM_LEN = 0x0,
984
+ RPMB_UNIT_DESC_PARAM_TYPE = 0x1,
985
+ RPMB_UNIT_DESC_PARAM_UNIT_INDEX = 0x2,
986
+ RPMB_UNIT_DESC_PARAM_LU_ENABLE = 0x3,
987
+ RPMB_UNIT_DESC_PARAM_BOOT_LUN_ID = 0x4,
988
+ RPMB_UNIT_DESC_PARAM_LU_WR_PROTECT = 0x5,
989
+ RPMB_UNIT_DESC_PARAM_LU_Q_DEPTH = 0x6,
990
+ RPMB_UNIT_DESC_PARAM_PSA_SENSITIVE = 0x7,
991
+ RPMB_UNIT_DESC_PARAM_MEM_TYPE = 0x8,
992
+ RPMB_UNIT_DESC_PARAM_REGION_EN = 0x9,
993
+ RPMB_UNIT_DESC_PARAM_LOGICAL_BLK_SIZE = 0xA,
994
+ RPMB_UNIT_DESC_PARAM_LOGICAL_BLK_COUNT = 0xB,
995
+ RPMB_UNIT_DESC_PARAM_REGION0_SIZE = 0x13,
996
+ RPMB_UNIT_DESC_PARAM_REGION1_SIZE = 0x14,
997
+ RPMB_UNIT_DESC_PARAM_REGION2_SIZE = 0x15,
998
+ RPMB_UNIT_DESC_PARAM_REGION3_SIZE = 0x16,
999
+ RPMB_UNIT_DESC_PARAM_PROVISIONING_TYPE = 0x17,
1000
+ RPMB_UNIT_DESC_PARAM_PHY_MEM_RSRC_CNT = 0x18,
1001
+};
1002
+
1003
+/* Device descriptor parameters offsets in bytes*/
1004
+enum device_desc_param {
1005
+ DEVICE_DESC_PARAM_LEN = 0x0,
1006
+ DEVICE_DESC_PARAM_TYPE = 0x1,
1007
+ DEVICE_DESC_PARAM_DEVICE_TYPE = 0x2,
1008
+ DEVICE_DESC_PARAM_DEVICE_CLASS = 0x3,
1009
+ DEVICE_DESC_PARAM_DEVICE_SUB_CLASS = 0x4,
1010
+ DEVICE_DESC_PARAM_PRTCL = 0x5,
1011
+ DEVICE_DESC_PARAM_NUM_LU = 0x6,
1012
+ DEVICE_DESC_PARAM_NUM_WLU = 0x7,
1013
+ DEVICE_DESC_PARAM_BOOT_ENBL = 0x8,
1014
+ DEVICE_DESC_PARAM_DESC_ACCSS_ENBL = 0x9,
1015
+ DEVICE_DESC_PARAM_INIT_PWR_MODE = 0xA,
1016
+ DEVICE_DESC_PARAM_HIGH_PR_LUN = 0xB,
1017
+ DEVICE_DESC_PARAM_SEC_RMV_TYPE = 0xC,
1018
+ DEVICE_DESC_PARAM_SEC_LU = 0xD,
1019
+ DEVICE_DESC_PARAM_BKOP_TERM_LT = 0xE,
1020
+ DEVICE_DESC_PARAM_ACTVE_ICC_LVL = 0xF,
1021
+ DEVICE_DESC_PARAM_SPEC_VER = 0x10,
1022
+ DEVICE_DESC_PARAM_MANF_DATE = 0x12,
1023
+ DEVICE_DESC_PARAM_MANF_NAME = 0x14,
1024
+ DEVICE_DESC_PARAM_PRDCT_NAME = 0x15,
1025
+ DEVICE_DESC_PARAM_SN = 0x16,
1026
+ DEVICE_DESC_PARAM_OEM_ID = 0x17,
1027
+ DEVICE_DESC_PARAM_MANF_ID = 0x18,
1028
+ DEVICE_DESC_PARAM_UD_OFFSET = 0x1A,
1029
+ DEVICE_DESC_PARAM_UD_LEN = 0x1B,
1030
+ DEVICE_DESC_PARAM_RTT_CAP = 0x1C,
1031
+ DEVICE_DESC_PARAM_FRQ_RTC = 0x1D,
1032
+ DEVICE_DESC_PARAM_UFS_FEAT = 0x1F,
1033
+ DEVICE_DESC_PARAM_FFU_TMT = 0x20,
1034
+ DEVICE_DESC_PARAM_Q_DPTH = 0x21,
1035
+ DEVICE_DESC_PARAM_DEV_VER = 0x22,
1036
+ DEVICE_DESC_PARAM_NUM_SEC_WPA = 0x24,
1037
+ DEVICE_DESC_PARAM_PSA_MAX_DATA = 0x25,
1038
+ DEVICE_DESC_PARAM_PSA_TMT = 0x29,
1039
+ DEVICE_DESC_PARAM_PRDCT_REV = 0x2A,
1040
+ DEVICE_DESC_PARAM_HPB_VER = 0x40,
1041
+ DEVICE_DESC_PARAM_HPB_CONTROL = 0x42,
1042
+ DEVICE_DESC_PARAM_EXT_UFS_FEATURE_SUP = 0x4F,
1043
+ DEVICE_DESC_PARAM_WB_PRESRV_USRSPC_EN = 0x53,
1044
+ DEVICE_DESC_PARAM_WB_TYPE = 0x54,
1045
+ DEVICE_DESC_PARAM_WB_SHARED_ALLOC_UNITS = 0x55,
1046
+};
1047
+
1048
+/* Interconnect descriptor parameters offsets in bytes*/
1049
+enum interconnect_desc_param {
1050
+ INTERCONNECT_DESC_PARAM_LEN = 0x0,
1051
+ INTERCONNECT_DESC_PARAM_TYPE = 0x1,
1052
+ INTERCONNECT_DESC_PARAM_UNIPRO_VER = 0x2,
1053
+ INTERCONNECT_DESC_PARAM_MPHY_VER = 0x4,
1054
+};
1055
+
1056
+/* Geometry descriptor parameters offsets in bytes*/
1057
+enum geometry_desc_param {
1058
+ GEOMETRY_DESC_PARAM_LEN = 0x0,
1059
+ GEOMETRY_DESC_PARAM_TYPE = 0x1,
1060
+ GEOMETRY_DESC_PARAM_DEV_CAP = 0x4,
1061
+ GEOMETRY_DESC_PARAM_MAX_NUM_LUN = 0xC,
1062
+ GEOMETRY_DESC_PARAM_SEG_SIZE = 0xD,
1063
+ GEOMETRY_DESC_PARAM_ALLOC_UNIT_SIZE = 0x11,
1064
+ GEOMETRY_DESC_PARAM_MIN_BLK_SIZE = 0x12,
1065
+ GEOMETRY_DESC_PARAM_OPT_RD_BLK_SIZE = 0x13,
1066
+ GEOMETRY_DESC_PARAM_OPT_WR_BLK_SIZE = 0x14,
1067
+ GEOMETRY_DESC_PARAM_MAX_IN_BUF_SIZE = 0x15,
1068
+ GEOMETRY_DESC_PARAM_MAX_OUT_BUF_SIZE = 0x16,
1069
+ GEOMETRY_DESC_PARAM_RPMB_RW_SIZE = 0x17,
1070
+ GEOMETRY_DESC_PARAM_DYN_CAP_RSRC_PLC = 0x18,
1071
+ GEOMETRY_DESC_PARAM_DATA_ORDER = 0x19,
1072
+ GEOMETRY_DESC_PARAM_MAX_NUM_CTX = 0x1A,
1073
+ GEOMETRY_DESC_PARAM_TAG_UNIT_SIZE = 0x1B,
1074
+ GEOMETRY_DESC_PARAM_TAG_RSRC_SIZE = 0x1C,
1075
+ GEOMETRY_DESC_PARAM_SEC_RM_TYPES = 0x1D,
1076
+ GEOMETRY_DESC_PARAM_MEM_TYPES = 0x1E,
1077
+ GEOMETRY_DESC_PARAM_SCM_MAX_NUM_UNITS = 0x20,
1078
+ GEOMETRY_DESC_PARAM_SCM_CAP_ADJ_FCTR = 0x24,
1079
+ GEOMETRY_DESC_PARAM_NPM_MAX_NUM_UNITS = 0x26,
1080
+ GEOMETRY_DESC_PARAM_NPM_CAP_ADJ_FCTR = 0x2A,
1081
+ GEOMETRY_DESC_PARAM_ENM1_MAX_NUM_UNITS = 0x2C,
1082
+ GEOMETRY_DESC_PARAM_ENM1_CAP_ADJ_FCTR = 0x30,
1083
+ GEOMETRY_DESC_PARAM_ENM2_MAX_NUM_UNITS = 0x32,
1084
+ GEOMETRY_DESC_PARAM_ENM2_CAP_ADJ_FCTR = 0x36,
1085
+ GEOMETRY_DESC_PARAM_ENM3_MAX_NUM_UNITS = 0x38,
1086
+ GEOMETRY_DESC_PARAM_ENM3_CAP_ADJ_FCTR = 0x3C,
1087
+ GEOMETRY_DESC_PARAM_ENM4_MAX_NUM_UNITS = 0x3E,
1088
+ GEOMETRY_DESC_PARAM_ENM4_CAP_ADJ_FCTR = 0x42,
1089
+ GEOMETRY_DESC_PARAM_OPT_LOG_BLK_SIZE = 0x44,
1090
+ GEOMETRY_DESC_PARAM_HPB_REGION_SIZE = 0x48,
1091
+ GEOMETRY_DESC_PARAM_HPB_NUMBER_LU = 0x49,
1092
+ GEOMETRY_DESC_PARAM_HPB_SUBREGION_SIZE = 0x4A,
1093
+ GEOMETRY_DESC_PARAM_HPB_MAX_ACTIVE_REGS = 0x4B,
1094
+ GEOMETRY_DESC_PARAM_WB_MAX_ALLOC_UNITS = 0x4F,
1095
+ GEOMETRY_DESC_PARAM_WB_MAX_WB_LUNS = 0x53,
1096
+ GEOMETRY_DESC_PARAM_WB_BUFF_CAP_ADJ = 0x54,
1097
+ GEOMETRY_DESC_PARAM_WB_SUP_RED_TYPE = 0x55,
1098
+ GEOMETRY_DESC_PARAM_WB_SUP_WB_TYPE = 0x56,
1099
+};
1100
+
1101
+/* Health descriptor parameters offsets in bytes*/
1102
+enum health_desc_param {
1103
+ HEALTH_DESC_PARAM_LEN = 0x0,
1104
+ HEALTH_DESC_PARAM_TYPE = 0x1,
1105
+ HEALTH_DESC_PARAM_EOL_INFO = 0x2,
1106
+ HEALTH_DESC_PARAM_LIFE_TIME_EST_A = 0x3,
1107
+ HEALTH_DESC_PARAM_LIFE_TIME_EST_B = 0x4,
1108
+};
1109
+
1110
+/* WriteBooster buffer mode */
1111
+enum {
1112
+ WB_BUF_MODE_LU_DEDICATED = 0x0,
1113
+ WB_BUF_MODE_SHARED = 0x1,
1114
+};
1115
+
1116
+/*
1117
+ * Logical Unit Write Protect
1118
+ * 00h: LU not write protected
1119
+ * 01h: LU write protected when fPowerOnWPEn =1
1120
+ * 02h: LU permanently write protected when fPermanentWPEn =1
1121
+ */
1122
+enum ufs_lu_wp_type {
1123
+ UFS_LU_NO_WP = 0x00,
1124
+ UFS_LU_POWER_ON_WP = 0x01,
1125
+ UFS_LU_PERM_WP = 0x02,
1126
+};
1127
+
1128
+/* UTP QUERY Transaction Specific Fields OpCode */
1129
+enum query_opcode {
1130
+ UPIU_QUERY_OPCODE_NOP = 0x0,
1131
+ UPIU_QUERY_OPCODE_READ_DESC = 0x1,
1132
+ UPIU_QUERY_OPCODE_WRITE_DESC = 0x2,
1133
+ UPIU_QUERY_OPCODE_READ_ATTR = 0x3,
1134
+ UPIU_QUERY_OPCODE_WRITE_ATTR = 0x4,
1135
+ UPIU_QUERY_OPCODE_READ_FLAG = 0x5,
1136
+ UPIU_QUERY_OPCODE_SET_FLAG = 0x6,
1137
+ UPIU_QUERY_OPCODE_CLEAR_FLAG = 0x7,
1138
+ UPIU_QUERY_OPCODE_TOGGLE_FLAG = 0x8,
1139
+};
1140
+
1141
+/* Query response result code */
1142
+typedef enum QueryRespCode {
1143
+ QUERY_RESULT_SUCCESS = 0x00,
1144
+ QUERY_RESULT_NOT_READABLE = 0xF6,
1145
+ QUERY_RESULT_NOT_WRITEABLE = 0xF7,
1146
+ QUERY_RESULT_ALREADY_WRITTEN = 0xF8,
1147
+ QUERY_RESULT_INVALID_LENGTH = 0xF9,
1148
+ QUERY_RESULT_INVALID_VALUE = 0xFA,
1149
+ QUERY_RESULT_INVALID_SELECTOR = 0xFB,
1150
+ QUERY_RESULT_INVALID_INDEX = 0xFC,
1151
+ QUERY_RESULT_INVALID_IDN = 0xFD,
1152
+ QUERY_RESULT_INVALID_OPCODE = 0xFE,
1153
+ QUERY_RESULT_GENERAL_FAILURE = 0xFF,
1154
+} QueryRespCode;
1155
+
1156
+/* UTP Transfer Request Command Type (CT) */
1157
+enum {
1158
+ UPIU_COMMAND_SET_TYPE_SCSI = 0x0,
1159
+ UPIU_COMMAND_SET_TYPE_UFS = 0x1,
1160
+ UPIU_COMMAND_SET_TYPE_QUERY = 0x2,
1161
+};
1162
+
1163
+/* Task management service response */
1164
+enum {
1165
+ UPIU_TASK_MANAGEMENT_FUNC_COMPL = 0x00,
1166
+ UPIU_TASK_MANAGEMENT_FUNC_NOT_SUPPORTED = 0x04,
1167
+ UPIU_TASK_MANAGEMENT_FUNC_SUCCEEDED = 0x08,
1168
+ UPIU_TASK_MANAGEMENT_FUNC_FAILED = 0x05,
1169
+ UPIU_INCORRECT_LOGICAL_UNIT_NO = 0x09,
1170
+};
1171
+
1172
+/* UFS device power modes */
1173
+enum ufs_dev_pwr_mode {
1174
+ UFS_ACTIVE_PWR_MODE = 1,
1175
+ UFS_SLEEP_PWR_MODE = 2,
1176
+ UFS_POWERDOWN_PWR_MODE = 3,
1177
+ UFS_DEEPSLEEP_PWR_MODE = 4,
1178
+};
1179
+
1180
+/*
1181
+ * struct UtpCmdRsp - Response UPIU structure
1182
+ * @residual_transfer_count: Residual transfer count DW-3
1183
+ * @reserved: Reserved double words DW-4 to DW-7
1184
+ * @sense_data_len: Sense data length DW-8 U16
1185
+ * @sense_data: Sense data field DW-8 to DW-12
1186
+ */
1187
+typedef struct QEMU_PACKED UtpCmdRsp {
1188
+ uint32_t residual_transfer_count;
1189
+ uint32_t reserved[4];
1190
+ uint16_t sense_data_len;
1191
+ uint8_t sense_data[UFS_SENSE_SIZE];
1192
+} UtpCmdRsp;
1193
+
1194
+/*
1195
+ * struct UtpUpiuRsp - general upiu response structure
1196
+ * @header: UPIU header structure DW-0 to DW-2
1197
+ * @sr: fields structure for scsi command DW-3 to DW-12
1198
+ * @qr: fields structure for query request DW-3 to DW-7
1199
+ */
1200
+typedef struct QEMU_PACKED UtpUpiuRsp {
1201
+ UtpUpiuHeader header;
1202
+ union {
1203
+ UtpCmdRsp sr;
1204
+ UtpUpiuQuery qr;
1205
+ };
1206
+} UtpUpiuRsp;
1207
+
1208
+static inline void _ufs_check_size(void)
1209
+{
1210
+ QEMU_BUILD_BUG_ON(sizeof(UfsReg) != 0x104);
1211
+ QEMU_BUILD_BUG_ON(sizeof(DeviceDescriptor) != 89);
1212
+ QEMU_BUILD_BUG_ON(sizeof(GeometryDescriptor) != 87);
1213
+ QEMU_BUILD_BUG_ON(sizeof(UnitDescriptor) != 45);
1214
+ QEMU_BUILD_BUG_ON(sizeof(RpmbUnitDescriptor) != 35);
1215
+ QEMU_BUILD_BUG_ON(sizeof(PowerParametersDescriptor) != 98);
1216
+ QEMU_BUILD_BUG_ON(sizeof(InterconnectDescriptor) != 6);
1217
+ QEMU_BUILD_BUG_ON(sizeof(StringDescriptor) != 254);
1218
+ QEMU_BUILD_BUG_ON(sizeof(DeviceHealthDescriptor) != 45);
1219
+ QEMU_BUILD_BUG_ON(sizeof(Flags) != 0x13);
1220
+ QEMU_BUILD_BUG_ON(sizeof(UtpUpiuHeader) != 12);
1221
+ QEMU_BUILD_BUG_ON(sizeof(UtpUpiuQuery) != 276);
1222
+ QEMU_BUILD_BUG_ON(sizeof(UtpUpiuCmd) != 20);
1223
+ QEMU_BUILD_BUG_ON(sizeof(UtpUpiuReq) != 288);
1224
+ QEMU_BUILD_BUG_ON(sizeof(UfshcdSgEntry) != 16);
1225
+ QEMU_BUILD_BUG_ON(sizeof(RequestDescHeader) != 16);
1226
+ QEMU_BUILD_BUG_ON(sizeof(UtpTransferReqDesc) != 32);
1227
+ QEMU_BUILD_BUG_ON(sizeof(UtpTaskReqDesc) != 80);
1228
+ QEMU_BUILD_BUG_ON(sizeof(UtpCmdRsp) != 40);
1229
+ QEMU_BUILD_BUG_ON(sizeof(UtpUpiuRsp) != 288);
1230
+}
1231
+#endif
1232
diff --git a/include/hw/pci/pci.h b/include/hw/pci/pci.h
1233
index XXXXXXX..XXXXXXX 100644
1234
--- a/include/hw/pci/pci.h
1235
+++ b/include/hw/pci/pci.h
1236
@@ -XXX,XX +XXX,XX @@ extern bool pci_available;
1237
#define PCI_DEVICE_ID_REDHAT_NVME 0x0010
1238
#define PCI_DEVICE_ID_REDHAT_PVPANIC 0x0011
1239
#define PCI_DEVICE_ID_REDHAT_ACPI_ERST 0x0012
1240
+#define PCI_DEVICE_ID_REDHAT_UFS 0x0013
1241
#define PCI_DEVICE_ID_REDHAT_QXL 0x0100
1242
1243
#define FMT_PCIBUS PRIx64
1244
diff --git a/include/hw/pci/pci_ids.h b/include/hw/pci/pci_ids.h
1245
index XXXXXXX..XXXXXXX 100644
1246
--- a/include/hw/pci/pci_ids.h
1247
+++ b/include/hw/pci/pci_ids.h
1248
@@ -XXX,XX +XXX,XX @@
1249
#define PCI_CLASS_STORAGE_SATA 0x0106
1250
#define PCI_CLASS_STORAGE_SAS 0x0107
1251
#define PCI_CLASS_STORAGE_EXPRESS 0x0108
1252
+#define PCI_CLASS_STORAGE_UFS 0x0109
1253
#define PCI_CLASS_STORAGE_OTHER 0x0180
1254
1255
#define PCI_BASE_CLASS_NETWORK 0x02
1256
diff --git a/hw/ufs/ufs.c b/hw/ufs/ufs.c
1257
new file mode 100644
1258
index XXXXXXX..XXXXXXX
1259
--- /dev/null
1260
+++ b/hw/ufs/ufs.c
1261
@@ -XXX,XX +XXX,XX @@
1262
+/*
1263
+ * QEMU Universal Flash Storage (UFS) Controller
1264
+ *
1265
+ * Copyright (c) 2023 Samsung Electronics Co., Ltd. All rights reserved.
1266
+ *
1267
+ * Written by Jeuk Kim <jeuk20.kim@samsung.com>
1268
+ *
1269
+ * SPDX-License-Identifier: GPL-2.0-or-later
1270
+ */
1271
+
91
+#include "qemu/osdep.h"
1272
+#include "qemu/osdep.h"
92
+#include "block/block.h"
93
+#include "vhost-user-blk-server.h"
94
+#include "qapi/error.h"
1273
+#include "qapi/error.h"
95
+#include "qom/object_interfaces.h"
1274
+#include "migration/vmstate.h"
96
+#include "sysemu/block-backend.h"
1275
+#include "trace.h"
97
+#include "util/block-helpers.h"
1276
+#include "ufs.h"
98
+
1277
+
99
+enum {
1278
+/* The QEMU-UFS device follows spec version 3.1 */
100
+ VHOST_USER_BLK_MAX_QUEUES = 1,
1279
+#define UFS_SPEC_VER 0x00000310
101
+};
1280
+#define UFS_MAX_NUTRS 32
102
+struct virtio_blk_inhdr {
1281
+#define UFS_MAX_NUTMRS 8
103
+ unsigned char status;
1282
+
104
+};
1283
+static void ufs_irq_check(UfsHc *u)
105
+
106
+typedef struct VuBlockReq {
107
+ VuVirtqElement *elem;
108
+ int64_t sector_num;
109
+ size_t size;
110
+ struct virtio_blk_inhdr *in;
111
+ struct virtio_blk_outhdr out;
112
+ VuServer *server;
113
+ struct VuVirtq *vq;
114
+} VuBlockReq;
115
+
116
+static void vu_block_req_complete(VuBlockReq *req)
117
+{
1284
+{
118
+ VuDev *vu_dev = &req->server->vu_dev;
1285
+ PCIDevice *pci = PCI_DEVICE(u);
119
+
1286
+
120
+ /* IO size with 1 extra status byte */
1287
+ if ((u->reg.is & UFS_INTR_MASK) & u->reg.ie) {
121
+ vu_queue_push(vu_dev, req->vq, req->elem, req->size + 1);
1288
+ trace_ufs_irq_raise();
122
+ vu_queue_notify(vu_dev, req->vq);
1289
+ pci_irq_assert(pci);
123
+
1290
+ } else {
124
+ if (req->elem) {
1291
+ trace_ufs_irq_lower();
125
+ free(req->elem);
1292
+ pci_irq_deassert(pci);
126
+ }
1293
+ }
127
+
128
+ g_free(req);
129
+}
1294
+}
130
+
1295
+
131
+static VuBlockDev *get_vu_block_device_by_server(VuServer *server)
1296
+static void ufs_process_uiccmd(UfsHc *u, uint32_t val)
132
+{
1297
+{
133
+ return container_of(server, VuBlockDev, vu_server);
1298
+ trace_ufs_process_uiccmd(val, u->reg.ucmdarg1, u->reg.ucmdarg2,
1299
+ u->reg.ucmdarg3);
1300
+ /*
1301
+ * Only the essential uic commands for running drivers on Linux and Windows
1302
+ * are implemented.
1303
+ */
1304
+ switch (val) {
1305
+ case UIC_CMD_DME_LINK_STARTUP:
1306
+ u->reg.hcs = FIELD_DP32(u->reg.hcs, HCS, DP, 1);
1307
+ u->reg.hcs = FIELD_DP32(u->reg.hcs, HCS, UTRLRDY, 1);
1308
+ u->reg.hcs = FIELD_DP32(u->reg.hcs, HCS, UTMRLRDY, 1);
1309
+ u->reg.ucmdarg2 = UIC_CMD_RESULT_SUCCESS;
1310
+ break;
1311
+ /* TODO: Revisit it when Power Management is implemented */
1312
+ case UIC_CMD_DME_HIBER_ENTER:
1313
+ u->reg.is = FIELD_DP32(u->reg.is, IS, UHES, 1);
1314
+ u->reg.hcs = FIELD_DP32(u->reg.hcs, HCS, UPMCRS, PWR_LOCAL);
1315
+ u->reg.ucmdarg2 = UIC_CMD_RESULT_SUCCESS;
1316
+ break;
1317
+ case UIC_CMD_DME_HIBER_EXIT:
1318
+ u->reg.is = FIELD_DP32(u->reg.is, IS, UHXS, 1);
1319
+ u->reg.hcs = FIELD_DP32(u->reg.hcs, HCS, UPMCRS, PWR_LOCAL);
1320
+ u->reg.ucmdarg2 = UIC_CMD_RESULT_SUCCESS;
1321
+ break;
1322
+ default:
1323
+ u->reg.ucmdarg2 = UIC_CMD_RESULT_FAILURE;
1324
+ }
1325
+
1326
+ u->reg.is = FIELD_DP32(u->reg.is, IS, UCCS, 1);
1327
+
1328
+ ufs_irq_check(u);
134
+}
1329
+}
135
+
1330
+
136
+static int coroutine_fn
1331
+static void ufs_write_reg(UfsHc *u, hwaddr offset, uint32_t data, unsigned size)
137
+vu_block_discard_write_zeroes(VuBlockReq *req, struct iovec *iov,
138
+ uint32_t iovcnt, uint32_t type)
139
+{
1332
+{
140
+ struct virtio_blk_discard_write_zeroes desc;
1333
+ switch (offset) {
141
+ ssize_t size = iov_to_buf(iov, iovcnt, 0, &desc, sizeof(desc));
1334
+ case A_IS:
142
+ if (unlikely(size != sizeof(desc))) {
1335
+ u->reg.is &= ~data;
143
+ error_report("Invalid size %zd, expect %zu", size, sizeof(desc));
1336
+ ufs_irq_check(u);
144
+ return -EINVAL;
1337
+ break;
1338
+ case A_IE:
1339
+ u->reg.ie = data;
1340
+ ufs_irq_check(u);
1341
+ break;
1342
+ case A_HCE:
1343
+ if (!FIELD_EX32(u->reg.hce, HCE, HCE) && FIELD_EX32(data, HCE, HCE)) {
1344
+ u->reg.hcs = FIELD_DP32(u->reg.hcs, HCS, UCRDY, 1);
1345
+ u->reg.hce = FIELD_DP32(u->reg.hce, HCE, HCE, 1);
1346
+ } else if (FIELD_EX32(u->reg.hce, HCE, HCE) &&
1347
+ !FIELD_EX32(data, HCE, HCE)) {
1348
+ u->reg.hcs = 0;
1349
+ u->reg.hce = FIELD_DP32(u->reg.hce, HCE, HCE, 0);
1350
+ }
1351
+ break;
1352
+ case A_UTRLBA:
1353
+ u->reg.utrlba = data & R_UTRLBA_UTRLBA_MASK;
1354
+ break;
1355
+ case A_UTRLBAU:
1356
+ u->reg.utrlbau = data;
1357
+ break;
1358
+ case A_UTRLDBR:
1359
+ /* Not yet supported */
1360
+ break;
1361
+ case A_UTRLRSR:
1362
+ u->reg.utrlrsr = data;
1363
+ break;
1364
+ case A_UTRLCNR:
1365
+ u->reg.utrlcnr &= ~data;
1366
+ break;
1367
+ case A_UTMRLBA:
1368
+ u->reg.utmrlba = data & R_UTMRLBA_UTMRLBA_MASK;
1369
+ break;
1370
+ case A_UTMRLBAU:
1371
+ u->reg.utmrlbau = data;
1372
+ break;
1373
+ case A_UICCMD:
1374
+ ufs_process_uiccmd(u, data);
1375
+ break;
1376
+ case A_UCMDARG1:
1377
+ u->reg.ucmdarg1 = data;
1378
+ break;
1379
+ case A_UCMDARG2:
1380
+ u->reg.ucmdarg2 = data;
1381
+ break;
1382
+ case A_UCMDARG3:
1383
+ u->reg.ucmdarg3 = data;
1384
+ break;
1385
+ case A_UTRLCLR:
1386
+ case A_UTMRLDBR:
1387
+ case A_UTMRLCLR:
1388
+ case A_UTMRLRSR:
1389
+ trace_ufs_err_unsupport_register_offset(offset);
1390
+ break;
1391
+ default:
1392
+ trace_ufs_err_invalid_register_offset(offset);
1393
+ break;
145
+ }
1394
+ }
146
+
1395
+}
147
+ VuBlockDev *vdev_blk = get_vu_block_device_by_server(req->server);
1396
+
148
+ uint64_t range[2] = { le64_to_cpu(desc.sector) << 9,
1397
+static uint64_t ufs_mmio_read(void *opaque, hwaddr addr, unsigned size)
149
+ le32_to_cpu(desc.num_sectors) << 9 };
1398
+{
150
+ if (type == VIRTIO_BLK_T_DISCARD) {
1399
+ UfsHc *u = (UfsHc *)opaque;
151
+ if (blk_co_pdiscard(vdev_blk->backend, range[0], range[1]) == 0) {
1400
+ uint8_t *ptr = (uint8_t *)&u->reg;
152
+ return 0;
1401
+ uint64_t value;
153
+ }
1402
+
154
+ } else if (type == VIRTIO_BLK_T_WRITE_ZEROES) {
1403
+ if (addr > sizeof(u->reg) - size) {
155
+ if (blk_co_pwrite_zeroes(vdev_blk->backend,
1404
+ trace_ufs_err_invalid_register_offset(addr);
156
+ range[0], range[1], 0) == 0) {
1405
+ return 0;
157
+ return 0;
158
+ }
159
+ }
1406
+ }
160
+
1407
+
161
+ return -EINVAL;
1408
+ value = *(uint32_t *)(ptr + addr);
1409
+ trace_ufs_mmio_read(addr, value, size);
1410
+ return value;
162
+}
1411
+}
163
+
1412
+
164
+static void coroutine_fn vu_block_flush(VuBlockReq *req)
1413
+static void ufs_mmio_write(void *opaque, hwaddr addr, uint64_t data,
1414
+ unsigned size)
165
+{
1415
+{
166
+ VuBlockDev *vdev_blk = get_vu_block_device_by_server(req->server);
1416
+ UfsHc *u = (UfsHc *)opaque;
167
+ BlockBackend *backend = vdev_blk->backend;
1417
+
168
+ blk_co_flush(backend);
1418
+ if (addr > sizeof(u->reg) - size) {
169
+}
1419
+ trace_ufs_err_invalid_register_offset(addr);
170
+
171
+struct req_data {
172
+ VuServer *server;
173
+ VuVirtq *vq;
174
+ VuVirtqElement *elem;
175
+};
176
+
177
+static void coroutine_fn vu_block_virtio_process_req(void *opaque)
178
+{
179
+ struct req_data *data = opaque;
180
+ VuServer *server = data->server;
181
+ VuVirtq *vq = data->vq;
182
+ VuVirtqElement *elem = data->elem;
183
+ uint32_t type;
184
+ VuBlockReq *req;
185
+
186
+ VuBlockDev *vdev_blk = get_vu_block_device_by_server(server);
187
+ BlockBackend *backend = vdev_blk->backend;
188
+
189
+ struct iovec *in_iov = elem->in_sg;
190
+ struct iovec *out_iov = elem->out_sg;
191
+ unsigned in_num = elem->in_num;
192
+ unsigned out_num = elem->out_num;
193
+ /* refer to hw/block/virtio_blk.c */
194
+ if (elem->out_num < 1 || elem->in_num < 1) {
195
+ error_report("virtio-blk request missing headers");
196
+ free(elem);
197
+ return;
1420
+ return;
198
+ }
1421
+ }
199
+
1422
+
200
+ req = g_new0(VuBlockReq, 1);
1423
+ trace_ufs_mmio_write(addr, data, size);
201
+ req->server = server;
1424
+ ufs_write_reg(u, addr, data, size);
202
+ req->vq = vq;
1425
+}
203
+ req->elem = elem;
1426
+
204
+
1427
+static const MemoryRegionOps ufs_mmio_ops = {
205
+ if (unlikely(iov_to_buf(out_iov, out_num, 0, &req->out,
1428
+ .read = ufs_mmio_read,
206
+ sizeof(req->out)) != sizeof(req->out))) {
1429
+ .write = ufs_mmio_write,
207
+ error_report("virtio-blk request outhdr too short");
1430
+ .endianness = DEVICE_LITTLE_ENDIAN,
208
+ goto err;
1431
+ .impl = {
1432
+ .min_access_size = 4,
1433
+ .max_access_size = 4,
1434
+ },
1435
+};
1436
+
1437
+static bool ufs_check_constraints(UfsHc *u, Error **errp)
1438
+{
1439
+ if (u->params.nutrs > UFS_MAX_NUTRS) {
1440
+ error_setg(errp, "nutrs must be less than or equal to %d",
1441
+ UFS_MAX_NUTRS);
1442
+ return false;
209
+ }
1443
+ }
210
+
1444
+
211
+ iov_discard_front(&out_iov, &out_num, sizeof(req->out));
1445
+ if (u->params.nutmrs > UFS_MAX_NUTMRS) {
212
+
1446
+ error_setg(errp, "nutmrs must be less than or equal to %d",
213
+ if (in_iov[in_num - 1].iov_len < sizeof(struct virtio_blk_inhdr)) {
1447
+ UFS_MAX_NUTMRS);
214
+ error_report("virtio-blk request inhdr too short");
1448
+ return false;
215
+ goto err;
216
+ }
1449
+ }
217
+
1450
+
218
+ /* We always touch the last byte, so just see how big in_iov is. */
1451
+ return true;
219
+ req->in = (void *)in_iov[in_num - 1].iov_base
220
+ + in_iov[in_num - 1].iov_len
221
+ - sizeof(struct virtio_blk_inhdr);
222
+ iov_discard_back(in_iov, &in_num, sizeof(struct virtio_blk_inhdr));
223
+
224
+ type = le32_to_cpu(req->out.type);
225
+ switch (type & ~VIRTIO_BLK_T_BARRIER) {
226
+ case VIRTIO_BLK_T_IN:
227
+ case VIRTIO_BLK_T_OUT: {
228
+ ssize_t ret = 0;
229
+ bool is_write = type & VIRTIO_BLK_T_OUT;
230
+ req->sector_num = le64_to_cpu(req->out.sector);
231
+
232
+ int64_t offset = req->sector_num * vdev_blk->blk_size;
233
+ QEMUIOVector qiov;
234
+ if (is_write) {
235
+ qemu_iovec_init_external(&qiov, out_iov, out_num);
236
+ ret = blk_co_pwritev(backend, offset, qiov.size,
237
+ &qiov, 0);
238
+ } else {
239
+ qemu_iovec_init_external(&qiov, in_iov, in_num);
240
+ ret = blk_co_preadv(backend, offset, qiov.size,
241
+ &qiov, 0);
242
+ }
243
+ if (ret >= 0) {
244
+ req->in->status = VIRTIO_BLK_S_OK;
245
+ } else {
246
+ req->in->status = VIRTIO_BLK_S_IOERR;
247
+ }
248
+ break;
249
+ }
250
+ case VIRTIO_BLK_T_FLUSH:
251
+ vu_block_flush(req);
252
+ req->in->status = VIRTIO_BLK_S_OK;
253
+ break;
254
+ case VIRTIO_BLK_T_GET_ID: {
255
+ size_t size = MIN(iov_size(&elem->in_sg[0], in_num),
256
+ VIRTIO_BLK_ID_BYTES);
257
+ snprintf(elem->in_sg[0].iov_base, size, "%s", "vhost_user_blk");
258
+ req->in->status = VIRTIO_BLK_S_OK;
259
+ req->size = elem->in_sg[0].iov_len;
260
+ break;
261
+ }
262
+ case VIRTIO_BLK_T_DISCARD:
263
+ case VIRTIO_BLK_T_WRITE_ZEROES: {
264
+ int rc;
265
+ rc = vu_block_discard_write_zeroes(req, &elem->out_sg[1],
266
+ out_num, type);
267
+ if (rc == 0) {
268
+ req->in->status = VIRTIO_BLK_S_OK;
269
+ } else {
270
+ req->in->status = VIRTIO_BLK_S_IOERR;
271
+ }
272
+ break;
273
+ }
274
+ default:
275
+ req->in->status = VIRTIO_BLK_S_UNSUPP;
276
+ break;
277
+ }
278
+
279
+ vu_block_req_complete(req);
280
+ return;
281
+
282
+err:
283
+ free(elem);
284
+ g_free(req);
285
+ return;
286
+}
1452
+}
287
+
1453
+
288
+static void vu_block_process_vq(VuDev *vu_dev, int idx)
1454
+static void ufs_init_pci(UfsHc *u, PCIDevice *pci_dev)
289
+{
1455
+{
290
+ VuServer *server;
1456
+ uint8_t *pci_conf = pci_dev->config;
291
+ VuVirtq *vq;
1457
+
292
+ struct req_data *req_data;
1458
+ pci_conf[PCI_INTERRUPT_PIN] = 1;
293
+
1459
+ pci_config_set_prog_interface(pci_conf, 0x1);
294
+ server = container_of(vu_dev, VuServer, vu_dev);
1460
+
295
+ assert(server);
1461
+ memory_region_init_io(&u->iomem, OBJECT(u), &ufs_mmio_ops, u, "ufs",
296
+
1462
+ u->reg_size);
297
+ vq = vu_get_queue(vu_dev, idx);
1463
+ pci_register_bar(pci_dev, 0, PCI_BASE_ADDRESS_SPACE_MEMORY, &u->iomem);
298
+ assert(vq);
1464
+ u->irq = pci_allocate_irq(pci_dev);
299
+ VuVirtqElement *elem;
300
+ while (1) {
301
+ elem = vu_queue_pop(vu_dev, vq, sizeof(VuVirtqElement) +
302
+ sizeof(VuBlockReq));
303
+ if (elem) {
304
+ req_data = g_new0(struct req_data, 1);
305
+ req_data->server = server;
306
+ req_data->vq = vq;
307
+ req_data->elem = elem;
308
+ Coroutine *co = qemu_coroutine_create(vu_block_virtio_process_req,
309
+ req_data);
310
+ aio_co_enter(server->ioc->ctx, co);
311
+ } else {
312
+ break;
313
+ }
314
+ }
315
+}
1465
+}
316
+
1466
+
317
+static void vu_block_queue_set_started(VuDev *vu_dev, int idx, bool started)
1467
+static void ufs_init_hc(UfsHc *u)
318
+{
1468
+{
319
+ VuVirtq *vq;
1469
+ uint32_t cap = 0;
320
+
1470
+
321
+ assert(vu_dev);
1471
+ u->reg_size = pow2ceil(sizeof(UfsReg));
322
+
1472
+
323
+ vq = vu_get_queue(vu_dev, idx);
1473
+ memset(&u->reg, 0, sizeof(u->reg));
324
+ vu_set_queue_handler(vu_dev, vq, started ? vu_block_process_vq : NULL);
1474
+ cap = FIELD_DP32(cap, CAP, NUTRS, (u->params.nutrs - 1));
1475
+ cap = FIELD_DP32(cap, CAP, RTT, 2);
1476
+ cap = FIELD_DP32(cap, CAP, NUTMRS, (u->params.nutmrs - 1));
1477
+ cap = FIELD_DP32(cap, CAP, AUTOH8, 0);
1478
+ cap = FIELD_DP32(cap, CAP, 64AS, 1);
1479
+ cap = FIELD_DP32(cap, CAP, OODDS, 0);
1480
+ cap = FIELD_DP32(cap, CAP, UICDMETMS, 0);
1481
+ cap = FIELD_DP32(cap, CAP, CS, 0);
1482
+ u->reg.cap = cap;
1483
+ u->reg.ver = UFS_SPEC_VER;
325
+}
1484
+}
326
+
1485
+
327
+static uint64_t vu_block_get_features(VuDev *dev)
1486
+static void ufs_realize(PCIDevice *pci_dev, Error **errp)
328
+{
1487
+{
329
+ uint64_t features;
1488
+ UfsHc *u = UFS(pci_dev);
330
+ VuServer *server = container_of(dev, VuServer, vu_dev);
1489
+
331
+ VuBlockDev *vdev_blk = get_vu_block_device_by_server(server);
1490
+ if (!ufs_check_constraints(u, errp)) {
332
+ features = 1ull << VIRTIO_BLK_F_SIZE_MAX |
333
+ 1ull << VIRTIO_BLK_F_SEG_MAX |
334
+ 1ull << VIRTIO_BLK_F_TOPOLOGY |
335
+ 1ull << VIRTIO_BLK_F_BLK_SIZE |
336
+ 1ull << VIRTIO_BLK_F_FLUSH |
337
+ 1ull << VIRTIO_BLK_F_DISCARD |
338
+ 1ull << VIRTIO_BLK_F_WRITE_ZEROES |
339
+ 1ull << VIRTIO_BLK_F_CONFIG_WCE |
340
+ 1ull << VIRTIO_F_VERSION_1 |
341
+ 1ull << VIRTIO_RING_F_INDIRECT_DESC |
342
+ 1ull << VIRTIO_RING_F_EVENT_IDX |
343
+ 1ull << VHOST_USER_F_PROTOCOL_FEATURES;
344
+
345
+ if (!vdev_blk->writable) {
346
+ features |= 1ull << VIRTIO_BLK_F_RO;
347
+ }
348
+
349
+ return features;
350
+}
351
+
352
+static uint64_t vu_block_get_protocol_features(VuDev *dev)
353
+{
354
+ return 1ull << VHOST_USER_PROTOCOL_F_CONFIG |
355
+ 1ull << VHOST_USER_PROTOCOL_F_INFLIGHT_SHMFD;
356
+}
357
+
358
+static int
359
+vu_block_get_config(VuDev *vu_dev, uint8_t *config, uint32_t len)
360
+{
361
+ VuServer *server = container_of(vu_dev, VuServer, vu_dev);
362
+ VuBlockDev *vdev_blk = get_vu_block_device_by_server(server);
363
+ memcpy(config, &vdev_blk->blkcfg, len);
364
+
365
+ return 0;
366
+}
367
+
368
+static int
369
+vu_block_set_config(VuDev *vu_dev, const uint8_t *data,
370
+ uint32_t offset, uint32_t size, uint32_t flags)
371
+{
372
+ VuServer *server = container_of(vu_dev, VuServer, vu_dev);
373
+ VuBlockDev *vdev_blk = get_vu_block_device_by_server(server);
374
+ uint8_t wce;
375
+
376
+ /* don't support live migration */
377
+ if (flags != VHOST_SET_CONFIG_TYPE_MASTER) {
378
+ return -EINVAL;
379
+ }
380
+
381
+ if (offset != offsetof(struct virtio_blk_config, wce) ||
382
+ size != 1) {
383
+ return -EINVAL;
384
+ }
385
+
386
+ wce = *data;
387
+ vdev_blk->blkcfg.wce = wce;
388
+ blk_set_enable_write_cache(vdev_blk->backend, wce);
389
+ return 0;
390
+}
391
+
392
+/*
393
+ * When the client disconnects, it sends a VHOST_USER_NONE request
394
+ * and vu_process_message will simple call exit which cause the VM
395
+ * to exit abruptly.
396
+ * To avoid this issue, process VHOST_USER_NONE request ahead
397
+ * of vu_process_message.
398
+ *
399
+ */
400
+static int vu_block_process_msg(VuDev *dev, VhostUserMsg *vmsg, int *do_reply)
401
+{
402
+ if (vmsg->request == VHOST_USER_NONE) {
403
+ dev->panic(dev, "disconnect");
404
+ return true;
405
+ }
406
+ return false;
407
+}
408
+
409
+static const VuDevIface vu_block_iface = {
410
+ .get_features = vu_block_get_features,
411
+ .queue_set_started = vu_block_queue_set_started,
412
+ .get_protocol_features = vu_block_get_protocol_features,
413
+ .get_config = vu_block_get_config,
414
+ .set_config = vu_block_set_config,
415
+ .process_msg = vu_block_process_msg,
416
+};
417
+
418
+static void blk_aio_attached(AioContext *ctx, void *opaque)
419
+{
420
+ VuBlockDev *vub_dev = opaque;
421
+ aio_context_acquire(ctx);
422
+ vhost_user_server_set_aio_context(&vub_dev->vu_server, ctx);
423
+ aio_context_release(ctx);
424
+}
425
+
426
+static void blk_aio_detach(void *opaque)
427
+{
428
+ VuBlockDev *vub_dev = opaque;
429
+ AioContext *ctx = vub_dev->vu_server.ctx;
430
+ aio_context_acquire(ctx);
431
+ vhost_user_server_set_aio_context(&vub_dev->vu_server, NULL);
432
+ aio_context_release(ctx);
433
+}
434
+
435
+static void
436
+vu_block_initialize_config(BlockDriverState *bs,
437
+ struct virtio_blk_config *config, uint32_t blk_size)
438
+{
439
+ config->capacity = bdrv_getlength(bs) >> BDRV_SECTOR_BITS;
440
+ config->blk_size = blk_size;
441
+ config->size_max = 0;
442
+ config->seg_max = 128 - 2;
443
+ config->min_io_size = 1;
444
+ config->opt_io_size = 1;
445
+ config->num_queues = VHOST_USER_BLK_MAX_QUEUES;
446
+ config->max_discard_sectors = 32768;
447
+ config->max_discard_seg = 1;
448
+ config->discard_sector_alignment = config->blk_size >> 9;
449
+ config->max_write_zeroes_sectors = 32768;
450
+ config->max_write_zeroes_seg = 1;
451
+}
452
+
453
+static VuBlockDev *vu_block_init(VuBlockDev *vu_block_device, Error **errp)
454
+{
455
+
456
+ BlockBackend *blk;
457
+ Error *local_error = NULL;
458
+ const char *node_name = vu_block_device->node_name;
459
+ bool writable = vu_block_device->writable;
460
+ uint64_t perm = BLK_PERM_CONSISTENT_READ;
461
+ int ret;
462
+
463
+ AioContext *ctx;
464
+
465
+ BlockDriverState *bs = bdrv_lookup_bs(node_name, node_name, &local_error);
466
+
467
+ if (!bs) {
468
+ error_propagate(errp, local_error);
469
+ return NULL;
470
+ }
471
+
472
+ if (bdrv_is_read_only(bs)) {
473
+ writable = false;
474
+ }
475
+
476
+ if (writable) {
477
+ perm |= BLK_PERM_WRITE;
478
+ }
479
+
480
+ ctx = bdrv_get_aio_context(bs);
481
+ aio_context_acquire(ctx);
482
+ bdrv_invalidate_cache(bs, NULL);
483
+ aio_context_release(ctx);
484
+
485
+ /*
486
+ * Don't allow resize while the vhost user server is running,
487
+ * otherwise we don't care what happens with the node.
488
+ */
489
+ blk = blk_new(bdrv_get_aio_context(bs), perm,
490
+ BLK_PERM_CONSISTENT_READ | BLK_PERM_WRITE_UNCHANGED |
491
+ BLK_PERM_WRITE | BLK_PERM_GRAPH_MOD);
492
+ ret = blk_insert_bs(blk, bs, errp);
493
+
494
+ if (ret < 0) {
495
+ goto fail;
496
+ }
497
+
498
+ blk_set_enable_write_cache(blk, false);
499
+
500
+ blk_set_allow_aio_context_change(blk, true);
501
+
502
+ vu_block_device->blkcfg.wce = 0;
503
+ vu_block_device->backend = blk;
504
+ if (!vu_block_device->blk_size) {
505
+ vu_block_device->blk_size = BDRV_SECTOR_SIZE;
506
+ }
507
+ vu_block_device->blkcfg.blk_size = vu_block_device->blk_size;
508
+ blk_set_guest_block_size(blk, vu_block_device->blk_size);
509
+ vu_block_initialize_config(bs, &vu_block_device->blkcfg,
510
+ vu_block_device->blk_size);
511
+ return vu_block_device;
512
+
513
+fail:
514
+ blk_unref(blk);
515
+ return NULL;
516
+}
517
+
518
+static void vu_block_deinit(VuBlockDev *vu_block_device)
519
+{
520
+ if (vu_block_device->backend) {
521
+ blk_remove_aio_context_notifier(vu_block_device->backend, blk_aio_attached,
522
+ blk_aio_detach, vu_block_device);
523
+ }
524
+
525
+ blk_unref(vu_block_device->backend);
526
+}
527
+
528
+static void vhost_user_blk_server_stop(VuBlockDev *vu_block_device)
529
+{
530
+ vhost_user_server_stop(&vu_block_device->vu_server);
531
+ vu_block_deinit(vu_block_device);
532
+}
533
+
534
+static void vhost_user_blk_server_start(VuBlockDev *vu_block_device,
535
+ Error **errp)
536
+{
537
+ AioContext *ctx;
538
+ SocketAddress *addr = vu_block_device->addr;
539
+
540
+ if (!vu_block_init(vu_block_device, errp)) {
541
+ return;
1491
+ return;
542
+ }
1492
+ }
543
+
1493
+
544
+ ctx = bdrv_get_aio_context(blk_bs(vu_block_device->backend));
1494
+ ufs_init_hc(u);
545
+
1495
+ ufs_init_pci(u, pci_dev);
546
+ if (!vhost_user_server_start(&vu_block_device->vu_server, addr, ctx,
547
+ VHOST_USER_BLK_MAX_QUEUES,
548
+ NULL, &vu_block_iface,
549
+ errp)) {
550
+ goto error;
551
+ }
552
+
553
+ blk_add_aio_context_notifier(vu_block_device->backend, blk_aio_attached,
554
+ blk_aio_detach, vu_block_device);
555
+ vu_block_device->running = true;
556
+ return;
557
+
558
+ error:
559
+ vu_block_deinit(vu_block_device);
560
+}
1496
+}
561
+
1497
+
562
+static bool vu_prop_modifiable(VuBlockDev *vus, Error **errp)
1498
+static Property ufs_props[] = {
1499
+ DEFINE_PROP_STRING("serial", UfsHc, params.serial),
1500
+ DEFINE_PROP_UINT8("nutrs", UfsHc, params.nutrs, 32),
1501
+ DEFINE_PROP_UINT8("nutmrs", UfsHc, params.nutmrs, 8),
1502
+ DEFINE_PROP_END_OF_LIST(),
1503
+};
1504
+
1505
+static const VMStateDescription ufs_vmstate = {
1506
+ .name = "ufs",
1507
+ .unmigratable = 1,
1508
+};
1509
+
1510
+static void ufs_class_init(ObjectClass *oc, void *data)
563
+{
1511
+{
564
+ if (vus->running) {
1512
+ DeviceClass *dc = DEVICE_CLASS(oc);
565
+ error_setg(errp, "The property can't be modified "
1513
+ PCIDeviceClass *pc = PCI_DEVICE_CLASS(oc);
566
+ "while the server is running");
1514
+
567
+ return false;
1515
+ pc->realize = ufs_realize;
568
+ }
1516
+ pc->vendor_id = PCI_VENDOR_ID_REDHAT;
569
+ return true;
1517
+ pc->device_id = PCI_DEVICE_ID_REDHAT_UFS;
1518
+ pc->class_id = PCI_CLASS_STORAGE_UFS;
1519
+
1520
+ set_bit(DEVICE_CATEGORY_STORAGE, dc->categories);
1521
+ dc->desc = "Universal Flash Storage";
1522
+ device_class_set_props(dc, ufs_props);
1523
+ dc->vmsd = &ufs_vmstate;
570
+}
1524
+}
571
+
1525
+
572
+static void vu_set_node_name(Object *obj, const char *value, Error **errp)
1526
+static const TypeInfo ufs_info = {
1527
+ .name = TYPE_UFS,
1528
+ .parent = TYPE_PCI_DEVICE,
1529
+ .class_init = ufs_class_init,
1530
+ .instance_size = sizeof(UfsHc),
1531
+ .interfaces = (InterfaceInfo[]){ { INTERFACE_PCIE_DEVICE }, {} },
1532
+};
1533
+
1534
+static void ufs_register_types(void)
573
+{
1535
+{
574
+ VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
1536
+ type_register_static(&ufs_info);
575
+
576
+ if (!vu_prop_modifiable(vus, errp)) {
577
+ return;
578
+ }
579
+
580
+ if (vus->node_name) {
581
+ g_free(vus->node_name);
582
+ }
583
+
584
+ vus->node_name = g_strdup(value);
585
+}
1537
+}
586
+
1538
+
587
+static char *vu_get_node_name(Object *obj, Error **errp)
1539
+type_init(ufs_register_types)
588
+{
1540
diff --git a/hw/Kconfig b/hw/Kconfig
589
+ VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
590
+ return g_strdup(vus->node_name);
591
+}
592
+
593
+static void free_socket_addr(SocketAddress *addr)
594
+{
595
+ g_free(addr->u.q_unix.path);
596
+ g_free(addr);
597
+}
598
+
599
+static void vu_set_unix_socket(Object *obj, const char *value,
600
+ Error **errp)
601
+{
602
+ VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
603
+
604
+ if (!vu_prop_modifiable(vus, errp)) {
605
+ return;
606
+ }
607
+
608
+ if (vus->addr) {
609
+ free_socket_addr(vus->addr);
610
+ }
611
+
612
+ SocketAddress *addr = g_new0(SocketAddress, 1);
613
+ addr->type = SOCKET_ADDRESS_TYPE_UNIX;
614
+ addr->u.q_unix.path = g_strdup(value);
615
+ vus->addr = addr;
616
+}
617
+
618
+static char *vu_get_unix_socket(Object *obj, Error **errp)
619
+{
620
+ VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
621
+ return g_strdup(vus->addr->u.q_unix.path);
622
+}
623
+
624
+static bool vu_get_block_writable(Object *obj, Error **errp)
625
+{
626
+ VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
627
+ return vus->writable;
628
+}
629
+
630
+static void vu_set_block_writable(Object *obj, bool value, Error **errp)
631
+{
632
+ VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
633
+
634
+ if (!vu_prop_modifiable(vus, errp)) {
635
+ return;
636
+ }
637
+
638
+ vus->writable = value;
639
+}
640
+
641
+static void vu_get_blk_size(Object *obj, Visitor *v, const char *name,
642
+ void *opaque, Error **errp)
643
+{
644
+ VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
645
+ uint32_t value = vus->blk_size;
646
+
647
+ visit_type_uint32(v, name, &value, errp);
648
+}
649
+
650
+static void vu_set_blk_size(Object *obj, Visitor *v, const char *name,
651
+ void *opaque, Error **errp)
652
+{
653
+ VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
654
+
655
+ Error *local_err = NULL;
656
+ uint32_t value;
657
+
658
+ if (!vu_prop_modifiable(vus, errp)) {
659
+ return;
660
+ }
661
+
662
+ visit_type_uint32(v, name, &value, &local_err);
663
+ if (local_err) {
664
+ goto out;
665
+ }
666
+
667
+ check_block_size(object_get_typename(obj), name, value, &local_err);
668
+ if (local_err) {
669
+ goto out;
670
+ }
671
+
672
+ vus->blk_size = value;
673
+
674
+out:
675
+ error_propagate(errp, local_err);
676
+}
677
+
678
+static void vhost_user_blk_server_instance_finalize(Object *obj)
679
+{
680
+ VuBlockDev *vub = VHOST_USER_BLK_SERVER(obj);
681
+
682
+ vhost_user_blk_server_stop(vub);
683
+
684
+ /*
685
+ * Unlike object_property_add_str, object_class_property_add_str
686
+ * doesn't have a release method. Thus manual memory freeing is
687
+ * needed.
688
+ */
689
+ free_socket_addr(vub->addr);
690
+ g_free(vub->node_name);
691
+}
692
+
693
+static void vhost_user_blk_server_complete(UserCreatable *obj, Error **errp)
694
+{
695
+ VuBlockDev *vub = VHOST_USER_BLK_SERVER(obj);
696
+
697
+ vhost_user_blk_server_start(vub, errp);
698
+}
699
+
700
+static void vhost_user_blk_server_class_init(ObjectClass *klass,
701
+ void *class_data)
702
+{
703
+ UserCreatableClass *ucc = USER_CREATABLE_CLASS(klass);
704
+ ucc->complete = vhost_user_blk_server_complete;
705
+
706
+ object_class_property_add_bool(klass, "writable",
707
+ vu_get_block_writable,
708
+ vu_set_block_writable);
709
+
710
+ object_class_property_add_str(klass, "node-name",
711
+ vu_get_node_name,
712
+ vu_set_node_name);
713
+
714
+ object_class_property_add_str(klass, "unix-socket",
715
+ vu_get_unix_socket,
716
+ vu_set_unix_socket);
717
+
718
+ object_class_property_add(klass, "logical-block-size", "uint32",
719
+ vu_get_blk_size, vu_set_blk_size,
720
+ NULL, NULL);
721
+}
722
+
723
+static const TypeInfo vhost_user_blk_server_info = {
724
+ .name = TYPE_VHOST_USER_BLK_SERVER,
725
+ .parent = TYPE_OBJECT,
726
+ .instance_size = sizeof(VuBlockDev),
727
+ .instance_finalize = vhost_user_blk_server_instance_finalize,
728
+ .class_init = vhost_user_blk_server_class_init,
729
+ .interfaces = (InterfaceInfo[]) {
730
+ {TYPE_USER_CREATABLE},
731
+ {}
732
+ },
733
+};
734
+
735
+static void vhost_user_blk_server_register_types(void)
736
+{
737
+ type_register_static(&vhost_user_blk_server_info);
738
+}
739
+
740
+type_init(vhost_user_blk_server_register_types)
741
diff --git a/softmmu/vl.c b/softmmu/vl.c
742
index XXXXXXX..XXXXXXX 100644
1541
index XXXXXXX..XXXXXXX 100644
743
--- a/softmmu/vl.c
1542
--- a/hw/Kconfig
744
+++ b/softmmu/vl.c
1543
+++ b/hw/Kconfig
745
@@ -XXX,XX +XXX,XX @@ static bool object_create_initial(const char *type, QemuOpts *opts)
1544
@@ -XXX,XX +XXX,XX @@ source smbios/Kconfig
746
}
1545
source ssi/Kconfig
747
#endif
1546
source timer/Kconfig
748
1547
source tpm/Kconfig
749
+ /* Reason: vhost-user-blk-server property "node-name" */
1548
+source ufs/Kconfig
750
+ if (g_str_equal(type, "vhost-user-blk-server")) {
1549
source usb/Kconfig
751
+ return false;
1550
source virtio/Kconfig
752
+ }
1551
source vfio/Kconfig
753
/*
1552
diff --git a/hw/meson.build b/hw/meson.build
754
* Reason: filter-* property "netdev" etc.
755
*/
756
diff --git a/block/meson.build b/block/meson.build
757
index XXXXXXX..XXXXXXX 100644
1553
index XXXXXXX..XXXXXXX 100644
758
--- a/block/meson.build
1554
--- a/hw/meson.build
759
+++ b/block/meson.build
1555
+++ b/hw/meson.build
760
@@ -XXX,XX +XXX,XX @@ block_ss.add(when: 'CONFIG_WIN32', if_true: files('file-win32.c', 'win32-aio.c')
1556
@@ -XXX,XX +XXX,XX @@ subdir('smbios')
761
block_ss.add(when: 'CONFIG_POSIX', if_true: [files('file-posix.c'), coref, iokit])
1557
subdir('ssi')
762
block_ss.add(when: 'CONFIG_LIBISCSI', if_true: files('iscsi-opts.c'))
1558
subdir('timer')
763
block_ss.add(when: 'CONFIG_LINUX', if_true: files('nvme.c'))
1559
subdir('tpm')
764
+block_ss.add(when: 'CONFIG_LINUX', if_true: files('export/vhost-user-blk-server.c', '../contrib/libvhost-user/libvhost-user.c'))
1560
+subdir('ufs')
765
block_ss.add(when: 'CONFIG_REPLICATION', if_true: files('replication.c'))
1561
subdir('usb')
766
block_ss.add(when: 'CONFIG_SHEEPDOG', if_true: files('sheepdog.c'))
1562
subdir('vfio')
767
block_ss.add(when: ['CONFIG_LINUX_AIO', libaio], if_true: files('linux-aio.c'))
1563
subdir('virtio')
1564
diff --git a/hw/ufs/Kconfig b/hw/ufs/Kconfig
1565
new file mode 100644
1566
index XXXXXXX..XXXXXXX
1567
--- /dev/null
1568
+++ b/hw/ufs/Kconfig
1569
@@ -XXX,XX +XXX,XX @@
1570
+config UFS_PCI
1571
+ bool
1572
+ default y if PCI_DEVICES
1573
+ depends on PCI
1574
diff --git a/hw/ufs/meson.build b/hw/ufs/meson.build
1575
new file mode 100644
1576
index XXXXXXX..XXXXXXX
1577
--- /dev/null
1578
+++ b/hw/ufs/meson.build
1579
@@ -0,0 +1 @@
1580
+system_ss.add(when: 'CONFIG_UFS_PCI', if_true: files('ufs.c'))
1581
diff --git a/hw/ufs/trace-events b/hw/ufs/trace-events
1582
new file mode 100644
1583
index XXXXXXX..XXXXXXX
1584
--- /dev/null
1585
+++ b/hw/ufs/trace-events
1586
@@ -XXX,XX +XXX,XX @@
1587
+# ufs.c
1588
+ufs_irq_raise(void) "INTx"
1589
+ufs_irq_lower(void) "INTx"
1590
+ufs_mmio_read(uint64_t addr, uint64_t data, unsigned size) "addr 0x%"PRIx64" data 0x%"PRIx64" size %d"
1591
+ufs_mmio_write(uint64_t addr, uint64_t data, unsigned size) "addr 0x%"PRIx64" data 0x%"PRIx64" size %d"
1592
+ufs_process_db(uint32_t slot) "UTRLDBR slot %"PRIu32""
1593
+ufs_process_req(uint32_t slot) "UTRLDBR slot %"PRIu32""
1594
+ufs_complete_req(uint32_t slot) "UTRLDBR slot %"PRIu32""
1595
+ufs_sendback_req(uint32_t slot) "UTRLDBR slot %"PRIu32""
1596
+ufs_exec_nop_cmd(uint32_t slot) "UTRLDBR slot %"PRIu32""
1597
+ufs_exec_scsi_cmd(uint32_t slot, uint8_t lun, uint8_t opcode) "slot %"PRIu32", lun 0x%"PRIx8", opcode 0x%"PRIx8""
1598
+ufs_exec_query_cmd(uint32_t slot, uint8_t opcode) "slot %"PRIu32", opcode 0x%"PRIx8""
1599
+ufs_process_uiccmd(uint32_t uiccmd, uint32_t ucmdarg1, uint32_t ucmdarg2, uint32_t ucmdarg3) "uiccmd 0x%"PRIx32", ucmdarg1 0x%"PRIx32", ucmdarg2 0x%"PRIx32", ucmdarg3 0x%"PRIx32""
1600
+
1601
+# error condition
1602
+ufs_err_dma_read_utrd(uint32_t slot, uint64_t addr) "failed to read utrd. UTRLDBR slot %"PRIu32", UTRD dma addr %"PRIu64""
1603
+ufs_err_dma_read_req_upiu(uint32_t slot, uint64_t addr) "failed to read req upiu. UTRLDBR slot %"PRIu32", request upiu addr %"PRIu64""
1604
+ufs_err_dma_read_prdt(uint32_t slot, uint64_t addr) "failed to read prdt. UTRLDBR slot %"PRIu32", prdt addr %"PRIu64""
1605
+ufs_err_dma_write_utrd(uint32_t slot, uint64_t addr) "failed to write utrd. UTRLDBR slot %"PRIu32", UTRD dma addr %"PRIu64""
1606
+ufs_err_dma_write_rsp_upiu(uint32_t slot, uint64_t addr) "failed to write rsp upiu. UTRLDBR slot %"PRIu32", response upiu addr %"PRIu64""
1607
+ufs_err_utrl_slot_busy(uint32_t slot) "UTRLDBR slot %"PRIu32" is busy"
1608
+ufs_err_unsupport_register_offset(uint32_t offset) "Register offset 0x%"PRIx32" is not yet supported"
1609
+ufs_err_invalid_register_offset(uint32_t offset) "Register offset 0x%"PRIx32" is invalid"
1610
+ufs_err_scsi_cmd_invalid_lun(uint8_t lun) "scsi command has invalid lun: 0x%"PRIx8""
1611
+ufs_err_query_flag_not_readable(uint8_t idn) "query flag idn 0x%"PRIx8" is denied to read"
1612
+ufs_err_query_flag_not_writable(uint8_t idn) "query flag idn 0x%"PRIx8" is denied to write"
1613
+ufs_err_query_attr_not_readable(uint8_t idn) "query attribute idn 0x%"PRIx8" is denied to read"
1614
+ufs_err_query_attr_not_writable(uint8_t idn) "query attribute idn 0x%"PRIx8" is denied to write"
1615
+ufs_err_query_invalid_opcode(uint8_t opcode) "query request has invalid opcode. opcode: 0x%"PRIx8""
1616
+ufs_err_query_invalid_idn(uint8_t opcode, uint8_t idn) "query request has invalid idn. opcode: 0x%"PRIx8", idn 0x%"PRIx8""
1617
+ufs_err_query_invalid_index(uint8_t opcode, uint8_t index) "query request has invalid index. opcode: 0x%"PRIx8", index 0x%"PRIx8""
1618
+ufs_err_invalid_trans_code(uint32_t slot, uint8_t trans_code) "request upiu has invalid transaction code. slot: %"PRIu32", trans_code: 0x%"PRIx8""
768
--
1619
--
769
2.26.2
1620
2.41.0
770
diff view generated by jsdifflib
1
The vu_client_trip() coroutine is leaked during AioContext switching. It
1
From: Jeuk Kim <jeuk20.kim@gmail.com>
2
is also unsafe to destroy the vu_dev in panic_cb() since its callers
3
still access it in some cases.
4
2
5
Rework the lifecycle to solve these safety issues.
3
This commit makes the UFS device support query
4
and nop out transfer requests.
6
5
7
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
6
The next patch would be support for UFS logical
8
Message-id: 20200924151549.913737-10-stefanha@redhat.com
7
unit and scsi command transfer request.
8
9
Signed-off-by: Jeuk Kim <jeuk20.kim@samsung.com>
10
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
11
Message-id: d06b440d660872092f70af1b8167bd5f4704c957.1691062912.git.jeuk20.kim@samsung.com
9
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
12
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
10
---
13
---
11
util/vhost-user-server.h | 29 ++--
14
hw/ufs/ufs.h | 46 +++
12
block/export/vhost-user-blk-server.c | 9 +-
15
hw/ufs/ufs.c | 980 +++++++++++++++++++++++++++++++++++++++++++-
13
util/vhost-user-server.c | 245 +++++++++++++++------------
16
hw/ufs/trace-events | 1 +
14
3 files changed, 155 insertions(+), 128 deletions(-)
17
3 files changed, 1025 insertions(+), 2 deletions(-)
15
18
16
diff --git a/util/vhost-user-server.h b/util/vhost-user-server.h
19
diff --git a/hw/ufs/ufs.h b/hw/ufs/ufs.h
17
index XXXXXXX..XXXXXXX 100644
20
index XXXXXXX..XXXXXXX 100644
18
--- a/util/vhost-user-server.h
21
--- a/hw/ufs/ufs.h
19
+++ b/util/vhost-user-server.h
22
+++ b/hw/ufs/ufs.h
20
@@ -XXX,XX +XXX,XX @@
23
@@ -XXX,XX +XXX,XX @@
21
#include "qapi/error.h"
24
#define UFS_MAX_LUS 32
22
#include "standard-headers/linux/virtio_blk.h"
25
#define UFS_BLOCK_SIZE 4096
23
26
24
+/* A kick fd that we monitor on behalf of libvhost-user */
27
+typedef enum UfsRequestState {
25
typedef struct VuFdWatch {
28
+ UFS_REQUEST_IDLE = 0,
26
VuDev *vu_dev;
29
+ UFS_REQUEST_READY = 1,
27
int fd; /*kick fd*/
30
+ UFS_REQUEST_RUNNING = 2,
28
void *pvt;
31
+ UFS_REQUEST_COMPLETE = 3,
29
vu_watch_cb cb;
32
+ UFS_REQUEST_ERROR = 4,
30
- bool processing;
33
+} UfsRequestState;
31
QTAILQ_ENTRY(VuFdWatch) next;
34
+
32
} VuFdWatch;
35
+typedef enum UfsReqResult {
33
36
+ UFS_REQUEST_SUCCESS = 0,
34
-typedef struct VuServer VuServer;
37
+ UFS_REQUEST_FAIL = 1,
35
-
38
+} UfsReqResult;
36
-struct VuServer {
39
+
37
+/**
40
+typedef struct UfsRequest {
38
+ * VuServer:
41
+ struct UfsHc *hc;
39
+ * A vhost-user server instance with user-defined VuDevIface callbacks.
42
+ UfsRequestState state;
40
+ * Vhost-user device backends can be implemented using VuServer. VuDevIface
43
+ int slot;
41
+ * callbacks and virtqueue kicks run in the given AioContext.
44
+
42
+ */
45
+ UtpTransferReqDesc utrd;
43
+typedef struct {
46
+ UtpUpiuReq req_upiu;
44
QIONetListener *listener;
47
+ UtpUpiuRsp rsp_upiu;
45
+ QEMUBH *restart_listener_bh;
48
+
46
AioContext *ctx;
49
+ /* for scsi command */
47
int max_queues;
50
+ QEMUSGList *sg;
48
const VuDevIface *vu_iface;
51
+} UfsRequest;
49
+
52
+
50
+ /* Protected by ctx lock */
53
typedef struct UfsParams {
51
VuDev vu_dev;
54
char *serial;
52
QIOChannel *ioc; /* The I/O channel with the client */
55
uint8_t nutrs; /* Number of UTP Transfer Request Slots */
53
QIOChannelSocket *sioc; /* The underlying data channel with the client */
56
@@ -XXX,XX +XXX,XX @@ typedef struct UfsHc {
54
- /* IOChannel for fd provided via VHOST_USER_SET_SLAVE_REQ_FD */
57
UfsReg reg;
55
- QIOChannel *ioc_slave;
58
UfsParams params;
56
- QIOChannelSocket *sioc_slave;
59
uint32_t reg_size;
57
- Coroutine *co_trip; /* coroutine for processing VhostUserMsg */
60
+ UfsRequest *req_list;
58
QTAILQ_HEAD(, VuFdWatch) vu_fd_watches;
61
+
59
- /* restart coroutine co_trip if AIOContext is changed */
62
+ DeviceDescriptor device_desc;
60
- bool aio_context_changed;
63
+ GeometryDescriptor geometry_desc;
61
- bool processing_msg;
64
+ Attributes attributes;
62
-};
65
+ Flags flags;
63
+
66
64
+ Coroutine *co_trip; /* coroutine for processing VhostUserMsg */
67
qemu_irq irq;
65
+} VuServer;
68
QEMUBH *doorbell_bh;
66
69
@@ -XXX,XX +XXX,XX @@ typedef struct UfsHc {
67
bool vhost_user_server_start(VuServer *server,
70
#define TYPE_UFS "ufs"
68
SocketAddress *unix_socket,
71
#define UFS(obj) OBJECT_CHECK(UfsHc, (obj), TYPE_UFS)
69
@@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server,
72
70
73
+typedef enum UfsQueryFlagPerm {
71
void vhost_user_server_stop(VuServer *server);
74
+ UFS_QUERY_FLAG_NONE = 0x0,
72
75
+ UFS_QUERY_FLAG_READ = 0x1,
73
-void vhost_user_server_set_aio_context(VuServer *server, AioContext *ctx);
76
+ UFS_QUERY_FLAG_SET = 0x2,
74
+void vhost_user_server_attach_aio_context(VuServer *server, AioContext *ctx);
77
+ UFS_QUERY_FLAG_CLEAR = 0x4,
75
+void vhost_user_server_detach_aio_context(VuServer *server);
78
+ UFS_QUERY_FLAG_TOGGLE = 0x8,
76
79
+} UfsQueryFlagPerm;
77
#endif /* VHOST_USER_SERVER_H */
80
+
78
diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c
81
+typedef enum UfsQueryAttrPerm {
82
+ UFS_QUERY_ATTR_NONE = 0x0,
83
+ UFS_QUERY_ATTR_READ = 0x1,
84
+ UFS_QUERY_ATTR_WRITE = 0x2,
85
+} UfsQueryAttrPerm;
86
+
87
#endif /* HW_UFS_UFS_H */
88
diff --git a/hw/ufs/ufs.c b/hw/ufs/ufs.c
79
index XXXXXXX..XXXXXXX 100644
89
index XXXXXXX..XXXXXXX 100644
80
--- a/block/export/vhost-user-blk-server.c
90
--- a/hw/ufs/ufs.c
81
+++ b/block/export/vhost-user-blk-server.c
91
+++ b/hw/ufs/ufs.c
82
@@ -XXX,XX +XXX,XX @@ static const VuDevIface vu_block_iface = {
92
@@ -XXX,XX +XXX,XX @@
83
static void blk_aio_attached(AioContext *ctx, void *opaque)
93
#include "ufs.h"
94
95
/* The QEMU-UFS device follows spec version 3.1 */
96
-#define UFS_SPEC_VER 0x00000310
97
+#define UFS_SPEC_VER 0x0310
98
#define UFS_MAX_NUTRS 32
99
#define UFS_MAX_NUTMRS 8
100
101
+static MemTxResult ufs_addr_read(UfsHc *u, hwaddr addr, void *buf, int size)
102
+{
103
+ hwaddr hi = addr + size - 1;
104
+
105
+ if (hi < addr) {
106
+ return MEMTX_DECODE_ERROR;
107
+ }
108
+
109
+ if (!FIELD_EX32(u->reg.cap, CAP, 64AS) && (hi >> 32)) {
110
+ return MEMTX_DECODE_ERROR;
111
+ }
112
+
113
+ return pci_dma_read(PCI_DEVICE(u), addr, buf, size);
114
+}
115
+
116
+static MemTxResult ufs_addr_write(UfsHc *u, hwaddr addr, const void *buf,
117
+ int size)
118
+{
119
+ hwaddr hi = addr + size - 1;
120
+ if (hi < addr) {
121
+ return MEMTX_DECODE_ERROR;
122
+ }
123
+
124
+ if (!FIELD_EX32(u->reg.cap, CAP, 64AS) && (hi >> 32)) {
125
+ return MEMTX_DECODE_ERROR;
126
+ }
127
+
128
+ return pci_dma_write(PCI_DEVICE(u), addr, buf, size);
129
+}
130
+
131
+static void ufs_complete_req(UfsRequest *req, UfsReqResult req_result);
132
+
133
+static inline hwaddr ufs_get_utrd_addr(UfsHc *u, uint32_t slot)
134
+{
135
+ hwaddr utrl_base_addr = (((hwaddr)u->reg.utrlbau) << 32) + u->reg.utrlba;
136
+ hwaddr utrd_addr = utrl_base_addr + slot * sizeof(UtpTransferReqDesc);
137
+
138
+ return utrd_addr;
139
+}
140
+
141
+static inline hwaddr ufs_get_req_upiu_base_addr(const UtpTransferReqDesc *utrd)
142
+{
143
+ uint32_t cmd_desc_base_addr_lo =
144
+ le32_to_cpu(utrd->command_desc_base_addr_lo);
145
+ uint32_t cmd_desc_base_addr_hi =
146
+ le32_to_cpu(utrd->command_desc_base_addr_hi);
147
+
148
+ return (((hwaddr)cmd_desc_base_addr_hi) << 32) + cmd_desc_base_addr_lo;
149
+}
150
+
151
+static inline hwaddr ufs_get_rsp_upiu_base_addr(const UtpTransferReqDesc *utrd)
152
+{
153
+ hwaddr req_upiu_base_addr = ufs_get_req_upiu_base_addr(utrd);
154
+ uint32_t rsp_upiu_byte_off =
155
+ le16_to_cpu(utrd->response_upiu_offset) * sizeof(uint32_t);
156
+ return req_upiu_base_addr + rsp_upiu_byte_off;
157
+}
158
+
159
+static MemTxResult ufs_dma_read_utrd(UfsRequest *req)
160
+{
161
+ UfsHc *u = req->hc;
162
+ hwaddr utrd_addr = ufs_get_utrd_addr(u, req->slot);
163
+ MemTxResult ret;
164
+
165
+ ret = ufs_addr_read(u, utrd_addr, &req->utrd, sizeof(req->utrd));
166
+ if (ret) {
167
+ trace_ufs_err_dma_read_utrd(req->slot, utrd_addr);
168
+ }
169
+ return ret;
170
+}
171
+
172
+static MemTxResult ufs_dma_read_req_upiu(UfsRequest *req)
173
+{
174
+ UfsHc *u = req->hc;
175
+ hwaddr req_upiu_base_addr = ufs_get_req_upiu_base_addr(&req->utrd);
176
+ UtpUpiuReq *req_upiu = &req->req_upiu;
177
+ uint32_t copy_size;
178
+ uint16_t data_segment_length;
179
+ MemTxResult ret;
180
+
181
+ /*
182
+ * To know the size of the req_upiu, we need to read the
183
+ * data_segment_length in the header first.
184
+ */
185
+ ret = ufs_addr_read(u, req_upiu_base_addr, &req_upiu->header,
186
+ sizeof(UtpUpiuHeader));
187
+ if (ret) {
188
+ trace_ufs_err_dma_read_req_upiu(req->slot, req_upiu_base_addr);
189
+ return ret;
190
+ }
191
+ data_segment_length = be16_to_cpu(req_upiu->header.data_segment_length);
192
+
193
+ copy_size = sizeof(UtpUpiuHeader) + UFS_TRANSACTION_SPECIFIC_FIELD_SIZE +
194
+ data_segment_length;
195
+
196
+ ret = ufs_addr_read(u, req_upiu_base_addr, &req->req_upiu, copy_size);
197
+ if (ret) {
198
+ trace_ufs_err_dma_read_req_upiu(req->slot, req_upiu_base_addr);
199
+ }
200
+ return ret;
201
+}
202
+
203
+static MemTxResult ufs_dma_read_prdt(UfsRequest *req)
204
+{
205
+ UfsHc *u = req->hc;
206
+ uint16_t prdt_len = le16_to_cpu(req->utrd.prd_table_length);
207
+ uint16_t prdt_byte_off =
208
+ le16_to_cpu(req->utrd.prd_table_offset) * sizeof(uint32_t);
209
+ uint32_t prdt_size = prdt_len * sizeof(UfshcdSgEntry);
210
+ g_autofree UfshcdSgEntry *prd_entries = NULL;
211
+ hwaddr req_upiu_base_addr, prdt_base_addr;
212
+ int err;
213
+
214
+ assert(!req->sg);
215
+
216
+ if (prdt_size == 0) {
217
+ return MEMTX_OK;
218
+ }
219
+ prd_entries = g_new(UfshcdSgEntry, prdt_size);
220
+
221
+ req_upiu_base_addr = ufs_get_req_upiu_base_addr(&req->utrd);
222
+ prdt_base_addr = req_upiu_base_addr + prdt_byte_off;
223
+
224
+ err = ufs_addr_read(u, prdt_base_addr, prd_entries, prdt_size);
225
+ if (err) {
226
+ trace_ufs_err_dma_read_prdt(req->slot, prdt_base_addr);
227
+ return err;
228
+ }
229
+
230
+ req->sg = g_malloc0(sizeof(QEMUSGList));
231
+ pci_dma_sglist_init(req->sg, PCI_DEVICE(u), prdt_len);
232
+
233
+ for (uint16_t i = 0; i < prdt_len; ++i) {
234
+ hwaddr data_dma_addr = le64_to_cpu(prd_entries[i].addr);
235
+ uint32_t data_byte_count = le32_to_cpu(prd_entries[i].size) + 1;
236
+ qemu_sglist_add(req->sg, data_dma_addr, data_byte_count);
237
+ }
238
+ return MEMTX_OK;
239
+}
240
+
241
+static MemTxResult ufs_dma_read_upiu(UfsRequest *req)
242
+{
243
+ MemTxResult ret;
244
+
245
+ ret = ufs_dma_read_utrd(req);
246
+ if (ret) {
247
+ return ret;
248
+ }
249
+
250
+ ret = ufs_dma_read_req_upiu(req);
251
+ if (ret) {
252
+ return ret;
253
+ }
254
+
255
+ ret = ufs_dma_read_prdt(req);
256
+ if (ret) {
257
+ return ret;
258
+ }
259
+
260
+ return 0;
261
+}
262
+
263
+static MemTxResult ufs_dma_write_utrd(UfsRequest *req)
264
+{
265
+ UfsHc *u = req->hc;
266
+ hwaddr utrd_addr = ufs_get_utrd_addr(u, req->slot);
267
+ MemTxResult ret;
268
+
269
+ ret = ufs_addr_write(u, utrd_addr, &req->utrd, sizeof(req->utrd));
270
+ if (ret) {
271
+ trace_ufs_err_dma_write_utrd(req->slot, utrd_addr);
272
+ }
273
+ return ret;
274
+}
275
+
276
+static MemTxResult ufs_dma_write_rsp_upiu(UfsRequest *req)
277
+{
278
+ UfsHc *u = req->hc;
279
+ hwaddr rsp_upiu_base_addr = ufs_get_rsp_upiu_base_addr(&req->utrd);
280
+ uint32_t rsp_upiu_byte_len =
281
+ le16_to_cpu(req->utrd.response_upiu_length) * sizeof(uint32_t);
282
+ uint16_t data_segment_length =
283
+ be16_to_cpu(req->rsp_upiu.header.data_segment_length);
284
+ uint32_t copy_size = sizeof(UtpUpiuHeader) +
285
+ UFS_TRANSACTION_SPECIFIC_FIELD_SIZE +
286
+ data_segment_length;
287
+ MemTxResult ret;
288
+
289
+ if (copy_size > rsp_upiu_byte_len) {
290
+ copy_size = rsp_upiu_byte_len;
291
+ }
292
+
293
+ ret = ufs_addr_write(u, rsp_upiu_base_addr, &req->rsp_upiu, copy_size);
294
+ if (ret) {
295
+ trace_ufs_err_dma_write_rsp_upiu(req->slot, rsp_upiu_base_addr);
296
+ }
297
+ return ret;
298
+}
299
+
300
+static MemTxResult ufs_dma_write_upiu(UfsRequest *req)
301
+{
302
+ MemTxResult ret;
303
+
304
+ ret = ufs_dma_write_rsp_upiu(req);
305
+ if (ret) {
306
+ return ret;
307
+ }
308
+
309
+ return ufs_dma_write_utrd(req);
310
+}
311
+
312
static void ufs_irq_check(UfsHc *u)
84
{
313
{
85
VuBlockDev *vub_dev = opaque;
314
PCIDevice *pci = PCI_DEVICE(u);
86
- aio_context_acquire(ctx);
315
@@ -XXX,XX +XXX,XX @@ static void ufs_irq_check(UfsHc *u)
87
- vhost_user_server_set_aio_context(&vub_dev->vu_server, ctx);
88
- aio_context_release(ctx);
89
+ vhost_user_server_attach_aio_context(&vub_dev->vu_server, ctx);
90
}
91
92
static void blk_aio_detach(void *opaque)
93
{
94
VuBlockDev *vub_dev = opaque;
95
- AioContext *ctx = vub_dev->vu_server.ctx;
96
- aio_context_acquire(ctx);
97
- vhost_user_server_set_aio_context(&vub_dev->vu_server, NULL);
98
- aio_context_release(ctx);
99
+ vhost_user_server_detach_aio_context(&vub_dev->vu_server);
100
}
101
102
static void
103
diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c
104
index XXXXXXX..XXXXXXX 100644
105
--- a/util/vhost-user-server.c
106
+++ b/util/vhost-user-server.c
107
@@ -XXX,XX +XXX,XX @@
108
*/
109
#include "qemu/osdep.h"
110
#include "qemu/main-loop.h"
111
+#include "block/aio-wait.h"
112
#include "vhost-user-server.h"
113
114
+/*
115
+ * Theory of operation:
116
+ *
117
+ * VuServer is started and stopped by vhost_user_server_start() and
118
+ * vhost_user_server_stop() from the main loop thread. Starting the server
119
+ * opens a vhost-user UNIX domain socket and listens for incoming connections.
120
+ * Only one connection is allowed at a time.
121
+ *
122
+ * The connection is handled by the vu_client_trip() coroutine in the
123
+ * VuServer->ctx AioContext. The coroutine consists of a vu_dispatch() loop
124
+ * where libvhost-user calls vu_message_read() to receive the next vhost-user
125
+ * protocol messages over the UNIX domain socket.
126
+ *
127
+ * When virtqueues are set up libvhost-user calls set_watch() to monitor kick
128
+ * fds. These fds are also handled in the VuServer->ctx AioContext.
129
+ *
130
+ * Both vu_client_trip() and kick fd monitoring can be stopped by shutting down
131
+ * the socket connection. Shutting down the socket connection causes
132
+ * vu_message_read() to fail since no more data can be received from the socket.
133
+ * After vu_dispatch() fails, vu_client_trip() calls vu_deinit() to stop
134
+ * libvhost-user before terminating the coroutine. vu_deinit() calls
135
+ * remove_watch() to stop monitoring kick fds and this stops virtqueue
136
+ * processing.
137
+ *
138
+ * When vu_client_trip() has finished cleaning up it schedules a BH in the main
139
+ * loop thread to accept the next client connection.
140
+ *
141
+ * When libvhost-user detects an error it calls panic_cb() and sets the
142
+ * dev->broken flag. Both vu_client_trip() and kick fd processing stop when
143
+ * the dev->broken flag is set.
144
+ *
145
+ * It is possible to switch AioContexts using
146
+ * vhost_user_server_detach_aio_context() and
147
+ * vhost_user_server_attach_aio_context(). They stop monitoring fds in the old
148
+ * AioContext and resume monitoring in the new AioContext. The vu_client_trip()
149
+ * coroutine remains in a yielded state during the switch. This is made
150
+ * possible by QIOChannel's support for spurious coroutine re-entry in
151
+ * qio_channel_yield(). The coroutine will restart I/O when re-entered from the
152
+ * new AioContext.
153
+ */
154
+
155
static void vmsg_close_fds(VhostUserMsg *vmsg)
156
{
157
int i;
158
@@ -XXX,XX +XXX,XX @@ static void vmsg_unblock_fds(VhostUserMsg *vmsg)
159
}
316
}
160
}
317
}
161
318
162
-static void vu_accept(QIONetListener *listener, QIOChannelSocket *sioc,
319
+static void ufs_process_db(UfsHc *u, uint32_t val)
163
- gpointer opaque);
320
+{
164
-
321
+ uint32_t slot;
165
-static void close_client(VuServer *server)
322
+ uint32_t nutrs = u->params.nutrs;
166
-{
323
+ UfsRequest *req;
167
- /*
324
+
168
- * Before closing the client
325
+ val &= ~u->reg.utrldbr;
169
- *
326
+ if (!val) {
170
- * 1. Let vu_client_trip stop processing new vhost-user msg
327
+ return;
171
- *
328
+ }
172
- * 2. remove kick_handler
329
+
173
- *
330
+ slot = find_first_bit((unsigned long *)&val, nutrs);
174
- * 3. wait for the kick handler to be finished
331
+
175
- *
332
+ while (slot < nutrs) {
176
- * 4. wait for the current vhost-user msg to be finished processing
333
+ req = &u->req_list[slot];
177
- */
334
+ if (req->state == UFS_REQUEST_ERROR) {
178
-
335
+ trace_ufs_err_utrl_slot_error(req->slot);
179
- QIOChannelSocket *sioc = server->sioc;
336
+ return;
180
- /* When this is set vu_client_trip will stop new processing vhost-user message */
337
+ }
181
- server->sioc = NULL;
338
+
182
-
339
+ if (req->state != UFS_REQUEST_IDLE) {
183
- while (server->processing_msg) {
340
+ trace_ufs_err_utrl_slot_busy(req->slot);
184
- if (server->ioc->read_coroutine) {
341
+ return;
185
- server->ioc->read_coroutine = NULL;
342
+ }
186
- qio_channel_set_aio_fd_handler(server->ioc, server->ioc->ctx, NULL,
343
+
187
- NULL, server->ioc);
344
+ trace_ufs_process_db(slot);
188
- server->processing_msg = false;
345
+ req->state = UFS_REQUEST_READY;
189
- }
346
+ slot = find_next_bit((unsigned long *)&val, nutrs, slot + 1);
190
- }
347
+ }
191
-
348
+
192
- vu_deinit(&server->vu_dev);
349
+ qemu_bh_schedule(u->doorbell_bh);
193
-
350
+}
194
- /* vu_deinit() should have called remove_watch() */
351
+
195
- assert(QTAILQ_EMPTY(&server->vu_fd_watches));
352
static void ufs_process_uiccmd(UfsHc *u, uint32_t val)
196
-
197
- object_unref(OBJECT(sioc));
198
- object_unref(OBJECT(server->ioc));
199
-}
200
-
201
static void panic_cb(VuDev *vu_dev, const char *buf)
202
{
353
{
203
- VuServer *server = container_of(vu_dev, VuServer, vu_dev);
354
trace_ufs_process_uiccmd(val, u->reg.ucmdarg1, u->reg.ucmdarg2,
204
-
355
@@ -XXX,XX +XXX,XX @@ static void ufs_write_reg(UfsHc *u, hwaddr offset, uint32_t data, unsigned size)
205
- /* avoid while loop in close_client */
356
u->reg.utrlbau = data;
206
- server->processing_msg = false;
357
break;
207
-
358
case A_UTRLDBR:
208
- if (buf) {
359
- /* Not yet supported */
209
- error_report("vu_panic: %s", buf);
360
+ ufs_process_db(u, data);
210
- }
361
+ u->reg.utrldbr |= data;
211
-
362
break;
212
- if (server->sioc) {
363
case A_UTRLRSR:
213
- close_client(server);
364
u->reg.utrlrsr = data;
214
- }
365
@@ -XXX,XX +XXX,XX @@ static const MemoryRegionOps ufs_mmio_ops = {
215
-
366
},
216
- /*
367
};
217
- * Set the callback function for network listener so another
368
218
- * vhost-user client can connect to this server
369
+static void ufs_build_upiu_header(UfsRequest *req, uint8_t trans_type,
219
- */
370
+ uint8_t flags, uint8_t response,
220
- qio_net_listener_set_client_func(server->listener,
371
+ uint8_t scsi_status,
221
- vu_accept,
372
+ uint16_t data_segment_length)
222
- server,
373
+{
223
- NULL);
374
+ memcpy(&req->rsp_upiu.header, &req->req_upiu.header, sizeof(UtpUpiuHeader));
224
+ error_report("vu_panic: %s", buf);
375
+ req->rsp_upiu.header.trans_type = trans_type;
376
+ req->rsp_upiu.header.flags = flags;
377
+ req->rsp_upiu.header.response = response;
378
+ req->rsp_upiu.header.scsi_status = scsi_status;
379
+ req->rsp_upiu.header.data_segment_length = cpu_to_be16(data_segment_length);
380
+}
381
+
382
+static UfsReqResult ufs_exec_nop_cmd(UfsRequest *req)
383
+{
384
+ trace_ufs_exec_nop_cmd(req->slot);
385
+ ufs_build_upiu_header(req, UPIU_TRANSACTION_NOP_IN, 0, 0, 0, 0);
386
+ return UFS_REQUEST_SUCCESS;
387
+}
388
+
389
+/*
390
+ * This defines the permission of flags based on their IDN. There are some
391
+ * things that are declared read-only, which is inconsistent with the ufs spec,
392
+ * because we want to return an error for features that are not yet supported.
393
+ */
394
+static const int flag_permission[QUERY_FLAG_IDN_COUNT] = {
395
+ [QUERY_FLAG_IDN_FDEVICEINIT] = UFS_QUERY_FLAG_READ | UFS_QUERY_FLAG_SET,
396
+ /* Write protection is not supported */
397
+ [QUERY_FLAG_IDN_PERMANENT_WPE] = UFS_QUERY_FLAG_READ,
398
+ [QUERY_FLAG_IDN_PWR_ON_WPE] = UFS_QUERY_FLAG_READ,
399
+ [QUERY_FLAG_IDN_BKOPS_EN] = UFS_QUERY_FLAG_READ | UFS_QUERY_FLAG_SET |
400
+ UFS_QUERY_FLAG_CLEAR | UFS_QUERY_FLAG_TOGGLE,
401
+ [QUERY_FLAG_IDN_LIFE_SPAN_MODE_ENABLE] =
402
+ UFS_QUERY_FLAG_READ | UFS_QUERY_FLAG_SET | UFS_QUERY_FLAG_CLEAR |
403
+ UFS_QUERY_FLAG_TOGGLE,
404
+ /* Purge Operation is not supported */
405
+ [QUERY_FLAG_IDN_PURGE_ENABLE] = UFS_QUERY_FLAG_NONE,
406
+ /* Refresh Operation is not supported */
407
+ [QUERY_FLAG_IDN_REFRESH_ENABLE] = UFS_QUERY_FLAG_NONE,
408
+ /* Physical Resource Removal is not supported */
409
+ [QUERY_FLAG_IDN_FPHYRESOURCEREMOVAL] = UFS_QUERY_FLAG_READ,
410
+ [QUERY_FLAG_IDN_BUSY_RTC] = UFS_QUERY_FLAG_READ,
411
+ [QUERY_FLAG_IDN_PERMANENTLY_DISABLE_FW_UPDATE] = UFS_QUERY_FLAG_READ,
412
+ /* Write Booster is not supported */
413
+ [QUERY_FLAG_IDN_WB_EN] = UFS_QUERY_FLAG_READ,
414
+ [QUERY_FLAG_IDN_WB_BUFF_FLUSH_EN] = UFS_QUERY_FLAG_READ,
415
+ [QUERY_FLAG_IDN_WB_BUFF_FLUSH_DURING_HIBERN8] = UFS_QUERY_FLAG_READ,
416
+};
417
+
418
+static inline QueryRespCode ufs_flag_check_idn_valid(uint8_t idn, int op)
419
+{
420
+ if (idn >= QUERY_FLAG_IDN_COUNT) {
421
+ return QUERY_RESULT_INVALID_IDN;
422
+ }
423
+
424
+ if (!(flag_permission[idn] & op)) {
425
+ if (op == UFS_QUERY_FLAG_READ) {
426
+ trace_ufs_err_query_flag_not_readable(idn);
427
+ return QUERY_RESULT_NOT_READABLE;
428
+ }
429
+ trace_ufs_err_query_flag_not_writable(idn);
430
+ return QUERY_RESULT_NOT_WRITEABLE;
431
+ }
432
+
433
+ return QUERY_RESULT_SUCCESS;
434
+}
435
+
436
+static const int attr_permission[QUERY_ATTR_IDN_COUNT] = {
437
+ /* booting is not supported */
438
+ [QUERY_ATTR_IDN_BOOT_LU_EN] = UFS_QUERY_ATTR_READ,
439
+ [QUERY_ATTR_IDN_POWER_MODE] = UFS_QUERY_ATTR_READ,
440
+ [QUERY_ATTR_IDN_ACTIVE_ICC_LVL] =
441
+ UFS_QUERY_ATTR_READ | UFS_QUERY_ATTR_WRITE,
442
+ [QUERY_ATTR_IDN_OOO_DATA_EN] = UFS_QUERY_ATTR_READ,
443
+ [QUERY_ATTR_IDN_BKOPS_STATUS] = UFS_QUERY_ATTR_READ,
444
+ [QUERY_ATTR_IDN_PURGE_STATUS] = UFS_QUERY_ATTR_READ,
445
+ [QUERY_ATTR_IDN_MAX_DATA_IN] = UFS_QUERY_ATTR_READ | UFS_QUERY_ATTR_WRITE,
446
+ [QUERY_ATTR_IDN_MAX_DATA_OUT] = UFS_QUERY_ATTR_READ | UFS_QUERY_ATTR_WRITE,
447
+ [QUERY_ATTR_IDN_DYN_CAP_NEEDED] = UFS_QUERY_ATTR_READ,
448
+ [QUERY_ATTR_IDN_REF_CLK_FREQ] = UFS_QUERY_ATTR_READ | UFS_QUERY_ATTR_WRITE,
449
+ [QUERY_ATTR_IDN_CONF_DESC_LOCK] = UFS_QUERY_ATTR_READ,
450
+ [QUERY_ATTR_IDN_MAX_NUM_OF_RTT] =
451
+ UFS_QUERY_ATTR_READ | UFS_QUERY_ATTR_WRITE,
452
+ [QUERY_ATTR_IDN_EE_CONTROL] = UFS_QUERY_ATTR_READ | UFS_QUERY_ATTR_WRITE,
453
+ [QUERY_ATTR_IDN_EE_STATUS] = UFS_QUERY_ATTR_READ,
454
+ [QUERY_ATTR_IDN_SECONDS_PASSED] = UFS_QUERY_ATTR_WRITE,
455
+ [QUERY_ATTR_IDN_CNTX_CONF] = UFS_QUERY_ATTR_READ,
456
+ [QUERY_ATTR_IDN_FFU_STATUS] = UFS_QUERY_ATTR_READ,
457
+ [QUERY_ATTR_IDN_PSA_STATE] = UFS_QUERY_ATTR_READ | UFS_QUERY_ATTR_WRITE,
458
+ [QUERY_ATTR_IDN_PSA_DATA_SIZE] = UFS_QUERY_ATTR_READ | UFS_QUERY_ATTR_WRITE,
459
+ [QUERY_ATTR_IDN_REF_CLK_GATING_WAIT_TIME] = UFS_QUERY_ATTR_READ,
460
+ [QUERY_ATTR_IDN_CASE_ROUGH_TEMP] = UFS_QUERY_ATTR_READ,
461
+ [QUERY_ATTR_IDN_HIGH_TEMP_BOUND] = UFS_QUERY_ATTR_READ,
462
+ [QUERY_ATTR_IDN_LOW_TEMP_BOUND] = UFS_QUERY_ATTR_READ,
463
+ [QUERY_ATTR_IDN_THROTTLING_STATUS] = UFS_QUERY_ATTR_READ,
464
+ [QUERY_ATTR_IDN_WB_FLUSH_STATUS] = UFS_QUERY_ATTR_READ,
465
+ [QUERY_ATTR_IDN_AVAIL_WB_BUFF_SIZE] = UFS_QUERY_ATTR_READ,
466
+ [QUERY_ATTR_IDN_WB_BUFF_LIFE_TIME_EST] = UFS_QUERY_ATTR_READ,
467
+ [QUERY_ATTR_IDN_CURR_WB_BUFF_SIZE] = UFS_QUERY_ATTR_READ,
468
+ /* refresh operation is not supported */
469
+ [QUERY_ATTR_IDN_REFRESH_STATUS] = UFS_QUERY_ATTR_READ,
470
+ [QUERY_ATTR_IDN_REFRESH_FREQ] = UFS_QUERY_ATTR_READ,
471
+ [QUERY_ATTR_IDN_REFRESH_UNIT] = UFS_QUERY_ATTR_READ,
472
+};
473
+
474
+static inline QueryRespCode ufs_attr_check_idn_valid(uint8_t idn, int op)
475
+{
476
+ if (idn >= QUERY_ATTR_IDN_COUNT) {
477
+ return QUERY_RESULT_INVALID_IDN;
478
+ }
479
+
480
+ if (!(attr_permission[idn] & op)) {
481
+ if (op == UFS_QUERY_ATTR_READ) {
482
+ trace_ufs_err_query_attr_not_readable(idn);
483
+ return QUERY_RESULT_NOT_READABLE;
484
+ }
485
+ trace_ufs_err_query_attr_not_writable(idn);
486
+ return QUERY_RESULT_NOT_WRITEABLE;
487
+ }
488
+
489
+ return QUERY_RESULT_SUCCESS;
490
+}
491
+
492
+static QueryRespCode ufs_exec_query_flag(UfsRequest *req, int op)
493
+{
494
+ UfsHc *u = req->hc;
495
+ uint8_t idn = req->req_upiu.qr.idn;
496
+ uint32_t value;
497
+ QueryRespCode ret;
498
+
499
+ ret = ufs_flag_check_idn_valid(idn, op);
500
+ if (ret) {
501
+ return ret;
502
+ }
503
+
504
+ if (idn == QUERY_FLAG_IDN_FDEVICEINIT) {
505
+ value = 0;
506
+ } else if (op == UFS_QUERY_FLAG_READ) {
507
+ value = *(((uint8_t *)&u->flags) + idn);
508
+ } else if (op == UFS_QUERY_FLAG_SET) {
509
+ value = 1;
510
+ } else if (op == UFS_QUERY_FLAG_CLEAR) {
511
+ value = 0;
512
+ } else if (op == UFS_QUERY_FLAG_TOGGLE) {
513
+ value = *(((uint8_t *)&u->flags) + idn);
514
+ value = !value;
515
+ } else {
516
+ trace_ufs_err_query_invalid_opcode(op);
517
+ return QUERY_RESULT_INVALID_OPCODE;
518
+ }
519
+
520
+ *(((uint8_t *)&u->flags) + idn) = value;
521
+ req->rsp_upiu.qr.value = cpu_to_be32(value);
522
+ return QUERY_RESULT_SUCCESS;
523
+}
524
+
525
+static uint32_t ufs_read_attr_value(UfsHc *u, uint8_t idn)
526
+{
527
+ switch (idn) {
528
+ case QUERY_ATTR_IDN_BOOT_LU_EN:
529
+ return u->attributes.boot_lun_en;
530
+ case QUERY_ATTR_IDN_POWER_MODE:
531
+ return u->attributes.current_power_mode;
532
+ case QUERY_ATTR_IDN_ACTIVE_ICC_LVL:
533
+ return u->attributes.active_icc_level;
534
+ case QUERY_ATTR_IDN_OOO_DATA_EN:
535
+ return u->attributes.out_of_order_data_en;
536
+ case QUERY_ATTR_IDN_BKOPS_STATUS:
537
+ return u->attributes.background_op_status;
538
+ case QUERY_ATTR_IDN_PURGE_STATUS:
539
+ return u->attributes.purge_status;
540
+ case QUERY_ATTR_IDN_MAX_DATA_IN:
541
+ return u->attributes.max_data_in_size;
542
+ case QUERY_ATTR_IDN_MAX_DATA_OUT:
543
+ return u->attributes.max_data_out_size;
544
+ case QUERY_ATTR_IDN_DYN_CAP_NEEDED:
545
+ return be32_to_cpu(u->attributes.dyn_cap_needed);
546
+ case QUERY_ATTR_IDN_REF_CLK_FREQ:
547
+ return u->attributes.ref_clk_freq;
548
+ case QUERY_ATTR_IDN_CONF_DESC_LOCK:
549
+ return u->attributes.config_descr_lock;
550
+ case QUERY_ATTR_IDN_MAX_NUM_OF_RTT:
551
+ return u->attributes.max_num_of_rtt;
552
+ case QUERY_ATTR_IDN_EE_CONTROL:
553
+ return be16_to_cpu(u->attributes.exception_event_control);
554
+ case QUERY_ATTR_IDN_EE_STATUS:
555
+ return be16_to_cpu(u->attributes.exception_event_status);
556
+ case QUERY_ATTR_IDN_SECONDS_PASSED:
557
+ return be32_to_cpu(u->attributes.seconds_passed);
558
+ case QUERY_ATTR_IDN_CNTX_CONF:
559
+ return be16_to_cpu(u->attributes.context_conf);
560
+ case QUERY_ATTR_IDN_FFU_STATUS:
561
+ return u->attributes.device_ffu_status;
562
+ case QUERY_ATTR_IDN_PSA_STATE:
563
+ return be32_to_cpu(u->attributes.psa_state);
564
+ case QUERY_ATTR_IDN_PSA_DATA_SIZE:
565
+ return be32_to_cpu(u->attributes.psa_data_size);
566
+ case QUERY_ATTR_IDN_REF_CLK_GATING_WAIT_TIME:
567
+ return u->attributes.ref_clk_gating_wait_time;
568
+ case QUERY_ATTR_IDN_CASE_ROUGH_TEMP:
569
+ return u->attributes.device_case_rough_temperaure;
570
+ case QUERY_ATTR_IDN_HIGH_TEMP_BOUND:
571
+ return u->attributes.device_too_high_temp_boundary;
572
+ case QUERY_ATTR_IDN_LOW_TEMP_BOUND:
573
+ return u->attributes.device_too_low_temp_boundary;
574
+ case QUERY_ATTR_IDN_THROTTLING_STATUS:
575
+ return u->attributes.throttling_status;
576
+ case QUERY_ATTR_IDN_WB_FLUSH_STATUS:
577
+ return u->attributes.wb_buffer_flush_status;
578
+ case QUERY_ATTR_IDN_AVAIL_WB_BUFF_SIZE:
579
+ return u->attributes.available_wb_buffer_size;
580
+ case QUERY_ATTR_IDN_WB_BUFF_LIFE_TIME_EST:
581
+ return u->attributes.wb_buffer_life_time_est;
582
+ case QUERY_ATTR_IDN_CURR_WB_BUFF_SIZE:
583
+ return be32_to_cpu(u->attributes.current_wb_buffer_size);
584
+ case QUERY_ATTR_IDN_REFRESH_STATUS:
585
+ return u->attributes.refresh_status;
586
+ case QUERY_ATTR_IDN_REFRESH_FREQ:
587
+ return u->attributes.refresh_freq;
588
+ case QUERY_ATTR_IDN_REFRESH_UNIT:
589
+ return u->attributes.refresh_unit;
590
+ }
591
+ return 0;
592
+}
593
+
594
+static void ufs_write_attr_value(UfsHc *u, uint8_t idn, uint32_t value)
595
+{
596
+ switch (idn) {
597
+ case QUERY_ATTR_IDN_ACTIVE_ICC_LVL:
598
+ u->attributes.active_icc_level = value;
599
+ break;
600
+ case QUERY_ATTR_IDN_MAX_DATA_IN:
601
+ u->attributes.max_data_in_size = value;
602
+ break;
603
+ case QUERY_ATTR_IDN_MAX_DATA_OUT:
604
+ u->attributes.max_data_out_size = value;
605
+ break;
606
+ case QUERY_ATTR_IDN_REF_CLK_FREQ:
607
+ u->attributes.ref_clk_freq = value;
608
+ break;
609
+ case QUERY_ATTR_IDN_MAX_NUM_OF_RTT:
610
+ u->attributes.max_num_of_rtt = value;
611
+ break;
612
+ case QUERY_ATTR_IDN_EE_CONTROL:
613
+ u->attributes.exception_event_control = cpu_to_be16(value);
614
+ break;
615
+ case QUERY_ATTR_IDN_SECONDS_PASSED:
616
+ u->attributes.seconds_passed = cpu_to_be32(value);
617
+ break;
618
+ case QUERY_ATTR_IDN_PSA_STATE:
619
+ u->attributes.psa_state = value;
620
+ break;
621
+ case QUERY_ATTR_IDN_PSA_DATA_SIZE:
622
+ u->attributes.psa_data_size = cpu_to_be32(value);
623
+ break;
624
+ }
625
+}
626
+
627
+static QueryRespCode ufs_exec_query_attr(UfsRequest *req, int op)
628
+{
629
+ UfsHc *u = req->hc;
630
+ uint8_t idn = req->req_upiu.qr.idn;
631
+ uint32_t value;
632
+ QueryRespCode ret;
633
+
634
+ ret = ufs_attr_check_idn_valid(idn, op);
635
+ if (ret) {
636
+ return ret;
637
+ }
638
+
639
+ if (op == UFS_QUERY_ATTR_READ) {
640
+ value = ufs_read_attr_value(u, idn);
641
+ } else {
642
+ value = be32_to_cpu(req->req_upiu.qr.value);
643
+ ufs_write_attr_value(u, idn, value);
644
+ }
645
+
646
+ req->rsp_upiu.qr.value = cpu_to_be32(value);
647
+ return QUERY_RESULT_SUCCESS;
648
+}
649
+
650
+static const RpmbUnitDescriptor rpmb_unit_desc = {
651
+ .length = sizeof(RpmbUnitDescriptor),
652
+ .descriptor_idn = 2,
653
+ .unit_index = UFS_UPIU_RPMB_WLUN,
654
+ .lu_enable = 0,
655
+};
656
+
657
+static QueryRespCode ufs_read_unit_desc(UfsRequest *req)
658
+{
659
+ uint8_t lun = req->req_upiu.qr.index;
660
+
661
+ if (lun != UFS_UPIU_RPMB_WLUN && lun > UFS_MAX_LUS) {
662
+ trace_ufs_err_query_invalid_index(req->req_upiu.qr.opcode, lun);
663
+ return QUERY_RESULT_INVALID_INDEX;
664
+ }
665
+
666
+ if (lun == UFS_UPIU_RPMB_WLUN) {
667
+ memcpy(&req->rsp_upiu.qr.data, &rpmb_unit_desc, rpmb_unit_desc.length);
668
+ } else {
669
+ /* unit descriptor is not yet supported */
670
+ return QUERY_RESULT_INVALID_INDEX;
671
+ }
672
+
673
+ return QUERY_RESULT_SUCCESS;
674
+}
675
+
676
+static inline StringDescriptor manufacturer_str_desc(void)
677
+{
678
+ StringDescriptor desc = {
679
+ .length = 0x12,
680
+ .descriptor_idn = QUERY_DESC_IDN_STRING,
681
+ };
682
+ desc.UC[0] = cpu_to_be16('R');
683
+ desc.UC[1] = cpu_to_be16('E');
684
+ desc.UC[2] = cpu_to_be16('D');
685
+ desc.UC[3] = cpu_to_be16('H');
686
+ desc.UC[4] = cpu_to_be16('A');
687
+ desc.UC[5] = cpu_to_be16('T');
688
+ return desc;
689
+}
690
+
691
+static inline StringDescriptor product_name_str_desc(void)
692
+{
693
+ StringDescriptor desc = {
694
+ .length = 0x22,
695
+ .descriptor_idn = QUERY_DESC_IDN_STRING,
696
+ };
697
+ desc.UC[0] = cpu_to_be16('Q');
698
+ desc.UC[1] = cpu_to_be16('E');
699
+ desc.UC[2] = cpu_to_be16('M');
700
+ desc.UC[3] = cpu_to_be16('U');
701
+ desc.UC[4] = cpu_to_be16(' ');
702
+ desc.UC[5] = cpu_to_be16('U');
703
+ desc.UC[6] = cpu_to_be16('F');
704
+ desc.UC[7] = cpu_to_be16('S');
705
+ return desc;
706
+}
707
+
708
+static inline StringDescriptor product_rev_level_str_desc(void)
709
+{
710
+ StringDescriptor desc = {
711
+ .length = 0x0a,
712
+ .descriptor_idn = QUERY_DESC_IDN_STRING,
713
+ };
714
+ desc.UC[0] = cpu_to_be16('0');
715
+ desc.UC[1] = cpu_to_be16('0');
716
+ desc.UC[2] = cpu_to_be16('0');
717
+ desc.UC[3] = cpu_to_be16('1');
718
+ return desc;
719
+}
720
+
721
+static const StringDescriptor null_str_desc = {
722
+ .length = 0x02,
723
+ .descriptor_idn = QUERY_DESC_IDN_STRING,
724
+};
725
+
726
+static QueryRespCode ufs_read_string_desc(UfsRequest *req)
727
+{
728
+ UfsHc *u = req->hc;
729
+ uint8_t index = req->req_upiu.qr.index;
730
+ StringDescriptor desc;
731
+
732
+ if (index == u->device_desc.manufacturer_name) {
733
+ desc = manufacturer_str_desc();
734
+ memcpy(&req->rsp_upiu.qr.data, &desc, desc.length);
735
+ } else if (index == u->device_desc.product_name) {
736
+ desc = product_name_str_desc();
737
+ memcpy(&req->rsp_upiu.qr.data, &desc, desc.length);
738
+ } else if (index == u->device_desc.serial_number) {
739
+ memcpy(&req->rsp_upiu.qr.data, &null_str_desc, null_str_desc.length);
740
+ } else if (index == u->device_desc.oem_id) {
741
+ memcpy(&req->rsp_upiu.qr.data, &null_str_desc, null_str_desc.length);
742
+ } else if (index == u->device_desc.product_revision_level) {
743
+ desc = product_rev_level_str_desc();
744
+ memcpy(&req->rsp_upiu.qr.data, &desc, desc.length);
745
+ } else {
746
+ trace_ufs_err_query_invalid_index(req->req_upiu.qr.opcode, index);
747
+ return QUERY_RESULT_INVALID_INDEX;
748
+ }
749
+ return QUERY_RESULT_SUCCESS;
750
+}
751
+
752
+static inline InterconnectDescriptor interconnect_desc(void)
753
+{
754
+ InterconnectDescriptor desc = {
755
+ .length = sizeof(InterconnectDescriptor),
756
+ .descriptor_idn = QUERY_DESC_IDN_INTERCONNECT,
757
+ };
758
+ desc.bcd_unipro_version = cpu_to_be16(0x180);
759
+ desc.bcd_mphy_version = cpu_to_be16(0x410);
760
+ return desc;
761
+}
762
+
763
+static QueryRespCode ufs_read_desc(UfsRequest *req)
764
+{
765
+ UfsHc *u = req->hc;
766
+ QueryRespCode status;
767
+ uint8_t idn = req->req_upiu.qr.idn;
768
+ uint16_t length = be16_to_cpu(req->req_upiu.qr.length);
769
+ InterconnectDescriptor desc;
770
+
771
+ switch (idn) {
772
+ case QUERY_DESC_IDN_DEVICE:
773
+ memcpy(&req->rsp_upiu.qr.data, &u->device_desc, sizeof(u->device_desc));
774
+ status = QUERY_RESULT_SUCCESS;
775
+ break;
776
+ case QUERY_DESC_IDN_UNIT:
777
+ status = ufs_read_unit_desc(req);
778
+ break;
779
+ case QUERY_DESC_IDN_GEOMETRY:
780
+ memcpy(&req->rsp_upiu.qr.data, &u->geometry_desc,
781
+ sizeof(u->geometry_desc));
782
+ status = QUERY_RESULT_SUCCESS;
783
+ break;
784
+ case QUERY_DESC_IDN_INTERCONNECT: {
785
+ desc = interconnect_desc();
786
+ memcpy(&req->rsp_upiu.qr.data, &desc, sizeof(InterconnectDescriptor));
787
+ status = QUERY_RESULT_SUCCESS;
788
+ break;
789
+ }
790
+ case QUERY_DESC_IDN_STRING:
791
+ status = ufs_read_string_desc(req);
792
+ break;
793
+ case QUERY_DESC_IDN_POWER:
794
+ /* mocking of power descriptor is not supported */
795
+ memset(&req->rsp_upiu.qr.data, 0, sizeof(PowerParametersDescriptor));
796
+ req->rsp_upiu.qr.data[0] = sizeof(PowerParametersDescriptor);
797
+ req->rsp_upiu.qr.data[1] = QUERY_DESC_IDN_POWER;
798
+ status = QUERY_RESULT_SUCCESS;
799
+ break;
800
+ case QUERY_DESC_IDN_HEALTH:
801
+ /* mocking of health descriptor is not supported */
802
+ memset(&req->rsp_upiu.qr.data, 0, sizeof(DeviceHealthDescriptor));
803
+ req->rsp_upiu.qr.data[0] = sizeof(DeviceHealthDescriptor);
804
+ req->rsp_upiu.qr.data[1] = QUERY_DESC_IDN_HEALTH;
805
+ status = QUERY_RESULT_SUCCESS;
806
+ break;
807
+ default:
808
+ length = 0;
809
+ trace_ufs_err_query_invalid_idn(req->req_upiu.qr.opcode, idn);
810
+ status = QUERY_RESULT_INVALID_IDN;
811
+ }
812
+
813
+ if (length > req->rsp_upiu.qr.data[0]) {
814
+ length = req->rsp_upiu.qr.data[0];
815
+ }
816
+ req->rsp_upiu.qr.opcode = req->req_upiu.qr.opcode;
817
+ req->rsp_upiu.qr.idn = req->req_upiu.qr.idn;
818
+ req->rsp_upiu.qr.index = req->req_upiu.qr.index;
819
+ req->rsp_upiu.qr.selector = req->req_upiu.qr.selector;
820
+ req->rsp_upiu.qr.length = cpu_to_be16(length);
821
+
822
+ return status;
823
+}
824
+
825
+static QueryRespCode ufs_exec_query_read(UfsRequest *req)
826
+{
827
+ QueryRespCode status;
828
+ switch (req->req_upiu.qr.opcode) {
829
+ case UPIU_QUERY_OPCODE_NOP:
830
+ status = QUERY_RESULT_SUCCESS;
831
+ break;
832
+ case UPIU_QUERY_OPCODE_READ_DESC:
833
+ status = ufs_read_desc(req);
834
+ break;
835
+ case UPIU_QUERY_OPCODE_READ_ATTR:
836
+ status = ufs_exec_query_attr(req, UFS_QUERY_ATTR_READ);
837
+ break;
838
+ case UPIU_QUERY_OPCODE_READ_FLAG:
839
+ status = ufs_exec_query_flag(req, UFS_QUERY_FLAG_READ);
840
+ break;
841
+ default:
842
+ trace_ufs_err_query_invalid_opcode(req->req_upiu.qr.opcode);
843
+ status = QUERY_RESULT_INVALID_OPCODE;
844
+ break;
845
+ }
846
+
847
+ return status;
848
+}
849
+
850
+static QueryRespCode ufs_exec_query_write(UfsRequest *req)
851
+{
852
+ QueryRespCode status;
853
+ switch (req->req_upiu.qr.opcode) {
854
+ case UPIU_QUERY_OPCODE_NOP:
855
+ status = QUERY_RESULT_SUCCESS;
856
+ break;
857
+ case UPIU_QUERY_OPCODE_WRITE_DESC:
858
+ /* write descriptor is not supported */
859
+ status = QUERY_RESULT_NOT_WRITEABLE;
860
+ break;
861
+ case UPIU_QUERY_OPCODE_WRITE_ATTR:
862
+ status = ufs_exec_query_attr(req, UFS_QUERY_ATTR_WRITE);
863
+ break;
864
+ case UPIU_QUERY_OPCODE_SET_FLAG:
865
+ status = ufs_exec_query_flag(req, UFS_QUERY_FLAG_SET);
866
+ break;
867
+ case UPIU_QUERY_OPCODE_CLEAR_FLAG:
868
+ status = ufs_exec_query_flag(req, UFS_QUERY_FLAG_CLEAR);
869
+ break;
870
+ case UPIU_QUERY_OPCODE_TOGGLE_FLAG:
871
+ status = ufs_exec_query_flag(req, UFS_QUERY_FLAG_TOGGLE);
872
+ break;
873
+ default:
874
+ trace_ufs_err_query_invalid_opcode(req->req_upiu.qr.opcode);
875
+ status = QUERY_RESULT_INVALID_OPCODE;
876
+ break;
877
+ }
878
+
879
+ return status;
880
+}
881
+
882
+static UfsReqResult ufs_exec_query_cmd(UfsRequest *req)
883
+{
884
+ uint8_t query_func = req->req_upiu.header.query_func;
885
+ uint16_t data_segment_length;
886
+ QueryRespCode status;
887
+
888
+ trace_ufs_exec_query_cmd(req->slot, req->req_upiu.qr.opcode);
889
+ if (query_func == UPIU_QUERY_FUNC_STANDARD_READ_REQUEST) {
890
+ status = ufs_exec_query_read(req);
891
+ } else if (query_func == UPIU_QUERY_FUNC_STANDARD_WRITE_REQUEST) {
892
+ status = ufs_exec_query_write(req);
893
+ } else {
894
+ status = QUERY_RESULT_GENERAL_FAILURE;
895
+ }
896
+
897
+ data_segment_length = be16_to_cpu(req->rsp_upiu.qr.length);
898
+ ufs_build_upiu_header(req, UPIU_TRANSACTION_QUERY_RSP, 0, status, 0,
899
+ data_segment_length);
900
+
901
+ if (status != QUERY_RESULT_SUCCESS) {
902
+ return UFS_REQUEST_FAIL;
903
+ }
904
+ return UFS_REQUEST_SUCCESS;
905
+}
906
+
907
+static void ufs_exec_req(UfsRequest *req)
908
+{
909
+ UfsReqResult req_result;
910
+
911
+ if (ufs_dma_read_upiu(req)) {
912
+ return;
913
+ }
914
+
915
+ switch (req->req_upiu.header.trans_type) {
916
+ case UPIU_TRANSACTION_NOP_OUT:
917
+ req_result = ufs_exec_nop_cmd(req);
918
+ break;
919
+ case UPIU_TRANSACTION_COMMAND:
920
+ /* Not yet implemented */
921
+ req_result = UFS_REQUEST_FAIL;
922
+ break;
923
+ case UPIU_TRANSACTION_QUERY_REQ:
924
+ req_result = ufs_exec_query_cmd(req);
925
+ break;
926
+ default:
927
+ trace_ufs_err_invalid_trans_code(req->slot,
928
+ req->req_upiu.header.trans_type);
929
+ req_result = UFS_REQUEST_FAIL;
930
+ }
931
+
932
+ ufs_complete_req(req, req_result);
933
+}
934
+
935
+static void ufs_process_req(void *opaque)
936
+{
937
+ UfsHc *u = opaque;
938
+ UfsRequest *req;
939
+ int slot;
940
+
941
+ for (slot = 0; slot < u->params.nutrs; slot++) {
942
+ req = &u->req_list[slot];
943
+
944
+ if (req->state != UFS_REQUEST_READY) {
945
+ continue;
946
+ }
947
+ trace_ufs_process_req(slot);
948
+ req->state = UFS_REQUEST_RUNNING;
949
+
950
+ ufs_exec_req(req);
951
+ }
952
+}
953
+
954
+static void ufs_complete_req(UfsRequest *req, UfsReqResult req_result)
955
+{
956
+ UfsHc *u = req->hc;
957
+ assert(req->state == UFS_REQUEST_RUNNING);
958
+
959
+ if (req_result == UFS_REQUEST_SUCCESS) {
960
+ req->utrd.header.dword_2 = cpu_to_le32(OCS_SUCCESS);
961
+ } else {
962
+ req->utrd.header.dword_2 = cpu_to_le32(OCS_INVALID_CMD_TABLE_ATTR);
963
+ }
964
+
965
+ trace_ufs_complete_req(req->slot);
966
+ req->state = UFS_REQUEST_COMPLETE;
967
+ qemu_bh_schedule(u->complete_bh);
968
+}
969
+
970
+static void ufs_clear_req(UfsRequest *req)
971
+{
972
+ if (req->sg != NULL) {
973
+ qemu_sglist_destroy(req->sg);
974
+ g_free(req->sg);
975
+ req->sg = NULL;
976
+ }
977
+
978
+ memset(&req->utrd, 0, sizeof(req->utrd));
979
+ memset(&req->req_upiu, 0, sizeof(req->req_upiu));
980
+ memset(&req->rsp_upiu, 0, sizeof(req->rsp_upiu));
981
+}
982
+
983
+static void ufs_sendback_req(void *opaque)
984
+{
985
+ UfsHc *u = opaque;
986
+ UfsRequest *req;
987
+ int slot;
988
+
989
+ for (slot = 0; slot < u->params.nutrs; slot++) {
990
+ req = &u->req_list[slot];
991
+
992
+ if (req->state != UFS_REQUEST_COMPLETE) {
993
+ continue;
994
+ }
995
+
996
+ if (ufs_dma_write_upiu(req)) {
997
+ req->state = UFS_REQUEST_ERROR;
998
+ continue;
999
+ }
1000
+
1001
+ /*
1002
+ * TODO: UTP Transfer Request Interrupt Aggregation Control is not yet
1003
+ * supported
1004
+ */
1005
+ if (le32_to_cpu(req->utrd.header.dword_2) != OCS_SUCCESS ||
1006
+ le32_to_cpu(req->utrd.header.dword_0) & UTP_REQ_DESC_INT_CMD) {
1007
+ u->reg.is = FIELD_DP32(u->reg.is, IS, UTRCS, 1);
1008
+ }
1009
+
1010
+ u->reg.utrldbr &= ~(1 << slot);
1011
+ u->reg.utrlcnr |= (1 << slot);
1012
+
1013
+ trace_ufs_sendback_req(req->slot);
1014
+
1015
+ ufs_clear_req(req);
1016
+ req->state = UFS_REQUEST_IDLE;
1017
+ }
1018
+
1019
+ ufs_irq_check(u);
1020
+}
1021
+
1022
static bool ufs_check_constraints(UfsHc *u, Error **errp)
1023
{
1024
if (u->params.nutrs > UFS_MAX_NUTRS) {
1025
@@ -XXX,XX +XXX,XX @@ static void ufs_init_pci(UfsHc *u, PCIDevice *pci_dev)
1026
u->irq = pci_allocate_irq(pci_dev);
225
}
1027
}
226
1028
227
static bool coroutine_fn
1029
+static void ufs_init_state(UfsHc *u)
228
@@ -XXX,XX +XXX,XX @@ fail:
1030
+{
229
return false;
1031
+ u->req_list = g_new0(UfsRequest, u->params.nutrs);
1032
+
1033
+ for (int i = 0; i < u->params.nutrs; i++) {
1034
+ u->req_list[i].hc = u;
1035
+ u->req_list[i].slot = i;
1036
+ u->req_list[i].sg = NULL;
1037
+ u->req_list[i].state = UFS_REQUEST_IDLE;
1038
+ }
1039
+
1040
+ u->doorbell_bh = qemu_bh_new_guarded(ufs_process_req, u,
1041
+ &DEVICE(u)->mem_reentrancy_guard);
1042
+ u->complete_bh = qemu_bh_new_guarded(ufs_sendback_req, u,
1043
+ &DEVICE(u)->mem_reentrancy_guard);
1044
+}
1045
+
1046
static void ufs_init_hc(UfsHc *u)
1047
{
1048
uint32_t cap = 0;
1049
@@ -XXX,XX +XXX,XX @@ static void ufs_init_hc(UfsHc *u)
1050
cap = FIELD_DP32(cap, CAP, CS, 0);
1051
u->reg.cap = cap;
1052
u->reg.ver = UFS_SPEC_VER;
1053
+
1054
+ memset(&u->device_desc, 0, sizeof(DeviceDescriptor));
1055
+ u->device_desc.length = sizeof(DeviceDescriptor);
1056
+ u->device_desc.descriptor_idn = QUERY_DESC_IDN_DEVICE;
1057
+ u->device_desc.device_sub_class = 0x01;
1058
+ u->device_desc.number_lu = 0x00;
1059
+ u->device_desc.number_wlu = 0x04;
1060
+ /* TODO: Revisit it when Power Management is implemented */
1061
+ u->device_desc.init_power_mode = 0x01; /* Active Mode */
1062
+ u->device_desc.high_priority_lun = 0x7F; /* Same Priority */
1063
+ u->device_desc.spec_version = cpu_to_be16(UFS_SPEC_VER);
1064
+ u->device_desc.manufacturer_name = 0x00;
1065
+ u->device_desc.product_name = 0x01;
1066
+ u->device_desc.serial_number = 0x02;
1067
+ u->device_desc.oem_id = 0x03;
1068
+ u->device_desc.ud_0_base_offset = 0x16;
1069
+ u->device_desc.ud_config_p_length = 0x1A;
1070
+ u->device_desc.device_rtt_cap = 0x02;
1071
+ u->device_desc.queue_depth = u->params.nutrs;
1072
+ u->device_desc.product_revision_level = 0x04;
1073
+
1074
+ memset(&u->geometry_desc, 0, sizeof(GeometryDescriptor));
1075
+ u->geometry_desc.length = sizeof(GeometryDescriptor);
1076
+ u->geometry_desc.descriptor_idn = QUERY_DESC_IDN_GEOMETRY;
1077
+ u->geometry_desc.max_number_lu = (UFS_MAX_LUS == 32) ? 0x1 : 0x0;
1078
+ u->geometry_desc.segment_size = cpu_to_be32(0x2000); /* 4KB */
1079
+ u->geometry_desc.allocation_unit_size = 0x1; /* 4KB */
1080
+ u->geometry_desc.min_addr_block_size = 0x8; /* 4KB */
1081
+ u->geometry_desc.max_in_buffer_size = 0x8;
1082
+ u->geometry_desc.max_out_buffer_size = 0x8;
1083
+ u->geometry_desc.rpmb_read_write_size = 0x40;
1084
+ u->geometry_desc.data_ordering =
1085
+ 0x0; /* out-of-order data transfer is not supported */
1086
+ u->geometry_desc.max_context_id_number = 0x5;
1087
+ u->geometry_desc.supported_memory_types = cpu_to_be16(0x8001);
1088
+
1089
+ memset(&u->attributes, 0, sizeof(u->attributes));
1090
+ u->attributes.max_data_in_size = 0x08;
1091
+ u->attributes.max_data_out_size = 0x08;
1092
+ u->attributes.ref_clk_freq = 0x01; /* 26 MHz */
1093
+ /* configure descriptor is not supported */
1094
+ u->attributes.config_descr_lock = 0x01;
1095
+ u->attributes.max_num_of_rtt = 0x02;
1096
+
1097
+ memset(&u->flags, 0, sizeof(u->flags));
1098
+ u->flags.permanently_disable_fw_update = 1;
230
}
1099
}
231
1100
232
-
1101
static void ufs_realize(PCIDevice *pci_dev, Error **errp)
233
-static void vu_client_start(VuServer *server);
1102
@@ -XXX,XX +XXX,XX @@ static void ufs_realize(PCIDevice *pci_dev, Error **errp)
234
static coroutine_fn void vu_client_trip(void *opaque)
235
{
236
VuServer *server = opaque;
237
+ VuDev *vu_dev = &server->vu_dev;
238
239
- while (!server->aio_context_changed && server->sioc) {
240
- server->processing_msg = true;
241
- vu_dispatch(&server->vu_dev);
242
- server->processing_msg = false;
243
+ while (!vu_dev->broken && vu_dispatch(vu_dev)) {
244
+ /* Keep running */
245
}
246
247
- if (server->aio_context_changed && server->sioc) {
248
- server->aio_context_changed = false;
249
- vu_client_start(server);
250
- }
251
-}
252
+ vu_deinit(vu_dev);
253
+
254
+ /* vu_deinit() should have called remove_watch() */
255
+ assert(QTAILQ_EMPTY(&server->vu_fd_watches));
256
+
257
+ object_unref(OBJECT(server->sioc));
258
+ server->sioc = NULL;
259
260
-static void vu_client_start(VuServer *server)
261
-{
262
- server->co_trip = qemu_coroutine_create(vu_client_trip, server);
263
- aio_co_enter(server->ctx, server->co_trip);
264
+ object_unref(OBJECT(server->ioc));
265
+ server->ioc = NULL;
266
+
267
+ server->co_trip = NULL;
268
+ if (server->restart_listener_bh) {
269
+ qemu_bh_schedule(server->restart_listener_bh);
270
+ }
271
+ aio_wait_kick();
272
}
273
274
/*
275
@@ -XXX,XX +XXX,XX @@ static void vu_client_start(VuServer *server)
276
static void kick_handler(void *opaque)
277
{
278
VuFdWatch *vu_fd_watch = opaque;
279
- vu_fd_watch->processing = true;
280
- vu_fd_watch->cb(vu_fd_watch->vu_dev, 0, vu_fd_watch->pvt);
281
- vu_fd_watch->processing = false;
282
+ VuDev *vu_dev = vu_fd_watch->vu_dev;
283
+
284
+ vu_fd_watch->cb(vu_dev, 0, vu_fd_watch->pvt);
285
+
286
+ /* Stop vu_client_trip() if an error occurred in vu_fd_watch->cb() */
287
+ if (vu_dev->broken) {
288
+ VuServer *server = container_of(vu_dev, VuServer, vu_dev);
289
+
290
+ qio_channel_shutdown(server->ioc, QIO_CHANNEL_SHUTDOWN_BOTH, NULL);
291
+ }
292
}
293
294
-
295
static VuFdWatch *find_vu_fd_watch(VuServer *server, int fd)
296
{
297
298
@@ -XXX,XX +XXX,XX @@ static void vu_accept(QIONetListener *listener, QIOChannelSocket *sioc,
299
qio_channel_set_name(QIO_CHANNEL(sioc), "vhost-user client");
300
server->ioc = QIO_CHANNEL(sioc);
301
object_ref(OBJECT(server->ioc));
302
- qio_channel_attach_aio_context(server->ioc, server->ctx);
303
+
304
+ /* TODO vu_message_write() spins if non-blocking! */
305
qio_channel_set_blocking(server->ioc, false, NULL);
306
- vu_client_start(server);
307
+
308
+ server->co_trip = qemu_coroutine_create(vu_client_trip, server);
309
+
310
+ aio_context_acquire(server->ctx);
311
+ vhost_user_server_attach_aio_context(server, server->ctx);
312
+ aio_context_release(server->ctx);
313
}
314
315
-
316
void vhost_user_server_stop(VuServer *server)
317
{
318
+ aio_context_acquire(server->ctx);
319
+
320
+ qemu_bh_delete(server->restart_listener_bh);
321
+ server->restart_listener_bh = NULL;
322
+
323
if (server->sioc) {
324
- close_client(server);
325
+ VuFdWatch *vu_fd_watch;
326
+
327
+ QTAILQ_FOREACH(vu_fd_watch, &server->vu_fd_watches, next) {
328
+ aio_set_fd_handler(server->ctx, vu_fd_watch->fd, true,
329
+ NULL, NULL, NULL, vu_fd_watch);
330
+ }
331
+
332
+ qio_channel_shutdown(server->ioc, QIO_CHANNEL_SHUTDOWN_BOTH, NULL);
333
+
334
+ AIO_WAIT_WHILE(server->ctx, server->co_trip);
335
}
336
337
+ aio_context_release(server->ctx);
338
+
339
if (server->listener) {
340
qio_net_listener_disconnect(server->listener);
341
object_unref(OBJECT(server->listener));
342
}
343
+}
344
+
345
+/*
346
+ * Allow the next client to connect to the server. Called from a BH in the main
347
+ * loop.
348
+ */
349
+static void restart_listener_bh(void *opaque)
350
+{
351
+ VuServer *server = opaque;
352
353
+ qio_net_listener_set_client_func(server->listener, vu_accept, server,
354
+ NULL);
355
}
356
357
-void vhost_user_server_set_aio_context(VuServer *server, AioContext *ctx)
358
+/* Called with ctx acquired */
359
+void vhost_user_server_attach_aio_context(VuServer *server, AioContext *ctx)
360
{
361
- VuFdWatch *vu_fd_watch, *next;
362
- void *opaque = NULL;
363
- IOHandler *io_read = NULL;
364
- bool attach;
365
+ VuFdWatch *vu_fd_watch;
366
367
- server->ctx = ctx ? ctx : qemu_get_aio_context();
368
+ server->ctx = ctx;
369
370
if (!server->sioc) {
371
- /* not yet serving any client*/
372
return;
1103
return;
373
}
1104
}
374
1105
375
- if (ctx) {
1106
+ ufs_init_state(u);
376
- qio_channel_attach_aio_context(server->ioc, ctx);
1107
ufs_init_hc(u);
377
- server->aio_context_changed = true;
1108
ufs_init_pci(u, pci_dev);
378
- io_read = kick_handler;
379
- attach = true;
380
- } else {
381
+ qio_channel_attach_aio_context(server->ioc, ctx);
382
+
383
+ QTAILQ_FOREACH(vu_fd_watch, &server->vu_fd_watches, next) {
384
+ aio_set_fd_handler(ctx, vu_fd_watch->fd, true, kick_handler, NULL,
385
+ NULL, vu_fd_watch);
386
+ }
387
+
388
+ aio_co_schedule(ctx, server->co_trip);
389
+}
390
+
391
+/* Called with server->ctx acquired */
392
+void vhost_user_server_detach_aio_context(VuServer *server)
393
+{
394
+ if (server->sioc) {
395
+ VuFdWatch *vu_fd_watch;
396
+
397
+ QTAILQ_FOREACH(vu_fd_watch, &server->vu_fd_watches, next) {
398
+ aio_set_fd_handler(server->ctx, vu_fd_watch->fd, true,
399
+ NULL, NULL, NULL, vu_fd_watch);
400
+ }
401
+
402
qio_channel_detach_aio_context(server->ioc);
403
- /* server->ioc->ctx keeps the old AioConext */
404
- ctx = server->ioc->ctx;
405
- attach = false;
406
}
407
408
- QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) {
409
- if (vu_fd_watch->cb) {
410
- opaque = attach ? vu_fd_watch : NULL;
411
- aio_set_fd_handler(ctx, vu_fd_watch->fd, true,
412
- io_read, NULL, NULL,
413
- opaque);
414
- }
415
- }
416
+ server->ctx = NULL;
417
}
1109
}
418
1110
419
-
1111
+static void ufs_exit(PCIDevice *pci_dev)
420
bool vhost_user_server_start(VuServer *server,
1112
+{
421
SocketAddress *socket_addr,
1113
+ UfsHc *u = UFS(pci_dev);
422
AioContext *ctx,
1114
+
423
@@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server,
1115
+ qemu_bh_delete(u->doorbell_bh);
424
const VuDevIface *vu_iface,
1116
+ qemu_bh_delete(u->complete_bh);
425
Error **errp)
1117
+
426
{
1118
+ for (int i = 0; i < u->params.nutrs; i++) {
427
+ QEMUBH *bh;
1119
+ ufs_clear_req(&u->req_list[i]);
428
QIONetListener *listener = qio_net_listener_new();
1120
+ }
429
if (qio_net_listener_open_sync(listener, socket_addr, 1,
1121
+ g_free(u->req_list);
430
errp) < 0) {
1122
+}
431
@@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server,
1123
+
432
return false;
1124
static Property ufs_props[] = {
433
}
1125
DEFINE_PROP_STRING("serial", UfsHc, params.serial),
434
1126
DEFINE_PROP_UINT8("nutrs", UfsHc, params.nutrs, 32),
435
+ bh = qemu_bh_new(restart_listener_bh, server);
1127
@@ -XXX,XX +XXX,XX @@ static void ufs_class_init(ObjectClass *oc, void *data)
436
+
1128
PCIDeviceClass *pc = PCI_DEVICE_CLASS(oc);
437
/* zero out unspecified fields */
1129
438
*server = (VuServer) {
1130
pc->realize = ufs_realize;
439
.listener = listener,
1131
+ pc->exit = ufs_exit;
440
+ .restart_listener_bh = bh,
1132
pc->vendor_id = PCI_VENDOR_ID_REDHAT;
441
.vu_iface = vu_iface,
1133
pc->device_id = PCI_DEVICE_ID_REDHAT_UFS;
442
.max_queues = max_queues,
1134
pc->class_id = PCI_CLASS_STORAGE_UFS;
443
.ctx = ctx,
1135
diff --git a/hw/ufs/trace-events b/hw/ufs/trace-events
1136
index XXXXXXX..XXXXXXX 100644
1137
--- a/hw/ufs/trace-events
1138
+++ b/hw/ufs/trace-events
1139
@@ -XXX,XX +XXX,XX @@ ufs_err_dma_read_req_upiu(uint32_t slot, uint64_t addr) "failed to read req upiu
1140
ufs_err_dma_read_prdt(uint32_t slot, uint64_t addr) "failed to read prdt. UTRLDBR slot %"PRIu32", prdt addr %"PRIu64""
1141
ufs_err_dma_write_utrd(uint32_t slot, uint64_t addr) "failed to write utrd. UTRLDBR slot %"PRIu32", UTRD dma addr %"PRIu64""
1142
ufs_err_dma_write_rsp_upiu(uint32_t slot, uint64_t addr) "failed to write rsp upiu. UTRLDBR slot %"PRIu32", response upiu addr %"PRIu64""
1143
+ufs_err_utrl_slot_error(uint32_t slot) "UTRLDBR slot %"PRIu32" is in error"
1144
ufs_err_utrl_slot_busy(uint32_t slot) "UTRLDBR slot %"PRIu32" is busy"
1145
ufs_err_unsupport_register_offset(uint32_t offset) "Register offset 0x%"PRIx32" is not yet supported"
1146
ufs_err_invalid_register_offset(uint32_t offset) "Register offset 0x%"PRIx32" is invalid"
444
--
1147
--
445
2.26.2
1148
2.41.0
446
diff view generated by jsdifflib
1
From: Coiby Xu <coiby.xu@gmail.com>
1
From: Jeuk Kim <jeuk20.kim@gmail.com>
2
2
3
Sharing QEMU devices via vhost-user protocol.
3
This commit adds support for ufs logical unit.
4
The LU handles processing for the SCSI command,
5
unit descriptor query request.
4
6
5
Only one vhost-user client can connect to the server one time.
7
This commit enables the UFS device to process
8
IO requests.
6
9
7
Suggested-by: Kevin Wolf <kwolf@redhat.com>
10
Signed-off-by: Jeuk Kim <jeuk20.kim@samsung.com>
8
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
9
Signed-off-by: Coiby Xu <coiby.xu@gmail.com>
10
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
11
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
11
Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com>
12
Message-id: 898cea923e819dc21a99597bf045a12d7983be28.1691062912.git.jeuk20.kim@samsung.com
12
Message-id: 20200918080912.321299-4-coiby.xu@gmail.com
13
[Fixed size_t %lu -> %zu format string compiler error.
14
--Stefan]
15
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
13
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
16
---
14
---
17
util/vhost-user-server.h | 65 ++++++
15
hw/ufs/ufs.h | 43 ++
18
util/vhost-user-server.c | 428 +++++++++++++++++++++++++++++++++++++++
16
include/scsi/constants.h | 1 +
19
util/meson.build | 1 +
17
hw/ufs/lu.c | 1445 ++++++++++++++++++++++++++++++++++++++
20
3 files changed, 494 insertions(+)
18
hw/ufs/ufs.c | 252 ++++++-
21
create mode 100644 util/vhost-user-server.h
19
hw/ufs/meson.build | 2 +-
22
create mode 100644 util/vhost-user-server.c
20
hw/ufs/trace-events | 25 +
21
6 files changed, 1761 insertions(+), 7 deletions(-)
22
create mode 100644 hw/ufs/lu.c
23
23
24
diff --git a/util/vhost-user-server.h b/util/vhost-user-server.h
24
diff --git a/hw/ufs/ufs.h b/hw/ufs/ufs.h
25
index XXXXXXX..XXXXXXX 100644
26
--- a/hw/ufs/ufs.h
27
+++ b/hw/ufs/ufs.h
28
@@ -XXX,XX +XXX,XX @@
29
#define UFS_MAX_LUS 32
30
#define UFS_BLOCK_SIZE 4096
31
32
+typedef struct UfsBusClass {
33
+ BusClass parent_class;
34
+ bool (*parent_check_address)(BusState *bus, DeviceState *dev, Error **errp);
35
+} UfsBusClass;
36
+
37
+typedef struct UfsBus {
38
+ SCSIBus parent_bus;
39
+} UfsBus;
40
+
41
+#define TYPE_UFS_BUS "ufs-bus"
42
+DECLARE_OBJ_CHECKERS(UfsBus, UfsBusClass, UFS_BUS, TYPE_UFS_BUS)
43
+
44
typedef enum UfsRequestState {
45
UFS_REQUEST_IDLE = 0,
46
UFS_REQUEST_READY = 1,
47
@@ -XXX,XX +XXX,XX @@ typedef enum UfsRequestState {
48
typedef enum UfsReqResult {
49
UFS_REQUEST_SUCCESS = 0,
50
UFS_REQUEST_FAIL = 1,
51
+ UFS_REQUEST_NO_COMPLETE = 2,
52
} UfsReqResult;
53
54
typedef struct UfsRequest {
55
@@ -XXX,XX +XXX,XX @@ typedef struct UfsRequest {
56
QEMUSGList *sg;
57
} UfsRequest;
58
59
+typedef struct UfsLu {
60
+ SCSIDevice qdev;
61
+ uint8_t lun;
62
+ UnitDescriptor unit_desc;
63
+} UfsLu;
64
+
65
+typedef struct UfsWLu {
66
+ SCSIDevice qdev;
67
+ uint8_t lun;
68
+} UfsWLu;
69
+
70
typedef struct UfsParams {
71
char *serial;
72
uint8_t nutrs; /* Number of UTP Transfer Request Slots */
73
@@ -XXX,XX +XXX,XX @@ typedef struct UfsParams {
74
75
typedef struct UfsHc {
76
PCIDevice parent_obj;
77
+ UfsBus bus;
78
MemoryRegion iomem;
79
UfsReg reg;
80
UfsParams params;
81
uint32_t reg_size;
82
UfsRequest *req_list;
83
84
+ UfsLu *lus[UFS_MAX_LUS];
85
+ UfsWLu *report_wlu;
86
+ UfsWLu *dev_wlu;
87
+ UfsWLu *boot_wlu;
88
+ UfsWLu *rpmb_wlu;
89
DeviceDescriptor device_desc;
90
GeometryDescriptor geometry_desc;
91
Attributes attributes;
92
@@ -XXX,XX +XXX,XX @@ typedef struct UfsHc {
93
#define TYPE_UFS "ufs"
94
#define UFS(obj) OBJECT_CHECK(UfsHc, (obj), TYPE_UFS)
95
96
+#define TYPE_UFS_LU "ufs-lu"
97
+#define UFSLU(obj) OBJECT_CHECK(UfsLu, (obj), TYPE_UFS_LU)
98
+
99
+#define TYPE_UFS_WLU "ufs-wlu"
100
+#define UFSWLU(obj) OBJECT_CHECK(UfsWLu, (obj), TYPE_UFS_WLU)
101
+
102
typedef enum UfsQueryFlagPerm {
103
UFS_QUERY_FLAG_NONE = 0x0,
104
UFS_QUERY_FLAG_READ = 0x1,
105
@@ -XXX,XX +XXX,XX @@ typedef enum UfsQueryAttrPerm {
106
UFS_QUERY_ATTR_WRITE = 0x2,
107
} UfsQueryAttrPerm;
108
109
+static inline bool is_wlun(uint8_t lun)
110
+{
111
+ return (lun == UFS_UPIU_REPORT_LUNS_WLUN ||
112
+ lun == UFS_UPIU_UFS_DEVICE_WLUN || lun == UFS_UPIU_BOOT_WLUN ||
113
+ lun == UFS_UPIU_RPMB_WLUN);
114
+}
115
+
116
#endif /* HW_UFS_UFS_H */
117
diff --git a/include/scsi/constants.h b/include/scsi/constants.h
118
index XXXXXXX..XXXXXXX 100644
119
--- a/include/scsi/constants.h
120
+++ b/include/scsi/constants.h
121
@@ -XXX,XX +XXX,XX @@
122
#define MODE_PAGE_FLEXIBLE_DISK_GEOMETRY 0x05
123
#define MODE_PAGE_CACHING 0x08
124
#define MODE_PAGE_AUDIO_CTL 0x0e
125
+#define MODE_PAGE_CONTROL 0x0a
126
#define MODE_PAGE_POWER 0x1a
127
#define MODE_PAGE_FAULT_FAIL 0x1c
128
#define MODE_PAGE_TO_PROTECT 0x1d
129
diff --git a/hw/ufs/lu.c b/hw/ufs/lu.c
25
new file mode 100644
130
new file mode 100644
26
index XXXXXXX..XXXXXXX
131
index XXXXXXX..XXXXXXX
27
--- /dev/null
132
--- /dev/null
28
+++ b/util/vhost-user-server.h
133
+++ b/hw/ufs/lu.c
29
@@ -XXX,XX +XXX,XX @@
134
@@ -XXX,XX +XXX,XX @@
30
+/*
135
+/*
31
+ * Sharing QEMU devices via vhost-user protocol
136
+ * QEMU UFS Logical Unit
32
+ *
137
+ *
33
+ * Copyright (c) Coiby Xu <coiby.xu@gmail.com>.
138
+ * Copyright (c) 2023 Samsung Electronics Co., Ltd. All rights reserved.
34
+ * Copyright (c) 2020 Red Hat, Inc.
35
+ *
139
+ *
36
+ * This work is licensed under the terms of the GNU GPL, version 2 or
140
+ * Written by Jeuk Kim <jeuk20.kim@samsung.com>
37
+ * later. See the COPYING file in the top-level directory.
141
+ *
142
+ * This code is licensed under the GNU GPL v2 or later.
38
+ */
143
+ */
39
+
144
+
40
+#ifndef VHOST_USER_SERVER_H
145
+#include "qemu/osdep.h"
41
+#define VHOST_USER_SERVER_H
146
+#include "qemu/units.h"
42
+
43
+#include "contrib/libvhost-user/libvhost-user.h"
44
+#include "io/channel-socket.h"
45
+#include "io/channel-file.h"
46
+#include "io/net-listener.h"
47
+#include "qemu/error-report.h"
48
+#include "qapi/error.h"
147
+#include "qapi/error.h"
49
+#include "standard-headers/linux/virtio_blk.h"
148
+#include "qemu/memalign.h"
50
+
149
+#include "hw/scsi/scsi.h"
51
+typedef struct VuFdWatch {
150
+#include "scsi/constants.h"
52
+ VuDev *vu_dev;
151
+#include "sysemu/block-backend.h"
53
+ int fd; /*kick fd*/
152
+#include "qemu/cutils.h"
54
+ void *pvt;
153
+#include "trace.h"
55
+ vu_watch_cb cb;
154
+#include "ufs.h"
56
+ bool processing;
155
+
57
+ QTAILQ_ENTRY(VuFdWatch) next;
156
+/*
58
+} VuFdWatch;
157
+ * The code below handling SCSI commands is copied from hw/scsi/scsi-disk.c,
59
+
158
+ * with minor adjustments to make it work for UFS.
60
+typedef struct VuServer VuServer;
159
+ */
61
+typedef void DevicePanicNotifierFn(VuServer *server);
160
+
62
+
161
+#define SCSI_DMA_BUF_SIZE (128 * KiB)
63
+struct VuServer {
162
+#define SCSI_MAX_INQUIRY_LEN 256
64
+ QIONetListener *listener;
163
+#define SCSI_INQUIRY_DATA_SIZE 36
65
+ AioContext *ctx;
164
+#define SCSI_MAX_MODE_LEN 256
66
+ DevicePanicNotifierFn *device_panic_notifier;
165
+
67
+ int max_queues;
166
+typedef struct UfsSCSIReq {
68
+ const VuDevIface *vu_iface;
167
+ SCSIRequest req;
69
+ VuDev vu_dev;
168
+ /* Both sector and sector_count are in terms of BDRV_SECTOR_SIZE bytes. */
70
+ QIOChannel *ioc; /* The I/O channel with the client */
169
+ uint64_t sector;
71
+ QIOChannelSocket *sioc; /* The underlying data channel with the client */
170
+ uint32_t sector_count;
72
+ /* IOChannel for fd provided via VHOST_USER_SET_SLAVE_REQ_FD */
171
+ uint32_t buflen;
73
+ QIOChannel *ioc_slave;
172
+ bool started;
74
+ QIOChannelSocket *sioc_slave;
173
+ bool need_fua_emulation;
75
+ Coroutine *co_trip; /* coroutine for processing VhostUserMsg */
174
+ struct iovec iov;
76
+ QTAILQ_HEAD(, VuFdWatch) vu_fd_watches;
175
+ QEMUIOVector qiov;
77
+ /* restart coroutine co_trip if AIOContext is changed */
176
+ BlockAcctCookie acct;
78
+ bool aio_context_changed;
177
+} UfsSCSIReq;
79
+ bool processing_msg;
178
+
179
+static void ufs_scsi_free_request(SCSIRequest *req)
180
+{
181
+ UfsSCSIReq *r = DO_UPCAST(UfsSCSIReq, req, req);
182
+
183
+ qemu_vfree(r->iov.iov_base);
184
+}
185
+
186
+static void scsi_check_condition(UfsSCSIReq *r, SCSISense sense)
187
+{
188
+ trace_ufs_scsi_check_condition(r->req.tag, sense.key, sense.asc,
189
+ sense.ascq);
190
+ scsi_req_build_sense(&r->req, sense);
191
+ scsi_req_complete(&r->req, CHECK_CONDITION);
192
+}
193
+
194
+static int ufs_scsi_emulate_vpd_page(SCSIRequest *req, uint8_t *outbuf,
195
+ uint32_t outbuf_len)
196
+{
197
+ UfsHc *u = UFS(req->bus->qbus.parent);
198
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, req->dev);
199
+ uint8_t page_code = req->cmd.buf[2];
200
+ int start, buflen = 0;
201
+
202
+ if (outbuf_len < SCSI_INQUIRY_DATA_SIZE) {
203
+ return -1;
204
+ }
205
+
206
+ outbuf[buflen++] = lu->qdev.type & 0x1f;
207
+ outbuf[buflen++] = page_code;
208
+ outbuf[buflen++] = 0x00;
209
+ outbuf[buflen++] = 0x00;
210
+ start = buflen;
211
+
212
+ switch (page_code) {
213
+ case 0x00: /* Supported page codes, mandatory */
214
+ {
215
+ trace_ufs_scsi_emulate_vpd_page_00(req->cmd.xfer);
216
+ outbuf[buflen++] = 0x00; /* list of supported pages (this page) */
217
+ if (u->params.serial) {
218
+ outbuf[buflen++] = 0x80; /* unit serial number */
219
+ }
220
+ outbuf[buflen++] = 0x87; /* mode page policy */
221
+ break;
222
+ }
223
+ case 0x80: /* Device serial number, optional */
224
+ {
225
+ int l;
226
+
227
+ if (!u->params.serial) {
228
+ trace_ufs_scsi_emulate_vpd_page_80_not_supported();
229
+ return -1;
230
+ }
231
+
232
+ l = strlen(u->params.serial);
233
+ if (l > SCSI_INQUIRY_DATA_SIZE) {
234
+ l = SCSI_INQUIRY_DATA_SIZE;
235
+ }
236
+
237
+ trace_ufs_scsi_emulate_vpd_page_80(req->cmd.xfer);
238
+ memcpy(outbuf + buflen, u->params.serial, l);
239
+ buflen += l;
240
+ break;
241
+ }
242
+ case 0x87: /* Mode Page Policy, mandatory */
243
+ {
244
+ trace_ufs_scsi_emulate_vpd_page_87(req->cmd.xfer);
245
+ outbuf[buflen++] = 0x3f; /* apply to all mode pages and subpages */
246
+ outbuf[buflen++] = 0xff;
247
+ outbuf[buflen++] = 0; /* shared */
248
+ outbuf[buflen++] = 0;
249
+ break;
250
+ }
251
+ default:
252
+ return -1;
253
+ }
254
+ /* done with EVPD */
255
+ assert(buflen - start <= 255);
256
+ outbuf[start - 1] = buflen - start;
257
+ return buflen;
258
+}
259
+
260
+static int ufs_scsi_emulate_inquiry(SCSIRequest *req, uint8_t *outbuf,
261
+ uint32_t outbuf_len)
262
+{
263
+ int buflen = 0;
264
+
265
+ if (outbuf_len < SCSI_INQUIRY_DATA_SIZE) {
266
+ return -1;
267
+ }
268
+
269
+ if (req->cmd.buf[1] & 0x1) {
270
+ /* Vital product data */
271
+ return ufs_scsi_emulate_vpd_page(req, outbuf, outbuf_len);
272
+ }
273
+
274
+ /* Standard INQUIRY data */
275
+ if (req->cmd.buf[2] != 0) {
276
+ return -1;
277
+ }
278
+
279
+ /* PAGE CODE == 0 */
280
+ buflen = req->cmd.xfer;
281
+ if (buflen > SCSI_MAX_INQUIRY_LEN) {
282
+ buflen = SCSI_MAX_INQUIRY_LEN;
283
+ }
284
+
285
+ if (is_wlun(req->lun)) {
286
+ outbuf[0] = TYPE_WLUN;
287
+ } else {
288
+ outbuf[0] = 0;
289
+ }
290
+ outbuf[1] = 0;
291
+
292
+ strpadcpy((char *)&outbuf[16], 16, "QEMU UFS", ' ');
293
+ strpadcpy((char *)&outbuf[8], 8, "QEMU", ' ');
294
+
295
+ memset(&outbuf[32], 0, 4);
296
+
297
+ outbuf[2] = 0x06; /* SPC-4 */
298
+ outbuf[3] = 0x2;
299
+
300
+ if (buflen > SCSI_INQUIRY_DATA_SIZE) {
301
+ outbuf[4] = buflen - 5; /* Additional Length = (Len - 1) - 4 */
302
+ } else {
303
+ /*
304
+ * If the allocation length of CDB is too small, the additional
305
+ * length is not adjusted
306
+ */
307
+ outbuf[4] = SCSI_INQUIRY_DATA_SIZE - 5;
308
+ }
309
+
310
+ /* Support TCQ. */
311
+ outbuf[7] = req->bus->info->tcq ? 0x02 : 0;
312
+ return buflen;
313
+}
314
+
315
+static int mode_sense_page(UfsLu *lu, int page, uint8_t **p_outbuf,
316
+ int page_control)
317
+{
318
+ static const int mode_sense_valid[0x3f] = {
319
+ [MODE_PAGE_CACHING] = 1,
320
+ [MODE_PAGE_R_W_ERROR] = 1,
321
+ [MODE_PAGE_CONTROL] = 1,
322
+ };
323
+
324
+ uint8_t *p = *p_outbuf + 2;
325
+ int length;
326
+
327
+ assert(page < ARRAY_SIZE(mode_sense_valid));
328
+ if ((mode_sense_valid[page]) == 0) {
329
+ return -1;
330
+ }
331
+
332
+ /*
333
+ * If Changeable Values are requested, a mask denoting those mode parameters
334
+ * that are changeable shall be returned. As we currently don't support
335
+ * parameter changes via MODE_SELECT all bits are returned set to zero.
336
+ * The buffer was already memset to zero by the caller of this function.
337
+ */
338
+ switch (page) {
339
+ case MODE_PAGE_CACHING:
340
+ length = 0x12;
341
+ if (page_control == 1 || /* Changeable Values */
342
+ blk_enable_write_cache(lu->qdev.conf.blk)) {
343
+ p[0] = 4; /* WCE */
344
+ }
345
+ break;
346
+
347
+ case MODE_PAGE_R_W_ERROR:
348
+ length = 10;
349
+ if (page_control == 1) { /* Changeable Values */
350
+ break;
351
+ }
352
+ p[0] = 0x80; /* Automatic Write Reallocation Enabled */
353
+ break;
354
+
355
+ case MODE_PAGE_CONTROL:
356
+ length = 10;
357
+ if (page_control == 1) { /* Changeable Values */
358
+ break;
359
+ }
360
+ p[1] = 0x10; /* Queue Algorithm modifier */
361
+ p[8] = 0xff; /* Busy Timeout Period */
362
+ p[9] = 0xff;
363
+ break;
364
+
365
+ default:
366
+ return -1;
367
+ }
368
+
369
+ assert(length < 256);
370
+ (*p_outbuf)[0] = page;
371
+ (*p_outbuf)[1] = length;
372
+ *p_outbuf += length + 2;
373
+ return length + 2;
374
+}
375
+
376
+static int ufs_scsi_emulate_mode_sense(UfsSCSIReq *r, uint8_t *outbuf)
377
+{
378
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
379
+ bool dbd;
380
+ int page, buflen, ret, page_control;
381
+ uint8_t *p;
382
+ uint8_t dev_specific_param = 0;
383
+
384
+ dbd = (r->req.cmd.buf[1] & 0x8) != 0;
385
+ if (!dbd) {
386
+ return -1;
387
+ }
388
+
389
+ page = r->req.cmd.buf[2] & 0x3f;
390
+ page_control = (r->req.cmd.buf[2] & 0xc0) >> 6;
391
+
392
+ trace_ufs_scsi_emulate_mode_sense((r->req.cmd.buf[0] == MODE_SENSE) ? 6 :
393
+ 10,
394
+ page, r->req.cmd.xfer, page_control);
395
+ memset(outbuf, 0, r->req.cmd.xfer);
396
+ p = outbuf;
397
+
398
+ if (!blk_is_writable(lu->qdev.conf.blk)) {
399
+ dev_specific_param |= 0x80; /* Readonly. */
400
+ }
401
+
402
+ p[2] = 0; /* Medium type. */
403
+ p[3] = dev_specific_param;
404
+ p[6] = p[7] = 0; /* Block descriptor length. */
405
+ p += 8;
406
+
407
+ if (page_control == 3) {
408
+ /* Saved Values */
409
+ scsi_check_condition(r, SENSE_CODE(SAVING_PARAMS_NOT_SUPPORTED));
410
+ return -1;
411
+ }
412
+
413
+ if (page == 0x3f) {
414
+ for (page = 0; page <= 0x3e; page++) {
415
+ mode_sense_page(lu, page, &p, page_control);
416
+ }
417
+ } else {
418
+ ret = mode_sense_page(lu, page, &p, page_control);
419
+ if (ret == -1) {
420
+ return -1;
421
+ }
422
+ }
423
+
424
+ buflen = p - outbuf;
425
+ /*
426
+ * The mode data length field specifies the length in bytes of the
427
+ * following data that is available to be transferred. The mode data
428
+ * length does not include itself.
429
+ */
430
+ outbuf[0] = ((buflen - 2) >> 8) & 0xff;
431
+ outbuf[1] = (buflen - 2) & 0xff;
432
+ return buflen;
433
+}
434
+
435
+/*
436
+ * scsi_handle_rw_error has two return values. False means that the error
437
+ * must be ignored, true means that the error has been processed and the
438
+ * caller should not do anything else for this request. Note that
439
+ * scsi_handle_rw_error always manages its reference counts, independent
440
+ * of the return value.
441
+ */
442
+static bool scsi_handle_rw_error(UfsSCSIReq *r, int ret, bool acct_failed)
443
+{
444
+ bool is_read = (r->req.cmd.mode == SCSI_XFER_FROM_DEV);
445
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
446
+ SCSISense sense = SENSE_CODE(NO_SENSE);
447
+ int error = 0;
448
+ bool req_has_sense = false;
449
+ BlockErrorAction action;
450
+ int status;
451
+
452
+ if (ret < 0) {
453
+ status = scsi_sense_from_errno(-ret, &sense);
454
+ error = -ret;
455
+ } else {
456
+ /* A passthrough command has completed with nonzero status. */
457
+ status = ret;
458
+ if (status == CHECK_CONDITION) {
459
+ req_has_sense = true;
460
+ error = scsi_sense_buf_to_errno(r->req.sense, sizeof(r->req.sense));
461
+ } else {
462
+ error = EINVAL;
463
+ }
464
+ }
465
+
466
+ /*
467
+ * Check whether the error has to be handled by the guest or should
468
+ * rather follow the rerror=/werror= settings. Guest-handled errors
469
+ * are usually retried immediately, so do not post them to QMP and
470
+ * do not account them as failed I/O.
471
+ */
472
+ if (req_has_sense && scsi_sense_buf_is_guest_recoverable(
473
+ r->req.sense, sizeof(r->req.sense))) {
474
+ action = BLOCK_ERROR_ACTION_REPORT;
475
+ acct_failed = false;
476
+ } else {
477
+ action = blk_get_error_action(lu->qdev.conf.blk, is_read, error);
478
+ blk_error_action(lu->qdev.conf.blk, action, is_read, error);
479
+ }
480
+
481
+ switch (action) {
482
+ case BLOCK_ERROR_ACTION_REPORT:
483
+ if (acct_failed) {
484
+ block_acct_failed(blk_get_stats(lu->qdev.conf.blk), &r->acct);
485
+ }
486
+ if (!req_has_sense && status == CHECK_CONDITION) {
487
+ scsi_req_build_sense(&r->req, sense);
488
+ }
489
+ scsi_req_complete(&r->req, status);
490
+ return true;
491
+
492
+ case BLOCK_ERROR_ACTION_IGNORE:
493
+ return false;
494
+
495
+ case BLOCK_ERROR_ACTION_STOP:
496
+ scsi_req_retry(&r->req);
497
+ return true;
498
+
499
+ default:
500
+ g_assert_not_reached();
501
+ }
502
+}
503
+
504
+static bool ufs_scsi_req_check_error(UfsSCSIReq *r, int ret, bool acct_failed)
505
+{
506
+ if (r->req.io_canceled) {
507
+ scsi_req_cancel_complete(&r->req);
508
+ return true;
509
+ }
510
+
511
+ if (ret < 0) {
512
+ return scsi_handle_rw_error(r, ret, acct_failed);
513
+ }
514
+
515
+ return false;
516
+}
517
+
518
+static void scsi_aio_complete(void *opaque, int ret)
519
+{
520
+ UfsSCSIReq *r = (UfsSCSIReq *)opaque;
521
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
522
+
523
+ assert(r->req.aiocb != NULL);
524
+ r->req.aiocb = NULL;
525
+ aio_context_acquire(blk_get_aio_context(lu->qdev.conf.blk));
526
+ if (ufs_scsi_req_check_error(r, ret, true)) {
527
+ goto done;
528
+ }
529
+
530
+ block_acct_done(blk_get_stats(lu->qdev.conf.blk), &r->acct);
531
+ scsi_req_complete(&r->req, GOOD);
532
+
533
+done:
534
+ aio_context_release(blk_get_aio_context(lu->qdev.conf.blk));
535
+ scsi_req_unref(&r->req);
536
+}
537
+
538
+static int32_t ufs_scsi_emulate_command(SCSIRequest *req, uint8_t *buf)
539
+{
540
+ UfsSCSIReq *r = DO_UPCAST(UfsSCSIReq, req, req);
541
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, req->dev);
542
+ uint32_t last_block = 0;
543
+ uint8_t *outbuf;
544
+ int buflen;
545
+
546
+ switch (req->cmd.buf[0]) {
547
+ case INQUIRY:
548
+ case MODE_SENSE_10:
549
+ case START_STOP:
550
+ case REQUEST_SENSE:
551
+ break;
552
+
553
+ default:
554
+ if (!blk_is_available(lu->qdev.conf.blk)) {
555
+ scsi_check_condition(r, SENSE_CODE(NO_MEDIUM));
556
+ return 0;
557
+ }
558
+ break;
559
+ }
560
+
561
+ /*
562
+ * FIXME: we shouldn't return anything bigger than 4k, but the code
563
+ * requires the buffer to be as big as req->cmd.xfer in several
564
+ * places. So, do not allow CDBs with a very large ALLOCATION
565
+ * LENGTH. The real fix would be to modify scsi_read_data and
566
+ * dma_buf_read, so that they return data beyond the buflen
567
+ * as all zeros.
568
+ */
569
+ if (req->cmd.xfer > 65536) {
570
+ goto illegal_request;
571
+ }
572
+ r->buflen = MAX(4096, req->cmd.xfer);
573
+
574
+ if (!r->iov.iov_base) {
575
+ r->iov.iov_base = blk_blockalign(lu->qdev.conf.blk, r->buflen);
576
+ }
577
+
578
+ outbuf = r->iov.iov_base;
579
+ memset(outbuf, 0, r->buflen);
580
+ switch (req->cmd.buf[0]) {
581
+ case TEST_UNIT_READY:
582
+ assert(blk_is_available(lu->qdev.conf.blk));
583
+ break;
584
+ case INQUIRY:
585
+ buflen = ufs_scsi_emulate_inquiry(req, outbuf, r->buflen);
586
+ if (buflen < 0) {
587
+ goto illegal_request;
588
+ }
589
+ break;
590
+ case MODE_SENSE_10:
591
+ buflen = ufs_scsi_emulate_mode_sense(r, outbuf);
592
+ if (buflen < 0) {
593
+ goto illegal_request;
594
+ }
595
+ break;
596
+ case READ_CAPACITY_10:
597
+ /* The normal LEN field for this command is zero. */
598
+ memset(outbuf, 0, 8);
599
+ if (lu->qdev.max_lba > 0) {
600
+ last_block = lu->qdev.max_lba - 1;
601
+ };
602
+ outbuf[0] = (last_block >> 24) & 0xff;
603
+ outbuf[1] = (last_block >> 16) & 0xff;
604
+ outbuf[2] = (last_block >> 8) & 0xff;
605
+ outbuf[3] = last_block & 0xff;
606
+ outbuf[4] = (lu->qdev.blocksize >> 24) & 0xff;
607
+ outbuf[5] = (lu->qdev.blocksize >> 16) & 0xff;
608
+ outbuf[6] = (lu->qdev.blocksize >> 8) & 0xff;
609
+ outbuf[7] = lu->qdev.blocksize & 0xff;
610
+ break;
611
+ case REQUEST_SENSE:
612
+ /* Just return "NO SENSE". */
613
+ buflen = scsi_convert_sense(NULL, 0, outbuf, r->buflen,
614
+ (req->cmd.buf[1] & 1) == 0);
615
+ if (buflen < 0) {
616
+ goto illegal_request;
617
+ }
618
+ break;
619
+ case SYNCHRONIZE_CACHE:
620
+ /* The request is used as the AIO opaque value, so add a ref. */
621
+ scsi_req_ref(&r->req);
622
+ block_acct_start(blk_get_stats(lu->qdev.conf.blk), &r->acct, 0,
623
+ BLOCK_ACCT_FLUSH);
624
+ r->req.aiocb = blk_aio_flush(lu->qdev.conf.blk, scsi_aio_complete, r);
625
+ return 0;
626
+ case VERIFY_10:
627
+ trace_ufs_scsi_emulate_command_VERIFY((req->cmd.buf[1] >> 1) & 3);
628
+ if (req->cmd.buf[1] & 6) {
629
+ goto illegal_request;
630
+ }
631
+ break;
632
+ case SERVICE_ACTION_IN_16:
633
+ /* Service Action In subcommands. */
634
+ if ((req->cmd.buf[1] & 31) == SAI_READ_CAPACITY_16) {
635
+ trace_ufs_scsi_emulate_command_SAI_16();
636
+ memset(outbuf, 0, req->cmd.xfer);
637
+
638
+ if (lu->qdev.max_lba > 0) {
639
+ last_block = lu->qdev.max_lba - 1;
640
+ };
641
+ outbuf[0] = 0;
642
+ outbuf[1] = 0;
643
+ outbuf[2] = 0;
644
+ outbuf[3] = 0;
645
+ outbuf[4] = (last_block >> 24) & 0xff;
646
+ outbuf[5] = (last_block >> 16) & 0xff;
647
+ outbuf[6] = (last_block >> 8) & 0xff;
648
+ outbuf[7] = last_block & 0xff;
649
+ outbuf[8] = (lu->qdev.blocksize >> 24) & 0xff;
650
+ outbuf[9] = (lu->qdev.blocksize >> 16) & 0xff;
651
+ outbuf[10] = (lu->qdev.blocksize >> 8) & 0xff;
652
+ outbuf[11] = lu->qdev.blocksize & 0xff;
653
+ outbuf[12] = 0;
654
+ outbuf[13] = get_physical_block_exp(&lu->qdev.conf);
655
+
656
+ if (lu->unit_desc.provisioning_type == 2 ||
657
+ lu->unit_desc.provisioning_type == 3) {
658
+ outbuf[14] = 0x80;
659
+ }
660
+ /* Protection, exponent and lowest lba field left blank. */
661
+ break;
662
+ }
663
+ trace_ufs_scsi_emulate_command_SAI_unsupported();
664
+ goto illegal_request;
665
+ case MODE_SELECT_10:
666
+ trace_ufs_scsi_emulate_command_MODE_SELECT_10(r->req.cmd.xfer);
667
+ break;
668
+ case START_STOP:
669
+ /*
670
+ * TODO: START_STOP is not yet implemented. It always returns success.
671
+ * Revisit it when ufs power management is implemented.
672
+ */
673
+ trace_ufs_scsi_emulate_command_START_STOP();
674
+ break;
675
+ case FORMAT_UNIT:
676
+ trace_ufs_scsi_emulate_command_FORMAT_UNIT();
677
+ break;
678
+ case SEND_DIAGNOSTIC:
679
+ trace_ufs_scsi_emulate_command_SEND_DIAGNOSTIC();
680
+ break;
681
+ default:
682
+ trace_ufs_scsi_emulate_command_UNKNOWN(buf[0],
683
+ scsi_command_name(buf[0]));
684
+ scsi_check_condition(r, SENSE_CODE(INVALID_OPCODE));
685
+ return 0;
686
+ }
687
+ assert(!r->req.aiocb);
688
+ r->iov.iov_len = MIN(r->buflen, req->cmd.xfer);
689
+ if (r->iov.iov_len == 0) {
690
+ scsi_req_complete(&r->req, GOOD);
691
+ }
692
+ if (r->req.cmd.mode == SCSI_XFER_TO_DEV) {
693
+ assert(r->iov.iov_len == req->cmd.xfer);
694
+ return -r->iov.iov_len;
695
+ } else {
696
+ return r->iov.iov_len;
697
+ }
698
+
699
+illegal_request:
700
+ if (r->req.status == -1) {
701
+ scsi_check_condition(r, SENSE_CODE(INVALID_FIELD));
702
+ }
703
+ return 0;
704
+}
705
+
706
+static void ufs_scsi_emulate_read_data(SCSIRequest *req)
707
+{
708
+ UfsSCSIReq *r = DO_UPCAST(UfsSCSIReq, req, req);
709
+ int buflen = r->iov.iov_len;
710
+
711
+ if (buflen) {
712
+ trace_ufs_scsi_emulate_read_data(buflen);
713
+ r->iov.iov_len = 0;
714
+ r->started = true;
715
+ scsi_req_data(&r->req, buflen);
716
+ return;
717
+ }
718
+
719
+ /* This also clears the sense buffer for REQUEST SENSE. */
720
+ scsi_req_complete(&r->req, GOOD);
721
+}
722
+
723
+static int ufs_scsi_check_mode_select(UfsLu *lu, int page, uint8_t *inbuf,
724
+ int inlen)
725
+{
726
+ uint8_t mode_current[SCSI_MAX_MODE_LEN];
727
+ uint8_t mode_changeable[SCSI_MAX_MODE_LEN];
728
+ uint8_t *p;
729
+ int len, expected_len, changeable_len, i;
730
+
731
+ /*
732
+ * The input buffer does not include the page header, so it is
733
+ * off by 2 bytes.
734
+ */
735
+ expected_len = inlen + 2;
736
+ if (expected_len > SCSI_MAX_MODE_LEN) {
737
+ return -1;
738
+ }
739
+
740
+ /* MODE_PAGE_ALLS is only valid for MODE SENSE commands */
741
+ if (page == MODE_PAGE_ALLS) {
742
+ return -1;
743
+ }
744
+
745
+ p = mode_current;
746
+ memset(mode_current, 0, inlen + 2);
747
+ len = mode_sense_page(lu, page, &p, 0);
748
+ if (len < 0 || len != expected_len) {
749
+ return -1;
750
+ }
751
+
752
+ p = mode_changeable;
753
+ memset(mode_changeable, 0, inlen + 2);
754
+ changeable_len = mode_sense_page(lu, page, &p, 1);
755
+ assert(changeable_len == len);
756
+
757
+ /*
758
+ * Check that unchangeable bits are the same as what MODE SENSE
759
+ * would return.
760
+ */
761
+ for (i = 2; i < len; i++) {
762
+ if (((mode_current[i] ^ inbuf[i - 2]) & ~mode_changeable[i]) != 0) {
763
+ return -1;
764
+ }
765
+ }
766
+ return 0;
767
+}
768
+
769
+static void ufs_scsi_apply_mode_select(UfsLu *lu, int page, uint8_t *p)
770
+{
771
+ switch (page) {
772
+ case MODE_PAGE_CACHING:
773
+ blk_set_enable_write_cache(lu->qdev.conf.blk, (p[0] & 4) != 0);
774
+ break;
775
+
776
+ default:
777
+ break;
778
+ }
779
+}
780
+
781
+static int mode_select_pages(UfsSCSIReq *r, uint8_t *p, int len, bool change)
782
+{
783
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
784
+
785
+ while (len > 0) {
786
+ int page, page_len;
787
+
788
+ page = p[0] & 0x3f;
789
+ if (p[0] & 0x40) {
790
+ goto invalid_param;
791
+ } else {
792
+ if (len < 2) {
793
+ goto invalid_param_len;
794
+ }
795
+ page_len = p[1];
796
+ p += 2;
797
+ len -= 2;
798
+ }
799
+
800
+ if (page_len > len) {
801
+ goto invalid_param_len;
802
+ }
803
+
804
+ if (!change) {
805
+ if (ufs_scsi_check_mode_select(lu, page, p, page_len) < 0) {
806
+ goto invalid_param;
807
+ }
808
+ } else {
809
+ ufs_scsi_apply_mode_select(lu, page, p);
810
+ }
811
+
812
+ p += page_len;
813
+ len -= page_len;
814
+ }
815
+ return 0;
816
+
817
+invalid_param:
818
+ scsi_check_condition(r, SENSE_CODE(INVALID_PARAM));
819
+ return -1;
820
+
821
+invalid_param_len:
822
+ scsi_check_condition(r, SENSE_CODE(INVALID_PARAM_LEN));
823
+ return -1;
824
+}
825
+
826
+static void ufs_scsi_emulate_mode_select(UfsSCSIReq *r, uint8_t *inbuf)
827
+{
828
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
829
+ uint8_t *p = inbuf;
830
+ int len = r->req.cmd.xfer;
831
+ int hdr_len = 8;
832
+ int bd_len;
833
+ int pass;
834
+
835
+ /* We only support PF=1, SP=0. */
836
+ if ((r->req.cmd.buf[1] & 0x11) != 0x10) {
837
+ goto invalid_field;
838
+ }
839
+
840
+ if (len < hdr_len) {
841
+ goto invalid_param_len;
842
+ }
843
+
844
+ bd_len = lduw_be_p(&p[6]);
845
+ if (bd_len != 0) {
846
+ goto invalid_param;
847
+ }
848
+
849
+ len -= hdr_len;
850
+ p += hdr_len;
851
+
852
+ /* Ensure no change is made if there is an error! */
853
+ for (pass = 0; pass < 2; pass++) {
854
+ if (mode_select_pages(r, p, len, pass == 1) < 0) {
855
+ assert(pass == 0);
856
+ return;
857
+ }
858
+ }
859
+
860
+ if (!blk_enable_write_cache(lu->qdev.conf.blk)) {
861
+ /* The request is used as the AIO opaque value, so add a ref. */
862
+ scsi_req_ref(&r->req);
863
+ block_acct_start(blk_get_stats(lu->qdev.conf.blk), &r->acct, 0,
864
+ BLOCK_ACCT_FLUSH);
865
+ r->req.aiocb = blk_aio_flush(lu->qdev.conf.blk, scsi_aio_complete, r);
866
+ return;
867
+ }
868
+
869
+ scsi_req_complete(&r->req, GOOD);
870
+ return;
871
+
872
+invalid_param:
873
+ scsi_check_condition(r, SENSE_CODE(INVALID_PARAM));
874
+ return;
875
+
876
+invalid_param_len:
877
+ scsi_check_condition(r, SENSE_CODE(INVALID_PARAM_LEN));
878
+ return;
879
+
880
+invalid_field:
881
+ scsi_check_condition(r, SENSE_CODE(INVALID_FIELD));
882
+}
883
+
884
+/* block_num and nb_blocks expected to be in qdev blocksize */
885
+static inline bool check_lba_range(UfsLu *lu, uint64_t block_num,
886
+ uint32_t nb_blocks)
887
+{
888
+ /*
889
+ * The first line tests that no overflow happens when computing the last
890
+ * block. The second line tests that the last accessed block is in
891
+ * range.
892
+ *
893
+ * Careful, the computations should not underflow for nb_blocks == 0,
894
+ * and a 0-block read to the first LBA beyond the end of device is
895
+ * valid.
896
+ */
897
+ return (block_num <= block_num + nb_blocks &&
898
+ block_num + nb_blocks <= lu->qdev.max_lba + 1);
899
+}
900
+
901
+static void ufs_scsi_emulate_write_data(SCSIRequest *req)
902
+{
903
+ UfsSCSIReq *r = DO_UPCAST(UfsSCSIReq, req, req);
904
+
905
+ if (r->iov.iov_len) {
906
+ int buflen = r->iov.iov_len;
907
+ trace_ufs_scsi_emulate_write_data(buflen);
908
+ r->iov.iov_len = 0;
909
+ scsi_req_data(&r->req, buflen);
910
+ return;
911
+ }
912
+
913
+ switch (req->cmd.buf[0]) {
914
+ case MODE_SELECT_10:
915
+ /* This also clears the sense buffer for REQUEST SENSE. */
916
+ ufs_scsi_emulate_mode_select(r, r->iov.iov_base);
917
+ break;
918
+ default:
919
+ abort();
920
+ }
921
+}
922
+
923
+/* Return a pointer to the data buffer. */
924
+static uint8_t *ufs_scsi_get_buf(SCSIRequest *req)
925
+{
926
+ UfsSCSIReq *r = DO_UPCAST(UfsSCSIReq, req, req);
927
+
928
+ return (uint8_t *)r->iov.iov_base;
929
+}
930
+
931
+static int32_t ufs_scsi_dma_command(SCSIRequest *req, uint8_t *buf)
932
+{
933
+ UfsSCSIReq *r = DO_UPCAST(UfsSCSIReq, req, req);
934
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, req->dev);
935
+ uint32_t len;
936
+ uint8_t command;
937
+
938
+ command = buf[0];
939
+
940
+ if (!blk_is_available(lu->qdev.conf.blk)) {
941
+ scsi_check_condition(r, SENSE_CODE(NO_MEDIUM));
942
+ return 0;
943
+ }
944
+
945
+ len = scsi_data_cdb_xfer(r->req.cmd.buf);
946
+ switch (command) {
947
+ case READ_6:
948
+ case READ_10:
949
+ trace_ufs_scsi_dma_command_READ(r->req.cmd.lba, len);
950
+ if (r->req.cmd.buf[1] & 0xe0) {
951
+ goto illegal_request;
952
+ }
953
+ if (!check_lba_range(lu, r->req.cmd.lba, len)) {
954
+ goto illegal_lba;
955
+ }
956
+ r->sector = r->req.cmd.lba * (lu->qdev.blocksize / BDRV_SECTOR_SIZE);
957
+ r->sector_count = len * (lu->qdev.blocksize / BDRV_SECTOR_SIZE);
958
+ break;
959
+ case WRITE_6:
960
+ case WRITE_10:
961
+ trace_ufs_scsi_dma_command_WRITE(r->req.cmd.lba, len);
962
+ if (!blk_is_writable(lu->qdev.conf.blk)) {
963
+ scsi_check_condition(r, SENSE_CODE(WRITE_PROTECTED));
964
+ return 0;
965
+ }
966
+ if (r->req.cmd.buf[1] & 0xe0) {
967
+ goto illegal_request;
968
+ }
969
+ if (!check_lba_range(lu, r->req.cmd.lba, len)) {
970
+ goto illegal_lba;
971
+ }
972
+ r->sector = r->req.cmd.lba * (lu->qdev.blocksize / BDRV_SECTOR_SIZE);
973
+ r->sector_count = len * (lu->qdev.blocksize / BDRV_SECTOR_SIZE);
974
+ break;
975
+ default:
976
+ abort();
977
+ illegal_request:
978
+ scsi_check_condition(r, SENSE_CODE(INVALID_FIELD));
979
+ return 0;
980
+ illegal_lba:
981
+ scsi_check_condition(r, SENSE_CODE(LBA_OUT_OF_RANGE));
982
+ return 0;
983
+ }
984
+ r->need_fua_emulation = ((r->req.cmd.buf[1] & 8) != 0);
985
+ if (r->sector_count == 0) {
986
+ scsi_req_complete(&r->req, GOOD);
987
+ }
988
+ assert(r->iov.iov_len == 0);
989
+ if (r->req.cmd.mode == SCSI_XFER_TO_DEV) {
990
+ return -r->sector_count * BDRV_SECTOR_SIZE;
991
+ } else {
992
+ return r->sector_count * BDRV_SECTOR_SIZE;
993
+ }
994
+}
995
+
996
+static void scsi_write_do_fua(UfsSCSIReq *r)
997
+{
998
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
999
+
1000
+ assert(r->req.aiocb == NULL);
1001
+ assert(!r->req.io_canceled);
1002
+
1003
+ if (r->need_fua_emulation) {
1004
+ block_acct_start(blk_get_stats(lu->qdev.conf.blk), &r->acct, 0,
1005
+ BLOCK_ACCT_FLUSH);
1006
+ r->req.aiocb = blk_aio_flush(lu->qdev.conf.blk, scsi_aio_complete, r);
1007
+ return;
1008
+ }
1009
+
1010
+ scsi_req_complete(&r->req, GOOD);
1011
+ scsi_req_unref(&r->req);
1012
+}
1013
+
1014
+static void scsi_dma_complete_noio(UfsSCSIReq *r, int ret)
1015
+{
1016
+ assert(r->req.aiocb == NULL);
1017
+ if (ufs_scsi_req_check_error(r, ret, false)) {
1018
+ goto done;
1019
+ }
1020
+
1021
+ r->sector += r->sector_count;
1022
+ r->sector_count = 0;
1023
+ if (r->req.cmd.mode == SCSI_XFER_TO_DEV) {
1024
+ scsi_write_do_fua(r);
1025
+ return;
1026
+ } else {
1027
+ scsi_req_complete(&r->req, GOOD);
1028
+ }
1029
+
1030
+done:
1031
+ scsi_req_unref(&r->req);
1032
+}
1033
+
1034
+static void scsi_dma_complete(void *opaque, int ret)
1035
+{
1036
+ UfsSCSIReq *r = (UfsSCSIReq *)opaque;
1037
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
1038
+
1039
+ assert(r->req.aiocb != NULL);
1040
+ r->req.aiocb = NULL;
1041
+
1042
+ aio_context_acquire(blk_get_aio_context(lu->qdev.conf.blk));
1043
+ if (ret < 0) {
1044
+ block_acct_failed(blk_get_stats(lu->qdev.conf.blk), &r->acct);
1045
+ } else {
1046
+ block_acct_done(blk_get_stats(lu->qdev.conf.blk), &r->acct);
1047
+ }
1048
+ scsi_dma_complete_noio(r, ret);
1049
+ aio_context_release(blk_get_aio_context(lu->qdev.conf.blk));
1050
+}
1051
+
1052
+static BlockAIOCB *scsi_dma_readv(int64_t offset, QEMUIOVector *iov,
1053
+ BlockCompletionFunc *cb, void *cb_opaque,
1054
+ void *opaque)
1055
+{
1056
+ UfsSCSIReq *r = opaque;
1057
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
1058
+ return blk_aio_preadv(lu->qdev.conf.blk, offset, iov, 0, cb, cb_opaque);
1059
+}
1060
+
1061
+static void scsi_init_iovec(UfsSCSIReq *r, size_t size)
1062
+{
1063
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
1064
+
1065
+ if (!r->iov.iov_base) {
1066
+ r->buflen = size;
1067
+ r->iov.iov_base = blk_blockalign(lu->qdev.conf.blk, r->buflen);
1068
+ }
1069
+ r->iov.iov_len = MIN(r->sector_count * BDRV_SECTOR_SIZE, r->buflen);
1070
+ qemu_iovec_init_external(&r->qiov, &r->iov, 1);
1071
+}
1072
+
1073
+static void scsi_read_complete_noio(UfsSCSIReq *r, int ret)
1074
+{
1075
+ uint32_t n;
1076
+
1077
+ assert(r->req.aiocb == NULL);
1078
+ if (ufs_scsi_req_check_error(r, ret, false)) {
1079
+ goto done;
1080
+ }
1081
+
1082
+ n = r->qiov.size / BDRV_SECTOR_SIZE;
1083
+ r->sector += n;
1084
+ r->sector_count -= n;
1085
+ scsi_req_data(&r->req, r->qiov.size);
1086
+
1087
+done:
1088
+ scsi_req_unref(&r->req);
1089
+}
1090
+
1091
+static void scsi_read_complete(void *opaque, int ret)
1092
+{
1093
+ UfsSCSIReq *r = (UfsSCSIReq *)opaque;
1094
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
1095
+
1096
+ assert(r->req.aiocb != NULL);
1097
+ r->req.aiocb = NULL;
1098
+ trace_ufs_scsi_read_data_count(r->sector_count);
1099
+ aio_context_acquire(blk_get_aio_context(lu->qdev.conf.blk));
1100
+ if (ret < 0) {
1101
+ block_acct_failed(blk_get_stats(lu->qdev.conf.blk), &r->acct);
1102
+ } else {
1103
+ block_acct_done(blk_get_stats(lu->qdev.conf.blk), &r->acct);
1104
+ trace_ufs_scsi_read_complete(r->req.tag, r->qiov.size);
1105
+ }
1106
+ scsi_read_complete_noio(r, ret);
1107
+ aio_context_release(blk_get_aio_context(lu->qdev.conf.blk));
1108
+}
1109
+
1110
+/* Actually issue a read to the block device. */
1111
+static void scsi_do_read(UfsSCSIReq *r, int ret)
1112
+{
1113
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
1114
+
1115
+ assert(r->req.aiocb == NULL);
1116
+ if (ufs_scsi_req_check_error(r, ret, false)) {
1117
+ goto done;
1118
+ }
1119
+
1120
+ /* The request is used as the AIO opaque value, so add a ref. */
1121
+ scsi_req_ref(&r->req);
1122
+
1123
+ if (r->req.sg) {
1124
+ dma_acct_start(lu->qdev.conf.blk, &r->acct, r->req.sg, BLOCK_ACCT_READ);
1125
+ r->req.residual -= r->req.sg->size;
1126
+ r->req.aiocb = dma_blk_io(
1127
+ blk_get_aio_context(lu->qdev.conf.blk), r->req.sg,
1128
+ r->sector << BDRV_SECTOR_BITS, BDRV_SECTOR_SIZE, scsi_dma_readv, r,
1129
+ scsi_dma_complete, r, DMA_DIRECTION_FROM_DEVICE);
1130
+ } else {
1131
+ scsi_init_iovec(r, SCSI_DMA_BUF_SIZE);
1132
+ block_acct_start(blk_get_stats(lu->qdev.conf.blk), &r->acct,
1133
+ r->qiov.size, BLOCK_ACCT_READ);
1134
+ r->req.aiocb = scsi_dma_readv(r->sector << BDRV_SECTOR_BITS, &r->qiov,
1135
+ scsi_read_complete, r, r);
1136
+ }
1137
+
1138
+done:
1139
+ scsi_req_unref(&r->req);
1140
+}
1141
+
1142
+static void scsi_do_read_cb(void *opaque, int ret)
1143
+{
1144
+ UfsSCSIReq *r = (UfsSCSIReq *)opaque;
1145
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
1146
+
1147
+ assert(r->req.aiocb != NULL);
1148
+ r->req.aiocb = NULL;
1149
+
1150
+ aio_context_acquire(blk_get_aio_context(lu->qdev.conf.blk));
1151
+ if (ret < 0) {
1152
+ block_acct_failed(blk_get_stats(lu->qdev.conf.blk), &r->acct);
1153
+ } else {
1154
+ block_acct_done(blk_get_stats(lu->qdev.conf.blk), &r->acct);
1155
+ }
1156
+ scsi_do_read(opaque, ret);
1157
+ aio_context_release(blk_get_aio_context(lu->qdev.conf.blk));
1158
+}
1159
+
1160
+/* Read more data from scsi device into buffer. */
1161
+static void scsi_read_data(SCSIRequest *req)
1162
+{
1163
+ UfsSCSIReq *r = DO_UPCAST(UfsSCSIReq, req, req);
1164
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
1165
+ bool first;
1166
+
1167
+ trace_ufs_scsi_read_data_count(r->sector_count);
1168
+ if (r->sector_count == 0) {
1169
+ /* This also clears the sense buffer for REQUEST SENSE. */
1170
+ scsi_req_complete(&r->req, GOOD);
1171
+ return;
1172
+ }
1173
+
1174
+ /* No data transfer may already be in progress */
1175
+ assert(r->req.aiocb == NULL);
1176
+
1177
+ /* The request is used as the AIO opaque value, so add a ref. */
1178
+ scsi_req_ref(&r->req);
1179
+ if (r->req.cmd.mode == SCSI_XFER_TO_DEV) {
1180
+ trace_ufs_scsi_read_data_invalid();
1181
+ scsi_read_complete_noio(r, -EINVAL);
1182
+ return;
1183
+ }
1184
+
1185
+ if (!blk_is_available(req->dev->conf.blk)) {
1186
+ scsi_read_complete_noio(r, -ENOMEDIUM);
1187
+ return;
1188
+ }
1189
+
1190
+ first = !r->started;
1191
+ r->started = true;
1192
+ if (first && r->need_fua_emulation) {
1193
+ block_acct_start(blk_get_stats(lu->qdev.conf.blk), &r->acct, 0,
1194
+ BLOCK_ACCT_FLUSH);
1195
+ r->req.aiocb = blk_aio_flush(lu->qdev.conf.blk, scsi_do_read_cb, r);
1196
+ } else {
1197
+ scsi_do_read(r, 0);
1198
+ }
1199
+}
1200
+
1201
+static void scsi_write_complete_noio(UfsSCSIReq *r, int ret)
1202
+{
1203
+ uint32_t n;
1204
+
1205
+ assert(r->req.aiocb == NULL);
1206
+ if (ufs_scsi_req_check_error(r, ret, false)) {
1207
+ goto done;
1208
+ }
1209
+
1210
+ n = r->qiov.size / BDRV_SECTOR_SIZE;
1211
+ r->sector += n;
1212
+ r->sector_count -= n;
1213
+ if (r->sector_count == 0) {
1214
+ scsi_write_do_fua(r);
1215
+ return;
1216
+ } else {
1217
+ scsi_init_iovec(r, SCSI_DMA_BUF_SIZE);
1218
+ trace_ufs_scsi_write_complete_noio(r->req.tag, r->qiov.size);
1219
+ scsi_req_data(&r->req, r->qiov.size);
1220
+ }
1221
+
1222
+done:
1223
+ scsi_req_unref(&r->req);
1224
+}
1225
+
1226
+static void scsi_write_complete(void *opaque, int ret)
1227
+{
1228
+ UfsSCSIReq *r = (UfsSCSIReq *)opaque;
1229
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
1230
+
1231
+ assert(r->req.aiocb != NULL);
1232
+ r->req.aiocb = NULL;
1233
+
1234
+ aio_context_acquire(blk_get_aio_context(lu->qdev.conf.blk));
1235
+ if (ret < 0) {
1236
+ block_acct_failed(blk_get_stats(lu->qdev.conf.blk), &r->acct);
1237
+ } else {
1238
+ block_acct_done(blk_get_stats(lu->qdev.conf.blk), &r->acct);
1239
+ }
1240
+ scsi_write_complete_noio(r, ret);
1241
+ aio_context_release(blk_get_aio_context(lu->qdev.conf.blk));
1242
+}
1243
+
1244
+static BlockAIOCB *scsi_dma_writev(int64_t offset, QEMUIOVector *iov,
1245
+ BlockCompletionFunc *cb, void *cb_opaque,
1246
+ void *opaque)
1247
+{
1248
+ UfsSCSIReq *r = opaque;
1249
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
1250
+ return blk_aio_pwritev(lu->qdev.conf.blk, offset, iov, 0, cb, cb_opaque);
1251
+}
1252
+
1253
+static void scsi_write_data(SCSIRequest *req)
1254
+{
1255
+ UfsSCSIReq *r = DO_UPCAST(UfsSCSIReq, req, req);
1256
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, r->req.dev);
1257
+
1258
+ /* No data transfer may already be in progress */
1259
+ assert(r->req.aiocb == NULL);
1260
+
1261
+ /* The request is used as the AIO opaque value, so add a ref. */
1262
+ scsi_req_ref(&r->req);
1263
+ if (r->req.cmd.mode != SCSI_XFER_TO_DEV) {
1264
+ trace_ufs_scsi_write_data_invalid();
1265
+ scsi_write_complete_noio(r, -EINVAL);
1266
+ return;
1267
+ }
1268
+
1269
+ if (!r->req.sg && !r->qiov.size) {
1270
+ /* Called for the first time. Ask the driver to send us more data. */
1271
+ r->started = true;
1272
+ scsi_write_complete_noio(r, 0);
1273
+ return;
1274
+ }
1275
+ if (!blk_is_available(req->dev->conf.blk)) {
1276
+ scsi_write_complete_noio(r, -ENOMEDIUM);
1277
+ return;
1278
+ }
1279
+
1280
+ if (r->req.sg) {
1281
+ dma_acct_start(lu->qdev.conf.blk, &r->acct, r->req.sg,
1282
+ BLOCK_ACCT_WRITE);
1283
+ r->req.residual -= r->req.sg->size;
1284
+ r->req.aiocb = dma_blk_io(
1285
+ blk_get_aio_context(lu->qdev.conf.blk), r->req.sg,
1286
+ r->sector << BDRV_SECTOR_BITS, BDRV_SECTOR_SIZE, scsi_dma_writev, r,
1287
+ scsi_dma_complete, r, DMA_DIRECTION_TO_DEVICE);
1288
+ } else {
1289
+ block_acct_start(blk_get_stats(lu->qdev.conf.blk), &r->acct,
1290
+ r->qiov.size, BLOCK_ACCT_WRITE);
1291
+ r->req.aiocb = scsi_dma_writev(r->sector << BDRV_SECTOR_BITS, &r->qiov,
1292
+ scsi_write_complete, r, r);
1293
+ }
1294
+}
1295
+
1296
+static const SCSIReqOps ufs_scsi_emulate_reqops = {
1297
+ .size = sizeof(UfsSCSIReq),
1298
+ .free_req = ufs_scsi_free_request,
1299
+ .send_command = ufs_scsi_emulate_command,
1300
+ .read_data = ufs_scsi_emulate_read_data,
1301
+ .write_data = ufs_scsi_emulate_write_data,
1302
+ .get_buf = ufs_scsi_get_buf,
80
+};
1303
+};
81
+
1304
+
82
+bool vhost_user_server_start(VuServer *server,
1305
+static const SCSIReqOps ufs_scsi_dma_reqops = {
83
+ SocketAddress *unix_socket,
1306
+ .size = sizeof(UfsSCSIReq),
84
+ AioContext *ctx,
1307
+ .free_req = ufs_scsi_free_request,
85
+ uint16_t max_queues,
1308
+ .send_command = ufs_scsi_dma_command,
86
+ DevicePanicNotifierFn *device_panic_notifier,
1309
+ .read_data = scsi_read_data,
87
+ const VuDevIface *vu_iface,
1310
+ .write_data = scsi_write_data,
88
+ Error **errp);
1311
+ .get_buf = ufs_scsi_get_buf,
89
+
1312
+};
90
+void vhost_user_server_stop(VuServer *server);
1313
+
91
+
1314
+/*
92
+void vhost_user_server_set_aio_context(VuServer *server, AioContext *ctx);
1315
+ * Following commands are not yet supported
93
+
1316
+ * PRE_FETCH(10),
94
+#endif /* VHOST_USER_SERVER_H */
1317
+ * UNMAP,
95
diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c
1318
+ * WRITE_BUFFER, READ_BUFFER,
96
new file mode 100644
1319
+ * SECURITY_PROTOCOL_IN, SECURITY_PROTOCOL_OUT
97
index XXXXXXX..XXXXXXX
1320
+ */
98
--- /dev/null
1321
+static const SCSIReqOps *const ufs_scsi_reqops_dispatch[256] = {
99
+++ b/util/vhost-user-server.c
1322
+ [TEST_UNIT_READY] = &ufs_scsi_emulate_reqops,
1323
+ [INQUIRY] = &ufs_scsi_emulate_reqops,
1324
+ [MODE_SENSE_10] = &ufs_scsi_emulate_reqops,
1325
+ [START_STOP] = &ufs_scsi_emulate_reqops,
1326
+ [READ_CAPACITY_10] = &ufs_scsi_emulate_reqops,
1327
+ [REQUEST_SENSE] = &ufs_scsi_emulate_reqops,
1328
+ [SYNCHRONIZE_CACHE] = &ufs_scsi_emulate_reqops,
1329
+ [MODE_SELECT_10] = &ufs_scsi_emulate_reqops,
1330
+ [VERIFY_10] = &ufs_scsi_emulate_reqops,
1331
+ [FORMAT_UNIT] = &ufs_scsi_emulate_reqops,
1332
+ [SERVICE_ACTION_IN_16] = &ufs_scsi_emulate_reqops,
1333
+ [SEND_DIAGNOSTIC] = &ufs_scsi_emulate_reqops,
1334
+
1335
+ [READ_6] = &ufs_scsi_dma_reqops,
1336
+ [READ_10] = &ufs_scsi_dma_reqops,
1337
+ [WRITE_6] = &ufs_scsi_dma_reqops,
1338
+ [WRITE_10] = &ufs_scsi_dma_reqops,
1339
+};
1340
+
1341
+static SCSIRequest *scsi_new_request(SCSIDevice *dev, uint32_t tag,
1342
+ uint32_t lun, uint8_t *buf,
1343
+ void *hba_private)
1344
+{
1345
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, dev);
1346
+ SCSIRequest *req;
1347
+ const SCSIReqOps *ops;
1348
+ uint8_t command;
1349
+
1350
+ command = buf[0];
1351
+ ops = ufs_scsi_reqops_dispatch[command];
1352
+ if (!ops) {
1353
+ ops = &ufs_scsi_emulate_reqops;
1354
+ }
1355
+ req = scsi_req_alloc(ops, &lu->qdev, tag, lun, hba_private);
1356
+
1357
+ return req;
1358
+}
1359
+
1360
+static Property ufs_lu_props[] = {
1361
+ DEFINE_PROP_DRIVE("drive", UfsLu, qdev.conf.blk),
1362
+ DEFINE_PROP_END_OF_LIST(),
1363
+};
1364
+
1365
+static bool ufs_lu_brdv_init(UfsLu *lu, Error **errp)
1366
+{
1367
+ SCSIDevice *dev = &lu->qdev;
1368
+ bool read_only;
1369
+
1370
+ if (!lu->qdev.conf.blk) {
1371
+ error_setg(errp, "drive property not set");
1372
+ return false;
1373
+ }
1374
+
1375
+ if (!blkconf_blocksizes(&lu->qdev.conf, errp)) {
1376
+ return false;
1377
+ }
1378
+
1379
+ if (blk_get_aio_context(lu->qdev.conf.blk) != qemu_get_aio_context() &&
1380
+ !lu->qdev.hba_supports_iothread) {
1381
+ error_setg(errp, "HBA does not support iothreads");
1382
+ return false;
1383
+ }
1384
+
1385
+ read_only = !blk_supports_write_perm(lu->qdev.conf.blk);
1386
+
1387
+ if (!blkconf_apply_backend_options(&dev->conf, read_only,
1388
+ dev->type == TYPE_DISK, errp)) {
1389
+ return false;
1390
+ }
1391
+
1392
+ if (blk_is_sg(lu->qdev.conf.blk)) {
1393
+ error_setg(errp, "unwanted /dev/sg*");
1394
+ return false;
1395
+ }
1396
+
1397
+ blk_iostatus_enable(lu->qdev.conf.blk);
1398
+ return true;
1399
+}
1400
+
1401
+static bool ufs_add_lu(UfsHc *u, UfsLu *lu, Error **errp)
1402
+{
1403
+ BlockBackend *blk = lu->qdev.conf.blk;
1404
+ int64_t brdv_len = blk_getlength(blk);
1405
+ uint64_t raw_dev_cap =
1406
+ be64_to_cpu(u->geometry_desc.total_raw_device_capacity);
1407
+
1408
+ if (u->device_desc.number_lu >= UFS_MAX_LUS) {
1409
+ error_setg(errp, "ufs host controller has too many logical units.");
1410
+ return false;
1411
+ }
1412
+
1413
+ if (u->lus[lu->lun] != NULL) {
1414
+ error_setg(errp, "ufs logical unit %d already exists.", lu->lun);
1415
+ return false;
1416
+ }
1417
+
1418
+ u->lus[lu->lun] = lu;
1419
+ u->device_desc.number_lu++;
1420
+ raw_dev_cap += (brdv_len >> UFS_GEOMETRY_CAPACITY_SHIFT);
1421
+ u->geometry_desc.total_raw_device_capacity = cpu_to_be64(raw_dev_cap);
1422
+ return true;
1423
+}
1424
+
1425
+static inline uint8_t ufs_log2(uint64_t input)
1426
+{
1427
+ int log = 0;
1428
+ while (input >>= 1) {
1429
+ log++;
1430
+ }
1431
+ return log;
1432
+}
1433
+
1434
+static void ufs_init_lu(UfsLu *lu)
1435
+{
1436
+ BlockBackend *blk = lu->qdev.conf.blk;
1437
+ int64_t brdv_len = blk_getlength(blk);
1438
+
1439
+ lu->lun = lu->qdev.lun;
1440
+ memset(&lu->unit_desc, 0, sizeof(lu->unit_desc));
1441
+ lu->unit_desc.length = sizeof(UnitDescriptor);
1442
+ lu->unit_desc.descriptor_idn = QUERY_DESC_IDN_UNIT;
1443
+ lu->unit_desc.lu_enable = 0x01;
1444
+ lu->unit_desc.logical_block_size = ufs_log2(lu->qdev.blocksize);
1445
+ lu->unit_desc.unit_index = lu->qdev.lun;
1446
+ lu->unit_desc.logical_block_count =
1447
+ cpu_to_be64(brdv_len / (1 << lu->unit_desc.logical_block_size));
1448
+}
1449
+
1450
+static bool ufs_lu_check_constraints(UfsLu *lu, Error **errp)
1451
+{
1452
+ if (!lu->qdev.conf.blk) {
1453
+ error_setg(errp, "drive property not set");
1454
+ return false;
1455
+ }
1456
+
1457
+ if (lu->qdev.channel != 0) {
1458
+ error_setg(errp, "ufs logical unit does not support channel");
1459
+ return false;
1460
+ }
1461
+
1462
+ if (lu->qdev.lun >= UFS_MAX_LUS) {
1463
+ error_setg(errp, "lun must be between 1 and %d", UFS_MAX_LUS - 1);
1464
+ return false;
1465
+ }
1466
+
1467
+ return true;
1468
+}
1469
+
1470
+static void ufs_lu_realize(SCSIDevice *dev, Error **errp)
1471
+{
1472
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, dev);
1473
+ BusState *s = qdev_get_parent_bus(&dev->qdev);
1474
+ UfsHc *u = UFS(s->parent);
1475
+ AioContext *ctx = NULL;
1476
+ uint64_t nb_sectors, nb_blocks;
1477
+
1478
+ if (!ufs_lu_check_constraints(lu, errp)) {
1479
+ return;
1480
+ }
1481
+
1482
+ if (lu->qdev.conf.blk) {
1483
+ ctx = blk_get_aio_context(lu->qdev.conf.blk);
1484
+ aio_context_acquire(ctx);
1485
+ if (!blkconf_blocksizes(&lu->qdev.conf, errp)) {
1486
+ goto out;
1487
+ }
1488
+ }
1489
+ lu->qdev.blocksize = UFS_BLOCK_SIZE;
1490
+ blk_get_geometry(lu->qdev.conf.blk, &nb_sectors);
1491
+ nb_blocks = nb_sectors / (lu->qdev.blocksize / BDRV_SECTOR_SIZE);
1492
+ if (nb_blocks > UINT32_MAX) {
1493
+ nb_blocks = UINT32_MAX;
1494
+ }
1495
+ lu->qdev.max_lba = nb_blocks;
1496
+ lu->qdev.type = TYPE_DISK;
1497
+
1498
+ ufs_init_lu(lu);
1499
+ if (!ufs_add_lu(u, lu, errp)) {
1500
+ goto out;
1501
+ }
1502
+
1503
+ ufs_lu_brdv_init(lu, errp);
1504
+out:
1505
+ if (ctx) {
1506
+ aio_context_release(ctx);
1507
+ }
1508
+}
1509
+
1510
+static void ufs_lu_unrealize(SCSIDevice *dev)
1511
+{
1512
+ UfsLu *lu = DO_UPCAST(UfsLu, qdev, dev);
1513
+
1514
+ blk_drain(lu->qdev.conf.blk);
1515
+}
1516
+
1517
+static void ufs_wlu_realize(DeviceState *qdev, Error **errp)
1518
+{
1519
+ UfsWLu *wlu = UFSWLU(qdev);
1520
+ SCSIDevice *dev = &wlu->qdev;
1521
+
1522
+ if (!is_wlun(dev->lun)) {
1523
+ error_setg(errp, "not well-known logical unit number");
1524
+ return;
1525
+ }
1526
+
1527
+ QTAILQ_INIT(&dev->requests);
1528
+}
1529
+
1530
+static void ufs_lu_class_init(ObjectClass *oc, void *data)
1531
+{
1532
+ DeviceClass *dc = DEVICE_CLASS(oc);
1533
+ SCSIDeviceClass *sc = SCSI_DEVICE_CLASS(oc);
1534
+
1535
+ sc->realize = ufs_lu_realize;
1536
+ sc->unrealize = ufs_lu_unrealize;
1537
+ sc->alloc_req = scsi_new_request;
1538
+ dc->bus_type = TYPE_UFS_BUS;
1539
+ device_class_set_props(dc, ufs_lu_props);
1540
+ dc->desc = "Virtual UFS logical unit";
1541
+}
1542
+
1543
+static void ufs_wlu_class_init(ObjectClass *oc, void *data)
1544
+{
1545
+ DeviceClass *dc = DEVICE_CLASS(oc);
1546
+ SCSIDeviceClass *sc = SCSI_DEVICE_CLASS(oc);
1547
+
1548
+ /*
1549
+ * The realize() function of TYPE_SCSI_DEVICE causes a segmentation fault
1550
+ * if a block drive does not exist. Define a new realize function for
1551
+ * well-known LUs that do not have a block drive.
1552
+ */
1553
+ dc->realize = ufs_wlu_realize;
1554
+ sc->alloc_req = scsi_new_request;
1555
+ dc->bus_type = TYPE_UFS_BUS;
1556
+ dc->desc = "Virtual UFS well-known logical unit";
1557
+}
1558
+
1559
+static const TypeInfo ufs_lu_info = {
1560
+ .name = TYPE_UFS_LU,
1561
+ .parent = TYPE_SCSI_DEVICE,
1562
+ .class_init = ufs_lu_class_init,
1563
+ .instance_size = sizeof(UfsLu),
1564
+};
1565
+
1566
+static const TypeInfo ufs_wlu_info = {
1567
+ .name = TYPE_UFS_WLU,
1568
+ .parent = TYPE_SCSI_DEVICE,
1569
+ .class_init = ufs_wlu_class_init,
1570
+ .instance_size = sizeof(UfsWLu),
1571
+};
1572
+
1573
+static void ufs_lu_register_types(void)
1574
+{
1575
+ type_register_static(&ufs_lu_info);
1576
+ type_register_static(&ufs_wlu_info);
1577
+}
1578
+
1579
+type_init(ufs_lu_register_types)
1580
diff --git a/hw/ufs/ufs.c b/hw/ufs/ufs.c
1581
index XXXXXXX..XXXXXXX 100644
1582
--- a/hw/ufs/ufs.c
1583
+++ b/hw/ufs/ufs.c
100
@@ -XXX,XX +XXX,XX @@
1584
@@ -XXX,XX +XXX,XX @@
101
+/*
1585
* SPDX-License-Identifier: GPL-2.0-or-later
102
+ * Sharing QEMU devices via vhost-user protocol
1586
*/
1587
1588
+/**
1589
+ * Reference Specs: https://www.jedec.org/, 3.1
103
+ *
1590
+ *
104
+ * Copyright (c) Coiby Xu <coiby.xu@gmail.com>.
1591
+ * Usage
105
+ * Copyright (c) 2020 Red Hat, Inc.
1592
+ * -----
106
+ *
1593
+ *
107
+ * This work is licensed under the terms of the GNU GPL, version 2 or
1594
+ * Add options:
108
+ * later. See the COPYING file in the top-level directory.
1595
+ * -drive file=<file>,if=none,id=<drive_id>
1596
+ * -device ufs,serial=<serial>,id=<bus_name>, \
1597
+ * nutrs=<N[optional]>,nutmrs=<N[optional]>
1598
+ * -device ufs-lu,drive=<drive_id>,bus=<bus_name>
109
+ */
1599
+ */
110
+#include "qemu/osdep.h"
1600
+
111
+#include "qemu/main-loop.h"
1601
#include "qemu/osdep.h"
112
+#include "vhost-user-server.h"
1602
#include "qapi/error.h"
113
+
1603
#include "migration/vmstate.h"
114
+static void vmsg_close_fds(VhostUserMsg *vmsg)
1604
@@ -XXX,XX +XXX,XX @@ static const MemoryRegionOps ufs_mmio_ops = {
115
+{
1605
},
116
+ int i;
1606
};
117
+ for (i = 0; i < vmsg->fd_num; i++) {
1607
118
+ close(vmsg->fds[i]);
1608
+static QEMUSGList *ufs_get_sg_list(SCSIRequest *scsi_req)
119
+ }
1609
+{
120
+}
1610
+ UfsRequest *req = scsi_req->hba_private;
121
+
1611
+ return req->sg;
122
+static void vmsg_unblock_fds(VhostUserMsg *vmsg)
1612
+}
123
+{
1613
+
124
+ int i;
1614
+static void ufs_build_upiu_sense_data(UfsRequest *req, SCSIRequest *scsi_req)
125
+ for (i = 0; i < vmsg->fd_num; i++) {
1615
+{
126
+ qemu_set_nonblock(vmsg->fds[i]);
1616
+ req->rsp_upiu.sr.sense_data_len = cpu_to_be16(scsi_req->sense_len);
127
+ }
1617
+ assert(scsi_req->sense_len <= SCSI_SENSE_LEN);
128
+}
1618
+ memcpy(req->rsp_upiu.sr.sense_data, scsi_req->sense, scsi_req->sense_len);
129
+
1619
+}
130
+static void vu_accept(QIONetListener *listener, QIOChannelSocket *sioc,
1620
+
131
+ gpointer opaque);
1621
static void ufs_build_upiu_header(UfsRequest *req, uint8_t trans_type,
132
+
1622
uint8_t flags, uint8_t response,
133
+static void close_client(VuServer *server)
1623
uint8_t scsi_status,
134
+{
1624
@@ -XXX,XX +XXX,XX @@ static void ufs_build_upiu_header(UfsRequest *req, uint8_t trans_type,
1625
req->rsp_upiu.header.data_segment_length = cpu_to_be16(data_segment_length);
1626
}
1627
1628
+static void ufs_scsi_command_complete(SCSIRequest *scsi_req, size_t resid)
1629
+{
1630
+ UfsRequest *req = scsi_req->hba_private;
1631
+ int16_t status = scsi_req->status;
1632
+ uint32_t expected_len = be32_to_cpu(req->req_upiu.sc.exp_data_transfer_len);
1633
+ uint32_t transfered_len = scsi_req->cmd.xfer - resid;
1634
+ uint8_t flags = 0, response = COMMAND_RESULT_SUCESS;
1635
+ uint16_t data_segment_length;
1636
+
1637
+ if (expected_len > transfered_len) {
1638
+ req->rsp_upiu.sr.residual_transfer_count =
1639
+ cpu_to_be32(expected_len - transfered_len);
1640
+ flags |= UFS_UPIU_FLAG_UNDERFLOW;
1641
+ } else if (expected_len < transfered_len) {
1642
+ req->rsp_upiu.sr.residual_transfer_count =
1643
+ cpu_to_be32(transfered_len - expected_len);
1644
+ flags |= UFS_UPIU_FLAG_OVERFLOW;
1645
+ }
1646
+
1647
+ if (status != 0) {
1648
+ ufs_build_upiu_sense_data(req, scsi_req);
1649
+ response = COMMAND_RESULT_FAIL;
1650
+ }
1651
+
1652
+ data_segment_length = cpu_to_be16(scsi_req->sense_len +
1653
+ sizeof(req->rsp_upiu.sr.sense_data_len));
1654
+ ufs_build_upiu_header(req, UPIU_TRANSACTION_RESPONSE, flags, response,
1655
+ status, data_segment_length);
1656
+
1657
+ ufs_complete_req(req, UFS_REQUEST_SUCCESS);
1658
+
1659
+ scsi_req->hba_private = NULL;
1660
+ scsi_req_unref(scsi_req);
1661
+}
1662
+
1663
+static const struct SCSIBusInfo ufs_scsi_info = {
1664
+ .tcq = true,
1665
+ .max_target = 0,
1666
+ .max_lun = UFS_MAX_LUS,
1667
+ .max_channel = 0,
1668
+
1669
+ .get_sg_list = ufs_get_sg_list,
1670
+ .complete = ufs_scsi_command_complete,
1671
+};
1672
+
1673
+static UfsReqResult ufs_exec_scsi_cmd(UfsRequest *req)
1674
+{
1675
+ UfsHc *u = req->hc;
1676
+ uint8_t lun = req->req_upiu.header.lun;
1677
+ uint8_t task_tag = req->req_upiu.header.task_tag;
1678
+ SCSIDevice *dev = NULL;
1679
+
1680
+ trace_ufs_exec_scsi_cmd(req->slot, lun, req->req_upiu.sc.cdb[0]);
1681
+
1682
+ if (!is_wlun(lun)) {
1683
+ if (lun >= u->device_desc.number_lu) {
1684
+ trace_ufs_err_scsi_cmd_invalid_lun(lun);
1685
+ return UFS_REQUEST_FAIL;
1686
+ } else if (u->lus[lun] == NULL) {
1687
+ trace_ufs_err_scsi_cmd_invalid_lun(lun);
1688
+ return UFS_REQUEST_FAIL;
1689
+ }
1690
+ }
1691
+
1692
+ switch (lun) {
1693
+ case UFS_UPIU_REPORT_LUNS_WLUN:
1694
+ dev = &u->report_wlu->qdev;
1695
+ break;
1696
+ case UFS_UPIU_UFS_DEVICE_WLUN:
1697
+ dev = &u->dev_wlu->qdev;
1698
+ break;
1699
+ case UFS_UPIU_BOOT_WLUN:
1700
+ dev = &u->boot_wlu->qdev;
1701
+ break;
1702
+ case UFS_UPIU_RPMB_WLUN:
1703
+ dev = &u->rpmb_wlu->qdev;
1704
+ break;
1705
+ default:
1706
+ dev = &u->lus[lun]->qdev;
1707
+ }
1708
+
1709
+ SCSIRequest *scsi_req = scsi_req_new(
1710
+ dev, task_tag, lun, req->req_upiu.sc.cdb, UFS_CDB_SIZE, req);
1711
+
1712
+ uint32_t len = scsi_req_enqueue(scsi_req);
1713
+ if (len) {
1714
+ scsi_req_continue(scsi_req);
1715
+ }
1716
+
1717
+ return UFS_REQUEST_NO_COMPLETE;
1718
+}
1719
+
1720
static UfsReqResult ufs_exec_nop_cmd(UfsRequest *req)
1721
{
1722
trace_ufs_exec_nop_cmd(req->slot);
1723
@@ -XXX,XX +XXX,XX @@ static const RpmbUnitDescriptor rpmb_unit_desc = {
1724
1725
static QueryRespCode ufs_read_unit_desc(UfsRequest *req)
1726
{
1727
+ UfsHc *u = req->hc;
1728
uint8_t lun = req->req_upiu.qr.index;
1729
1730
- if (lun != UFS_UPIU_RPMB_WLUN && lun > UFS_MAX_LUS) {
1731
+ if (lun != UFS_UPIU_RPMB_WLUN &&
1732
+ (lun > UFS_MAX_LUS || u->lus[lun] == NULL)) {
1733
trace_ufs_err_query_invalid_index(req->req_upiu.qr.opcode, lun);
1734
return QUERY_RESULT_INVALID_INDEX;
1735
}
1736
@@ -XXX,XX +XXX,XX @@ static QueryRespCode ufs_read_unit_desc(UfsRequest *req)
1737
if (lun == UFS_UPIU_RPMB_WLUN) {
1738
memcpy(&req->rsp_upiu.qr.data, &rpmb_unit_desc, rpmb_unit_desc.length);
1739
} else {
1740
- /* unit descriptor is not yet supported */
1741
- return QUERY_RESULT_INVALID_INDEX;
1742
+ memcpy(&req->rsp_upiu.qr.data, &u->lus[lun]->unit_desc,
1743
+ sizeof(u->lus[lun]->unit_desc));
1744
}
1745
1746
return QUERY_RESULT_SUCCESS;
1747
@@ -XXX,XX +XXX,XX @@ static void ufs_exec_req(UfsRequest *req)
1748
req_result = ufs_exec_nop_cmd(req);
1749
break;
1750
case UPIU_TRANSACTION_COMMAND:
1751
- /* Not yet implemented */
1752
- req_result = UFS_REQUEST_FAIL;
1753
+ req_result = ufs_exec_scsi_cmd(req);
1754
break;
1755
case UPIU_TRANSACTION_QUERY_REQ:
1756
req_result = ufs_exec_query_cmd(req);
1757
@@ -XXX,XX +XXX,XX @@ static void ufs_exec_req(UfsRequest *req)
1758
req_result = UFS_REQUEST_FAIL;
1759
}
1760
1761
- ufs_complete_req(req, req_result);
135
+ /*
1762
+ /*
136
+ * Before closing the client
1763
+ * The ufs_complete_req for scsi commands is handled by the
137
+ *
1764
+ * ufs_scsi_command_complete() callback function. Therefore, to avoid
138
+ * 1. Let vu_client_trip stop processing new vhost-user msg
1765
+ * duplicate processing, ufs_complete_req() is not called for scsi commands.
139
+ *
140
+ * 2. remove kick_handler
141
+ *
142
+ * 3. wait for the kick handler to be finished
143
+ *
144
+ * 4. wait for the current vhost-user msg to be finished processing
145
+ */
1766
+ */
146
+
1767
+ if (req_result != UFS_REQUEST_NO_COMPLETE) {
147
+ QIOChannelSocket *sioc = server->sioc;
1768
+ ufs_complete_req(req, req_result);
148
+ /* When this is set vu_client_trip will stop new processing vhost-user message */
1769
+ }
149
+ server->sioc = NULL;
1770
}
150
+
1771
151
+ VuFdWatch *vu_fd_watch, *next;
1772
static void ufs_process_req(void *opaque)
152
+ QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) {
1773
@@ -XXX,XX +XXX,XX @@ static void ufs_init_hc(UfsHc *u)
153
+ aio_set_fd_handler(server->ioc->ctx, vu_fd_watch->fd, true, NULL,
1774
u->flags.permanently_disable_fw_update = 1;
154
+ NULL, NULL, NULL);
1775
}
155
+ }
1776
156
+
1777
+static bool ufs_init_wlu(UfsHc *u, UfsWLu **wlu, uint8_t wlun, Error **errp)
157
+ while (!QTAILQ_EMPTY(&server->vu_fd_watches)) {
1778
+{
158
+ QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) {
1779
+ UfsWLu *new_wlu = UFSWLU(qdev_new(TYPE_UFS_WLU));
159
+ if (!vu_fd_watch->processing) {
1780
+
160
+ QTAILQ_REMOVE(&server->vu_fd_watches, vu_fd_watch, next);
1781
+ qdev_prop_set_uint32(DEVICE(new_wlu), "lun", wlun);
161
+ g_free(vu_fd_watch);
162
+ }
163
+ }
164
+ }
165
+
166
+ while (server->processing_msg) {
167
+ if (server->ioc->read_coroutine) {
168
+ server->ioc->read_coroutine = NULL;
169
+ qio_channel_set_aio_fd_handler(server->ioc, server->ioc->ctx, NULL,
170
+ NULL, server->ioc);
171
+ server->processing_msg = false;
172
+ }
173
+ }
174
+
175
+ vu_deinit(&server->vu_dev);
176
+ object_unref(OBJECT(sioc));
177
+ object_unref(OBJECT(server->ioc));
178
+}
179
+
180
+static void panic_cb(VuDev *vu_dev, const char *buf)
181
+{
182
+ VuServer *server = container_of(vu_dev, VuServer, vu_dev);
183
+
184
+ /* avoid while loop in close_client */
185
+ server->processing_msg = false;
186
+
187
+ if (buf) {
188
+ error_report("vu_panic: %s", buf);
189
+ }
190
+
191
+ if (server->sioc) {
192
+ close_client(server);
193
+ }
194
+
195
+ if (server->device_panic_notifier) {
196
+ server->device_panic_notifier(server);
197
+ }
198
+
1782
+
199
+ /*
1783
+ /*
200
+ * Set the callback function for network listener so another
1784
+ * The well-known lu shares the same bus as the normal lu. If the well-known
201
+ * vhost-user client can connect to this server
1785
+ * lu writes the same channel value as the normal lu, the report will be
1786
+ * made not only for the normal lu but also for the well-known lu at
1787
+ * REPORT_LUN time. To prevent this, the channel value of normal lu is fixed
1788
+ * to 0 and the channel value of well-known lu is fixed to 1.
202
+ */
1789
+ */
203
+ qio_net_listener_set_client_func(server->listener,
1790
+ qdev_prop_set_uint32(DEVICE(new_wlu), "channel", 1);
204
+ vu_accept,
1791
+ if (!qdev_realize_and_unref(DEVICE(new_wlu), BUS(&u->bus), errp)) {
205
+ server,
1792
+ return false;
206
+ NULL);
1793
+ }
207
+}
1794
+
208
+
1795
+ *wlu = new_wlu;
209
+static bool coroutine_fn
210
+vu_message_read(VuDev *vu_dev, int conn_fd, VhostUserMsg *vmsg)
211
+{
212
+ struct iovec iov = {
213
+ .iov_base = (char *)vmsg,
214
+ .iov_len = VHOST_USER_HDR_SIZE,
215
+ };
216
+ int rc, read_bytes = 0;
217
+ Error *local_err = NULL;
218
+ /*
219
+ * Store fds/nfds returned from qio_channel_readv_full into
220
+ * temporary variables.
221
+ *
222
+ * VhostUserMsg is a packed structure, gcc will complain about passing
223
+ * pointer to a packed structure member if we pass &VhostUserMsg.fd_num
224
+ * and &VhostUserMsg.fds directly when calling qio_channel_readv_full,
225
+ * thus two temporary variables nfds and fds are used here.
226
+ */
227
+ size_t nfds = 0, nfds_t = 0;
228
+ const size_t max_fds = G_N_ELEMENTS(vmsg->fds);
229
+ int *fds_t = NULL;
230
+ VuServer *server = container_of(vu_dev, VuServer, vu_dev);
231
+ QIOChannel *ioc = server->ioc;
232
+
233
+ if (!ioc) {
234
+ error_report_err(local_err);
235
+ goto fail;
236
+ }
237
+
238
+ assert(qemu_in_coroutine());
239
+ do {
240
+ /*
241
+ * qio_channel_readv_full may have short reads, keeping calling it
242
+ * until getting VHOST_USER_HDR_SIZE or 0 bytes in total
243
+ */
244
+ rc = qio_channel_readv_full(ioc, &iov, 1, &fds_t, &nfds_t, &local_err);
245
+ if (rc < 0) {
246
+ if (rc == QIO_CHANNEL_ERR_BLOCK) {
247
+ qio_channel_yield(ioc, G_IO_IN);
248
+ continue;
249
+ } else {
250
+ error_report_err(local_err);
251
+ return false;
252
+ }
253
+ }
254
+ read_bytes += rc;
255
+ if (nfds_t > 0) {
256
+ if (nfds + nfds_t > max_fds) {
257
+ error_report("A maximum of %zu fds are allowed, "
258
+ "however got %zu fds now",
259
+ max_fds, nfds + nfds_t);
260
+ goto fail;
261
+ }
262
+ memcpy(vmsg->fds + nfds, fds_t,
263
+ nfds_t *sizeof(vmsg->fds[0]));
264
+ nfds += nfds_t;
265
+ g_free(fds_t);
266
+ }
267
+ if (read_bytes == VHOST_USER_HDR_SIZE || rc == 0) {
268
+ break;
269
+ }
270
+ iov.iov_base = (char *)vmsg + read_bytes;
271
+ iov.iov_len = VHOST_USER_HDR_SIZE - read_bytes;
272
+ } while (true);
273
+
274
+ vmsg->fd_num = nfds;
275
+ /* qio_channel_readv_full will make socket fds blocking, unblock them */
276
+ vmsg_unblock_fds(vmsg);
277
+ if (vmsg->size > sizeof(vmsg->payload)) {
278
+ error_report("Error: too big message request: %d, "
279
+ "size: vmsg->size: %u, "
280
+ "while sizeof(vmsg->payload) = %zu",
281
+ vmsg->request, vmsg->size, sizeof(vmsg->payload));
282
+ goto fail;
283
+ }
284
+
285
+ struct iovec iov_payload = {
286
+ .iov_base = (char *)&vmsg->payload,
287
+ .iov_len = vmsg->size,
288
+ };
289
+ if (vmsg->size) {
290
+ rc = qio_channel_readv_all_eof(ioc, &iov_payload, 1, &local_err);
291
+ if (rc == -1) {
292
+ error_report_err(local_err);
293
+ goto fail;
294
+ }
295
+ }
296
+
297
+ return true;
1796
+ return true;
298
+
1797
+}
299
+fail:
1798
+
300
+ vmsg_close_fds(vmsg);
1799
static void ufs_realize(PCIDevice *pci_dev, Error **errp)
301
+
1800
{
302
+ return false;
1801
UfsHc *u = UFS(pci_dev);
303
+}
1802
@@ -XXX,XX +XXX,XX @@ static void ufs_realize(PCIDevice *pci_dev, Error **errp)
304
+
1803
return;
305
+
1804
}
306
+static void vu_client_start(VuServer *server);
1805
307
+static coroutine_fn void vu_client_trip(void *opaque)
1806
+ qbus_init(&u->bus, sizeof(UfsBus), TYPE_UFS_BUS, &pci_dev->qdev,
308
+{
1807
+ u->parent_obj.qdev.id);
309
+ VuServer *server = opaque;
1808
+ u->bus.parent_bus.info = &ufs_scsi_info;
310
+
1809
+
311
+ while (!server->aio_context_changed && server->sioc) {
1810
ufs_init_state(u);
312
+ server->processing_msg = true;
1811
ufs_init_hc(u);
313
+ vu_dispatch(&server->vu_dev);
1812
ufs_init_pci(u, pci_dev);
314
+ server->processing_msg = false;
1813
+
315
+ }
1814
+ if (!ufs_init_wlu(u, &u->report_wlu, UFS_UPIU_REPORT_LUNS_WLUN, errp)) {
316
+
317
+ if (server->aio_context_changed && server->sioc) {
318
+ server->aio_context_changed = false;
319
+ vu_client_start(server);
320
+ }
321
+}
322
+
323
+static void vu_client_start(VuServer *server)
324
+{
325
+ server->co_trip = qemu_coroutine_create(vu_client_trip, server);
326
+ aio_co_enter(server->ctx, server->co_trip);
327
+}
328
+
329
+/*
330
+ * a wrapper for vu_kick_cb
331
+ *
332
+ * since aio_dispatch can only pass one user data pointer to the
333
+ * callback function, pack VuDev and pvt into a struct. Then unpack it
334
+ * and pass them to vu_kick_cb
335
+ */
336
+static void kick_handler(void *opaque)
337
+{
338
+ VuFdWatch *vu_fd_watch = opaque;
339
+ vu_fd_watch->processing = true;
340
+ vu_fd_watch->cb(vu_fd_watch->vu_dev, 0, vu_fd_watch->pvt);
341
+ vu_fd_watch->processing = false;
342
+}
343
+
344
+
345
+static VuFdWatch *find_vu_fd_watch(VuServer *server, int fd)
346
+{
347
+
348
+ VuFdWatch *vu_fd_watch, *next;
349
+ QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) {
350
+ if (vu_fd_watch->fd == fd) {
351
+ return vu_fd_watch;
352
+ }
353
+ }
354
+ return NULL;
355
+}
356
+
357
+static void
358
+set_watch(VuDev *vu_dev, int fd, int vu_evt,
359
+ vu_watch_cb cb, void *pvt)
360
+{
361
+
362
+ VuServer *server = container_of(vu_dev, VuServer, vu_dev);
363
+ g_assert(vu_dev);
364
+ g_assert(fd >= 0);
365
+ g_assert(cb);
366
+
367
+ VuFdWatch *vu_fd_watch = find_vu_fd_watch(server, fd);
368
+
369
+ if (!vu_fd_watch) {
370
+ VuFdWatch *vu_fd_watch = g_new0(VuFdWatch, 1);
371
+
372
+ QTAILQ_INSERT_TAIL(&server->vu_fd_watches, vu_fd_watch, next);
373
+
374
+ vu_fd_watch->fd = fd;
375
+ vu_fd_watch->cb = cb;
376
+ qemu_set_nonblock(fd);
377
+ aio_set_fd_handler(server->ioc->ctx, fd, true, kick_handler,
378
+ NULL, NULL, vu_fd_watch);
379
+ vu_fd_watch->vu_dev = vu_dev;
380
+ vu_fd_watch->pvt = pvt;
381
+ }
382
+}
383
+
384
+
385
+static void remove_watch(VuDev *vu_dev, int fd)
386
+{
387
+ VuServer *server;
388
+ g_assert(vu_dev);
389
+ g_assert(fd >= 0);
390
+
391
+ server = container_of(vu_dev, VuServer, vu_dev);
392
+
393
+ VuFdWatch *vu_fd_watch = find_vu_fd_watch(server, fd);
394
+
395
+ if (!vu_fd_watch) {
396
+ return;
1815
+ return;
397
+ }
1816
+ }
398
+ aio_set_fd_handler(server->ioc->ctx, fd, true, NULL, NULL, NULL, NULL);
1817
+
399
+
1818
+ if (!ufs_init_wlu(u, &u->dev_wlu, UFS_UPIU_UFS_DEVICE_WLUN, errp)) {
400
+ QTAILQ_REMOVE(&server->vu_fd_watches, vu_fd_watch, next);
401
+ g_free(vu_fd_watch);
402
+}
403
+
404
+
405
+static void vu_accept(QIONetListener *listener, QIOChannelSocket *sioc,
406
+ gpointer opaque)
407
+{
408
+ VuServer *server = opaque;
409
+
410
+ if (server->sioc) {
411
+ warn_report("Only one vhost-user client is allowed to "
412
+ "connect the server one time");
413
+ return;
1819
+ return;
414
+ }
1820
+ }
415
+
1821
+
416
+ if (!vu_init(&server->vu_dev, server->max_queues, sioc->fd, panic_cb,
1822
+ if (!ufs_init_wlu(u, &u->boot_wlu, UFS_UPIU_BOOT_WLUN, errp)) {
417
+ vu_message_read, set_watch, remove_watch, server->vu_iface)) {
418
+ error_report("Failed to initialize libvhost-user");
419
+ return;
1823
+ return;
420
+ }
1824
+ }
421
+
1825
+
422
+ /*
1826
+ if (!ufs_init_wlu(u, &u->rpmb_wlu, UFS_UPIU_RPMB_WLUN, errp)) {
423
+ * Unset the callback function for network listener to make another
424
+ * vhost-user client keeping waiting until this client disconnects
425
+ */
426
+ qio_net_listener_set_client_func(server->listener,
427
+ NULL,
428
+ NULL,
429
+ NULL);
430
+ server->sioc = sioc;
431
+ /*
432
+ * Increase the object reference, so sioc will not freed by
433
+ * qio_net_listener_channel_func which will call object_unref(OBJECT(sioc))
434
+ */
435
+ object_ref(OBJECT(server->sioc));
436
+ qio_channel_set_name(QIO_CHANNEL(sioc), "vhost-user client");
437
+ server->ioc = QIO_CHANNEL(sioc);
438
+ object_ref(OBJECT(server->ioc));
439
+ qio_channel_attach_aio_context(server->ioc, server->ctx);
440
+ qio_channel_set_blocking(QIO_CHANNEL(server->sioc), false, NULL);
441
+ vu_client_start(server);
442
+}
443
+
444
+
445
+void vhost_user_server_stop(VuServer *server)
446
+{
447
+ if (server->sioc) {
448
+ close_client(server);
449
+ }
450
+
451
+ if (server->listener) {
452
+ qio_net_listener_disconnect(server->listener);
453
+ object_unref(OBJECT(server->listener));
454
+ }
455
+
456
+}
457
+
458
+void vhost_user_server_set_aio_context(VuServer *server, AioContext *ctx)
459
+{
460
+ VuFdWatch *vu_fd_watch, *next;
461
+ void *opaque = NULL;
462
+ IOHandler *io_read = NULL;
463
+ bool attach;
464
+
465
+ server->ctx = ctx ? ctx : qemu_get_aio_context();
466
+
467
+ if (!server->sioc) {
468
+ /* not yet serving any client*/
469
+ return;
1827
+ return;
470
+ }
1828
+ }
471
+
1829
}
472
+ if (ctx) {
1830
473
+ qio_channel_attach_aio_context(server->ioc, ctx);
1831
static void ufs_exit(PCIDevice *pci_dev)
474
+ server->aio_context_changed = true;
1832
{
475
+ io_read = kick_handler;
1833
UfsHc *u = UFS(pci_dev);
476
+ attach = true;
1834
477
+ } else {
1835
+ if (u->dev_wlu) {
478
+ qio_channel_detach_aio_context(server->ioc);
1836
+ object_unref(OBJECT(u->dev_wlu));
479
+ /* server->ioc->ctx keeps the old AioConext */
1837
+ u->dev_wlu = NULL;
480
+ ctx = server->ioc->ctx;
1838
+ }
481
+ attach = false;
1839
+
482
+ }
1840
+ if (u->report_wlu) {
483
+
1841
+ object_unref(OBJECT(u->report_wlu));
484
+ QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) {
1842
+ u->report_wlu = NULL;
485
+ if (vu_fd_watch->cb) {
1843
+ }
486
+ opaque = attach ? vu_fd_watch : NULL;
1844
+
487
+ aio_set_fd_handler(ctx, vu_fd_watch->fd, true,
1845
+ if (u->rpmb_wlu) {
488
+ io_read, NULL, NULL,
1846
+ object_unref(OBJECT(u->rpmb_wlu));
489
+ opaque);
1847
+ u->rpmb_wlu = NULL;
490
+ }
1848
+ }
491
+ }
1849
+
492
+}
1850
+ if (u->boot_wlu) {
493
+
1851
+ object_unref(OBJECT(u->boot_wlu));
494
+
1852
+ u->boot_wlu = NULL;
495
+bool vhost_user_server_start(VuServer *server,
1853
+ }
496
+ SocketAddress *socket_addr,
1854
+
497
+ AioContext *ctx,
1855
qemu_bh_delete(u->doorbell_bh);
498
+ uint16_t max_queues,
1856
qemu_bh_delete(u->complete_bh);
499
+ DevicePanicNotifierFn *device_panic_notifier,
1857
500
+ const VuDevIface *vu_iface,
1858
@@ -XXX,XX +XXX,XX @@ static void ufs_class_init(ObjectClass *oc, void *data)
501
+ Error **errp)
1859
dc->vmsd = &ufs_vmstate;
502
+{
1860
}
503
+ QIONetListener *listener = qio_net_listener_new();
1861
504
+ if (qio_net_listener_open_sync(listener, socket_addr, 1,
1862
+static bool ufs_bus_check_address(BusState *qbus, DeviceState *qdev,
505
+ errp) < 0) {
1863
+ Error **errp)
506
+ object_unref(OBJECT(listener));
1864
+{
1865
+ SCSIDevice *dev = SCSI_DEVICE(qdev);
1866
+ UfsBusClass *ubc = UFS_BUS_GET_CLASS(qbus);
1867
+ UfsHc *u = UFS(qbus->parent);
1868
+
1869
+ if (strcmp(object_get_typename(OBJECT(dev)), TYPE_UFS_WLU) == 0) {
1870
+ if (dev->lun != UFS_UPIU_REPORT_LUNS_WLUN &&
1871
+ dev->lun != UFS_UPIU_UFS_DEVICE_WLUN &&
1872
+ dev->lun != UFS_UPIU_BOOT_WLUN && dev->lun != UFS_UPIU_RPMB_WLUN) {
1873
+ error_setg(errp, "bad well-known lun: %d", dev->lun);
1874
+ return false;
1875
+ }
1876
+
1877
+ if ((dev->lun == UFS_UPIU_REPORT_LUNS_WLUN && u->report_wlu != NULL) ||
1878
+ (dev->lun == UFS_UPIU_UFS_DEVICE_WLUN && u->dev_wlu != NULL) ||
1879
+ (dev->lun == UFS_UPIU_BOOT_WLUN && u->boot_wlu != NULL) ||
1880
+ (dev->lun == UFS_UPIU_RPMB_WLUN && u->rpmb_wlu != NULL)) {
1881
+ error_setg(errp, "well-known lun %d already exists", dev->lun);
1882
+ return false;
1883
+ }
1884
+
1885
+ return true;
1886
+ }
1887
+
1888
+ if (strcmp(object_get_typename(OBJECT(dev)), TYPE_UFS_LU) != 0) {
1889
+ error_setg(errp, "%s cannot be connected to ufs-bus",
1890
+ object_get_typename(OBJECT(dev)));
507
+ return false;
1891
+ return false;
508
+ }
1892
+ }
509
+
1893
+
510
+ /* zero out unspecified fileds */
1894
+ return ubc->parent_check_address(qbus, qdev, errp);
511
+ *server = (VuServer) {
1895
+}
512
+ .listener = listener,
1896
+
513
+ .vu_iface = vu_iface,
1897
+static void ufs_bus_class_init(ObjectClass *class, void *data)
514
+ .max_queues = max_queues,
1898
+{
515
+ .ctx = ctx,
1899
+ BusClass *bc = BUS_CLASS(class);
516
+ .device_panic_notifier = device_panic_notifier,
1900
+ UfsBusClass *ubc = UFS_BUS_CLASS(class);
517
+ };
1901
+ ubc->parent_check_address = bc->check_address;
518
+
1902
+ bc->check_address = ufs_bus_check_address;
519
+ qio_net_listener_set_name(server->listener, "vhost-user-backend-listener");
1903
+}
520
+
1904
+
521
+ qio_net_listener_set_client_func(server->listener,
1905
static const TypeInfo ufs_info = {
522
+ vu_accept,
1906
.name = TYPE_UFS,
523
+ server,
1907
.parent = TYPE_PCI_DEVICE,
524
+ NULL);
1908
@@ -XXX,XX +XXX,XX @@ static const TypeInfo ufs_info = {
525
+
1909
.interfaces = (InterfaceInfo[]){ { INTERFACE_PCIE_DEVICE }, {} },
526
+ QTAILQ_INIT(&server->vu_fd_watches);
1910
};
527
+ return true;
1911
528
+}
1912
+static const TypeInfo ufs_bus_info = {
529
diff --git a/util/meson.build b/util/meson.build
1913
+ .name = TYPE_UFS_BUS,
1914
+ .parent = TYPE_SCSI_BUS,
1915
+ .class_init = ufs_bus_class_init,
1916
+ .class_size = sizeof(UfsBusClass),
1917
+ .instance_size = sizeof(UfsBus),
1918
+};
1919
+
1920
static void ufs_register_types(void)
1921
{
1922
type_register_static(&ufs_info);
1923
+ type_register_static(&ufs_bus_info);
1924
}
1925
1926
type_init(ufs_register_types)
1927
diff --git a/hw/ufs/meson.build b/hw/ufs/meson.build
530
index XXXXXXX..XXXXXXX 100644
1928
index XXXXXXX..XXXXXXX 100644
531
--- a/util/meson.build
1929
--- a/hw/ufs/meson.build
532
+++ b/util/meson.build
1930
+++ b/hw/ufs/meson.build
533
@@ -XXX,XX +XXX,XX @@ if have_block
1931
@@ -1 +1 @@
534
util_ss.add(files('main-loop.c'))
1932
-system_ss.add(when: 'CONFIG_UFS_PCI', if_true: files('ufs.c'))
535
util_ss.add(files('nvdimm-utils.c'))
1933
+system_ss.add(when: 'CONFIG_UFS_PCI', if_true: files('ufs.c', 'lu.c'))
536
util_ss.add(files('qemu-coroutine.c', 'qemu-coroutine-lock.c', 'qemu-coroutine-io.c'))
1934
diff --git a/hw/ufs/trace-events b/hw/ufs/trace-events
537
+ util_ss.add(when: 'CONFIG_LINUX', if_true: files('vhost-user-server.c'))
1935
index XXXXXXX..XXXXXXX 100644
538
util_ss.add(files('qemu-coroutine-sleep.c'))
1936
--- a/hw/ufs/trace-events
539
util_ss.add(files('qemu-co-shared-resource.c'))
1937
+++ b/hw/ufs/trace-events
540
util_ss.add(files('thread-pool.c', 'qemu-timer.c'))
1938
@@ -XXX,XX +XXX,XX @@ ufs_exec_scsi_cmd(uint32_t slot, uint8_t lun, uint8_t opcode) "slot %"PRIu32", l
1939
ufs_exec_query_cmd(uint32_t slot, uint8_t opcode) "slot %"PRIu32", opcode 0x%"PRIx8""
1940
ufs_process_uiccmd(uint32_t uiccmd, uint32_t ucmdarg1, uint32_t ucmdarg2, uint32_t ucmdarg3) "uiccmd 0x%"PRIx32", ucmdarg1 0x%"PRIx32", ucmdarg2 0x%"PRIx32", ucmdarg3 0x%"PRIx32""
1941
1942
+# lu.c
1943
+ufs_scsi_check_condition(uint32_t tag, uint8_t key, uint8_t asc, uint8_t ascq) "Command complete tag=0x%x sense=%d/%d/%d"
1944
+ufs_scsi_read_complete(uint32_t tag, size_t size) "Data ready tag=0x%x len=%zd"
1945
+ufs_scsi_read_data_count(uint32_t sector_count) "Read sector_count=%d"
1946
+ufs_scsi_read_data_invalid(void) "Data transfer direction invalid"
1947
+ufs_scsi_write_complete_noio(uint32_t tag, size_t size) "Write complete tag=0x%x more=%zd"
1948
+ufs_scsi_write_data_invalid(void) "Data transfer direction invalid"
1949
+ufs_scsi_emulate_vpd_page_00(size_t xfer) "Inquiry EVPD[Supported pages] buffer size %zd"
1950
+ufs_scsi_emulate_vpd_page_80_not_supported(void) "Inquiry EVPD[Serial number] not supported"
1951
+ufs_scsi_emulate_vpd_page_80(size_t xfer) "Inquiry EVPD[Serial number] buffer size %zd"
1952
+ufs_scsi_emulate_vpd_page_87(size_t xfer) "Inquiry EVPD[Mode Page Policy] buffer size %zd"
1953
+ufs_scsi_emulate_mode_sense(int cmd, int page, size_t xfer, int control) "Mode Sense(%d) (page %d, xfer %zd, page_control %d)"
1954
+ufs_scsi_emulate_read_data(int buflen) "Read buf_len=%d"
1955
+ufs_scsi_emulate_write_data(int buflen) "Write buf_len=%d"
1956
+ufs_scsi_emulate_command_START_STOP(void) "START STOP UNIT"
1957
+ufs_scsi_emulate_command_FORMAT_UNIT(void) "FORMAT UNIT"
1958
+ufs_scsi_emulate_command_SEND_DIAGNOSTIC(void) "SEND DIAGNOSTIC"
1959
+ufs_scsi_emulate_command_SAI_16(void) "SAI READ CAPACITY(16)"
1960
+ufs_scsi_emulate_command_SAI_unsupported(void) "Unsupported Service Action In"
1961
+ufs_scsi_emulate_command_MODE_SELECT_10(size_t xfer) "Mode Select(10) (len %zd)"
1962
+ufs_scsi_emulate_command_VERIFY(int bytchk) "Verify (bytchk %d)"
1963
+ufs_scsi_emulate_command_UNKNOWN(int cmd, const char *name) "Unknown SCSI command (0x%2.2x=%s)"
1964
+ufs_scsi_dma_command_READ(uint64_t lba, uint32_t len) "Read (block %" PRIu64 ", count %u)"
1965
+ufs_scsi_dma_command_WRITE(uint64_t lba, int len) "Write (block %" PRIu64 ", count %u)"
1966
+
1967
# error condition
1968
ufs_err_dma_read_utrd(uint32_t slot, uint64_t addr) "failed to read utrd. UTRLDBR slot %"PRIu32", UTRD dma addr %"PRIu64""
1969
ufs_err_dma_read_req_upiu(uint32_t slot, uint64_t addr) "failed to read req upiu. UTRLDBR slot %"PRIu32", request upiu addr %"PRIu64""
541
--
1970
--
542
2.26.2
1971
2.41.0
543
diff view generated by jsdifflib
1
From: Coiby Xu <coiby.xu@gmail.com>
1
From: Jeuk Kim <jeuk20.kim@gmail.com>
2
2
3
Move the constants from hw/core/qdev-properties.c to
3
This patch includes the following tests
4
util/block-helpers.h so that knowledge of the min/max values is
4
Test mmio read
5
Test ufs device initialization and ufs-lu recognition
6
Test I/O (Performs a write followed by a read to verify)
5
7
6
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
8
Signed-off-by: Jeuk Kim <jeuk20.kim@samsung.com>
7
Signed-off-by: Coiby Xu <coiby.xu@gmail.com>
9
Acked-by: Thomas Huth <thuth@redhat.com>
8
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
10
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
9
Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com>
11
Message-id: af6b8d54c049490b3533a784a0aeac4798bb9217.1691062912.git.jeuk20.kim@samsung.com
10
Acked-by: Eduardo Habkost <ehabkost@redhat.com>
11
Message-id: 20200918080912.321299-5-coiby.xu@gmail.com
12
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
12
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
13
---
13
---
14
util/block-helpers.h | 19 +++++++++++++
14
MAINTAINERS | 1 +
15
hw/core/qdev-properties-system.c | 31 ++++-----------------
15
tests/qtest/ufs-test.c | 584 ++++++++++++++++++++++++++++++++++++++++
16
util/block-helpers.c | 46 ++++++++++++++++++++++++++++++++
16
tests/qtest/meson.build | 1 +
17
util/meson.build | 1 +
17
3 files changed, 586 insertions(+)
18
4 files changed, 71 insertions(+), 26 deletions(-)
18
create mode 100644 tests/qtest/ufs-test.c
19
create mode 100644 util/block-helpers.h
20
create mode 100644 util/block-helpers.c
21
19
22
diff --git a/util/block-helpers.h b/util/block-helpers.h
20
diff --git a/MAINTAINERS b/MAINTAINERS
21
index XXXXXXX..XXXXXXX 100644
22
--- a/MAINTAINERS
23
+++ b/MAINTAINERS
24
@@ -XXX,XX +XXX,XX @@ M: Jeuk Kim <jeuk20.kim@samsung.com>
25
S: Supported
26
F: hw/ufs/*
27
F: include/block/ufs.h
28
+F: tests/qtest/ufs-test.c
29
30
megasas
31
M: Hannes Reinecke <hare@suse.com>
32
diff --git a/tests/qtest/ufs-test.c b/tests/qtest/ufs-test.c
23
new file mode 100644
33
new file mode 100644
24
index XXXXXXX..XXXXXXX
34
index XXXXXXX..XXXXXXX
25
--- /dev/null
35
--- /dev/null
26
+++ b/util/block-helpers.h
36
+++ b/tests/qtest/ufs-test.c
27
@@ -XXX,XX +XXX,XX @@
28
+#ifndef BLOCK_HELPERS_H
29
+#define BLOCK_HELPERS_H
30
+
31
+#include "qemu/units.h"
32
+
33
+/* lower limit is sector size */
34
+#define MIN_BLOCK_SIZE INT64_C(512)
35
+#define MIN_BLOCK_SIZE_STR "512 B"
36
+/*
37
+ * upper limit is arbitrary, 2 MiB looks sufficient for all sensible uses, and
38
+ * matches qcow2 cluster size limit
39
+ */
40
+#define MAX_BLOCK_SIZE (2 * MiB)
41
+#define MAX_BLOCK_SIZE_STR "2 MiB"
42
+
43
+void check_block_size(const char *id, const char *name, int64_t value,
44
+ Error **errp);
45
+
46
+#endif /* BLOCK_HELPERS_H */
47
diff --git a/hw/core/qdev-properties-system.c b/hw/core/qdev-properties-system.c
48
index XXXXXXX..XXXXXXX 100644
49
--- a/hw/core/qdev-properties-system.c
50
+++ b/hw/core/qdev-properties-system.c
51
@@ -XXX,XX +XXX,XX @@
52
#include "sysemu/blockdev.h"
53
#include "net/net.h"
54
#include "hw/pci/pci.h"
55
+#include "util/block-helpers.h"
56
57
static bool check_prop_still_unset(DeviceState *dev, const char *name,
58
const void *old_val, const char *new_val,
59
@@ -XXX,XX +XXX,XX @@ const PropertyInfo qdev_prop_losttickpolicy = {
60
61
/* --- blocksize --- */
62
63
-/* lower limit is sector size */
64
-#define MIN_BLOCK_SIZE 512
65
-#define MIN_BLOCK_SIZE_STR "512 B"
66
-/*
67
- * upper limit is arbitrary, 2 MiB looks sufficient for all sensible uses, and
68
- * matches qcow2 cluster size limit
69
- */
70
-#define MAX_BLOCK_SIZE (2 * MiB)
71
-#define MAX_BLOCK_SIZE_STR "2 MiB"
72
-
73
static void set_blocksize(Object *obj, Visitor *v, const char *name,
74
void *opaque, Error **errp)
75
{
76
@@ -XXX,XX +XXX,XX @@ static void set_blocksize(Object *obj, Visitor *v, const char *name,
77
Property *prop = opaque;
78
uint32_t *ptr = qdev_get_prop_ptr(dev, prop);
79
uint64_t value;
80
+ Error *local_err = NULL;
81
82
if (dev->realized) {
83
qdev_prop_set_after_realize(dev, name, errp);
84
@@ -XXX,XX +XXX,XX @@ static void set_blocksize(Object *obj, Visitor *v, const char *name,
85
if (!visit_type_size(v, name, &value, errp)) {
86
return;
87
}
88
- /* value of 0 means "unset" */
89
- if (value && (value < MIN_BLOCK_SIZE || value > MAX_BLOCK_SIZE)) {
90
- error_setg(errp,
91
- "Property %s.%s doesn't take value %" PRIu64
92
- " (minimum: " MIN_BLOCK_SIZE_STR
93
- ", maximum: " MAX_BLOCK_SIZE_STR ")",
94
- dev->id ? : "", name, value);
95
+ check_block_size(dev->id ? : "", name, value, &local_err);
96
+ if (local_err) {
97
+ error_propagate(errp, local_err);
98
return;
99
}
100
-
101
- /* We rely on power-of-2 blocksizes for bitmasks */
102
- if ((value & (value - 1)) != 0) {
103
- error_setg(errp,
104
- "Property %s.%s doesn't take value '%" PRId64 "', "
105
- "it's not a power of 2", dev->id ?: "", name, (int64_t)value);
106
- return;
107
- }
108
-
109
*ptr = value;
110
}
111
112
diff --git a/util/block-helpers.c b/util/block-helpers.c
113
new file mode 100644
114
index XXXXXXX..XXXXXXX
115
--- /dev/null
116
+++ b/util/block-helpers.c
117
@@ -XXX,XX +XXX,XX @@
37
@@ -XXX,XX +XXX,XX @@
118
+/*
38
+/*
119
+ * Block utility functions
39
+ * QTest testcase for UFS
120
+ *
40
+ *
121
+ * Copyright IBM, Corp. 2011
41
+ * Copyright (c) 2023 Samsung Electronics Co., Ltd. All rights reserved.
122
+ * Copyright (c) 2020 Coiby Xu <coiby.xu@gmail.com>
123
+ *
42
+ *
124
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
43
+ * SPDX-License-Identifier: GPL-2.0-or-later
125
+ * See the COPYING file in the top-level directory.
126
+ */
44
+ */
127
+
45
+
128
+#include "qemu/osdep.h"
46
+#include "qemu/osdep.h"
129
+#include "qapi/error.h"
47
+#include "qemu/module.h"
130
+#include "qapi/qmp/qerror.h"
48
+#include "qemu/units.h"
131
+#include "block-helpers.h"
49
+#include "libqtest.h"
50
+#include "libqos/qgraph.h"
51
+#include "libqos/pci.h"
52
+#include "scsi/constants.h"
53
+#include "include/block/ufs.h"
54
+
55
+/* Test images sizes in Bytes */
56
+#define TEST_IMAGE_SIZE (64 * 1024 * 1024)
57
+/* Timeout for various operations, in seconds. */
58
+#define TIMEOUT_SECONDS 10
59
+/* Maximum PRD entry count */
60
+#define MAX_PRD_ENTRY_COUNT 10
61
+#define PRD_ENTRY_DATA_SIZE 4096
62
+/* Constants to build upiu */
63
+#define UTP_COMMAND_DESCRIPTOR_SIZE 4096
64
+#define UTP_RESPONSE_UPIU_OFFSET 1024
65
+#define UTP_PRDT_UPIU_OFFSET 2048
66
+
67
+typedef struct QUfs QUfs;
68
+
69
+struct QUfs {
70
+ QOSGraphObject obj;
71
+ QPCIDevice dev;
72
+ QPCIBar bar;
73
+
74
+ uint64_t utrlba;
75
+ uint64_t utmrlba;
76
+ uint64_t cmd_desc_addr;
77
+ uint64_t data_buffer_addr;
78
+
79
+ bool enabled;
80
+};
81
+
82
+static inline uint32_t ufs_rreg(QUfs *ufs, size_t offset)
83
+{
84
+ return qpci_io_readl(&ufs->dev, ufs->bar, offset);
85
+}
86
+
87
+static inline void ufs_wreg(QUfs *ufs, size_t offset, uint32_t value)
88
+{
89
+ qpci_io_writel(&ufs->dev, ufs->bar, offset, value);
90
+}
91
+
92
+static void ufs_wait_for_irq(QUfs *ufs)
93
+{
94
+ uint64_t end_time;
95
+ uint32_t is;
96
+ /* Wait for device to reset as the linux driver does. */
97
+ end_time = g_get_monotonic_time() + TIMEOUT_SECONDS * G_TIME_SPAN_SECOND;
98
+ do {
99
+ qtest_clock_step(ufs->dev.bus->qts, 100);
100
+ is = ufs_rreg(ufs, A_IS);
101
+ } while (is == 0 && g_get_monotonic_time() < end_time);
102
+}
103
+
104
+static UtpTransferReqDesc ufs_build_req_utrd(uint64_t cmd_desc_addr,
105
+ uint8_t slot,
106
+ uint32_t data_direction,
107
+ uint16_t prd_table_length)
108
+{
109
+ UtpTransferReqDesc req = { 0 };
110
+ uint64_t command_desc_base_addr =
111
+ cmd_desc_addr + slot * UTP_COMMAND_DESCRIPTOR_SIZE;
112
+
113
+ req.header.dword_0 =
114
+ cpu_to_le32(1 << 28 | data_direction | UTP_REQ_DESC_INT_CMD);
115
+ req.header.dword_2 = cpu_to_le32(OCS_INVALID_COMMAND_STATUS);
116
+
117
+ req.command_desc_base_addr_hi = cpu_to_le32(command_desc_base_addr >> 32);
118
+ req.command_desc_base_addr_lo =
119
+ cpu_to_le32(command_desc_base_addr & 0xffffffff);
120
+ req.response_upiu_offset =
121
+ cpu_to_le16(UTP_RESPONSE_UPIU_OFFSET / sizeof(uint32_t));
122
+ req.response_upiu_length = cpu_to_le16(sizeof(UtpUpiuRsp));
123
+ req.prd_table_offset = cpu_to_le16(UTP_PRDT_UPIU_OFFSET / sizeof(uint32_t));
124
+ req.prd_table_length = cpu_to_le16(prd_table_length);
125
+ return req;
126
+}
127
+
128
+static void ufs_send_nop_out(QUfs *ufs, uint8_t slot,
129
+ UtpTransferReqDesc *utrd_out, UtpUpiuRsp *rsp_out)
130
+{
131
+ /* Build up utp transfer request descriptor */
132
+ UtpTransferReqDesc utrd =
133
+ ufs_build_req_utrd(ufs->cmd_desc_addr, slot, UTP_NO_DATA_TRANSFER, 0);
134
+ uint64_t utrd_addr = ufs->utrlba + slot * sizeof(UtpTransferReqDesc);
135
+ uint64_t req_upiu_addr =
136
+ ufs->cmd_desc_addr + slot * UTP_COMMAND_DESCRIPTOR_SIZE;
137
+ uint64_t rsp_upiu_addr = req_upiu_addr + UTP_RESPONSE_UPIU_OFFSET;
138
+ qtest_memwrite(ufs->dev.bus->qts, utrd_addr, &utrd, sizeof(utrd));
139
+
140
+ /* Build up request upiu */
141
+ UtpUpiuReq req_upiu = { 0 };
142
+ req_upiu.header.trans_type = UPIU_TRANSACTION_NOP_OUT;
143
+ req_upiu.header.task_tag = slot;
144
+ qtest_memwrite(ufs->dev.bus->qts, req_upiu_addr, &req_upiu,
145
+ sizeof(req_upiu));
146
+
147
+ /* Ring Doorbell */
148
+ ufs_wreg(ufs, A_UTRLDBR, 1);
149
+ ufs_wait_for_irq(ufs);
150
+ g_assert_true(FIELD_EX32(ufs_rreg(ufs, A_IS), IS, UTRCS));
151
+ ufs_wreg(ufs, A_IS, FIELD_DP32(0, IS, UTRCS, 1));
152
+
153
+ qtest_memread(ufs->dev.bus->qts, utrd_addr, utrd_out, sizeof(*utrd_out));
154
+ qtest_memread(ufs->dev.bus->qts, rsp_upiu_addr, rsp_out, sizeof(*rsp_out));
155
+}
156
+
157
+static void ufs_send_query(QUfs *ufs, uint8_t slot, uint8_t query_function,
158
+ uint8_t query_opcode, uint8_t idn, uint8_t index,
159
+ UtpTransferReqDesc *utrd_out, UtpUpiuRsp *rsp_out)
160
+{
161
+ /* Build up utp transfer request descriptor */
162
+ UtpTransferReqDesc utrd =
163
+ ufs_build_req_utrd(ufs->cmd_desc_addr, slot, UTP_NO_DATA_TRANSFER, 0);
164
+ uint64_t utrd_addr = ufs->utrlba + slot * sizeof(UtpTransferReqDesc);
165
+ uint64_t req_upiu_addr =
166
+ ufs->cmd_desc_addr + slot * UTP_COMMAND_DESCRIPTOR_SIZE;
167
+ uint64_t rsp_upiu_addr = req_upiu_addr + UTP_RESPONSE_UPIU_OFFSET;
168
+ qtest_memwrite(ufs->dev.bus->qts, utrd_addr, &utrd, sizeof(utrd));
169
+
170
+ /* Build up request upiu */
171
+ UtpUpiuReq req_upiu = { 0 };
172
+ req_upiu.header.trans_type = UPIU_TRANSACTION_QUERY_REQ;
173
+ req_upiu.header.query_func = query_function;
174
+ req_upiu.header.task_tag = slot;
175
+ /*
176
+ * QEMU UFS does not currently support Write descriptor and Write attribute,
177
+ * so the value of data_segment_length is always 0.
178
+ */
179
+ req_upiu.header.data_segment_length = 0;
180
+ req_upiu.qr.opcode = query_opcode;
181
+ req_upiu.qr.idn = idn;
182
+ req_upiu.qr.index = index;
183
+ qtest_memwrite(ufs->dev.bus->qts, req_upiu_addr, &req_upiu,
184
+ sizeof(req_upiu));
185
+
186
+ /* Ring Doorbell */
187
+ ufs_wreg(ufs, A_UTRLDBR, 1);
188
+ ufs_wait_for_irq(ufs);
189
+ g_assert_true(FIELD_EX32(ufs_rreg(ufs, A_IS), IS, UTRCS));
190
+ ufs_wreg(ufs, A_IS, FIELD_DP32(0, IS, UTRCS, 1));
191
+
192
+ qtest_memread(ufs->dev.bus->qts, utrd_addr, utrd_out, sizeof(*utrd_out));
193
+ qtest_memread(ufs->dev.bus->qts, rsp_upiu_addr, rsp_out, sizeof(*rsp_out));
194
+}
195
+
196
+static void ufs_send_scsi_command(QUfs *ufs, uint8_t slot, uint8_t lun,
197
+ const uint8_t *cdb, const uint8_t *data_in,
198
+ size_t data_in_len, uint8_t *data_out,
199
+ size_t data_out_len,
200
+ UtpTransferReqDesc *utrd_out,
201
+ UtpUpiuRsp *rsp_out)
202
+
203
+{
204
+ /* Build up PRDT */
205
+ UfshcdSgEntry entries[MAX_PRD_ENTRY_COUNT] = {
206
+ 0,
207
+ };
208
+ uint8_t flags;
209
+ uint16_t prd_table_length, i;
210
+ uint32_t data_direction, data_len;
211
+ uint64_t req_upiu_addr =
212
+ ufs->cmd_desc_addr + slot * UTP_COMMAND_DESCRIPTOR_SIZE;
213
+ uint64_t prdt_addr = req_upiu_addr + UTP_PRDT_UPIU_OFFSET;
214
+
215
+ g_assert_true(data_in_len < MAX_PRD_ENTRY_COUNT * PRD_ENTRY_DATA_SIZE);
216
+ g_assert_true(data_out_len < MAX_PRD_ENTRY_COUNT * PRD_ENTRY_DATA_SIZE);
217
+ if (data_in_len > 0) {
218
+ g_assert_nonnull(data_in);
219
+ data_direction = UTP_HOST_TO_DEVICE;
220
+ data_len = data_in_len;
221
+ flags = UPIU_CMD_FLAGS_WRITE;
222
+ } else if (data_out_len > 0) {
223
+ g_assert_nonnull(data_out);
224
+ data_direction = UTP_DEVICE_TO_HOST;
225
+ data_len = data_out_len;
226
+ flags = UPIU_CMD_FLAGS_READ;
227
+ } else {
228
+ data_direction = UTP_NO_DATA_TRANSFER;
229
+ data_len = 0;
230
+ flags = UPIU_CMD_FLAGS_NONE;
231
+ }
232
+ prd_table_length = DIV_ROUND_UP(data_len, PRD_ENTRY_DATA_SIZE);
233
+
234
+ qtest_memset(ufs->dev.bus->qts, ufs->data_buffer_addr, 0,
235
+ MAX_PRD_ENTRY_COUNT * PRD_ENTRY_DATA_SIZE);
236
+ if (data_in_len) {
237
+ qtest_memwrite(ufs->dev.bus->qts, ufs->data_buffer_addr, data_in,
238
+ data_in_len);
239
+ }
240
+
241
+ for (i = 0; i < prd_table_length; i++) {
242
+ entries[i].addr =
243
+ cpu_to_le64(ufs->data_buffer_addr + i * sizeof(UfshcdSgEntry));
244
+ if (i + 1 != prd_table_length) {
245
+ entries[i].size = cpu_to_le32(PRD_ENTRY_DATA_SIZE - 1);
246
+ } else {
247
+ entries[i].size = cpu_to_le32(
248
+ data_len - (PRD_ENTRY_DATA_SIZE * (prd_table_length - 1)) - 1);
249
+ }
250
+ }
251
+ qtest_memwrite(ufs->dev.bus->qts, prdt_addr, entries,
252
+ prd_table_length * sizeof(UfshcdSgEntry));
253
+
254
+ /* Build up utp transfer request descriptor */
255
+ UtpTransferReqDesc utrd = ufs_build_req_utrd(
256
+ ufs->cmd_desc_addr, slot, data_direction, prd_table_length);
257
+ uint64_t utrd_addr = ufs->utrlba + slot * sizeof(UtpTransferReqDesc);
258
+ uint64_t rsp_upiu_addr = req_upiu_addr + UTP_RESPONSE_UPIU_OFFSET;
259
+ qtest_memwrite(ufs->dev.bus->qts, utrd_addr, &utrd, sizeof(utrd));
260
+
261
+ /* Build up request upiu */
262
+ UtpUpiuReq req_upiu = { 0 };
263
+ req_upiu.header.trans_type = UPIU_TRANSACTION_COMMAND;
264
+ req_upiu.header.flags = flags;
265
+ req_upiu.header.lun = lun;
266
+ req_upiu.header.task_tag = slot;
267
+ req_upiu.sc.exp_data_transfer_len = cpu_to_be32(data_len);
268
+ memcpy(req_upiu.sc.cdb, cdb, UFS_CDB_SIZE);
269
+ qtest_memwrite(ufs->dev.bus->qts, req_upiu_addr, &req_upiu,
270
+ sizeof(req_upiu));
271
+
272
+ /* Ring Doorbell */
273
+ ufs_wreg(ufs, A_UTRLDBR, 1);
274
+ ufs_wait_for_irq(ufs);
275
+ g_assert_true(FIELD_EX32(ufs_rreg(ufs, A_IS), IS, UTRCS));
276
+ ufs_wreg(ufs, A_IS, FIELD_DP32(0, IS, UTRCS, 1));
277
+
278
+ qtest_memread(ufs->dev.bus->qts, utrd_addr, utrd_out, sizeof(*utrd_out));
279
+ qtest_memread(ufs->dev.bus->qts, rsp_upiu_addr, rsp_out, sizeof(*rsp_out));
280
+ if (data_out_len) {
281
+ qtest_memread(ufs->dev.bus->qts, ufs->data_buffer_addr, data_out,
282
+ data_out_len);
283
+ }
284
+}
132
+
285
+
133
+/**
286
+/**
134
+ * check_block_size:
287
+ * Initialize Ufs host controller and logical unit.
135
+ * @id: The unique ID of the object
288
+ * After running this function, you can make a transfer request to the UFS.
136
+ * @name: The name of the property being validated
137
+ * @value: The block size in bytes
138
+ * @errp: A pointer to an area to store an error
139
+ *
140
+ * This function checks that the block size meets the following conditions:
141
+ * 1. At least MIN_BLOCK_SIZE
142
+ * 2. No larger than MAX_BLOCK_SIZE
143
+ * 3. A power of 2
144
+ */
289
+ */
145
+void check_block_size(const char *id, const char *name, int64_t value,
290
+static void ufs_init(QUfs *ufs, QGuestAllocator *alloc)
146
+ Error **errp)
291
+{
147
+{
292
+ uint64_t end_time;
148
+ /* value of 0 means "unset" */
293
+ uint32_t nutrs, nutmrs;
149
+ if (value && (value < MIN_BLOCK_SIZE || value > MAX_BLOCK_SIZE)) {
294
+ uint32_t hcs, is, ucmdarg2, cap;
150
+ error_setg(errp, QERR_PROPERTY_VALUE_OUT_OF_RANGE,
295
+ uint32_t hce = 0, ie = 0;
151
+ id, name, value, MIN_BLOCK_SIZE, MAX_BLOCK_SIZE);
296
+ UtpTransferReqDesc utrd;
297
+ UtpUpiuRsp rsp_upiu;
298
+
299
+ ufs->bar = qpci_iomap(&ufs->dev, 0, NULL);
300
+ qpci_device_enable(&ufs->dev);
301
+
302
+ /* Start host controller initialization */
303
+ hce = FIELD_DP32(hce, HCE, HCE, 1);
304
+ ufs_wreg(ufs, A_HCE, hce);
305
+
306
+ /* Wait for device to reset */
307
+ end_time = g_get_monotonic_time() + TIMEOUT_SECONDS * G_TIME_SPAN_SECOND;
308
+ do {
309
+ qtest_clock_step(ufs->dev.bus->qts, 100);
310
+ hce = FIELD_EX32(ufs_rreg(ufs, A_HCE), HCE, HCE);
311
+ } while (hce == 0 && g_get_monotonic_time() < end_time);
312
+ g_assert_cmpuint(hce, ==, 1);
313
+
314
+ /* Enable interrupt */
315
+ ie = FIELD_DP32(ie, IE, UCCE, 1);
316
+ ie = FIELD_DP32(ie, IE, UHESE, 1);
317
+ ie = FIELD_DP32(ie, IE, UHXSE, 1);
318
+ ie = FIELD_DP32(ie, IE, UPMSE, 1);
319
+ ufs_wreg(ufs, A_IE, ie);
320
+
321
+ /* Send DME_LINK_STARTUP uic command */
322
+ hcs = ufs_rreg(ufs, A_HCS);
323
+ g_assert_true(FIELD_EX32(hcs, HCS, UCRDY));
324
+
325
+ ufs_wreg(ufs, A_UCMDARG1, 0);
326
+ ufs_wreg(ufs, A_UCMDARG2, 0);
327
+ ufs_wreg(ufs, A_UCMDARG3, 0);
328
+ ufs_wreg(ufs, A_UICCMD, UIC_CMD_DME_LINK_STARTUP);
329
+
330
+ is = ufs_rreg(ufs, A_IS);
331
+ g_assert_true(FIELD_EX32(is, IS, UCCS));
332
+ ufs_wreg(ufs, A_IS, FIELD_DP32(0, IS, UCCS, 1));
333
+
334
+ ucmdarg2 = ufs_rreg(ufs, A_UCMDARG2);
335
+ g_assert_cmpuint(ucmdarg2, ==, 0);
336
+ is = ufs_rreg(ufs, A_IS);
337
+ g_assert_cmpuint(is, ==, 0);
338
+ hcs = ufs_rreg(ufs, A_HCS);
339
+ g_assert_true(FIELD_EX32(hcs, HCS, DP));
340
+ g_assert_true(FIELD_EX32(hcs, HCS, UTRLRDY));
341
+ g_assert_true(FIELD_EX32(hcs, HCS, UTMRLRDY));
342
+ g_assert_true(FIELD_EX32(hcs, HCS, UCRDY));
343
+
344
+ /* Enable all interrupt functions */
345
+ ie = FIELD_DP32(ie, IE, UTRCE, 1);
346
+ ie = FIELD_DP32(ie, IE, UEE, 1);
347
+ ie = FIELD_DP32(ie, IE, UPMSE, 1);
348
+ ie = FIELD_DP32(ie, IE, UHXSE, 1);
349
+ ie = FIELD_DP32(ie, IE, UHESE, 1);
350
+ ie = FIELD_DP32(ie, IE, UTMRCE, 1);
351
+ ie = FIELD_DP32(ie, IE, UCCE, 1);
352
+ ie = FIELD_DP32(ie, IE, DFEE, 1);
353
+ ie = FIELD_DP32(ie, IE, HCFEE, 1);
354
+ ie = FIELD_DP32(ie, IE, SBFEE, 1);
355
+ ie = FIELD_DP32(ie, IE, CEFEE, 1);
356
+ ufs_wreg(ufs, A_IE, ie);
357
+ ufs_wreg(ufs, A_UTRIACR, 0);
358
+
359
+ /* Enable tranfer request and task management request */
360
+ cap = ufs_rreg(ufs, A_CAP);
361
+ nutrs = FIELD_EX32(cap, CAP, NUTRS) + 1;
362
+ nutmrs = FIELD_EX32(cap, CAP, NUTMRS) + 1;
363
+ ufs->cmd_desc_addr =
364
+ guest_alloc(alloc, nutrs * UTP_COMMAND_DESCRIPTOR_SIZE);
365
+ ufs->data_buffer_addr =
366
+ guest_alloc(alloc, MAX_PRD_ENTRY_COUNT * PRD_ENTRY_DATA_SIZE);
367
+ ufs->utrlba = guest_alloc(alloc, nutrs * sizeof(UtpTransferReqDesc));
368
+ ufs->utmrlba = guest_alloc(alloc, nutmrs * sizeof(UtpTaskReqDesc));
369
+
370
+ ufs_wreg(ufs, A_UTRLBA, ufs->utrlba & 0xffffffff);
371
+ ufs_wreg(ufs, A_UTRLBAU, ufs->utrlba >> 32);
372
+ ufs_wreg(ufs, A_UTMRLBA, ufs->utmrlba & 0xffffffff);
373
+ ufs_wreg(ufs, A_UTMRLBAU, ufs->utmrlba >> 32);
374
+ ufs_wreg(ufs, A_UTRLRSR, 1);
375
+ ufs_wreg(ufs, A_UTMRLRSR, 1);
376
+
377
+ /* Send nop out to test transfer request */
378
+ ufs_send_nop_out(ufs, 0, &utrd, &rsp_upiu);
379
+ g_assert_cmpuint(le32_to_cpu(utrd.header.dword_2), ==, OCS_SUCCESS);
380
+
381
+ /* Set fDeviceInit flag via query request */
382
+ ufs_send_query(ufs, 0, UPIU_QUERY_FUNC_STANDARD_WRITE_REQUEST,
383
+ UPIU_QUERY_OPCODE_SET_FLAG, QUERY_FLAG_IDN_FDEVICEINIT, 0,
384
+ &utrd, &rsp_upiu);
385
+ g_assert_cmpuint(le32_to_cpu(utrd.header.dword_2), ==, OCS_SUCCESS);
386
+
387
+ /* Wait for device to reset */
388
+ end_time = g_get_monotonic_time() + TIMEOUT_SECONDS * G_TIME_SPAN_SECOND;
389
+ do {
390
+ qtest_clock_step(ufs->dev.bus->qts, 100);
391
+ ufs_send_query(ufs, 0, UPIU_QUERY_FUNC_STANDARD_READ_REQUEST,
392
+ UPIU_QUERY_OPCODE_READ_FLAG, QUERY_FLAG_IDN_FDEVICEINIT,
393
+ 0, &utrd, &rsp_upiu);
394
+ } while (be32_to_cpu(rsp_upiu.qr.value) != 0 &&
395
+ g_get_monotonic_time() < end_time);
396
+ g_assert_cmpuint(be32_to_cpu(rsp_upiu.qr.value), ==, 0);
397
+
398
+ ufs->enabled = true;
399
+}
400
+
401
+static void ufs_exit(QUfs *ufs, QGuestAllocator *alloc)
402
+{
403
+ if (ufs->enabled) {
404
+ guest_free(alloc, ufs->utrlba);
405
+ guest_free(alloc, ufs->utmrlba);
406
+ guest_free(alloc, ufs->cmd_desc_addr);
407
+ guest_free(alloc, ufs->data_buffer_addr);
408
+ }
409
+
410
+ qpci_iounmap(&ufs->dev, ufs->bar);
411
+}
412
+
413
+static void *ufs_get_driver(void *obj, const char *interface)
414
+{
415
+ QUfs *ufs = obj;
416
+
417
+ if (!g_strcmp0(interface, "pci-device")) {
418
+ return &ufs->dev;
419
+ }
420
+
421
+ fprintf(stderr, "%s not present in ufs\n", interface);
422
+ g_assert_not_reached();
423
+}
424
+
425
+static void *ufs_create(void *pci_bus, QGuestAllocator *alloc, void *addr)
426
+{
427
+ QUfs *ufs = g_new0(QUfs, 1);
428
+ QPCIBus *bus = pci_bus;
429
+
430
+ qpci_device_init(&ufs->dev, bus, addr);
431
+ ufs->obj.get_driver = ufs_get_driver;
432
+
433
+ return &ufs->obj;
434
+}
435
+
436
+static void ufstest_reg_read(void *obj, void *data, QGuestAllocator *alloc)
437
+{
438
+ QUfs *ufs = obj;
439
+ uint32_t cap;
440
+
441
+ ufs->bar = qpci_iomap(&ufs->dev, 0, NULL);
442
+ qpci_device_enable(&ufs->dev);
443
+
444
+ cap = ufs_rreg(ufs, A_CAP);
445
+ g_assert_cmpuint(FIELD_EX32(cap, CAP, NUTRS), ==, 31);
446
+ g_assert_cmpuint(FIELD_EX32(cap, CAP, NUTMRS), ==, 7);
447
+ g_assert_cmpuint(FIELD_EX32(cap, CAP, 64AS), ==, 1);
448
+
449
+ qpci_iounmap(&ufs->dev, ufs->bar);
450
+}
451
+
452
+static void ufstest_init(void *obj, void *data, QGuestAllocator *alloc)
453
+{
454
+ QUfs *ufs = obj;
455
+
456
+ uint8_t buf[4096] = { 0 };
457
+ const uint8_t report_luns_cdb[UFS_CDB_SIZE] = {
458
+ /* allocation length 4096 */
459
+ REPORT_LUNS, 0x00, 0x00, 0x00, 0x00, 0x00,
460
+ 0x00, 0x00, 0x10, 0x00, 0x00, 0x00
461
+ };
462
+ const uint8_t test_unit_ready_cdb[UFS_CDB_SIZE] = {
463
+ TEST_UNIT_READY,
464
+ };
465
+ UtpTransferReqDesc utrd;
466
+ UtpUpiuRsp rsp_upiu;
467
+
468
+ ufs_init(ufs, alloc);
469
+
470
+ /* Check REPORT_LUNS */
471
+ ufs_send_scsi_command(ufs, 0, 0, report_luns_cdb, NULL, 0, buf, sizeof(buf),
472
+ &utrd, &rsp_upiu);
473
+ g_assert_cmpuint(le32_to_cpu(utrd.header.dword_2), ==, OCS_SUCCESS);
474
+ g_assert_cmpuint(rsp_upiu.header.scsi_status, ==, GOOD);
475
+ /* LUN LIST LENGTH should be 8, in big endian */
476
+ g_assert_cmpuint(buf[3], ==, 8);
477
+ /* There is one logical unit whose lun is 0 */
478
+ g_assert_cmpuint(buf[9], ==, 0);
479
+
480
+ /* Check TEST_UNIT_READY */
481
+ ufs_send_scsi_command(ufs, 0, 0, test_unit_ready_cdb, NULL, 0, NULL, 0,
482
+ &utrd, &rsp_upiu);
483
+ g_assert_cmpuint(le32_to_cpu(utrd.header.dword_2), ==, OCS_SUCCESS);
484
+ g_assert_cmpuint(rsp_upiu.header.scsi_status, ==, GOOD);
485
+
486
+ ufs_exit(ufs, alloc);
487
+}
488
+
489
+static void ufstest_read_write(void *obj, void *data, QGuestAllocator *alloc)
490
+{
491
+ QUfs *ufs = obj;
492
+ uint8_t read_buf[4096] = { 0 };
493
+ uint8_t write_buf[4096] = { 0 };
494
+ const uint8_t read_capacity_cdb[UFS_CDB_SIZE] = {
495
+ /* allocation length 4096 */
496
+ SERVICE_ACTION_IN_16,
497
+ SAI_READ_CAPACITY_16,
498
+ 0x00,
499
+ 0x00,
500
+ 0x00,
501
+ 0x00,
502
+ 0x00,
503
+ 0x00,
504
+ 0x00,
505
+ 0x00,
506
+ 0x00,
507
+ 0x00,
508
+ 0x10,
509
+ 0x00,
510
+ 0x00,
511
+ 0x00
512
+ };
513
+ const uint8_t read_cdb[UFS_CDB_SIZE] = {
514
+ /* READ(10) to LBA 0, transfer length 1 */
515
+ READ_10, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, 0x00
516
+ };
517
+ const uint8_t write_cdb[UFS_CDB_SIZE] = {
518
+ /* WRITE(10) to LBA 0, transfer length 1 */
519
+ WRITE_10, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, 0x00
520
+ };
521
+ uint32_t block_size;
522
+ UtpTransferReqDesc utrd;
523
+ UtpUpiuRsp rsp_upiu;
524
+
525
+ ufs_init(ufs, alloc);
526
+
527
+ /* Read capacity */
528
+ ufs_send_scsi_command(ufs, 0, 1, read_capacity_cdb, NULL, 0, read_buf,
529
+ sizeof(read_buf), &utrd, &rsp_upiu);
530
+ g_assert_cmpuint(le32_to_cpu(utrd.header.dword_2), ==, OCS_SUCCESS);
531
+ g_assert_cmpuint(rsp_upiu.header.scsi_status, ==, COMMAND_RESULT_SUCESS);
532
+ block_size = ldl_be_p(&read_buf[8]);
533
+ g_assert_cmpuint(block_size, ==, 4096);
534
+
535
+ /* Write data */
536
+ memset(write_buf, rand() % 255 + 1, block_size);
537
+ ufs_send_scsi_command(ufs, 0, 1, write_cdb, write_buf, block_size, NULL, 0,
538
+ &utrd, &rsp_upiu);
539
+ g_assert_cmpuint(le32_to_cpu(utrd.header.dword_2), ==, OCS_SUCCESS);
540
+ g_assert_cmpuint(rsp_upiu.header.scsi_status, ==, COMMAND_RESULT_SUCESS);
541
+
542
+ /* Read data and verify */
543
+ ufs_send_scsi_command(ufs, 0, 1, read_cdb, NULL, 0, read_buf, block_size,
544
+ &utrd, &rsp_upiu);
545
+ g_assert_cmpuint(le32_to_cpu(utrd.header.dword_2), ==, OCS_SUCCESS);
546
+ g_assert_cmpuint(rsp_upiu.header.scsi_status, ==, COMMAND_RESULT_SUCESS);
547
+ g_assert_cmpint(memcmp(read_buf, write_buf, block_size), ==, 0);
548
+
549
+ ufs_exit(ufs, alloc);
550
+}
551
+
552
+static void drive_destroy(void *path)
553
+{
554
+ unlink(path);
555
+ g_free(path);
556
+ qos_invalidate_command_line();
557
+}
558
+
559
+static char *drive_create(void)
560
+{
561
+ int fd, ret;
562
+ char *t_path;
563
+
564
+ /* Create a temporary raw image */
565
+ fd = g_file_open_tmp("qtest-ufs.XXXXXX", &t_path, NULL);
566
+ g_assert_cmpint(fd, >=, 0);
567
+ ret = ftruncate(fd, TEST_IMAGE_SIZE);
568
+ g_assert_cmpint(ret, ==, 0);
569
+ close(fd);
570
+
571
+ g_test_queue_destroy(drive_destroy, t_path);
572
+ return t_path;
573
+}
574
+
575
+static void *ufs_blk_test_setup(GString *cmd_line, void *arg)
576
+{
577
+ char *tmp_path = drive_create();
578
+
579
+ g_string_append_printf(cmd_line,
580
+ " -blockdev file,filename=%s,node-name=drv1 "
581
+ "-device ufs-lu,bus=ufs0,drive=drv1,lun=1 ",
582
+ tmp_path);
583
+
584
+ return arg;
585
+}
586
+
587
+static void ufs_register_nodes(void)
588
+{
589
+ const char *arch;
590
+ QOSGraphEdgeOptions edge_opts = {
591
+ .before_cmd_line = "-blockdev null-co,node-name=drv0,read-zeroes=on",
592
+ .after_cmd_line = "-device ufs-lu,bus=ufs0,drive=drv0,lun=0",
593
+ .extra_device_opts = "addr=04.0,id=ufs0,nutrs=32,nutmrs=8"
594
+ };
595
+
596
+ QOSGraphTestOptions io_test_opts = {
597
+ .before = ufs_blk_test_setup,
598
+ };
599
+
600
+ add_qpci_address(&edge_opts, &(QPCIAddress){ .devfn = QPCI_DEVFN(4, 0) });
601
+
602
+ qos_node_create_driver("ufs", ufs_create);
603
+ qos_node_consumes("ufs", "pci-bus", &edge_opts);
604
+ qos_node_produces("ufs", "pci-device");
605
+
606
+ qos_add_test("reg-read", "ufs", ufstest_reg_read, NULL);
607
+
608
+ /*
609
+ * Check architecture
610
+ * TODO: Enable ufs io tests for ppc64
611
+ */
612
+ arch = qtest_get_arch();
613
+ if (!strcmp(arch, "ppc64")) {
614
+ g_test_message("Skipping ufs io tests for ppc64");
152
+ return;
615
+ return;
153
+ }
616
+ }
154
+
617
+ qos_add_test("init", "ufs", ufstest_init, NULL);
155
+ /* We rely on power-of-2 blocksizes for bitmasks */
618
+ qos_add_test("read-write", "ufs", ufstest_read_write, &io_test_opts);
156
+ if ((value & (value - 1)) != 0) {
619
+}
157
+ error_setg(errp,
620
+
158
+ "Property %s.%s doesn't take value '%" PRId64
621
+libqos_init(ufs_register_nodes);
159
+ "', it's not a power of 2",
622
diff --git a/tests/qtest/meson.build b/tests/qtest/meson.build
160
+ id, name, value);
161
+ return;
162
+ }
163
+}
164
diff --git a/util/meson.build b/util/meson.build
165
index XXXXXXX..XXXXXXX 100644
623
index XXXXXXX..XXXXXXX 100644
166
--- a/util/meson.build
624
--- a/tests/qtest/meson.build
167
+++ b/util/meson.build
625
+++ b/tests/qtest/meson.build
168
@@ -XXX,XX +XXX,XX @@ if have_block
626
@@ -XXX,XX +XXX,XX @@ qos_test_ss.add(
169
util_ss.add(files('nvdimm-utils.c'))
627
'virtio-iommu-test.c',
170
util_ss.add(files('qemu-coroutine.c', 'qemu-coroutine-lock.c', 'qemu-coroutine-io.c'))
628
'vmxnet3-test.c',
171
util_ss.add(when: 'CONFIG_LINUX', if_true: files('vhost-user-server.c'))
629
'igb-test.c',
172
+ util_ss.add(files('block-helpers.c'))
630
+ 'ufs-test.c',
173
util_ss.add(files('qemu-coroutine-sleep.c'))
631
)
174
util_ss.add(files('qemu-co-shared-resource.c'))
632
175
util_ss.add(files('thread-pool.c', 'qemu-timer.c'))
633
if config_all_devices.has_key('CONFIG_VIRTIO_SERIAL')
176
--
634
--
177
2.26.2
635
2.41.0
178
diff view generated by jsdifflib
Deleted patch
1
From: Coiby Xu <coiby.xu@gmail.com>
2
1
3
Suggested-by: Stefano Garzarella <sgarzare@redhat.com>
4
Signed-off-by: Coiby Xu <coiby.xu@gmail.com>
5
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
6
Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com>
7
Message-id: 20200918080912.321299-8-coiby.xu@gmail.com
8
[Removed reference to vhost-user-blk-test.c, it will be sent in a
9
separate pull request.
10
--Stefan]
11
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
12
---
13
MAINTAINERS | 7 +++++++
14
1 file changed, 7 insertions(+)
15
16
diff --git a/MAINTAINERS b/MAINTAINERS
17
index XXXXXXX..XXXXXXX 100644
18
--- a/MAINTAINERS
19
+++ b/MAINTAINERS
20
@@ -XXX,XX +XXX,XX @@ L: qemu-block@nongnu.org
21
S: Supported
22
F: tests/image-fuzzer/
23
24
+Vhost-user block device backend server
25
+M: Coiby Xu <Coiby.Xu@gmail.com>
26
+S: Maintained
27
+F: block/export/vhost-user-blk-server.c
28
+F: util/vhost-user-server.c
29
+F: tests/qtest/libqos/vhost-user-blk.c
30
+
31
Replication
32
M: Wen Congyang <wencongyang2@huawei.com>
33
M: Xie Changlong <xiechanglong.d@gmail.com>
34
--
35
2.26.2
36
diff view generated by jsdifflib
Deleted patch
1
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
2
Message-id: 20200924151549.913737-3-stefanha@redhat.com
3
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
4
---
5
util/vhost-user-server.c | 2 +-
6
1 file changed, 1 insertion(+), 1 deletion(-)
7
1
8
diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c
9
index XXXXXXX..XXXXXXX 100644
10
--- a/util/vhost-user-server.c
11
+++ b/util/vhost-user-server.c
12
@@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server,
13
return false;
14
}
15
16
- /* zero out unspecified fileds */
17
+ /* zero out unspecified fields */
18
*server = (VuServer) {
19
.listener = listener,
20
.vu_iface = vu_iface,
21
--
22
2.26.2
23
diff view generated by jsdifflib
Deleted patch
1
We already have access to the value with the correct type (ioc and sioc
2
are the same QIOChannel).
3
1
4
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
5
Message-id: 20200924151549.913737-4-stefanha@redhat.com
6
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
7
---
8
util/vhost-user-server.c | 2 +-
9
1 file changed, 1 insertion(+), 1 deletion(-)
10
11
diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c
12
index XXXXXXX..XXXXXXX 100644
13
--- a/util/vhost-user-server.c
14
+++ b/util/vhost-user-server.c
15
@@ -XXX,XX +XXX,XX @@ static void vu_accept(QIONetListener *listener, QIOChannelSocket *sioc,
16
server->ioc = QIO_CHANNEL(sioc);
17
object_ref(OBJECT(server->ioc));
18
qio_channel_attach_aio_context(server->ioc, server->ctx);
19
- qio_channel_set_blocking(QIO_CHANNEL(server->sioc), false, NULL);
20
+ qio_channel_set_blocking(server->ioc, false, NULL);
21
vu_client_start(server);
22
}
23
24
--
25
2.26.2
26
diff view generated by jsdifflib
Deleted patch
1
Explicitly deleting watches is not necessary since libvhost-user calls
2
remove_watch() during vu_deinit(). Add an assertion to check this
3
though.
4
1
5
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
6
Message-id: 20200924151549.913737-5-stefanha@redhat.com
7
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
8
---
9
util/vhost-user-server.c | 19 ++++---------------
10
1 file changed, 4 insertions(+), 15 deletions(-)
11
12
diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c
13
index XXXXXXX..XXXXXXX 100644
14
--- a/util/vhost-user-server.c
15
+++ b/util/vhost-user-server.c
16
@@ -XXX,XX +XXX,XX @@ static void close_client(VuServer *server)
17
/* When this is set vu_client_trip will stop new processing vhost-user message */
18
server->sioc = NULL;
19
20
- VuFdWatch *vu_fd_watch, *next;
21
- QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) {
22
- aio_set_fd_handler(server->ioc->ctx, vu_fd_watch->fd, true, NULL,
23
- NULL, NULL, NULL);
24
- }
25
-
26
- while (!QTAILQ_EMPTY(&server->vu_fd_watches)) {
27
- QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) {
28
- if (!vu_fd_watch->processing) {
29
- QTAILQ_REMOVE(&server->vu_fd_watches, vu_fd_watch, next);
30
- g_free(vu_fd_watch);
31
- }
32
- }
33
- }
34
-
35
while (server->processing_msg) {
36
if (server->ioc->read_coroutine) {
37
server->ioc->read_coroutine = NULL;
38
@@ -XXX,XX +XXX,XX @@ static void close_client(VuServer *server)
39
}
40
41
vu_deinit(&server->vu_dev);
42
+
43
+ /* vu_deinit() should have called remove_watch() */
44
+ assert(QTAILQ_EMPTY(&server->vu_fd_watches));
45
+
46
object_unref(OBJECT(sioc));
47
object_unref(OBJECT(server->ioc));
48
}
49
--
50
2.26.2
51
diff view generated by jsdifflib
1
Only one struct is needed per request. Drop req_data and the separate
1
From: Fabiano Rosas <farosas@suse.de>
2
VuBlockReq instance. Instead let vu_queue_pop() allocate everything at
3
once.
4
2
5
This fixes the req_data memory leak in vu_block_virtio_process_req().
3
We can fail the blk_insert_bs() at init_blk_migration(), leaving the
4
BlkMigDevState without a dirty_bitmap and BlockDriverState. Account
5
for the possibly missing elements when doing cleanup.
6
6
7
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
7
Fix the following crashes:
8
Message-id: 20200924151549.913737-6-stefanha@redhat.com
8
9
Thread 1 "qemu-system-x86" received signal SIGSEGV, Segmentation fault.
10
0x0000555555ec83ef in bdrv_release_dirty_bitmap (bitmap=0x0) at ../block/dirty-bitmap.c:359
11
359 BlockDriverState *bs = bitmap->bs;
12
#0 0x0000555555ec83ef in bdrv_release_dirty_bitmap (bitmap=0x0) at ../block/dirty-bitmap.c:359
13
#1 0x0000555555bba331 in unset_dirty_tracking () at ../migration/block.c:371
14
#2 0x0000555555bbad98 in block_migration_cleanup_bmds () at ../migration/block.c:681
15
16
Thread 1 "qemu-system-x86" received signal SIGSEGV, Segmentation fault.
17
0x0000555555e971ff in bdrv_op_unblock (bs=0x0, op=BLOCK_OP_TYPE_BACKUP_SOURCE, reason=0x0) at ../block.c:7073
18
7073 QLIST_FOREACH_SAFE(blocker, &bs->op_blockers[op], list, next) {
19
#0 0x0000555555e971ff in bdrv_op_unblock (bs=0x0, op=BLOCK_OP_TYPE_BACKUP_SOURCE, reason=0x0) at ../block.c:7073
20
#1 0x0000555555e9734a in bdrv_op_unblock_all (bs=0x0, reason=0x0) at ../block.c:7095
21
#2 0x0000555555bbae13 in block_migration_cleanup_bmds () at ../migration/block.c:690
22
23
Signed-off-by: Fabiano Rosas <farosas@suse.de>
24
Message-id: 20230731203338.27581-1-farosas@suse.de
9
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
25
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
10
---
26
---
11
block/export/vhost-user-blk-server.c | 68 +++++++++-------------------
27
migration/block.c | 11 +++++++++--
12
1 file changed, 21 insertions(+), 47 deletions(-)
28
1 file changed, 9 insertions(+), 2 deletions(-)
13
29
14
diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c
30
diff --git a/migration/block.c b/migration/block.c
15
index XXXXXXX..XXXXXXX 100644
31
index XXXXXXX..XXXXXXX 100644
16
--- a/block/export/vhost-user-blk-server.c
32
--- a/migration/block.c
17
+++ b/block/export/vhost-user-blk-server.c
33
+++ b/migration/block.c
18
@@ -XXX,XX +XXX,XX @@ struct virtio_blk_inhdr {
34
@@ -XXX,XX +XXX,XX @@ static void unset_dirty_tracking(void)
19
};
35
BlkMigDevState *bmds;
20
36
21
typedef struct VuBlockReq {
37
QSIMPLEQ_FOREACH(bmds, &block_mig_state.bmds_list, entry) {
22
- VuVirtqElement *elem;
38
- bdrv_release_dirty_bitmap(bmds->dirty_bitmap);
23
+ VuVirtqElement elem;
39
+ if (bmds->dirty_bitmap) {
24
int64_t sector_num;
40
+ bdrv_release_dirty_bitmap(bmds->dirty_bitmap);
25
size_t size;
41
+ }
26
struct virtio_blk_inhdr *in;
27
@@ -XXX,XX +XXX,XX @@ static void vu_block_req_complete(VuBlockReq *req)
28
VuDev *vu_dev = &req->server->vu_dev;
29
30
/* IO size with 1 extra status byte */
31
- vu_queue_push(vu_dev, req->vq, req->elem, req->size + 1);
32
+ vu_queue_push(vu_dev, req->vq, &req->elem, req->size + 1);
33
vu_queue_notify(vu_dev, req->vq);
34
35
- if (req->elem) {
36
- free(req->elem);
37
- }
38
-
39
- g_free(req);
40
+ free(req);
41
}
42
43
static VuBlockDev *get_vu_block_device_by_server(VuServer *server)
44
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_flush(VuBlockReq *req)
45
blk_co_flush(backend);
46
}
47
48
-struct req_data {
49
- VuServer *server;
50
- VuVirtq *vq;
51
- VuVirtqElement *elem;
52
-};
53
-
54
static void coroutine_fn vu_block_virtio_process_req(void *opaque)
55
{
56
- struct req_data *data = opaque;
57
- VuServer *server = data->server;
58
- VuVirtq *vq = data->vq;
59
- VuVirtqElement *elem = data->elem;
60
+ VuBlockReq *req = opaque;
61
+ VuServer *server = req->server;
62
+ VuVirtqElement *elem = &req->elem;
63
uint32_t type;
64
- VuBlockReq *req;
65
66
VuBlockDev *vdev_blk = get_vu_block_device_by_server(server);
67
BlockBackend *backend = vdev_blk->backend;
68
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque)
69
struct iovec *out_iov = elem->out_sg;
70
unsigned in_num = elem->in_num;
71
unsigned out_num = elem->out_num;
72
+
73
/* refer to hw/block/virtio_blk.c */
74
if (elem->out_num < 1 || elem->in_num < 1) {
75
error_report("virtio-blk request missing headers");
76
- free(elem);
77
- return;
78
+ goto err;
79
}
80
81
- req = g_new0(VuBlockReq, 1);
82
- req->server = server;
83
- req->vq = vq;
84
- req->elem = elem;
85
-
86
if (unlikely(iov_to_buf(out_iov, out_num, 0, &req->out,
87
sizeof(req->out)) != sizeof(req->out))) {
88
error_report("virtio-blk request outhdr too short");
89
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque)
90
91
err:
92
free(elem);
93
- g_free(req);
94
- return;
95
}
96
97
static void vu_block_process_vq(VuDev *vu_dev, int idx)
98
{
99
- VuServer *server;
100
- VuVirtq *vq;
101
- struct req_data *req_data;
102
+ VuServer *server = container_of(vu_dev, VuServer, vu_dev);
103
+ VuVirtq *vq = vu_get_queue(vu_dev, idx);
104
105
- server = container_of(vu_dev, VuServer, vu_dev);
106
- assert(server);
107
-
108
- vq = vu_get_queue(vu_dev, idx);
109
- assert(vq);
110
- VuVirtqElement *elem;
111
while (1) {
112
- elem = vu_queue_pop(vu_dev, vq, sizeof(VuVirtqElement) +
113
- sizeof(VuBlockReq));
114
- if (elem) {
115
- req_data = g_new0(struct req_data, 1);
116
- req_data->server = server;
117
- req_data->vq = vq;
118
- req_data->elem = elem;
119
- Coroutine *co = qemu_coroutine_create(vu_block_virtio_process_req,
120
- req_data);
121
- aio_co_enter(server->ioc->ctx, co);
122
- } else {
123
+ VuBlockReq *req;
124
+
125
+ req = vu_queue_pop(vu_dev, vq, sizeof(VuBlockReq));
126
+ if (!req) {
127
break;
128
}
129
+
130
+ req->server = server;
131
+ req->vq = vq;
132
+
133
+ Coroutine *co =
134
+ qemu_coroutine_create(vu_block_virtio_process_req, req);
135
+ qemu_coroutine_enter(co);
136
}
42
}
137
}
43
}
138
44
45
@@ -XXX,XX +XXX,XX @@ static int64_t get_remaining_dirty(void)
46
static void block_migration_cleanup_bmds(void)
47
{
48
BlkMigDevState *bmds;
49
+ BlockDriverState *bs;
50
AioContext *ctx;
51
52
unset_dirty_tracking();
53
54
while ((bmds = QSIMPLEQ_FIRST(&block_mig_state.bmds_list)) != NULL) {
55
QSIMPLEQ_REMOVE_HEAD(&block_mig_state.bmds_list, entry);
56
- bdrv_op_unblock_all(blk_bs(bmds->blk), bmds->blocker);
57
+
58
+ bs = blk_bs(bmds->blk);
59
+ if (bs) {
60
+ bdrv_op_unblock_all(bs, bmds->blocker);
61
+ }
62
error_free(bmds->blocker);
63
64
/* Save ctx, because bmds->blk can disappear during blk_unref. */
139
--
65
--
140
2.26.2
66
2.41.0
141
diff view generated by jsdifflib
Deleted patch
1
The device panic notifier callback is not used. Drop it.
2
1
3
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
4
Message-id: 20200924151549.913737-7-stefanha@redhat.com
5
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
6
---
7
util/vhost-user-server.h | 3 ---
8
block/export/vhost-user-blk-server.c | 3 +--
9
util/vhost-user-server.c | 6 ------
10
3 files changed, 1 insertion(+), 11 deletions(-)
11
12
diff --git a/util/vhost-user-server.h b/util/vhost-user-server.h
13
index XXXXXXX..XXXXXXX 100644
14
--- a/util/vhost-user-server.h
15
+++ b/util/vhost-user-server.h
16
@@ -XXX,XX +XXX,XX @@ typedef struct VuFdWatch {
17
} VuFdWatch;
18
19
typedef struct VuServer VuServer;
20
-typedef void DevicePanicNotifierFn(VuServer *server);
21
22
struct VuServer {
23
QIONetListener *listener;
24
AioContext *ctx;
25
- DevicePanicNotifierFn *device_panic_notifier;
26
int max_queues;
27
const VuDevIface *vu_iface;
28
VuDev vu_dev;
29
@@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server,
30
SocketAddress *unix_socket,
31
AioContext *ctx,
32
uint16_t max_queues,
33
- DevicePanicNotifierFn *device_panic_notifier,
34
const VuDevIface *vu_iface,
35
Error **errp);
36
37
diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c
38
index XXXXXXX..XXXXXXX 100644
39
--- a/block/export/vhost-user-blk-server.c
40
+++ b/block/export/vhost-user-blk-server.c
41
@@ -XXX,XX +XXX,XX @@ static void vhost_user_blk_server_start(VuBlockDev *vu_block_device,
42
ctx = bdrv_get_aio_context(blk_bs(vu_block_device->backend));
43
44
if (!vhost_user_server_start(&vu_block_device->vu_server, addr, ctx,
45
- VHOST_USER_BLK_MAX_QUEUES,
46
- NULL, &vu_block_iface,
47
+ VHOST_USER_BLK_MAX_QUEUES, &vu_block_iface,
48
errp)) {
49
goto error;
50
}
51
diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c
52
index XXXXXXX..XXXXXXX 100644
53
--- a/util/vhost-user-server.c
54
+++ b/util/vhost-user-server.c
55
@@ -XXX,XX +XXX,XX @@ static void panic_cb(VuDev *vu_dev, const char *buf)
56
close_client(server);
57
}
58
59
- if (server->device_panic_notifier) {
60
- server->device_panic_notifier(server);
61
- }
62
-
63
/*
64
* Set the callback function for network listener so another
65
* vhost-user client can connect to this server
66
@@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server,
67
SocketAddress *socket_addr,
68
AioContext *ctx,
69
uint16_t max_queues,
70
- DevicePanicNotifierFn *device_panic_notifier,
71
const VuDevIface *vu_iface,
72
Error **errp)
73
{
74
@@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server,
75
.vu_iface = vu_iface,
76
.max_queues = max_queues,
77
.ctx = ctx,
78
- .device_panic_notifier = device_panic_notifier,
79
};
80
81
qio_net_listener_set_name(server->listener, "vhost-user-backend-listener");
82
--
83
2.26.2
84
diff view generated by jsdifflib
Deleted patch
1
fds[] is leaked when qio_channel_readv_full() fails.
2
1
3
Use vmsg->fds[] instead of keeping a local fds[] array. Then we can
4
reuse goto fail to clean up fds. vmsg->fd_num must be zeroed before the
5
loop to make this safe.
6
7
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
8
Message-id: 20200924151549.913737-8-stefanha@redhat.com
9
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
10
---
11
util/vhost-user-server.c | 50 ++++++++++++++++++----------------------
12
1 file changed, 23 insertions(+), 27 deletions(-)
13
14
diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c
15
index XXXXXXX..XXXXXXX 100644
16
--- a/util/vhost-user-server.c
17
+++ b/util/vhost-user-server.c
18
@@ -XXX,XX +XXX,XX @@ vu_message_read(VuDev *vu_dev, int conn_fd, VhostUserMsg *vmsg)
19
};
20
int rc, read_bytes = 0;
21
Error *local_err = NULL;
22
- /*
23
- * Store fds/nfds returned from qio_channel_readv_full into
24
- * temporary variables.
25
- *
26
- * VhostUserMsg is a packed structure, gcc will complain about passing
27
- * pointer to a packed structure member if we pass &VhostUserMsg.fd_num
28
- * and &VhostUserMsg.fds directly when calling qio_channel_readv_full,
29
- * thus two temporary variables nfds and fds are used here.
30
- */
31
- size_t nfds = 0, nfds_t = 0;
32
const size_t max_fds = G_N_ELEMENTS(vmsg->fds);
33
- int *fds_t = NULL;
34
VuServer *server = container_of(vu_dev, VuServer, vu_dev);
35
QIOChannel *ioc = server->ioc;
36
37
+ vmsg->fd_num = 0;
38
if (!ioc) {
39
error_report_err(local_err);
40
goto fail;
41
@@ -XXX,XX +XXX,XX @@ vu_message_read(VuDev *vu_dev, int conn_fd, VhostUserMsg *vmsg)
42
43
assert(qemu_in_coroutine());
44
do {
45
+ size_t nfds = 0;
46
+ int *fds = NULL;
47
+
48
/*
49
* qio_channel_readv_full may have short reads, keeping calling it
50
* until getting VHOST_USER_HDR_SIZE or 0 bytes in total
51
*/
52
- rc = qio_channel_readv_full(ioc, &iov, 1, &fds_t, &nfds_t, &local_err);
53
+ rc = qio_channel_readv_full(ioc, &iov, 1, &fds, &nfds, &local_err);
54
if (rc < 0) {
55
if (rc == QIO_CHANNEL_ERR_BLOCK) {
56
+ assert(local_err == NULL);
57
qio_channel_yield(ioc, G_IO_IN);
58
continue;
59
} else {
60
error_report_err(local_err);
61
- return false;
62
+ goto fail;
63
}
64
}
65
- read_bytes += rc;
66
- if (nfds_t > 0) {
67
- if (nfds + nfds_t > max_fds) {
68
+
69
+ if (nfds > 0) {
70
+ if (vmsg->fd_num + nfds > max_fds) {
71
error_report("A maximum of %zu fds are allowed, "
72
"however got %zu fds now",
73
- max_fds, nfds + nfds_t);
74
+ max_fds, vmsg->fd_num + nfds);
75
+ g_free(fds);
76
goto fail;
77
}
78
- memcpy(vmsg->fds + nfds, fds_t,
79
- nfds_t *sizeof(vmsg->fds[0]));
80
- nfds += nfds_t;
81
- g_free(fds_t);
82
+ memcpy(vmsg->fds + vmsg->fd_num, fds, nfds * sizeof(vmsg->fds[0]));
83
+ vmsg->fd_num += nfds;
84
+ g_free(fds);
85
}
86
- if (read_bytes == VHOST_USER_HDR_SIZE || rc == 0) {
87
- break;
88
+
89
+ if (rc == 0) { /* socket closed */
90
+ goto fail;
91
}
92
- iov.iov_base = (char *)vmsg + read_bytes;
93
- iov.iov_len = VHOST_USER_HDR_SIZE - read_bytes;
94
- } while (true);
95
96
- vmsg->fd_num = nfds;
97
+ iov.iov_base += rc;
98
+ iov.iov_len -= rc;
99
+ read_bytes += rc;
100
+ } while (read_bytes != VHOST_USER_HDR_SIZE);
101
+
102
/* qio_channel_readv_full will make socket fds blocking, unblock them */
103
vmsg_unblock_fds(vmsg);
104
if (vmsg->size > sizeof(vmsg->payload)) {
105
--
106
2.26.2
107
diff view generated by jsdifflib
Deleted patch
1
Unexpected EOF is an error that must be reported.
2
1
3
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
4
Message-id: 20200924151549.913737-9-stefanha@redhat.com
5
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
6
---
7
util/vhost-user-server.c | 6 ++++--
8
1 file changed, 4 insertions(+), 2 deletions(-)
9
10
diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c
11
index XXXXXXX..XXXXXXX 100644
12
--- a/util/vhost-user-server.c
13
+++ b/util/vhost-user-server.c
14
@@ -XXX,XX +XXX,XX @@ vu_message_read(VuDev *vu_dev, int conn_fd, VhostUserMsg *vmsg)
15
};
16
if (vmsg->size) {
17
rc = qio_channel_readv_all_eof(ioc, &iov_payload, 1, &local_err);
18
- if (rc == -1) {
19
- error_report_err(local_err);
20
+ if (rc != 1) {
21
+ if (local_err) {
22
+ error_report_err(local_err);
23
+ }
24
goto fail;
25
}
26
}
27
--
28
2.26.2
29
diff view generated by jsdifflib
Deleted patch
1
Propagate the flush return value since errors are possible.
2
1
3
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
4
Message-id: 20200924151549.913737-11-stefanha@redhat.com
5
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
6
---
7
block/export/vhost-user-blk-server.c | 11 +++++++----
8
1 file changed, 7 insertions(+), 4 deletions(-)
9
10
diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c
11
index XXXXXXX..XXXXXXX 100644
12
--- a/block/export/vhost-user-blk-server.c
13
+++ b/block/export/vhost-user-blk-server.c
14
@@ -XXX,XX +XXX,XX @@ vu_block_discard_write_zeroes(VuBlockReq *req, struct iovec *iov,
15
return -EINVAL;
16
}
17
18
-static void coroutine_fn vu_block_flush(VuBlockReq *req)
19
+static int coroutine_fn vu_block_flush(VuBlockReq *req)
20
{
21
VuBlockDev *vdev_blk = get_vu_block_device_by_server(req->server);
22
BlockBackend *backend = vdev_blk->backend;
23
- blk_co_flush(backend);
24
+ return blk_co_flush(backend);
25
}
26
27
static void coroutine_fn vu_block_virtio_process_req(void *opaque)
28
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque)
29
break;
30
}
31
case VIRTIO_BLK_T_FLUSH:
32
- vu_block_flush(req);
33
- req->in->status = VIRTIO_BLK_S_OK;
34
+ if (vu_block_flush(req) == 0) {
35
+ req->in->status = VIRTIO_BLK_S_OK;
36
+ } else {
37
+ req->in->status = VIRTIO_BLK_S_IOERR;
38
+ }
39
break;
40
case VIRTIO_BLK_T_GET_ID: {
41
size_t size = MIN(iov_size(&elem->in_sg[0], in_num),
42
--
43
2.26.2
44
diff view generated by jsdifflib
Deleted patch
1
Use the new QAPI block exports API instead of defining our own QOM
2
objects.
3
1
4
This is a large change because the lifecycle of VuBlockDev needs to
5
follow BlockExportDriver. QOM properties are replaced by QAPI options
6
objects.
7
8
VuBlockDev is renamed VuBlkExport and contains a BlockExport field.
9
Several fields can be dropped since BlockExport already has equivalents.
10
11
The file names and meson build integration will be adjusted in a future
12
patch. libvhost-user should probably be built as a static library that
13
is linked into QEMU instead of as a .c file that results in duplicate
14
compilation.
15
16
The new command-line syntax is:
17
18
$ qemu-storage-daemon \
19
--blockdev file,node-name=drive0,filename=test.img \
20
--export vhost-user-blk,node-name=drive0,id=export0,unix-socket=/tmp/vhost-user-blk.sock
21
22
Note that unix-socket is optional because we may wish to accept chardevs
23
too in the future.
24
25
Markus noted that supported address families are not explicit in the
26
QAPI schema. It is unlikely that support for more address families will
27
be added since file descriptor passing is required and few address
28
families support it. If a new address family needs to be added, then the
29
QAPI 'features' syntax can be used to advertize them.
30
31
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
32
Acked-by: Markus Armbruster <armbru@redhat.com>
33
Message-id: 20200924151549.913737-12-stefanha@redhat.com
34
[Skip test on big-endian host architectures because this device doesn't
35
support them yet (as already mentioned in a code comment).
36
--Stefan]
37
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
38
---
39
qapi/block-export.json | 21 +-
40
block/export/vhost-user-blk-server.h | 23 +-
41
block/export/export.c | 6 +
42
block/export/vhost-user-blk-server.c | 452 +++++++--------------------
43
util/vhost-user-server.c | 10 +-
44
block/export/meson.build | 1 +
45
block/meson.build | 1 -
46
7 files changed, 156 insertions(+), 358 deletions(-)
47
48
diff --git a/qapi/block-export.json b/qapi/block-export.json
49
index XXXXXXX..XXXXXXX 100644
50
--- a/qapi/block-export.json
51
+++ b/qapi/block-export.json
52
@@ -XXX,XX +XXX,XX @@
53
'data': { '*name': 'str', '*description': 'str',
54
'*bitmap': 'str' } }
55
56
+##
57
+# @BlockExportOptionsVhostUserBlk:
58
+#
59
+# A vhost-user-blk block export.
60
+#
61
+# @addr: The vhost-user socket on which to listen. Both 'unix' and 'fd'
62
+# SocketAddress types are supported. Passed fds must be UNIX domain
63
+# sockets.
64
+# @logical-block-size: Logical block size in bytes. Defaults to 512 bytes.
65
+#
66
+# Since: 5.2
67
+##
68
+{ 'struct': 'BlockExportOptionsVhostUserBlk',
69
+ 'data': { 'addr': 'SocketAddress', '*logical-block-size': 'size' } }
70
+
71
##
72
# @NbdServerAddOptions:
73
#
74
@@ -XXX,XX +XXX,XX @@
75
# An enumeration of block export types
76
#
77
# @nbd: NBD export
78
+# @vhost-user-blk: vhost-user-blk export (since 5.2)
79
#
80
# Since: 4.2
81
##
82
{ 'enum': 'BlockExportType',
83
- 'data': [ 'nbd' ] }
84
+ 'data': [ 'nbd', 'vhost-user-blk' ] }
85
86
##
87
# @BlockExportOptions:
88
@@ -XXX,XX +XXX,XX @@
89
'*writethrough': 'bool' },
90
'discriminator': 'type',
91
'data': {
92
- 'nbd': 'BlockExportOptionsNbd'
93
+ 'nbd': 'BlockExportOptionsNbd',
94
+ 'vhost-user-blk': 'BlockExportOptionsVhostUserBlk'
95
} }
96
97
##
98
diff --git a/block/export/vhost-user-blk-server.h b/block/export/vhost-user-blk-server.h
99
index XXXXXXX..XXXXXXX 100644
100
--- a/block/export/vhost-user-blk-server.h
101
+++ b/block/export/vhost-user-blk-server.h
102
@@ -XXX,XX +XXX,XX @@
103
104
#ifndef VHOST_USER_BLK_SERVER_H
105
#define VHOST_USER_BLK_SERVER_H
106
-#include "util/vhost-user-server.h"
107
108
-typedef struct VuBlockDev VuBlockDev;
109
-#define TYPE_VHOST_USER_BLK_SERVER "vhost-user-blk-server"
110
-#define VHOST_USER_BLK_SERVER(obj) \
111
- OBJECT_CHECK(VuBlockDev, obj, TYPE_VHOST_USER_BLK_SERVER)
112
+#include "block/export.h"
113
114
-/* vhost user block device */
115
-struct VuBlockDev {
116
- Object parent_obj;
117
- char *node_name;
118
- SocketAddress *addr;
119
- AioContext *ctx;
120
- VuServer vu_server;
121
- bool running;
122
- uint32_t blk_size;
123
- BlockBackend *backend;
124
- QIOChannelSocket *sioc;
125
- QTAILQ_ENTRY(VuBlockDev) next;
126
- struct virtio_blk_config blkcfg;
127
- bool writable;
128
-};
129
+/* For block/export/export.c */
130
+extern const BlockExportDriver blk_exp_vhost_user_blk;
131
132
#endif /* VHOST_USER_BLK_SERVER_H */
133
diff --git a/block/export/export.c b/block/export/export.c
134
index XXXXXXX..XXXXXXX 100644
135
--- a/block/export/export.c
136
+++ b/block/export/export.c
137
@@ -XXX,XX +XXX,XX @@
138
#include "sysemu/block-backend.h"
139
#include "block/export.h"
140
#include "block/nbd.h"
141
+#if CONFIG_LINUX
142
+#include "block/export/vhost-user-blk-server.h"
143
+#endif
144
#include "qapi/error.h"
145
#include "qapi/qapi-commands-block-export.h"
146
#include "qapi/qapi-events-block-export.h"
147
@@ -XXX,XX +XXX,XX @@
148
149
static const BlockExportDriver *blk_exp_drivers[] = {
150
&blk_exp_nbd,
151
+#if CONFIG_LINUX
152
+ &blk_exp_vhost_user_blk,
153
+#endif
154
};
155
156
/* Only accessed from the main thread */
157
diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c
158
index XXXXXXX..XXXXXXX 100644
159
--- a/block/export/vhost-user-blk-server.c
160
+++ b/block/export/vhost-user-blk-server.c
161
@@ -XXX,XX +XXX,XX @@
162
*/
163
#include "qemu/osdep.h"
164
#include "block/block.h"
165
+#include "contrib/libvhost-user/libvhost-user.h"
166
+#include "standard-headers/linux/virtio_blk.h"
167
+#include "util/vhost-user-server.h"
168
#include "vhost-user-blk-server.h"
169
#include "qapi/error.h"
170
#include "qom/object_interfaces.h"
171
@@ -XXX,XX +XXX,XX @@ struct virtio_blk_inhdr {
172
unsigned char status;
173
};
174
175
-typedef struct VuBlockReq {
176
+typedef struct VuBlkReq {
177
VuVirtqElement elem;
178
int64_t sector_num;
179
size_t size;
180
@@ -XXX,XX +XXX,XX @@ typedef struct VuBlockReq {
181
struct virtio_blk_outhdr out;
182
VuServer *server;
183
struct VuVirtq *vq;
184
-} VuBlockReq;
185
+} VuBlkReq;
186
187
-static void vu_block_req_complete(VuBlockReq *req)
188
+/* vhost user block device */
189
+typedef struct {
190
+ BlockExport export;
191
+ VuServer vu_server;
192
+ uint32_t blk_size;
193
+ QIOChannelSocket *sioc;
194
+ struct virtio_blk_config blkcfg;
195
+ bool writable;
196
+} VuBlkExport;
197
+
198
+static void vu_blk_req_complete(VuBlkReq *req)
199
{
200
VuDev *vu_dev = &req->server->vu_dev;
201
202
@@ -XXX,XX +XXX,XX @@ static void vu_block_req_complete(VuBlockReq *req)
203
free(req);
204
}
205
206
-static VuBlockDev *get_vu_block_device_by_server(VuServer *server)
207
-{
208
- return container_of(server, VuBlockDev, vu_server);
209
-}
210
-
211
static int coroutine_fn
212
-vu_block_discard_write_zeroes(VuBlockReq *req, struct iovec *iov,
213
- uint32_t iovcnt, uint32_t type)
214
+vu_blk_discard_write_zeroes(BlockBackend *blk, struct iovec *iov,
215
+ uint32_t iovcnt, uint32_t type)
216
{
217
struct virtio_blk_discard_write_zeroes desc;
218
ssize_t size = iov_to_buf(iov, iovcnt, 0, &desc, sizeof(desc));
219
@@ -XXX,XX +XXX,XX @@ vu_block_discard_write_zeroes(VuBlockReq *req, struct iovec *iov,
220
return -EINVAL;
221
}
222
223
- VuBlockDev *vdev_blk = get_vu_block_device_by_server(req->server);
224
uint64_t range[2] = { le64_to_cpu(desc.sector) << 9,
225
le32_to_cpu(desc.num_sectors) << 9 };
226
if (type == VIRTIO_BLK_T_DISCARD) {
227
- if (blk_co_pdiscard(vdev_blk->backend, range[0], range[1]) == 0) {
228
+ if (blk_co_pdiscard(blk, range[0], range[1]) == 0) {
229
return 0;
230
}
231
} else if (type == VIRTIO_BLK_T_WRITE_ZEROES) {
232
- if (blk_co_pwrite_zeroes(vdev_blk->backend,
233
- range[0], range[1], 0) == 0) {
234
+ if (blk_co_pwrite_zeroes(blk, range[0], range[1], 0) == 0) {
235
return 0;
236
}
237
}
238
@@ -XXX,XX +XXX,XX @@ vu_block_discard_write_zeroes(VuBlockReq *req, struct iovec *iov,
239
return -EINVAL;
240
}
241
242
-static int coroutine_fn vu_block_flush(VuBlockReq *req)
243
+static void coroutine_fn vu_blk_virtio_process_req(void *opaque)
244
{
245
- VuBlockDev *vdev_blk = get_vu_block_device_by_server(req->server);
246
- BlockBackend *backend = vdev_blk->backend;
247
- return blk_co_flush(backend);
248
-}
249
-
250
-static void coroutine_fn vu_block_virtio_process_req(void *opaque)
251
-{
252
- VuBlockReq *req = opaque;
253
+ VuBlkReq *req = opaque;
254
VuServer *server = req->server;
255
VuVirtqElement *elem = &req->elem;
256
uint32_t type;
257
258
- VuBlockDev *vdev_blk = get_vu_block_device_by_server(server);
259
- BlockBackend *backend = vdev_blk->backend;
260
+ VuBlkExport *vexp = container_of(server, VuBlkExport, vu_server);
261
+ BlockBackend *blk = vexp->export.blk;
262
263
struct iovec *in_iov = elem->in_sg;
264
struct iovec *out_iov = elem->out_sg;
265
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque)
266
bool is_write = type & VIRTIO_BLK_T_OUT;
267
req->sector_num = le64_to_cpu(req->out.sector);
268
269
- int64_t offset = req->sector_num * vdev_blk->blk_size;
270
+ if (is_write && !vexp->writable) {
271
+ req->in->status = VIRTIO_BLK_S_IOERR;
272
+ break;
273
+ }
274
+
275
+ int64_t offset = req->sector_num * vexp->blk_size;
276
QEMUIOVector qiov;
277
if (is_write) {
278
qemu_iovec_init_external(&qiov, out_iov, out_num);
279
- ret = blk_co_pwritev(backend, offset, qiov.size,
280
- &qiov, 0);
281
+ ret = blk_co_pwritev(blk, offset, qiov.size, &qiov, 0);
282
} else {
283
qemu_iovec_init_external(&qiov, in_iov, in_num);
284
- ret = blk_co_preadv(backend, offset, qiov.size,
285
- &qiov, 0);
286
+ ret = blk_co_preadv(blk, offset, qiov.size, &qiov, 0);
287
}
288
if (ret >= 0) {
289
req->in->status = VIRTIO_BLK_S_OK;
290
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque)
291
break;
292
}
293
case VIRTIO_BLK_T_FLUSH:
294
- if (vu_block_flush(req) == 0) {
295
+ if (blk_co_flush(blk) == 0) {
296
req->in->status = VIRTIO_BLK_S_OK;
297
} else {
298
req->in->status = VIRTIO_BLK_S_IOERR;
299
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque)
300
case VIRTIO_BLK_T_DISCARD:
301
case VIRTIO_BLK_T_WRITE_ZEROES: {
302
int rc;
303
- rc = vu_block_discard_write_zeroes(req, &elem->out_sg[1],
304
- out_num, type);
305
+
306
+ if (!vexp->writable) {
307
+ req->in->status = VIRTIO_BLK_S_IOERR;
308
+ break;
309
+ }
310
+
311
+ rc = vu_blk_discard_write_zeroes(blk, &elem->out_sg[1], out_num, type);
312
if (rc == 0) {
313
req->in->status = VIRTIO_BLK_S_OK;
314
} else {
315
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque)
316
break;
317
}
318
319
- vu_block_req_complete(req);
320
+ vu_blk_req_complete(req);
321
return;
322
323
err:
324
- free(elem);
325
+ free(req);
326
}
327
328
-static void vu_block_process_vq(VuDev *vu_dev, int idx)
329
+static void vu_blk_process_vq(VuDev *vu_dev, int idx)
330
{
331
VuServer *server = container_of(vu_dev, VuServer, vu_dev);
332
VuVirtq *vq = vu_get_queue(vu_dev, idx);
333
334
while (1) {
335
- VuBlockReq *req;
336
+ VuBlkReq *req;
337
338
- req = vu_queue_pop(vu_dev, vq, sizeof(VuBlockReq));
339
+ req = vu_queue_pop(vu_dev, vq, sizeof(VuBlkReq));
340
if (!req) {
341
break;
342
}
343
@@ -XXX,XX +XXX,XX @@ static void vu_block_process_vq(VuDev *vu_dev, int idx)
344
req->vq = vq;
345
346
Coroutine *co =
347
- qemu_coroutine_create(vu_block_virtio_process_req, req);
348
+ qemu_coroutine_create(vu_blk_virtio_process_req, req);
349
qemu_coroutine_enter(co);
350
}
351
}
352
353
-static void vu_block_queue_set_started(VuDev *vu_dev, int idx, bool started)
354
+static void vu_blk_queue_set_started(VuDev *vu_dev, int idx, bool started)
355
{
356
VuVirtq *vq;
357
358
assert(vu_dev);
359
360
vq = vu_get_queue(vu_dev, idx);
361
- vu_set_queue_handler(vu_dev, vq, started ? vu_block_process_vq : NULL);
362
+ vu_set_queue_handler(vu_dev, vq, started ? vu_blk_process_vq : NULL);
363
}
364
365
-static uint64_t vu_block_get_features(VuDev *dev)
366
+static uint64_t vu_blk_get_features(VuDev *dev)
367
{
368
uint64_t features;
369
VuServer *server = container_of(dev, VuServer, vu_dev);
370
- VuBlockDev *vdev_blk = get_vu_block_device_by_server(server);
371
+ VuBlkExport *vexp = container_of(server, VuBlkExport, vu_server);
372
features = 1ull << VIRTIO_BLK_F_SIZE_MAX |
373
1ull << VIRTIO_BLK_F_SEG_MAX |
374
1ull << VIRTIO_BLK_F_TOPOLOGY |
375
@@ -XXX,XX +XXX,XX @@ static uint64_t vu_block_get_features(VuDev *dev)
376
1ull << VIRTIO_RING_F_EVENT_IDX |
377
1ull << VHOST_USER_F_PROTOCOL_FEATURES;
378
379
- if (!vdev_blk->writable) {
380
+ if (!vexp->writable) {
381
features |= 1ull << VIRTIO_BLK_F_RO;
382
}
383
384
return features;
385
}
386
387
-static uint64_t vu_block_get_protocol_features(VuDev *dev)
388
+static uint64_t vu_blk_get_protocol_features(VuDev *dev)
389
{
390
return 1ull << VHOST_USER_PROTOCOL_F_CONFIG |
391
1ull << VHOST_USER_PROTOCOL_F_INFLIGHT_SHMFD;
392
}
393
394
static int
395
-vu_block_get_config(VuDev *vu_dev, uint8_t *config, uint32_t len)
396
+vu_blk_get_config(VuDev *vu_dev, uint8_t *config, uint32_t len)
397
{
398
+ /* TODO blkcfg must be little-endian for VIRTIO 1.0 */
399
VuServer *server = container_of(vu_dev, VuServer, vu_dev);
400
- VuBlockDev *vdev_blk = get_vu_block_device_by_server(server);
401
- memcpy(config, &vdev_blk->blkcfg, len);
402
-
403
+ VuBlkExport *vexp = container_of(server, VuBlkExport, vu_server);
404
+ memcpy(config, &vexp->blkcfg, len);
405
return 0;
406
}
407
408
static int
409
-vu_block_set_config(VuDev *vu_dev, const uint8_t *data,
410
+vu_blk_set_config(VuDev *vu_dev, const uint8_t *data,
411
uint32_t offset, uint32_t size, uint32_t flags)
412
{
413
VuServer *server = container_of(vu_dev, VuServer, vu_dev);
414
- VuBlockDev *vdev_blk = get_vu_block_device_by_server(server);
415
+ VuBlkExport *vexp = container_of(server, VuBlkExport, vu_server);
416
uint8_t wce;
417
418
/* don't support live migration */
419
@@ -XXX,XX +XXX,XX @@ vu_block_set_config(VuDev *vu_dev, const uint8_t *data,
420
}
421
422
wce = *data;
423
- vdev_blk->blkcfg.wce = wce;
424
- blk_set_enable_write_cache(vdev_blk->backend, wce);
425
+ vexp->blkcfg.wce = wce;
426
+ blk_set_enable_write_cache(vexp->export.blk, wce);
427
return 0;
428
}
429
430
@@ -XXX,XX +XXX,XX @@ vu_block_set_config(VuDev *vu_dev, const uint8_t *data,
431
* of vu_process_message.
432
*
433
*/
434
-static int vu_block_process_msg(VuDev *dev, VhostUserMsg *vmsg, int *do_reply)
435
+static int vu_blk_process_msg(VuDev *dev, VhostUserMsg *vmsg, int *do_reply)
436
{
437
if (vmsg->request == VHOST_USER_NONE) {
438
dev->panic(dev, "disconnect");
439
@@ -XXX,XX +XXX,XX @@ static int vu_block_process_msg(VuDev *dev, VhostUserMsg *vmsg, int *do_reply)
440
return false;
441
}
442
443
-static const VuDevIface vu_block_iface = {
444
- .get_features = vu_block_get_features,
445
- .queue_set_started = vu_block_queue_set_started,
446
- .get_protocol_features = vu_block_get_protocol_features,
447
- .get_config = vu_block_get_config,
448
- .set_config = vu_block_set_config,
449
- .process_msg = vu_block_process_msg,
450
+static const VuDevIface vu_blk_iface = {
451
+ .get_features = vu_blk_get_features,
452
+ .queue_set_started = vu_blk_queue_set_started,
453
+ .get_protocol_features = vu_blk_get_protocol_features,
454
+ .get_config = vu_blk_get_config,
455
+ .set_config = vu_blk_set_config,
456
+ .process_msg = vu_blk_process_msg,
457
};
458
459
static void blk_aio_attached(AioContext *ctx, void *opaque)
460
{
461
- VuBlockDev *vub_dev = opaque;
462
- vhost_user_server_attach_aio_context(&vub_dev->vu_server, ctx);
463
+ VuBlkExport *vexp = opaque;
464
+ vhost_user_server_attach_aio_context(&vexp->vu_server, ctx);
465
}
466
467
static void blk_aio_detach(void *opaque)
468
{
469
- VuBlockDev *vub_dev = opaque;
470
- vhost_user_server_detach_aio_context(&vub_dev->vu_server);
471
+ VuBlkExport *vexp = opaque;
472
+ vhost_user_server_detach_aio_context(&vexp->vu_server);
473
}
474
475
static void
476
-vu_block_initialize_config(BlockDriverState *bs,
477
+vu_blk_initialize_config(BlockDriverState *bs,
478
struct virtio_blk_config *config, uint32_t blk_size)
479
{
480
config->capacity = bdrv_getlength(bs) >> BDRV_SECTOR_BITS;
481
@@ -XXX,XX +XXX,XX @@ vu_block_initialize_config(BlockDriverState *bs,
482
config->max_write_zeroes_seg = 1;
483
}
484
485
-static VuBlockDev *vu_block_init(VuBlockDev *vu_block_device, Error **errp)
486
+static void vu_blk_exp_request_shutdown(BlockExport *exp)
487
{
488
+ VuBlkExport *vexp = container_of(exp, VuBlkExport, export);
489
490
- BlockBackend *blk;
491
- Error *local_error = NULL;
492
- const char *node_name = vu_block_device->node_name;
493
- bool writable = vu_block_device->writable;
494
- uint64_t perm = BLK_PERM_CONSISTENT_READ;
495
- int ret;
496
-
497
- AioContext *ctx;
498
-
499
- BlockDriverState *bs = bdrv_lookup_bs(node_name, node_name, &local_error);
500
-
501
- if (!bs) {
502
- error_propagate(errp, local_error);
503
- return NULL;
504
- }
505
-
506
- if (bdrv_is_read_only(bs)) {
507
- writable = false;
508
- }
509
-
510
- if (writable) {
511
- perm |= BLK_PERM_WRITE;
512
- }
513
-
514
- ctx = bdrv_get_aio_context(bs);
515
- aio_context_acquire(ctx);
516
- bdrv_invalidate_cache(bs, NULL);
517
- aio_context_release(ctx);
518
-
519
- /*
520
- * Don't allow resize while the vhost user server is running,
521
- * otherwise we don't care what happens with the node.
522
- */
523
- blk = blk_new(bdrv_get_aio_context(bs), perm,
524
- BLK_PERM_CONSISTENT_READ | BLK_PERM_WRITE_UNCHANGED |
525
- BLK_PERM_WRITE | BLK_PERM_GRAPH_MOD);
526
- ret = blk_insert_bs(blk, bs, errp);
527
-
528
- if (ret < 0) {
529
- goto fail;
530
- }
531
-
532
- blk_set_enable_write_cache(blk, false);
533
-
534
- blk_set_allow_aio_context_change(blk, true);
535
-
536
- vu_block_device->blkcfg.wce = 0;
537
- vu_block_device->backend = blk;
538
- if (!vu_block_device->blk_size) {
539
- vu_block_device->blk_size = BDRV_SECTOR_SIZE;
540
- }
541
- vu_block_device->blkcfg.blk_size = vu_block_device->blk_size;
542
- blk_set_guest_block_size(blk, vu_block_device->blk_size);
543
- vu_block_initialize_config(bs, &vu_block_device->blkcfg,
544
- vu_block_device->blk_size);
545
- return vu_block_device;
546
-
547
-fail:
548
- blk_unref(blk);
549
- return NULL;
550
-}
551
-
552
-static void vu_block_deinit(VuBlockDev *vu_block_device)
553
-{
554
- if (vu_block_device->backend) {
555
- blk_remove_aio_context_notifier(vu_block_device->backend, blk_aio_attached,
556
- blk_aio_detach, vu_block_device);
557
- }
558
-
559
- blk_unref(vu_block_device->backend);
560
-}
561
-
562
-static void vhost_user_blk_server_stop(VuBlockDev *vu_block_device)
563
-{
564
- vhost_user_server_stop(&vu_block_device->vu_server);
565
- vu_block_deinit(vu_block_device);
566
-}
567
-
568
-static void vhost_user_blk_server_start(VuBlockDev *vu_block_device,
569
- Error **errp)
570
-{
571
- AioContext *ctx;
572
- SocketAddress *addr = vu_block_device->addr;
573
-
574
- if (!vu_block_init(vu_block_device, errp)) {
575
- return;
576
- }
577
-
578
- ctx = bdrv_get_aio_context(blk_bs(vu_block_device->backend));
579
-
580
- if (!vhost_user_server_start(&vu_block_device->vu_server, addr, ctx,
581
- VHOST_USER_BLK_MAX_QUEUES, &vu_block_iface,
582
- errp)) {
583
- goto error;
584
- }
585
-
586
- blk_add_aio_context_notifier(vu_block_device->backend, blk_aio_attached,
587
- blk_aio_detach, vu_block_device);
588
- vu_block_device->running = true;
589
- return;
590
-
591
- error:
592
- vu_block_deinit(vu_block_device);
593
-}
594
-
595
-static bool vu_prop_modifiable(VuBlockDev *vus, Error **errp)
596
-{
597
- if (vus->running) {
598
- error_setg(errp, "The property can't be modified "
599
- "while the server is running");
600
- return false;
601
- }
602
- return true;
603
-}
604
-
605
-static void vu_set_node_name(Object *obj, const char *value, Error **errp)
606
-{
607
- VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
608
-
609
- if (!vu_prop_modifiable(vus, errp)) {
610
- return;
611
- }
612
-
613
- if (vus->node_name) {
614
- g_free(vus->node_name);
615
- }
616
-
617
- vus->node_name = g_strdup(value);
618
-}
619
-
620
-static char *vu_get_node_name(Object *obj, Error **errp)
621
-{
622
- VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
623
- return g_strdup(vus->node_name);
624
-}
625
-
626
-static void free_socket_addr(SocketAddress *addr)
627
-{
628
- g_free(addr->u.q_unix.path);
629
- g_free(addr);
630
-}
631
-
632
-static void vu_set_unix_socket(Object *obj, const char *value,
633
- Error **errp)
634
-{
635
- VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
636
-
637
- if (!vu_prop_modifiable(vus, errp)) {
638
- return;
639
- }
640
-
641
- if (vus->addr) {
642
- free_socket_addr(vus->addr);
643
- }
644
-
645
- SocketAddress *addr = g_new0(SocketAddress, 1);
646
- addr->type = SOCKET_ADDRESS_TYPE_UNIX;
647
- addr->u.q_unix.path = g_strdup(value);
648
- vus->addr = addr;
649
+ vhost_user_server_stop(&vexp->vu_server);
650
}
651
652
-static char *vu_get_unix_socket(Object *obj, Error **errp)
653
+static int vu_blk_exp_create(BlockExport *exp, BlockExportOptions *opts,
654
+ Error **errp)
655
{
656
- VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
657
- return g_strdup(vus->addr->u.q_unix.path);
658
-}
659
-
660
-static bool vu_get_block_writable(Object *obj, Error **errp)
661
-{
662
- VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
663
- return vus->writable;
664
-}
665
-
666
-static void vu_set_block_writable(Object *obj, bool value, Error **errp)
667
-{
668
- VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
669
-
670
- if (!vu_prop_modifiable(vus, errp)) {
671
- return;
672
- }
673
-
674
- vus->writable = value;
675
-}
676
-
677
-static void vu_get_blk_size(Object *obj, Visitor *v, const char *name,
678
- void *opaque, Error **errp)
679
-{
680
- VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
681
- uint32_t value = vus->blk_size;
682
-
683
- visit_type_uint32(v, name, &value, errp);
684
-}
685
-
686
-static void vu_set_blk_size(Object *obj, Visitor *v, const char *name,
687
- void *opaque, Error **errp)
688
-{
689
- VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj);
690
-
691
+ VuBlkExport *vexp = container_of(exp, VuBlkExport, export);
692
+ BlockExportOptionsVhostUserBlk *vu_opts = &opts->u.vhost_user_blk;
693
Error *local_err = NULL;
694
- uint32_t value;
695
+ uint64_t logical_block_size;
696
697
- if (!vu_prop_modifiable(vus, errp)) {
698
- return;
699
- }
700
+ vexp->writable = opts->writable;
701
+ vexp->blkcfg.wce = 0;
702
703
- visit_type_uint32(v, name, &value, &local_err);
704
- if (local_err) {
705
- goto out;
706
+ if (vu_opts->has_logical_block_size) {
707
+ logical_block_size = vu_opts->logical_block_size;
708
+ } else {
709
+ logical_block_size = BDRV_SECTOR_SIZE;
710
}
711
-
712
- check_block_size(object_get_typename(obj), name, value, &local_err);
713
+ check_block_size(exp->id, "logical-block-size", logical_block_size,
714
+ &local_err);
715
if (local_err) {
716
- goto out;
717
+ error_propagate(errp, local_err);
718
+ return -EINVAL;
719
+ }
720
+ vexp->blk_size = logical_block_size;
721
+ blk_set_guest_block_size(exp->blk, logical_block_size);
722
+ vu_blk_initialize_config(blk_bs(exp->blk), &vexp->blkcfg,
723
+ logical_block_size);
724
+
725
+ blk_set_allow_aio_context_change(exp->blk, true);
726
+ blk_add_aio_context_notifier(exp->blk, blk_aio_attached, blk_aio_detach,
727
+ vexp);
728
+
729
+ if (!vhost_user_server_start(&vexp->vu_server, vu_opts->addr, exp->ctx,
730
+ VHOST_USER_BLK_MAX_QUEUES, &vu_blk_iface,
731
+ errp)) {
732
+ blk_remove_aio_context_notifier(exp->blk, blk_aio_attached,
733
+ blk_aio_detach, vexp);
734
+ return -EADDRNOTAVAIL;
735
}
736
737
- vus->blk_size = value;
738
-
739
-out:
740
- error_propagate(errp, local_err);
741
-}
742
-
743
-static void vhost_user_blk_server_instance_finalize(Object *obj)
744
-{
745
- VuBlockDev *vub = VHOST_USER_BLK_SERVER(obj);
746
-
747
- vhost_user_blk_server_stop(vub);
748
-
749
- /*
750
- * Unlike object_property_add_str, object_class_property_add_str
751
- * doesn't have a release method. Thus manual memory freeing is
752
- * needed.
753
- */
754
- free_socket_addr(vub->addr);
755
- g_free(vub->node_name);
756
-}
757
-
758
-static void vhost_user_blk_server_complete(UserCreatable *obj, Error **errp)
759
-{
760
- VuBlockDev *vub = VHOST_USER_BLK_SERVER(obj);
761
-
762
- vhost_user_blk_server_start(vub, errp);
763
+ return 0;
764
}
765
766
-static void vhost_user_blk_server_class_init(ObjectClass *klass,
767
- void *class_data)
768
+static void vu_blk_exp_delete(BlockExport *exp)
769
{
770
- UserCreatableClass *ucc = USER_CREATABLE_CLASS(klass);
771
- ucc->complete = vhost_user_blk_server_complete;
772
-
773
- object_class_property_add_bool(klass, "writable",
774
- vu_get_block_writable,
775
- vu_set_block_writable);
776
-
777
- object_class_property_add_str(klass, "node-name",
778
- vu_get_node_name,
779
- vu_set_node_name);
780
-
781
- object_class_property_add_str(klass, "unix-socket",
782
- vu_get_unix_socket,
783
- vu_set_unix_socket);
784
+ VuBlkExport *vexp = container_of(exp, VuBlkExport, export);
785
786
- object_class_property_add(klass, "logical-block-size", "uint32",
787
- vu_get_blk_size, vu_set_blk_size,
788
- NULL, NULL);
789
+ blk_remove_aio_context_notifier(exp->blk, blk_aio_attached, blk_aio_detach,
790
+ vexp);
791
}
792
793
-static const TypeInfo vhost_user_blk_server_info = {
794
- .name = TYPE_VHOST_USER_BLK_SERVER,
795
- .parent = TYPE_OBJECT,
796
- .instance_size = sizeof(VuBlockDev),
797
- .instance_finalize = vhost_user_blk_server_instance_finalize,
798
- .class_init = vhost_user_blk_server_class_init,
799
- .interfaces = (InterfaceInfo[]) {
800
- {TYPE_USER_CREATABLE},
801
- {}
802
- },
803
+const BlockExportDriver blk_exp_vhost_user_blk = {
804
+ .type = BLOCK_EXPORT_TYPE_VHOST_USER_BLK,
805
+ .instance_size = sizeof(VuBlkExport),
806
+ .create = vu_blk_exp_create,
807
+ .delete = vu_blk_exp_delete,
808
+ .request_shutdown = vu_blk_exp_request_shutdown,
809
};
810
-
811
-static void vhost_user_blk_server_register_types(void)
812
-{
813
- type_register_static(&vhost_user_blk_server_info);
814
-}
815
-
816
-type_init(vhost_user_blk_server_register_types)
817
diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c
818
index XXXXXXX..XXXXXXX 100644
819
--- a/util/vhost-user-server.c
820
+++ b/util/vhost-user-server.c
821
@@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server,
822
Error **errp)
823
{
824
QEMUBH *bh;
825
- QIONetListener *listener = qio_net_listener_new();
826
+ QIONetListener *listener;
827
+
828
+ if (socket_addr->type != SOCKET_ADDRESS_TYPE_UNIX &&
829
+ socket_addr->type != SOCKET_ADDRESS_TYPE_FD) {
830
+ error_setg(errp, "Only socket address types 'unix' and 'fd' are supported");
831
+ return false;
832
+ }
833
+
834
+ listener = qio_net_listener_new();
835
if (qio_net_listener_open_sync(listener, socket_addr, 1,
836
errp) < 0) {
837
object_unref(OBJECT(listener));
838
diff --git a/block/export/meson.build b/block/export/meson.build
839
index XXXXXXX..XXXXXXX 100644
840
--- a/block/export/meson.build
841
+++ b/block/export/meson.build
842
@@ -1 +1,2 @@
843
block_ss.add(files('export.c'))
844
+block_ss.add(when: 'CONFIG_LINUX', if_true: files('vhost-user-blk-server.c', '../../contrib/libvhost-user/libvhost-user.c'))
845
diff --git a/block/meson.build b/block/meson.build
846
index XXXXXXX..XXXXXXX 100644
847
--- a/block/meson.build
848
+++ b/block/meson.build
849
@@ -XXX,XX +XXX,XX @@ block_ss.add(when: 'CONFIG_WIN32', if_true: files('file-win32.c', 'win32-aio.c')
850
block_ss.add(when: 'CONFIG_POSIX', if_true: [files('file-posix.c'), coref, iokit])
851
block_ss.add(when: 'CONFIG_LIBISCSI', if_true: files('iscsi-opts.c'))
852
block_ss.add(when: 'CONFIG_LINUX', if_true: files('nvme.c'))
853
-block_ss.add(when: 'CONFIG_LINUX', if_true: files('export/vhost-user-blk-server.c', '../contrib/libvhost-user/libvhost-user.c'))
854
block_ss.add(when: 'CONFIG_REPLICATION', if_true: files('replication.c'))
855
block_ss.add(when: 'CONFIG_SHEEPDOG', if_true: files('sheepdog.c'))
856
block_ss.add(when: ['CONFIG_LINUX_AIO', libaio], if_true: files('linux-aio.c'))
857
--
858
2.26.2
859
diff view generated by jsdifflib
Deleted patch
1
Headers used by other subsystems are located in include/. Also add the
2
vhost-user-server and vhost-user-blk-server headers to MAINTAINERS.
3
1
4
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
5
Message-id: 20200924151549.913737-13-stefanha@redhat.com
6
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
7
---
8
MAINTAINERS | 4 +++-
9
{util => include/qemu}/vhost-user-server.h | 0
10
block/export/vhost-user-blk-server.c | 2 +-
11
util/vhost-user-server.c | 2 +-
12
4 files changed, 5 insertions(+), 3 deletions(-)
13
rename {util => include/qemu}/vhost-user-server.h (100%)
14
15
diff --git a/MAINTAINERS b/MAINTAINERS
16
index XXXXXXX..XXXXXXX 100644
17
--- a/MAINTAINERS
18
+++ b/MAINTAINERS
19
@@ -XXX,XX +XXX,XX @@ Vhost-user block device backend server
20
M: Coiby Xu <Coiby.Xu@gmail.com>
21
S: Maintained
22
F: block/export/vhost-user-blk-server.c
23
-F: util/vhost-user-server.c
24
+F: block/export/vhost-user-blk-server.h
25
+F: include/qemu/vhost-user-server.h
26
F: tests/qtest/libqos/vhost-user-blk.c
27
+F: util/vhost-user-server.c
28
29
Replication
30
M: Wen Congyang <wencongyang2@huawei.com>
31
diff --git a/util/vhost-user-server.h b/include/qemu/vhost-user-server.h
32
similarity index 100%
33
rename from util/vhost-user-server.h
34
rename to include/qemu/vhost-user-server.h
35
diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c
36
index XXXXXXX..XXXXXXX 100644
37
--- a/block/export/vhost-user-blk-server.c
38
+++ b/block/export/vhost-user-blk-server.c
39
@@ -XXX,XX +XXX,XX @@
40
#include "block/block.h"
41
#include "contrib/libvhost-user/libvhost-user.h"
42
#include "standard-headers/linux/virtio_blk.h"
43
-#include "util/vhost-user-server.h"
44
+#include "qemu/vhost-user-server.h"
45
#include "vhost-user-blk-server.h"
46
#include "qapi/error.h"
47
#include "qom/object_interfaces.h"
48
diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c
49
index XXXXXXX..XXXXXXX 100644
50
--- a/util/vhost-user-server.c
51
+++ b/util/vhost-user-server.c
52
@@ -XXX,XX +XXX,XX @@
53
*/
54
#include "qemu/osdep.h"
55
#include "qemu/main-loop.h"
56
+#include "qemu/vhost-user-server.h"
57
#include "block/aio-wait.h"
58
-#include "vhost-user-server.h"
59
60
/*
61
* Theory of operation:
62
--
63
2.26.2
64
diff view generated by jsdifflib
Deleted patch
1
Don't compile contrib/libvhost-user/libvhost-user.c again. Instead build
2
the static library once and then reuse it throughout QEMU.
3
1
4
Also switch from CONFIG_LINUX to CONFIG_VHOST_USER, which is what the
5
vhost-user tools (vhost-user-gpu, etc) do.
6
7
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
8
Message-id: 20200924151549.913737-14-stefanha@redhat.com
9
[Added CONFIG_LINUX again because libvhost-user doesn't build on macOS.
10
--Stefan]
11
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
12
---
13
block/export/export.c | 8 ++++----
14
block/export/meson.build | 2 +-
15
contrib/libvhost-user/meson.build | 1 +
16
meson.build | 6 +++++-
17
util/meson.build | 4 +++-
18
5 files changed, 14 insertions(+), 7 deletions(-)
19
20
diff --git a/block/export/export.c b/block/export/export.c
21
index XXXXXXX..XXXXXXX 100644
22
--- a/block/export/export.c
23
+++ b/block/export/export.c
24
@@ -XXX,XX +XXX,XX @@
25
#include "sysemu/block-backend.h"
26
#include "block/export.h"
27
#include "block/nbd.h"
28
-#if CONFIG_LINUX
29
-#include "block/export/vhost-user-blk-server.h"
30
-#endif
31
#include "qapi/error.h"
32
#include "qapi/qapi-commands-block-export.h"
33
#include "qapi/qapi-events-block-export.h"
34
#include "qemu/id.h"
35
+#ifdef CONFIG_VHOST_USER
36
+#include "vhost-user-blk-server.h"
37
+#endif
38
39
static const BlockExportDriver *blk_exp_drivers[] = {
40
&blk_exp_nbd,
41
-#if CONFIG_LINUX
42
+#ifdef CONFIG_VHOST_USER
43
&blk_exp_vhost_user_blk,
44
#endif
45
};
46
diff --git a/block/export/meson.build b/block/export/meson.build
47
index XXXXXXX..XXXXXXX 100644
48
--- a/block/export/meson.build
49
+++ b/block/export/meson.build
50
@@ -XXX,XX +XXX,XX @@
51
block_ss.add(files('export.c'))
52
-block_ss.add(when: 'CONFIG_LINUX', if_true: files('vhost-user-blk-server.c', '../../contrib/libvhost-user/libvhost-user.c'))
53
+block_ss.add(when: ['CONFIG_LINUX', 'CONFIG_VHOST_USER'], if_true: files('vhost-user-blk-server.c'))
54
diff --git a/contrib/libvhost-user/meson.build b/contrib/libvhost-user/meson.build
55
index XXXXXXX..XXXXXXX 100644
56
--- a/contrib/libvhost-user/meson.build
57
+++ b/contrib/libvhost-user/meson.build
58
@@ -XXX,XX +XXX,XX @@
59
libvhost_user = static_library('vhost-user',
60
files('libvhost-user.c', 'libvhost-user-glib.c'),
61
build_by_default: false)
62
+vhost_user = declare_dependency(link_with: libvhost_user)
63
diff --git a/meson.build b/meson.build
64
index XXXXXXX..XXXXXXX 100644
65
--- a/meson.build
66
+++ b/meson.build
67
@@ -XXX,XX +XXX,XX @@ trace_events_subdirs += [
68
'util',
69
]
70
71
+vhost_user = not_found
72
+if 'CONFIG_VHOST_USER' in config_host
73
+ subdir('contrib/libvhost-user')
74
+endif
75
+
76
subdir('qapi')
77
subdir('qobject')
78
subdir('stubs')
79
@@ -XXX,XX +XXX,XX @@ if have_tools
80
install: true)
81
82
if 'CONFIG_VHOST_USER' in config_host
83
- subdir('contrib/libvhost-user')
84
subdir('contrib/vhost-user-blk')
85
subdir('contrib/vhost-user-gpu')
86
subdir('contrib/vhost-user-input')
87
diff --git a/util/meson.build b/util/meson.build
88
index XXXXXXX..XXXXXXX 100644
89
--- a/util/meson.build
90
+++ b/util/meson.build
91
@@ -XXX,XX +XXX,XX @@ if have_block
92
util_ss.add(files('main-loop.c'))
93
util_ss.add(files('nvdimm-utils.c'))
94
util_ss.add(files('qemu-coroutine.c', 'qemu-coroutine-lock.c', 'qemu-coroutine-io.c'))
95
- util_ss.add(when: 'CONFIG_LINUX', if_true: files('vhost-user-server.c'))
96
+ util_ss.add(when: ['CONFIG_LINUX', 'CONFIG_VHOST_USER'], if_true: [
97
+ files('vhost-user-server.c'), vhost_user
98
+ ])
99
util_ss.add(files('block-helpers.c'))
100
util_ss.add(files('qemu-coroutine-sleep.c'))
101
util_ss.add(files('qemu-co-shared-resource.c'))
102
--
103
2.26.2
104
diff view generated by jsdifflib
Deleted patch
1
Introduce libblkdev.fa to avoid recompiling blockdev_ss twice.
2
1
3
Suggested-by: Paolo Bonzini <pbonzini@redhat.com>
4
Reviewed-by: Paolo Bonzini <pbonzini@redhat.com>
5
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
6
Message-id: 20200929125516.186715-3-stefanha@redhat.com
7
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
8
---
9
meson.build | 12 ++++++++++--
10
storage-daemon/meson.build | 3 +--
11
2 files changed, 11 insertions(+), 4 deletions(-)
12
13
diff --git a/meson.build b/meson.build
14
index XXXXXXX..XXXXXXX 100644
15
--- a/meson.build
16
+++ b/meson.build
17
@@ -XXX,XX +XXX,XX @@ blockdev_ss.add(files(
18
# os-win32.c does not
19
blockdev_ss.add(when: 'CONFIG_POSIX', if_true: files('os-posix.c'))
20
softmmu_ss.add(when: 'CONFIG_WIN32', if_true: [files('os-win32.c')])
21
-softmmu_ss.add_all(blockdev_ss)
22
23
common_ss.add(files('cpus-common.c'))
24
25
@@ -XXX,XX +XXX,XX @@ block = declare_dependency(link_whole: [libblock],
26
link_args: '@block.syms',
27
dependencies: [crypto, io])
28
29
+blockdev_ss = blockdev_ss.apply(config_host, strict: false)
30
+libblockdev = static_library('blockdev', blockdev_ss.sources() + genh,
31
+ dependencies: blockdev_ss.dependencies(),
32
+ name_suffix: 'fa',
33
+ build_by_default: false)
34
+
35
+blockdev = declare_dependency(link_whole: [libblockdev],
36
+ dependencies: [block])
37
+
38
qmp_ss = qmp_ss.apply(config_host, strict: false)
39
libqmp = static_library('qmp', qmp_ss.sources() + genh,
40
dependencies: qmp_ss.dependencies(),
41
@@ -XXX,XX +XXX,XX @@ foreach m : block_mods + softmmu_mods
42
install_dir: config_host['qemu_moddir'])
43
endforeach
44
45
-softmmu_ss.add(authz, block, chardev, crypto, io, qmp)
46
+softmmu_ss.add(authz, blockdev, chardev, crypto, io, qmp)
47
common_ss.add(qom, qemuutil)
48
49
common_ss.add_all(when: 'CONFIG_SOFTMMU', if_true: [softmmu_ss])
50
diff --git a/storage-daemon/meson.build b/storage-daemon/meson.build
51
index XXXXXXX..XXXXXXX 100644
52
--- a/storage-daemon/meson.build
53
+++ b/storage-daemon/meson.build
54
@@ -XXX,XX +XXX,XX @@
55
qsd_ss = ss.source_set()
56
qsd_ss.add(files('qemu-storage-daemon.c'))
57
-qsd_ss.add(block, chardev, qmp, qom, qemuutil)
58
-qsd_ss.add_all(blockdev_ss)
59
+qsd_ss.add(blockdev, chardev, qmp, qom, qemuutil)
60
61
subdir('qapi')
62
63
--
64
2.26.2
65
diff view generated by jsdifflib
Deleted patch
1
Block exports are used by softmmu, qemu-storage-daemon, and qemu-nbd.
2
They are not used by other programs and are not otherwise needed in
3
libblock.
4
1
5
Undo the recent move of blockdev-nbd.c from blockdev_ss into block_ss.
6
Since bdrv_close_all() (libblock) calls blk_exp_close_all()
7
(libblockdev) a stub function is required..
8
9
Make qemu-nbd.c use signal handling utility functions instead of
10
duplicating the code. This helps because os-posix.c is in libblockdev
11
and it depends on a qemu_system_killed() symbol that qemu-nbd.c lacks.
12
Once we use the signal handling utility functions we also end up
13
providing the necessary symbol.
14
15
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
16
Reviewed-by: Paolo Bonzini <pbonzini@redhat.com>
17
Reviewed-by: Eric Blake <eblake@redhat.com>
18
Message-id: 20200929125516.186715-4-stefanha@redhat.com
19
[Fixed s/ndb/nbd/ typo in commit description as suggested by Eric Blake
20
--Stefan]
21
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
22
---
23
qemu-nbd.c | 21 ++++++++-------------
24
stubs/blk-exp-close-all.c | 7 +++++++
25
block/export/meson.build | 4 ++--
26
meson.build | 4 ++--
27
nbd/meson.build | 2 ++
28
stubs/meson.build | 1 +
29
6 files changed, 22 insertions(+), 17 deletions(-)
30
create mode 100644 stubs/blk-exp-close-all.c
31
32
diff --git a/qemu-nbd.c b/qemu-nbd.c
33
index XXXXXXX..XXXXXXX 100644
34
--- a/qemu-nbd.c
35
+++ b/qemu-nbd.c
36
@@ -XXX,XX +XXX,XX @@
37
#include "qapi/error.h"
38
#include "qemu/cutils.h"
39
#include "sysemu/block-backend.h"
40
+#include "sysemu/runstate.h" /* for qemu_system_killed() prototype */
41
#include "block/block_int.h"
42
#include "block/nbd.h"
43
#include "qemu/main-loop.h"
44
@@ -XXX,XX +XXX,XX @@ QEMU_COPYRIGHT "\n"
45
}
46
47
#ifdef CONFIG_POSIX
48
-static void termsig_handler(int signum)
49
+/*
50
+ * The client thread uses SIGTERM to interrupt the server. A signal
51
+ * handler ensures that "qemu-nbd -v -c" exits with a nice status code.
52
+ */
53
+void qemu_system_killed(int signum, pid_t pid)
54
{
55
qatomic_cmpxchg(&state, RUNNING, TERMINATE);
56
qemu_notify_event();
57
@@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv)
58
BlockExportOptions *export_opts;
59
60
#ifdef CONFIG_POSIX
61
- /*
62
- * Exit gracefully on various signals, which includes SIGTERM used
63
- * by 'qemu-nbd -v -c'.
64
- */
65
- struct sigaction sa_sigterm;
66
- memset(&sa_sigterm, 0, sizeof(sa_sigterm));
67
- sa_sigterm.sa_handler = termsig_handler;
68
- sigaction(SIGTERM, &sa_sigterm, NULL);
69
- sigaction(SIGINT, &sa_sigterm, NULL);
70
- sigaction(SIGHUP, &sa_sigterm, NULL);
71
-
72
- signal(SIGPIPE, SIG_IGN);
73
+ os_setup_early_signal_handling();
74
+ os_setup_signal_handling();
75
#endif
76
77
socket_init();
78
diff --git a/stubs/blk-exp-close-all.c b/stubs/blk-exp-close-all.c
79
new file mode 100644
80
index XXXXXXX..XXXXXXX
81
--- /dev/null
82
+++ b/stubs/blk-exp-close-all.c
83
@@ -XXX,XX +XXX,XX @@
84
+#include "qemu/osdep.h"
85
+#include "block/export.h"
86
+
87
+/* Only used in programs that support block exports (libblockdev.fa) */
88
+void blk_exp_close_all(void)
89
+{
90
+}
91
diff --git a/block/export/meson.build b/block/export/meson.build
92
index XXXXXXX..XXXXXXX 100644
93
--- a/block/export/meson.build
94
+++ b/block/export/meson.build
95
@@ -XXX,XX +XXX,XX @@
96
-block_ss.add(files('export.c'))
97
-block_ss.add(when: ['CONFIG_LINUX', 'CONFIG_VHOST_USER'], if_true: files('vhost-user-blk-server.c'))
98
+blockdev_ss.add(files('export.c'))
99
+blockdev_ss.add(when: ['CONFIG_LINUX', 'CONFIG_VHOST_USER'], if_true: files('vhost-user-blk-server.c'))
100
diff --git a/meson.build b/meson.build
101
index XXXXXXX..XXXXXXX 100644
102
--- a/meson.build
103
+++ b/meson.build
104
@@ -XXX,XX +XXX,XX @@ subdir('dump')
105
106
block_ss.add(files(
107
'block.c',
108
- 'blockdev-nbd.c',
109
'blockjob.c',
110
'job.c',
111
'qemu-io-cmds.c',
112
@@ -XXX,XX +XXX,XX @@ subdir('block')
113
114
blockdev_ss.add(files(
115
'blockdev.c',
116
+ 'blockdev-nbd.c',
117
'iothread.c',
118
'job-qmp.c',
119
))
120
@@ -XXX,XX +XXX,XX @@ if have_tools
121
qemu_io = executable('qemu-io', files('qemu-io.c'),
122
dependencies: [block, qemuutil], install: true)
123
qemu_nbd = executable('qemu-nbd', files('qemu-nbd.c'),
124
- dependencies: [block, qemuutil], install: true)
125
+ dependencies: [blockdev, qemuutil], install: true)
126
127
subdir('storage-daemon')
128
subdir('contrib/rdmacm-mux')
129
diff --git a/nbd/meson.build b/nbd/meson.build
130
index XXXXXXX..XXXXXXX 100644
131
--- a/nbd/meson.build
132
+++ b/nbd/meson.build
133
@@ -XXX,XX +XXX,XX @@
134
block_ss.add(files(
135
'client.c',
136
'common.c',
137
+))
138
+blockdev_ss.add(files(
139
'server.c',
140
))
141
diff --git a/stubs/meson.build b/stubs/meson.build
142
index XXXXXXX..XXXXXXX 100644
143
--- a/stubs/meson.build
144
+++ b/stubs/meson.build
145
@@ -XXX,XX +XXX,XX @@
146
stub_ss.add(files('arch_type.c'))
147
stub_ss.add(files('bdrv-next-monitor-owned.c'))
148
stub_ss.add(files('blk-commit-all.c'))
149
+stub_ss.add(files('blk-exp-close-all.c'))
150
stub_ss.add(files('blockdev-close-all-bdrv-states.c'))
151
stub_ss.add(files('change-state-handler.c'))
152
stub_ss.add(files('cmos.c'))
153
--
154
2.26.2
155
diff view generated by jsdifflib
Deleted patch
1
Make it possible to specify the iothread where the export will run. By
2
default the block node can be moved to other AioContexts later and the
3
export will follow. The fixed-iothread option forces strict behavior
4
that prevents changing AioContext while the export is active. See the
5
QAPI docs for details.
6
1
7
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
8
Message-id: 20200929125516.186715-5-stefanha@redhat.com
9
[Fix stray '#' character in block-export.json and add missing "(since:
10
5.2)" as suggested by Eric Blake.
11
--Stefan]
12
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
13
---
14
qapi/block-export.json | 11 ++++++++++
15
block/export/export.c | 31 +++++++++++++++++++++++++++-
16
block/export/vhost-user-blk-server.c | 5 ++++-
17
nbd/server.c | 2 --
18
4 files changed, 45 insertions(+), 4 deletions(-)
19
20
diff --git a/qapi/block-export.json b/qapi/block-export.json
21
index XXXXXXX..XXXXXXX 100644
22
--- a/qapi/block-export.json
23
+++ b/qapi/block-export.json
24
@@ -XXX,XX +XXX,XX @@
25
# export before completion is signalled. (since: 5.2;
26
# default: false)
27
#
28
+# @iothread: The name of the iothread object where the export will run. The
29
+# default is to use the thread currently associated with the
30
+# block node. (since: 5.2)
31
+#
32
+# @fixed-iothread: True prevents the block node from being moved to another
33
+# thread while the export is active. If true and @iothread is
34
+# given, export creation fails if the block node cannot be
35
+# moved to the iothread. The default is false. (since: 5.2)
36
+#
37
# Since: 4.2
38
##
39
{ 'union': 'BlockExportOptions',
40
'base': { 'type': 'BlockExportType',
41
'id': 'str',
42
+     '*fixed-iothread': 'bool',
43
+     '*iothread': 'str',
44
'node-name': 'str',
45
'*writable': 'bool',
46
'*writethrough': 'bool' },
47
diff --git a/block/export/export.c b/block/export/export.c
48
index XXXXXXX..XXXXXXX 100644
49
--- a/block/export/export.c
50
+++ b/block/export/export.c
51
@@ -XXX,XX +XXX,XX @@
52
53
#include "block/block.h"
54
#include "sysemu/block-backend.h"
55
+#include "sysemu/iothread.h"
56
#include "block/export.h"
57
#include "block/nbd.h"
58
#include "qapi/error.h"
59
@@ -XXX,XX +XXX,XX @@ static const BlockExportDriver *blk_exp_find_driver(BlockExportType type)
60
61
BlockExport *blk_exp_add(BlockExportOptions *export, Error **errp)
62
{
63
+ bool fixed_iothread = export->has_fixed_iothread && export->fixed_iothread;
64
const BlockExportDriver *drv;
65
BlockExport *exp = NULL;
66
BlockDriverState *bs;
67
- BlockBackend *blk;
68
+ BlockBackend *blk = NULL;
69
AioContext *ctx;
70
uint64_t perm;
71
int ret;
72
@@ -XXX,XX +XXX,XX @@ BlockExport *blk_exp_add(BlockExportOptions *export, Error **errp)
73
ctx = bdrv_get_aio_context(bs);
74
aio_context_acquire(ctx);
75
76
+ if (export->has_iothread) {
77
+ IOThread *iothread;
78
+ AioContext *new_ctx;
79
+
80
+ iothread = iothread_by_id(export->iothread);
81
+ if (!iothread) {
82
+ error_setg(errp, "iothread \"%s\" not found", export->iothread);
83
+ goto fail;
84
+ }
85
+
86
+ new_ctx = iothread_get_aio_context(iothread);
87
+
88
+ ret = bdrv_try_set_aio_context(bs, new_ctx, errp);
89
+ if (ret == 0) {
90
+ aio_context_release(ctx);
91
+ aio_context_acquire(new_ctx);
92
+ ctx = new_ctx;
93
+ } else if (fixed_iothread) {
94
+ goto fail;
95
+ }
96
+ }
97
+
98
/*
99
* Block exports are used for non-shared storage migration. Make sure
100
* that BDRV_O_INACTIVE is cleared and the image is ready for write
101
@@ -XXX,XX +XXX,XX @@ BlockExport *blk_exp_add(BlockExportOptions *export, Error **errp)
102
}
103
104
blk = blk_new(ctx, perm, BLK_PERM_ALL);
105
+
106
+ if (!fixed_iothread) {
107
+ blk_set_allow_aio_context_change(blk, true);
108
+ }
109
+
110
ret = blk_insert_bs(blk, bs, errp);
111
if (ret < 0) {
112
goto fail;
113
diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c
114
index XXXXXXX..XXXXXXX 100644
115
--- a/block/export/vhost-user-blk-server.c
116
+++ b/block/export/vhost-user-blk-server.c
117
@@ -XXX,XX +XXX,XX @@ static const VuDevIface vu_blk_iface = {
118
static void blk_aio_attached(AioContext *ctx, void *opaque)
119
{
120
VuBlkExport *vexp = opaque;
121
+
122
+ vexp->export.ctx = ctx;
123
vhost_user_server_attach_aio_context(&vexp->vu_server, ctx);
124
}
125
126
static void blk_aio_detach(void *opaque)
127
{
128
VuBlkExport *vexp = opaque;
129
+
130
vhost_user_server_detach_aio_context(&vexp->vu_server);
131
+ vexp->export.ctx = NULL;
132
}
133
134
static void
135
@@ -XXX,XX +XXX,XX @@ static int vu_blk_exp_create(BlockExport *exp, BlockExportOptions *opts,
136
vu_blk_initialize_config(blk_bs(exp->blk), &vexp->blkcfg,
137
logical_block_size);
138
139
- blk_set_allow_aio_context_change(exp->blk, true);
140
blk_add_aio_context_notifier(exp->blk, blk_aio_attached, blk_aio_detach,
141
vexp);
142
143
diff --git a/nbd/server.c b/nbd/server.c
144
index XXXXXXX..XXXXXXX 100644
145
--- a/nbd/server.c
146
+++ b/nbd/server.c
147
@@ -XXX,XX +XXX,XX @@ static int nbd_export_create(BlockExport *blk_exp, BlockExportOptions *exp_args,
148
return ret;
149
}
150
151
- blk_set_allow_aio_context_change(blk, true);
152
-
153
QTAILQ_INIT(&exp->clients);
154
exp->name = g_strdup(arg->name);
155
exp->description = g_strdup(arg->description);
156
--
157
2.26.2
158
diff view generated by jsdifflib
Deleted patch
1
Allow the number of queues to be configured using --export
2
vhost-user-blk,num-queues=N. This setting should match the QEMU --device
3
vhost-user-blk-pci,num-queues=N setting but QEMU vhost-user-blk.c lowers
4
its own value if the vhost-user-blk backend offers fewer queues than
5
QEMU.
6
1
7
The vhost-user-blk-server.c code is already capable of multi-queue. All
8
virtqueue processing runs in the same AioContext. No new locking is
9
needed.
10
11
Add the num-queues=N option and set the VIRTIO_BLK_F_MQ feature bit.
12
Note that the feature bit only announces the presence of the num_queues
13
configuration space field. It does not promise that there is more than 1
14
virtqueue, so we can set it unconditionally.
15
16
I tested multi-queue by running a random read fio test with numjobs=4 on
17
an -smp 4 guest. After the benchmark finished the guest /proc/interrupts
18
file showed activity on all 4 virtio-blk MSI-X. The /sys/block/vda/mq/
19
directory shows that Linux blk-mq has 4 queues configured.
20
21
An automated test is included in the next commit.
22
23
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
24
Acked-by: Markus Armbruster <armbru@redhat.com>
25
Message-id: 20201001144604.559733-2-stefanha@redhat.com
26
[Fixed accidental tab characters as suggested by Markus Armbruster
27
--Stefan]
28
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
29
---
30
qapi/block-export.json | 10 +++++++---
31
block/export/vhost-user-blk-server.c | 24 ++++++++++++++++++------
32
2 files changed, 25 insertions(+), 9 deletions(-)
33
34
diff --git a/qapi/block-export.json b/qapi/block-export.json
35
index XXXXXXX..XXXXXXX 100644
36
--- a/qapi/block-export.json
37
+++ b/qapi/block-export.json
38
@@ -XXX,XX +XXX,XX @@
39
# SocketAddress types are supported. Passed fds must be UNIX domain
40
# sockets.
41
# @logical-block-size: Logical block size in bytes. Defaults to 512 bytes.
42
+# @num-queues: Number of request virtqueues. Must be greater than 0. Defaults
43
+# to 1.
44
#
45
# Since: 5.2
46
##
47
{ 'struct': 'BlockExportOptionsVhostUserBlk',
48
- 'data': { 'addr': 'SocketAddress', '*logical-block-size': 'size' } }
49
+ 'data': { 'addr': 'SocketAddress',
50
+     '*logical-block-size': 'size',
51
+ '*num-queues': 'uint16'} }
52
53
##
54
# @NbdServerAddOptions:
55
@@ -XXX,XX +XXX,XX @@
56
{ 'union': 'BlockExportOptions',
57
'base': { 'type': 'BlockExportType',
58
'id': 'str',
59
-     '*fixed-iothread': 'bool',
60
-     '*iothread': 'str',
61
+ '*fixed-iothread': 'bool',
62
+ '*iothread': 'str',
63
'node-name': 'str',
64
'*writable': 'bool',
65
'*writethrough': 'bool' },
66
diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c
67
index XXXXXXX..XXXXXXX 100644
68
--- a/block/export/vhost-user-blk-server.c
69
+++ b/block/export/vhost-user-blk-server.c
70
@@ -XXX,XX +XXX,XX @@
71
#include "util/block-helpers.h"
72
73
enum {
74
- VHOST_USER_BLK_MAX_QUEUES = 1,
75
+ VHOST_USER_BLK_NUM_QUEUES_DEFAULT = 1,
76
};
77
struct virtio_blk_inhdr {
78
unsigned char status;
79
@@ -XXX,XX +XXX,XX @@ static uint64_t vu_blk_get_features(VuDev *dev)
80
1ull << VIRTIO_BLK_F_DISCARD |
81
1ull << VIRTIO_BLK_F_WRITE_ZEROES |
82
1ull << VIRTIO_BLK_F_CONFIG_WCE |
83
+ 1ull << VIRTIO_BLK_F_MQ |
84
1ull << VIRTIO_F_VERSION_1 |
85
1ull << VIRTIO_RING_F_INDIRECT_DESC |
86
1ull << VIRTIO_RING_F_EVENT_IDX |
87
@@ -XXX,XX +XXX,XX @@ static void blk_aio_detach(void *opaque)
88
89
static void
90
vu_blk_initialize_config(BlockDriverState *bs,
91
- struct virtio_blk_config *config, uint32_t blk_size)
92
+ struct virtio_blk_config *config,
93
+ uint32_t blk_size,
94
+ uint16_t num_queues)
95
{
96
config->capacity = bdrv_getlength(bs) >> BDRV_SECTOR_BITS;
97
config->blk_size = blk_size;
98
@@ -XXX,XX +XXX,XX @@ vu_blk_initialize_config(BlockDriverState *bs,
99
config->seg_max = 128 - 2;
100
config->min_io_size = 1;
101
config->opt_io_size = 1;
102
- config->num_queues = VHOST_USER_BLK_MAX_QUEUES;
103
+ config->num_queues = num_queues;
104
config->max_discard_sectors = 32768;
105
config->max_discard_seg = 1;
106
config->discard_sector_alignment = config->blk_size >> 9;
107
@@ -XXX,XX +XXX,XX @@ static int vu_blk_exp_create(BlockExport *exp, BlockExportOptions *opts,
108
BlockExportOptionsVhostUserBlk *vu_opts = &opts->u.vhost_user_blk;
109
Error *local_err = NULL;
110
uint64_t logical_block_size;
111
+ uint16_t num_queues = VHOST_USER_BLK_NUM_QUEUES_DEFAULT;
112
113
vexp->writable = opts->writable;
114
vexp->blkcfg.wce = 0;
115
@@ -XXX,XX +XXX,XX @@ static int vu_blk_exp_create(BlockExport *exp, BlockExportOptions *opts,
116
}
117
vexp->blk_size = logical_block_size;
118
blk_set_guest_block_size(exp->blk, logical_block_size);
119
+
120
+ if (vu_opts->has_num_queues) {
121
+ num_queues = vu_opts->num_queues;
122
+ }
123
+ if (num_queues == 0) {
124
+ error_setg(errp, "num-queues must be greater than 0");
125
+ return -EINVAL;
126
+ }
127
+
128
vu_blk_initialize_config(blk_bs(exp->blk), &vexp->blkcfg,
129
- logical_block_size);
130
+ logical_block_size, num_queues);
131
132
blk_add_aio_context_notifier(exp->blk, blk_aio_attached, blk_aio_detach,
133
vexp);
134
135
if (!vhost_user_server_start(&vexp->vu_server, vu_opts->addr, exp->ctx,
136
- VHOST_USER_BLK_MAX_QUEUES, &vu_blk_iface,
137
- errp)) {
138
+ num_queues, &vu_blk_iface, errp)) {
139
blk_remove_aio_context_notifier(exp->blk, blk_aio_attached,
140
blk_aio_detach, vexp);
141
return -EADDRNOTAVAIL;
142
--
143
2.26.2
144
diff view generated by jsdifflib
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
1
From: Andrey Drobyshev <andrey.drobyshev@virtuozzo.com>
2
2
3
bdrv_co_block_status_above has several design problems with handling
3
This is going to be used in the subsequent commit as requests alignment
4
short backing files:
4
(in particular, during copy-on-read). This value only makes sense for
5
the formats which support subclusters (currently QCOW2 only). If this
6
field isn't set by driver's own bdrv_get_info() implementation, we
7
simply set it equal to the cluster size thus treating each cluster as
8
having a single subcluster.
5
9
6
1. With want_zeros=true, it may return ret with BDRV_BLOCK_ZERO but
10
Reviewed-by: Eric Blake <eblake@redhat.com>
7
without BDRV_BLOCK_ALLOCATED flag, when actually short backing file
11
Reviewed-by: Denis V. Lunev <den@openvz.org>
8
which produces these after-EOF zeros is inside requested backing
12
Signed-off-by: Andrey Drobyshev <andrey.drobyshev@virtuozzo.com>
9
sequence.
13
Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru>
14
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
15
Message-ID: <20230711172553.234055-2-andrey.drobyshev@virtuozzo.com>
16
---
17
include/block/block-common.h | 5 +++++
18
block.c | 7 +++++++
19
block/qcow2.c | 1 +
20
3 files changed, 13 insertions(+)
10
21
11
2. With want_zero=false, it may return pnum=0 prior to actual EOF,
22
diff --git a/include/block/block-common.h b/include/block/block-common.h
12
because of EOF of short backing file.
13
14
Fix these things, making logic about short backing files clearer.
15
16
With fixed bdrv_block_status_above we also have to improve is_zero in
17
qcow2 code, otherwise iotest 154 will fail, because with this patch we
18
stop to merge zeros of different types (produced by fully unallocated
19
in the whole backing chain regions vs produced by short backing files).
20
21
Note also, that this patch leaves for another day the general problem
22
around block-status: misuse of BDRV_BLOCK_ALLOCATED as is-fs-allocated
23
vs go-to-backing.
24
25
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
26
Reviewed-by: Alberto Garcia <berto@igalia.com>
27
Reviewed-by: Eric Blake <eblake@redhat.com>
28
Message-id: 20200924194003.22080-2-vsementsov@virtuozzo.com
29
[Fix s/comes/come/ as suggested by Eric Blake
30
--Stefan]
31
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
32
---
33
block/io.c | 68 ++++++++++++++++++++++++++++++++++++++++-----------
34
block/qcow2.c | 16 ++++++++++--
35
2 files changed, 68 insertions(+), 16 deletions(-)
36
37
diff --git a/block/io.c b/block/io.c
38
index XXXXXXX..XXXXXXX 100644
23
index XXXXXXX..XXXXXXX 100644
39
--- a/block/io.c
24
--- a/include/block/block-common.h
40
+++ b/block/io.c
25
+++ b/include/block/block-common.h
41
@@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs,
26
@@ -XXX,XX +XXX,XX @@ typedef struct BlockZoneWps {
42
int64_t *map,
27
typedef struct BlockDriverInfo {
43
BlockDriverState **file)
28
/* in bytes, 0 if irrelevant */
44
{
29
int cluster_size;
45
+ int ret;
30
+ /*
46
BlockDriverState *p;
31
+ * A fraction of cluster_size, if supported (currently QCOW2 only); if
47
- int ret = 0;
32
+ * disabled or unsupported, set equal to cluster_size.
48
- bool first = true;
33
+ */
49
+ int64_t eof = 0;
34
+ int subcluster_size;
50
35
/* offset at which the VM state can be saved (0 if not possible) */
51
assert(bs != base);
36
int64_t vm_state_offset;
52
- for (p = bs; p != base; p = bdrv_filter_or_cow_bs(p)) {
37
bool is_dirty;
53
+
38
diff --git a/block.c b/block.c
54
+ ret = bdrv_co_block_status(bs, want_zero, offset, bytes, pnum, map, file);
39
index XXXXXXX..XXXXXXX 100644
55
+ if (ret < 0 || *pnum == 0 || ret & BDRV_BLOCK_ALLOCATED) {
40
--- a/block.c
56
+ return ret;
41
+++ b/block.c
42
@@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_co_get_info(BlockDriverState *bs, BlockDriverInfo *bdi)
43
}
44
memset(bdi, 0, sizeof(*bdi));
45
ret = drv->bdrv_co_get_info(bs, bdi);
46
+ if (bdi->subcluster_size == 0) {
47
+ /*
48
+ * If the driver left this unset, subclusters are not supported.
49
+ * Then it is safe to treat each cluster as having only one subcluster.
50
+ */
51
+ bdi->subcluster_size = bdi->cluster_size;
57
+ }
52
+ }
58
+
53
if (ret < 0) {
59
+ if (ret & BDRV_BLOCK_EOF) {
54
return ret;
60
+ eof = offset + *pnum;
61
+ }
62
+
63
+ assert(*pnum <= bytes);
64
+ bytes = *pnum;
65
+
66
+ for (p = bdrv_filter_or_cow_bs(bs); p != base;
67
+ p = bdrv_filter_or_cow_bs(p))
68
+ {
69
ret = bdrv_co_block_status(p, want_zero, offset, bytes, pnum, map,
70
file);
71
if (ret < 0) {
72
- break;
73
+ return ret;
74
}
75
- if (ret & BDRV_BLOCK_ZERO && ret & BDRV_BLOCK_EOF && !first) {
76
+ if (*pnum == 0) {
77
/*
78
- * Reading beyond the end of the file continues to read
79
- * zeroes, but we can only widen the result to the
80
- * unallocated length we learned from an earlier
81
- * iteration.
82
+ * The top layer deferred to this layer, and because this layer is
83
+ * short, any zeroes that we synthesize beyond EOF behave as if they
84
+ * were allocated at this layer.
85
+ *
86
+ * We don't include BDRV_BLOCK_EOF into ret, as upper layer may be
87
+ * larger. We'll add BDRV_BLOCK_EOF if needed at function end, see
88
+ * below.
89
*/
90
+ assert(ret & BDRV_BLOCK_EOF);
91
*pnum = bytes;
92
+ if (file) {
93
+ *file = p;
94
+ }
95
+ ret = BDRV_BLOCK_ZERO | BDRV_BLOCK_ALLOCATED;
96
+ break;
97
}
98
- if (ret & (BDRV_BLOCK_ZERO | BDRV_BLOCK_DATA)) {
99
+ if (ret & BDRV_BLOCK_ALLOCATED) {
100
+ /*
101
+ * We've found the node and the status, we must break.
102
+ *
103
+ * Drop BDRV_BLOCK_EOF, as it's not for upper layer, which may be
104
+ * larger. We'll add BDRV_BLOCK_EOF if needed at function end, see
105
+ * below.
106
+ */
107
+ ret &= ~BDRV_BLOCK_EOF;
108
break;
109
}
110
- /* [offset, pnum] unallocated on this layer, which could be only
111
- * the first part of [offset, bytes]. */
112
- bytes = MIN(bytes, *pnum);
113
- first = false;
114
+
115
+ /*
116
+ * OK, [offset, offset + *pnum) region is unallocated on this layer,
117
+ * let's continue the diving.
118
+ */
119
+ assert(*pnum <= bytes);
120
+ bytes = *pnum;
121
+ }
122
+
123
+ if (offset + *pnum == eof) {
124
+ ret |= BDRV_BLOCK_EOF;
125
}
55
}
126
+
127
return ret;
128
}
129
130
diff --git a/block/qcow2.c b/block/qcow2.c
56
diff --git a/block/qcow2.c b/block/qcow2.c
131
index XXXXXXX..XXXXXXX 100644
57
index XXXXXXX..XXXXXXX 100644
132
--- a/block/qcow2.c
58
--- a/block/qcow2.c
133
+++ b/block/qcow2.c
59
+++ b/block/qcow2.c
134
@@ -XXX,XX +XXX,XX @@ static bool is_zero(BlockDriverState *bs, int64_t offset, int64_t bytes)
60
@@ -XXX,XX +XXX,XX @@ qcow2_co_get_info(BlockDriverState *bs, BlockDriverInfo *bdi)
135
if (!bytes) {
61
{
136
return true;
62
BDRVQcow2State *s = bs->opaque;
137
}
63
bdi->cluster_size = s->cluster_size;
138
- res = bdrv_block_status_above(bs, NULL, offset, bytes, &nr, NULL, NULL);
64
+ bdi->subcluster_size = s->subcluster_size;
139
- return res >= 0 && (res & BDRV_BLOCK_ZERO) && nr == bytes;
65
bdi->vm_state_offset = qcow2_vm_state_offset(s);
140
+
66
bdi->is_dirty = s->incompatible_features & QCOW2_INCOMPAT_DIRTY;
141
+ /*
67
return 0;
142
+ * bdrv_block_status_above doesn't merge different types of zeros, for
143
+ * example, zeros which come from the region which is unallocated in
144
+ * the whole backing chain, and zeros which come because of a short
145
+ * backing file. So, we need a loop.
146
+ */
147
+ do {
148
+ res = bdrv_block_status_above(bs, NULL, offset, bytes, &nr, NULL, NULL);
149
+ offset += nr;
150
+ bytes -= nr;
151
+ } while (res >= 0 && (res & BDRV_BLOCK_ZERO) && nr && bytes);
152
+
153
+ return res >= 0 && (res & BDRV_BLOCK_ZERO) && bytes == 0;
154
}
155
156
static coroutine_fn int qcow2_co_pwrite_zeroes(BlockDriverState *bs,
157
--
68
--
158
2.26.2
69
2.41.0
159
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
In order to reuse bdrv_common_block_status_above in
4
bdrv_is_allocated_above, let's support include_base parameter.
5
6
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
7
Reviewed-by: Alberto Garcia <berto@igalia.com>
8
Reviewed-by: Eric Blake <eblake@redhat.com>
9
Message-id: 20200924194003.22080-3-vsementsov@virtuozzo.com
10
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
11
---
12
block/coroutines.h | 2 ++
13
block/io.c | 21 ++++++++++++++-------
14
2 files changed, 16 insertions(+), 7 deletions(-)
15
16
diff --git a/block/coroutines.h b/block/coroutines.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/block/coroutines.h
19
+++ b/block/coroutines.h
20
@@ -XXX,XX +XXX,XX @@ bdrv_pwritev(BdrvChild *child, int64_t offset, unsigned int bytes,
21
int coroutine_fn
22
bdrv_co_common_block_status_above(BlockDriverState *bs,
23
BlockDriverState *base,
24
+ bool include_base,
25
bool want_zero,
26
int64_t offset,
27
int64_t bytes,
28
@@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs,
29
int generated_co_wrapper
30
bdrv_common_block_status_above(BlockDriverState *bs,
31
BlockDriverState *base,
32
+ bool include_base,
33
bool want_zero,
34
int64_t offset,
35
int64_t bytes,
36
diff --git a/block/io.c b/block/io.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/block/io.c
39
+++ b/block/io.c
40
@@ -XXX,XX +XXX,XX @@ early_out:
41
int coroutine_fn
42
bdrv_co_common_block_status_above(BlockDriverState *bs,
43
BlockDriverState *base,
44
+ bool include_base,
45
bool want_zero,
46
int64_t offset,
47
int64_t bytes,
48
@@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs,
49
BlockDriverState *p;
50
int64_t eof = 0;
51
52
- assert(bs != base);
53
+ assert(include_base || bs != base);
54
+ assert(!include_base || base); /* Can't include NULL base */
55
56
ret = bdrv_co_block_status(bs, want_zero, offset, bytes, pnum, map, file);
57
- if (ret < 0 || *pnum == 0 || ret & BDRV_BLOCK_ALLOCATED) {
58
+ if (ret < 0 || *pnum == 0 || ret & BDRV_BLOCK_ALLOCATED || bs == base) {
59
return ret;
60
}
61
62
@@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs,
63
assert(*pnum <= bytes);
64
bytes = *pnum;
65
66
- for (p = bdrv_filter_or_cow_bs(bs); p != base;
67
+ for (p = bdrv_filter_or_cow_bs(bs); include_base || p != base;
68
p = bdrv_filter_or_cow_bs(p))
69
{
70
ret = bdrv_co_block_status(p, want_zero, offset, bytes, pnum, map,
71
@@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs,
72
break;
73
}
74
75
+ if (p == base) {
76
+ assert(include_base);
77
+ break;
78
+ }
79
+
80
/*
81
* OK, [offset, offset + *pnum) region is unallocated on this layer,
82
* let's continue the diving.
83
@@ -XXX,XX +XXX,XX @@ int bdrv_block_status_above(BlockDriverState *bs, BlockDriverState *base,
84
int64_t offset, int64_t bytes, int64_t *pnum,
85
int64_t *map, BlockDriverState **file)
86
{
87
- return bdrv_common_block_status_above(bs, base, true, offset, bytes,
88
+ return bdrv_common_block_status_above(bs, base, false, true, offset, bytes,
89
pnum, map, file);
90
}
91
92
@@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_is_allocated(BlockDriverState *bs, int64_t offset,
93
int ret;
94
int64_t dummy;
95
96
- ret = bdrv_common_block_status_above(bs, bdrv_filter_or_cow_bs(bs), false,
97
- offset, bytes, pnum ? pnum : &dummy,
98
- NULL, NULL);
99
+ ret = bdrv_common_block_status_above(bs, bs, true, false, offset,
100
+ bytes, pnum ? pnum : &dummy, NULL,
101
+ NULL);
102
if (ret < 0) {
103
return ret;
104
}
105
--
106
2.26.2
107
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
We are going to reuse bdrv_common_block_status_above in
4
bdrv_is_allocated_above. bdrv_is_allocated_above may be called with
5
include_base == false and still bs == base (for ex. from img_rebase()).
6
7
So, support this corner case.
8
9
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
10
Reviewed-by: Kevin Wolf <kwolf@redhat.com>
11
Reviewed-by: Eric Blake <eblake@redhat.com>
12
Reviewed-by: Alberto Garcia <berto@igalia.com>
13
Message-id: 20200924194003.22080-4-vsementsov@virtuozzo.com
14
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
15
---
16
block/io.c | 6 +++++-
17
1 file changed, 5 insertions(+), 1 deletion(-)
18
19
diff --git a/block/io.c b/block/io.c
20
index XXXXXXX..XXXXXXX 100644
21
--- a/block/io.c
22
+++ b/block/io.c
23
@@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs,
24
BlockDriverState *p;
25
int64_t eof = 0;
26
27
- assert(include_base || bs != base);
28
assert(!include_base || base); /* Can't include NULL base */
29
30
+ if (!include_base && bs == base) {
31
+ *pnum = bytes;
32
+ return 0;
33
+ }
34
+
35
ret = bdrv_co_block_status(bs, want_zero, offset, bytes, pnum, map, file);
36
if (ret < 0 || *pnum == 0 || ret & BDRV_BLOCK_ALLOCATED || bs == base) {
37
return ret;
38
--
39
2.26.2
40
diff view generated by jsdifflib
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
1
From: Andrey Drobyshev <andrey.drobyshev@virtuozzo.com>
2
2
3
bdrv_is_allocated_above wrongly handles short backing files: it reports
3
When target image is using subclusters, and we align the request during
4
after-EOF space as UNALLOCATED which is wrong, as on read the data is
4
copy-on-read, it makes sense to align to subcluster_size rather than
5
generated on the level of short backing file (if all overlays have
5
cluster_size. Otherwise we end up with unnecessary allocations.
6
unallocated areas at that place).
6
7
7
This commit renames bdrv_round_to_clusters() to bdrv_round_to_subclusters()
8
Reusing bdrv_common_block_status_above fixes the issue and unifies code
8
and utilizes subcluster_size field of BlockDriverInfo to make necessary
9
path.
9
alignments. It affects copy-on-read as well as mirror job (which is
10
10
using bdrv_round_to_clusters()).
11
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
11
12
This change also fixes the following bug with failing assert (covered by
13
the test in the subsequent commit):
14
15
qemu-img create -f qcow2 base.qcow2 64K
16
qemu-img create -f qcow2 -o extended_l2=on,backing_file=base.qcow2,backing_fmt=qcow2 img.qcow2 64K
17
qemu-io -c "write -P 0xaa 0 2K" img.qcow2
18
qemu-io -C -c "read -P 0x00 2K 62K" img.qcow2
19
20
qemu-io: ../block/io.c:1236: bdrv_co_do_copy_on_readv: Assertion `skip_bytes < pnum' failed.
21
12
Reviewed-by: Eric Blake <eblake@redhat.com>
22
Reviewed-by: Eric Blake <eblake@redhat.com>
13
Reviewed-by: Alberto Garcia <berto@igalia.com>
23
Reviewed-by: Denis V. Lunev <den@openvz.org>
14
Message-id: 20200924194003.22080-5-vsementsov@virtuozzo.com
24
Signed-off-by: Andrey Drobyshev <andrey.drobyshev@virtuozzo.com>
15
[Fix s/has/have/ as suggested by Eric Blake. Fix s/area/areas/.
25
Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru>
16
--Stefan]
17
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
26
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
27
Message-ID: <20230711172553.234055-3-andrey.drobyshev@virtuozzo.com>
18
---
28
---
19
block/io.c | 43 +++++--------------------------------------
29
include/block/block-io.h | 8 +++----
20
1 file changed, 5 insertions(+), 38 deletions(-)
30
block/io.c | 50 ++++++++++++++++++++--------------------
21
31
block/mirror.c | 8 +++----
32
3 files changed, 33 insertions(+), 33 deletions(-)
33
34
diff --git a/include/block/block-io.h b/include/block/block-io.h
35
index XXXXXXX..XXXXXXX 100644
36
--- a/include/block/block-io.h
37
+++ b/include/block/block-io.h
38
@@ -XXX,XX +XXX,XX @@ bdrv_get_info(BlockDriverState *bs, BlockDriverInfo *bdi);
39
ImageInfoSpecific *bdrv_get_specific_info(BlockDriverState *bs,
40
Error **errp);
41
BlockStatsSpecific *bdrv_get_specific_stats(BlockDriverState *bs);
42
-void bdrv_round_to_clusters(BlockDriverState *bs,
43
- int64_t offset, int64_t bytes,
44
- int64_t *cluster_offset,
45
- int64_t *cluster_bytes);
46
+void bdrv_round_to_subclusters(BlockDriverState *bs,
47
+ int64_t offset, int64_t bytes,
48
+ int64_t *cluster_offset,
49
+ int64_t *cluster_bytes);
50
51
void bdrv_get_backing_filename(BlockDriverState *bs,
52
char *filename, int filename_size);
22
diff --git a/block/io.c b/block/io.c
53
diff --git a/block/io.c b/block/io.c
23
index XXXXXXX..XXXXXXX 100644
54
index XXXXXXX..XXXXXXX 100644
24
--- a/block/io.c
55
--- a/block/io.c
25
+++ b/block/io.c
56
+++ b/block/io.c
26
@@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_is_allocated(BlockDriverState *bs, int64_t offset,
57
@@ -XXX,XX +XXX,XX @@ BdrvTrackedRequest *coroutine_fn bdrv_co_get_self_request(BlockDriverState *bs)
27
* at 'offset + *pnum' may return the same allocation status (in other
58
}
28
* words, the result is not necessarily the maximum possible range);
59
29
* but 'pnum' will only be 0 when end of file is reached.
60
/**
30
- *
61
- * Round a region to cluster boundaries
62
+ * Round a region to subcluster (if supported) or cluster boundaries
31
*/
63
*/
32
int bdrv_is_allocated_above(BlockDriverState *top,
64
void coroutine_fn GRAPH_RDLOCK
33
BlockDriverState *base,
65
-bdrv_round_to_clusters(BlockDriverState *bs, int64_t offset, int64_t bytes,
34
bool include_base, int64_t offset,
66
- int64_t *cluster_offset, int64_t *cluster_bytes)
35
int64_t bytes, int64_t *pnum)
67
+bdrv_round_to_subclusters(BlockDriverState *bs, int64_t offset, int64_t bytes,
68
+ int64_t *align_offset, int64_t *align_bytes)
36
{
69
{
37
- BlockDriverState *intermediate;
70
BlockDriverInfo bdi;
38
- int ret;
71
IO_CODE();
39
- int64_t n = bytes;
72
- if (bdrv_co_get_info(bs, &bdi) < 0 || bdi.cluster_size == 0) {
40
-
73
- *cluster_offset = offset;
41
- assert(base || !include_base);
74
- *cluster_bytes = bytes;
42
-
75
+ if (bdrv_co_get_info(bs, &bdi) < 0 || bdi.subcluster_size == 0) {
43
- intermediate = top;
76
+ *align_offset = offset;
44
- while (include_base || intermediate != base) {
77
+ *align_bytes = bytes;
45
- int64_t pnum_inter;
78
} else {
46
- int64_t size_inter;
79
- int64_t c = bdi.cluster_size;
47
-
80
- *cluster_offset = QEMU_ALIGN_DOWN(offset, c);
48
- assert(intermediate);
81
- *cluster_bytes = QEMU_ALIGN_UP(offset - *cluster_offset + bytes, c);
49
- ret = bdrv_is_allocated(intermediate, offset, bytes, &pnum_inter);
82
+ int64_t c = bdi.subcluster_size;
50
- if (ret < 0) {
83
+ *align_offset = QEMU_ALIGN_DOWN(offset, c);
51
- return ret;
84
+ *align_bytes = QEMU_ALIGN_UP(offset - *align_offset + bytes, c);
52
- }
53
- if (ret) {
54
- *pnum = pnum_inter;
55
- return 1;
56
- }
57
-
58
- size_inter = bdrv_getlength(intermediate);
59
- if (size_inter < 0) {
60
- return size_inter;
61
- }
62
- if (n > pnum_inter &&
63
- (intermediate == top || offset + pnum_inter < size_inter)) {
64
- n = pnum_inter;
65
- }
66
-
67
- if (intermediate == base) {
68
- break;
69
- }
70
-
71
- intermediate = bdrv_filter_or_cow_bs(intermediate);
72
+ int ret = bdrv_common_block_status_above(top, base, include_base, false,
73
+ offset, bytes, pnum, NULL, NULL);
74
+ if (ret < 0) {
75
+ return ret;
76
}
85
}
77
78
- *pnum = n;
79
- return 0;
80
+ return !!(ret & BDRV_BLOCK_ALLOCATED);
81
}
86
}
82
87
83
int coroutine_fn
88
@@ -XXX,XX +XXX,XX @@ bdrv_co_do_copy_on_readv(BdrvChild *child, int64_t offset, int64_t bytes,
89
void *bounce_buffer = NULL;
90
91
BlockDriver *drv = bs->drv;
92
- int64_t cluster_offset;
93
- int64_t cluster_bytes;
94
+ int64_t align_offset;
95
+ int64_t align_bytes;
96
int64_t skip_bytes;
97
int ret;
98
int max_transfer = MIN_NON_ZERO(bs->bl.max_transfer,
99
@@ -XXX,XX +XXX,XX @@ bdrv_co_do_copy_on_readv(BdrvChild *child, int64_t offset, int64_t bytes,
100
* BDRV_REQUEST_MAX_BYTES (even when the original read did not), which
101
* is one reason we loop rather than doing it all at once.
102
*/
103
- bdrv_round_to_clusters(bs, offset, bytes, &cluster_offset, &cluster_bytes);
104
- skip_bytes = offset - cluster_offset;
105
+ bdrv_round_to_subclusters(bs, offset, bytes, &align_offset, &align_bytes);
106
+ skip_bytes = offset - align_offset;
107
108
trace_bdrv_co_do_copy_on_readv(bs, offset, bytes,
109
- cluster_offset, cluster_bytes);
110
+ align_offset, align_bytes);
111
112
- while (cluster_bytes) {
113
+ while (align_bytes) {
114
int64_t pnum;
115
116
if (skip_write) {
117
ret = 1; /* "already allocated", so nothing will be copied */
118
- pnum = MIN(cluster_bytes, max_transfer);
119
+ pnum = MIN(align_bytes, max_transfer);
120
} else {
121
- ret = bdrv_is_allocated(bs, cluster_offset,
122
- MIN(cluster_bytes, max_transfer), &pnum);
123
+ ret = bdrv_is_allocated(bs, align_offset,
124
+ MIN(align_bytes, max_transfer), &pnum);
125
if (ret < 0) {
126
/*
127
* Safe to treat errors in querying allocation as if
128
* unallocated; we'll probably fail again soon on the
129
* read, but at least that will set a decent errno.
130
*/
131
- pnum = MIN(cluster_bytes, max_transfer);
132
+ pnum = MIN(align_bytes, max_transfer);
133
}
134
135
/* Stop at EOF if the image ends in the middle of the cluster */
136
@@ -XXX,XX +XXX,XX @@ bdrv_co_do_copy_on_readv(BdrvChild *child, int64_t offset, int64_t bytes,
137
/* Must copy-on-read; use the bounce buffer */
138
pnum = MIN(pnum, MAX_BOUNCE_BUFFER);
139
if (!bounce_buffer) {
140
- int64_t max_we_need = MAX(pnum, cluster_bytes - pnum);
141
+ int64_t max_we_need = MAX(pnum, align_bytes - pnum);
142
int64_t max_allowed = MIN(max_transfer, MAX_BOUNCE_BUFFER);
143
int64_t bounce_buffer_len = MIN(max_we_need, max_allowed);
144
145
@@ -XXX,XX +XXX,XX @@ bdrv_co_do_copy_on_readv(BdrvChild *child, int64_t offset, int64_t bytes,
146
}
147
qemu_iovec_init_buf(&local_qiov, bounce_buffer, pnum);
148
149
- ret = bdrv_driver_preadv(bs, cluster_offset, pnum,
150
+ ret = bdrv_driver_preadv(bs, align_offset, pnum,
151
&local_qiov, 0, 0);
152
if (ret < 0) {
153
goto err;
154
@@ -XXX,XX +XXX,XX @@ bdrv_co_do_copy_on_readv(BdrvChild *child, int64_t offset, int64_t bytes,
155
/* FIXME: Should we (perhaps conditionally) be setting
156
* BDRV_REQ_MAY_UNMAP, if it will allow for a sparser copy
157
* that still correctly reads as zero? */
158
- ret = bdrv_co_do_pwrite_zeroes(bs, cluster_offset, pnum,
159
+ ret = bdrv_co_do_pwrite_zeroes(bs, align_offset, pnum,
160
BDRV_REQ_WRITE_UNCHANGED);
161
} else {
162
/* This does not change the data on the disk, it is not
163
* necessary to flush even in cache=writethrough mode.
164
*/
165
- ret = bdrv_driver_pwritev(bs, cluster_offset, pnum,
166
+ ret = bdrv_driver_pwritev(bs, align_offset, pnum,
167
&local_qiov, 0,
168
BDRV_REQ_WRITE_UNCHANGED);
169
}
170
@@ -XXX,XX +XXX,XX @@ bdrv_co_do_copy_on_readv(BdrvChild *child, int64_t offset, int64_t bytes,
171
}
172
}
173
174
- cluster_offset += pnum;
175
- cluster_bytes -= pnum;
176
+ align_offset += pnum;
177
+ align_bytes -= pnum;
178
progress += pnum - skip_bytes;
179
skip_bytes = 0;
180
}
181
diff --git a/block/mirror.c b/block/mirror.c
182
index XXXXXXX..XXXXXXX 100644
183
--- a/block/mirror.c
184
+++ b/block/mirror.c
185
@@ -XXX,XX +XXX,XX @@ static int coroutine_fn mirror_cow_align(MirrorBlockJob *s, int64_t *offset,
186
need_cow |= !test_bit((*offset + *bytes - 1) / s->granularity,
187
s->cow_bitmap);
188
if (need_cow) {
189
- bdrv_round_to_clusters(blk_bs(s->target), *offset, *bytes,
190
- &align_offset, &align_bytes);
191
+ bdrv_round_to_subclusters(blk_bs(s->target), *offset, *bytes,
192
+ &align_offset, &align_bytes);
193
}
194
195
if (align_bytes > max_bytes) {
196
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn mirror_iteration(MirrorBlockJob *s)
197
int64_t target_offset;
198
int64_t target_bytes;
199
WITH_GRAPH_RDLOCK_GUARD() {
200
- bdrv_round_to_clusters(blk_bs(s->target), offset, io_bytes,
201
- &target_offset, &target_bytes);
202
+ bdrv_round_to_subclusters(blk_bs(s->target), offset, io_bytes,
203
+ &target_offset, &target_bytes);
204
}
205
if (target_offset == offset &&
206
target_bytes == io_bytes) {
84
--
207
--
85
2.26.2
208
2.41.0
86
diff view generated by jsdifflib
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
1
From: Andrey Drobyshev <andrey.drobyshev@virtuozzo.com>
2
2
3
These cases are fixed by previous patches around block_status and
3
Add testcase which checks that allocations during copy-on-read are
4
is_allocated.
4
performed on the subcluster basis when subclusters are enabled in target
5
image.
5
6
6
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
7
This testcase also triggers the following assert with previous commit
8
not being applied, so we check that as well:
9
10
qemu-io: ../block/io.c:1236: bdrv_co_do_copy_on_readv: Assertion `skip_bytes < pnum' failed.
11
7
Reviewed-by: Eric Blake <eblake@redhat.com>
12
Reviewed-by: Eric Blake <eblake@redhat.com>
8
Reviewed-by: Alberto Garcia <berto@igalia.com>
13
Reviewed-by: Denis V. Lunev <den@openvz.org>
9
Message-id: 20200924194003.22080-6-vsementsov@virtuozzo.com
14
Signed-off-by: Andrey Drobyshev <andrey.drobyshev@virtuozzo.com>
15
Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru>
10
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
16
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
17
Message-ID: <20230711172553.234055-4-andrey.drobyshev@virtuozzo.com>
11
---
18
---
12
tests/qemu-iotests/274 | 20 +++++++++++
19
tests/qemu-iotests/197 | 29 +++++++++++++++++++++++++++++
13
tests/qemu-iotests/274.out | 68 ++++++++++++++++++++++++++++++++++++++
20
tests/qemu-iotests/197.out | 24 ++++++++++++++++++++++++
14
2 files changed, 88 insertions(+)
21
2 files changed, 53 insertions(+)
15
22
16
diff --git a/tests/qemu-iotests/274 b/tests/qemu-iotests/274
23
diff --git a/tests/qemu-iotests/197 b/tests/qemu-iotests/197
17
index XXXXXXX..XXXXXXX 100755
24
index XXXXXXX..XXXXXXX 100755
18
--- a/tests/qemu-iotests/274
25
--- a/tests/qemu-iotests/197
19
+++ b/tests/qemu-iotests/274
26
+++ b/tests/qemu-iotests/197
20
@@ -XXX,XX +XXX,XX @@ with iotests.FilePath('base') as base, \
27
@@ -XXX,XX +XXX,XX @@ $QEMU_IO -f qcow2 -C -c 'read 0 1024' "$TEST_WRAP" | _filter_qemu_io
21
iotests.qemu_io_log('-c', 'read -P 1 0 %d' % size_short, mid)
28
$QEMU_IO -f qcow2 -c map "$TEST_WRAP"
22
iotests.qemu_io_log('-c', 'read -P 0 %d %d' % (size_short, size_diff), mid)
29
_check_test_img
23
30
24
+ iotests.log('=== Testing qemu-img commit (top -> base) ===')
31
+echo
32
+echo '=== Copy-on-read with subclusters ==='
33
+echo
25
+
34
+
26
+ create_chain()
35
+# Create base and top images 64K (1 cluster) each. Make subclusters enabled
27
+ iotests.qemu_img_log('commit', '-b', base, top)
36
+# for the top image
28
+ iotests.img_info_log(base)
37
+_make_test_img 64K
29
+ iotests.qemu_io_log('-c', 'read -P 1 0 %d' % size_short, base)
38
+IMGPROTO=file IMGFMT=qcow2 TEST_IMG_FILE="$TEST_WRAP" \
30
+ iotests.qemu_io_log('-c', 'read -P 0 %d %d' % (size_short, size_diff), base)
39
+ _make_test_img --no-opts -o extended_l2=true -F "$IMGFMT" -b "$TEST_IMG" \
40
+ 64K | _filter_img_create
31
+
41
+
32
+ iotests.log('=== Testing QMP active commit (top -> base) ===')
42
+$QEMU_IO -c "write -P 0xaa 0 64k" "$TEST_IMG" | _filter_qemu_io
33
+
43
+
34
+ create_chain()
44
+# Allocate individual subclusters in the top image, and not the whole cluster
35
+ with create_vm() as vm:
45
+$QEMU_IO -c "write -P 0xbb 28K 2K" -c "write -P 0xcc 34K 2K" "$TEST_WRAP" \
36
+ vm.launch()
46
+ | _filter_qemu_io
37
+ vm.qmp_log('block-commit', device='top', base_node='base',
38
+ job_id='job0', auto_dismiss=False)
39
+ vm.run_job('job0', wait=5)
40
+
47
+
41
+ iotests.img_info_log(mid)
48
+# Only 2 subclusters should be allocated in the top image at this point
42
+ iotests.qemu_io_log('-c', 'read -P 1 0 %d' % size_short, base)
49
+$QEMU_IMG map "$TEST_WRAP" | _filter_qemu_img_map
43
+ iotests.qemu_io_log('-c', 'read -P 0 %d %d' % (size_short, size_diff), base)
50
+
44
51
+# Actual copy-on-read operation
45
iotests.log('== Resize tests ==')
52
+$QEMU_IO -C -c "read -P 0xaa 30K 4K" "$TEST_WRAP" | _filter_qemu_io
46
53
+
47
diff --git a/tests/qemu-iotests/274.out b/tests/qemu-iotests/274.out
54
+# And here we should have 4 subclusters allocated right in the middle of the
55
+# top image. Make sure the whole cluster remains unallocated
56
+$QEMU_IMG map "$TEST_WRAP" | _filter_qemu_img_map
57
+
58
+_check_test_img
59
+
60
# success, all done
61
echo '*** done'
62
status=0
63
diff --git a/tests/qemu-iotests/197.out b/tests/qemu-iotests/197.out
48
index XXXXXXX..XXXXXXX 100644
64
index XXXXXXX..XXXXXXX 100644
49
--- a/tests/qemu-iotests/274.out
65
--- a/tests/qemu-iotests/197.out
50
+++ b/tests/qemu-iotests/274.out
66
+++ b/tests/qemu-iotests/197.out
51
@@ -XXX,XX +XXX,XX @@ read 1048576/1048576 bytes at offset 0
67
@@ -XXX,XX +XXX,XX @@ read 1024/1024 bytes at offset 0
52
read 1048576/1048576 bytes at offset 1048576
68
1 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
53
1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
69
1 KiB (0x400) bytes allocated at offset 0 bytes (0x0)
54
70
No errors were found on the image.
55
+=== Testing qemu-img commit (top -> base) ===
56
+Formatting 'TEST_DIR/PID-base', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=2097152 lazy_refcounts=off refcount_bits=16
57
+
71
+
58
+Formatting 'TEST_DIR/PID-mid', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=1048576 backing_file=TEST_DIR/PID-base backing_fmt=qcow2 lazy_refcounts=off refcount_bits=16
72
+=== Copy-on-read with subclusters ===
59
+
73
+
60
+Formatting 'TEST_DIR/PID-top', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=2097152 backing_file=TEST_DIR/PID-mid backing_fmt=qcow2 lazy_refcounts=off refcount_bits=16
74
+Formatting 'TEST_DIR/t.IMGFMT', fmt=IMGFMT size=65536
61
+
75
+Formatting 'TEST_DIR/t.wrap.IMGFMT', fmt=IMGFMT size=65536 backing_file=TEST_DIR/t.IMGFMT backing_fmt=IMGFMT
62
+wrote 2097152/2097152 bytes at offset 0
76
+wrote 65536/65536 bytes at offset 0
63
+2 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
77
+64 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
64
+
78
+wrote 2048/2048 bytes at offset 28672
65
+Image committed.
79
+2 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
66
+
80
+wrote 2048/2048 bytes at offset 34816
67
+image: TEST_IMG
81
+2 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
68
+file format: IMGFMT
82
+Offset Length File
69
+virtual size: 2 MiB (2097152 bytes)
83
+0 0x7000 TEST_DIR/t.IMGFMT
70
+cluster_size: 65536
84
+0x7000 0x800 TEST_DIR/t.wrap.IMGFMT
71
+Format specific information:
85
+0x7800 0x1000 TEST_DIR/t.IMGFMT
72
+ compat: 1.1
86
+0x8800 0x800 TEST_DIR/t.wrap.IMGFMT
73
+ compression type: zlib
87
+0x9000 0x7000 TEST_DIR/t.IMGFMT
74
+ lazy refcounts: false
88
+read 4096/4096 bytes at offset 30720
75
+ refcount bits: 16
89
+4 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
76
+ corrupt: false
90
+Offset Length File
77
+ extended l2: false
91
+0 0x7000 TEST_DIR/t.IMGFMT
78
+
92
+0x7000 0x2000 TEST_DIR/t.wrap.IMGFMT
79
+read 1048576/1048576 bytes at offset 0
93
+0x9000 0x7000 TEST_DIR/t.IMGFMT
80
+1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
94
+No errors were found on the image.
81
+
95
*** done
82
+read 1048576/1048576 bytes at offset 1048576
83
+1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
84
+
85
+=== Testing QMP active commit (top -> base) ===
86
+Formatting 'TEST_DIR/PID-base', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=2097152 lazy_refcounts=off refcount_bits=16
87
+
88
+Formatting 'TEST_DIR/PID-mid', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=1048576 backing_file=TEST_DIR/PID-base backing_fmt=qcow2 lazy_refcounts=off refcount_bits=16
89
+
90
+Formatting 'TEST_DIR/PID-top', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=2097152 backing_file=TEST_DIR/PID-mid backing_fmt=qcow2 lazy_refcounts=off refcount_bits=16
91
+
92
+wrote 2097152/2097152 bytes at offset 0
93
+2 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
94
+
95
+{"execute": "block-commit", "arguments": {"auto-dismiss": false, "base-node": "base", "device": "top", "job-id": "job0"}}
96
+{"return": {}}
97
+{"execute": "job-complete", "arguments": {"id": "job0"}}
98
+{"return": {}}
99
+{"data": {"device": "job0", "len": 1048576, "offset": 1048576, "speed": 0, "type": "commit"}, "event": "BLOCK_JOB_READY", "timestamp": {"microseconds": "USECS", "seconds": "SECS"}}
100
+{"data": {"device": "job0", "len": 1048576, "offset": 1048576, "speed": 0, "type": "commit"}, "event": "BLOCK_JOB_COMPLETED", "timestamp": {"microseconds": "USECS", "seconds": "SECS"}}
101
+{"execute": "job-dismiss", "arguments": {"id": "job0"}}
102
+{"return": {}}
103
+image: TEST_IMG
104
+file format: IMGFMT
105
+virtual size: 1 MiB (1048576 bytes)
106
+cluster_size: 65536
107
+backing file: TEST_DIR/PID-base
108
+backing file format: IMGFMT
109
+Format specific information:
110
+ compat: 1.1
111
+ compression type: zlib
112
+ lazy refcounts: false
113
+ refcount bits: 16
114
+ corrupt: false
115
+ extended l2: false
116
+
117
+read 1048576/1048576 bytes at offset 0
118
+1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
119
+
120
+read 1048576/1048576 bytes at offset 1048576
121
+1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
122
+
123
== Resize tests ==
124
=== preallocation=off ===
125
Formatting 'TEST_DIR/PID-base', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=6442450944 lazy_refcounts=off refcount_bits=16
126
--
96
--
127
2.26.2
97
2.41.0
128
diff view generated by jsdifflib