Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 1 | /* |
| 2 | * QEMU Enhanced Disk Format Table I/O |
| 3 | * |
| 4 | * Copyright IBM, Corp. 2010 |
| 5 | * |
| 6 | * Authors: |
| 7 | * Stefan Hajnoczi <stefanha@linux.vnet.ibm.com> |
| 8 | * Anthony Liguori <aliguori@us.ibm.com> |
| 9 | * |
| 10 | * This work is licensed under the terms of the GNU LGPL, version 2 or later. |
| 11 | * See the COPYING.LIB file in the top-level directory. |
| 12 | * |
| 13 | */ |
| 14 | |
Peter Maydell | 80c71a2 | 2016-01-18 18:01:42 +0000 | [diff] [blame] | 15 | #include "qemu/osdep.h" |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 16 | #include "trace.h" |
Paolo Bonzini | 1de7afc | 2012-12-17 18:20:00 +0100 | [diff] [blame] | 17 | #include "qemu/sockets.h" /* for EINPROGRESS on Windows */ |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 18 | #include "qed.h" |
Paolo Bonzini | 58369e2 | 2016-03-15 17:22:36 +0100 | [diff] [blame] | 19 | #include "qemu/bswap.h" |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 20 | |
Paolo Bonzini | 2fd6163 | 2018-03-01 17:36:19 +0100 | [diff] [blame] | 21 | /* Called with table_lock held. */ |
Kevin Wolf | f651352 | 2016-11-14 16:08:44 +0100 | [diff] [blame] | 22 | static int qed_read_table(BDRVQEDState *s, uint64_t offset, QEDTable *table) |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 23 | { |
Kevin Wolf | 1127307 | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 24 | QEMUIOVector qiov; |
| 25 | int noffsets; |
| 26 | int i, ret; |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 27 | |
Kevin Wolf | 1127307 | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 28 | struct iovec iov = { |
| 29 | .iov_base = table->offsets, |
| 30 | .iov_len = s->header.cluster_size * s->header.table_size, |
| 31 | }; |
| 32 | qemu_iovec_init_external(&qiov, &iov, 1); |
| 33 | |
| 34 | trace_qed_read_table(s, offset, table); |
| 35 | |
Paolo Bonzini | 2fd6163 | 2018-03-01 17:36:19 +0100 | [diff] [blame] | 36 | qemu_co_mutex_unlock(&s->table_lock); |
Kevin Wolf | 1127307 | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 37 | ret = bdrv_preadv(s->bs->file, offset, &qiov); |
Paolo Bonzini | 2fd6163 | 2018-03-01 17:36:19 +0100 | [diff] [blame] | 38 | qemu_co_mutex_lock(&s->table_lock); |
Kevin Wolf | 1127307 | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 39 | if (ret < 0) { |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 40 | goto out; |
| 41 | } |
| 42 | |
| 43 | /* Byteswap offsets */ |
Kevin Wolf | 1127307 | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 44 | noffsets = qiov.size / sizeof(uint64_t); |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 45 | for (i = 0; i < noffsets; i++) { |
| 46 | table->offsets[i] = le64_to_cpu(table->offsets[i]); |
| 47 | } |
| 48 | |
Kevin Wolf | 1127307 | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 49 | ret = 0; |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 50 | out: |
| 51 | /* Completion */ |
Kevin Wolf | 1127307 | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 52 | trace_qed_read_table_cb(s, table, ret); |
Kevin Wolf | f651352 | 2016-11-14 16:08:44 +0100 | [diff] [blame] | 53 | return ret; |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 54 | } |
| 55 | |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 56 | /** |
| 57 | * Write out an updated part or all of a table |
| 58 | * |
| 59 | * @s: QED state |
| 60 | * @offset: Offset of table in image file, in bytes |
| 61 | * @table: Table |
| 62 | * @index: Index of first element |
| 63 | * @n: Number of elements |
| 64 | * @flush: Whether or not to sync to disk |
Paolo Bonzini | 1f01e50 | 2017-06-29 15:27:47 +0200 | [diff] [blame] | 65 | * |
Paolo Bonzini | 2fd6163 | 2018-03-01 17:36:19 +0100 | [diff] [blame] | 66 | * Called with table_lock held. |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 67 | */ |
Kevin Wolf | 453e53e | 2016-11-15 11:14:01 +0100 | [diff] [blame] | 68 | static int qed_write_table(BDRVQEDState *s, uint64_t offset, QEDTable *table, |
| 69 | unsigned int index, unsigned int n, bool flush) |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 70 | { |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 71 | unsigned int sector_mask = BDRV_SECTOR_SIZE / sizeof(uint64_t) - 1; |
| 72 | unsigned int start, end, i; |
Kevin Wolf | 602b57f | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 73 | QEDTable *new_table; |
| 74 | struct iovec iov; |
| 75 | QEMUIOVector qiov; |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 76 | size_t len_bytes; |
Kevin Wolf | 602b57f | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 77 | int ret; |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 78 | |
| 79 | trace_qed_write_table(s, offset, table, index, n); |
| 80 | |
| 81 | /* Calculate indices of the first and one after last elements */ |
| 82 | start = index & ~sector_mask; |
| 83 | end = (index + n + sector_mask) & ~sector_mask; |
| 84 | |
| 85 | len_bytes = (end - start) * sizeof(uint64_t); |
| 86 | |
Kevin Wolf | 602b57f | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 87 | new_table = qemu_blockalign(s->bs, len_bytes); |
| 88 | iov = (struct iovec) { |
| 89 | .iov_base = new_table->offsets, |
| 90 | .iov_len = len_bytes, |
| 91 | }; |
| 92 | qemu_iovec_init_external(&qiov, &iov, 1); |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 93 | |
| 94 | /* Byteswap table */ |
| 95 | for (i = start; i < end; i++) { |
| 96 | uint64_t le_offset = cpu_to_le64(table->offsets[i]); |
Kevin Wolf | 602b57f | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 97 | new_table->offsets[i - start] = le_offset; |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 98 | } |
| 99 | |
| 100 | /* Adjust for offset into table */ |
| 101 | offset += start * sizeof(uint64_t); |
| 102 | |
Paolo Bonzini | 2fd6163 | 2018-03-01 17:36:19 +0100 | [diff] [blame] | 103 | qemu_co_mutex_unlock(&s->table_lock); |
Kevin Wolf | 602b57f | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 104 | ret = bdrv_pwritev(s->bs->file, offset, &qiov); |
Paolo Bonzini | 2fd6163 | 2018-03-01 17:36:19 +0100 | [diff] [blame] | 105 | qemu_co_mutex_lock(&s->table_lock); |
Kevin Wolf | 602b57f | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 106 | trace_qed_write_table_cb(s, table, flush, ret); |
| 107 | if (ret < 0) { |
| 108 | goto out; |
| 109 | } |
| 110 | |
| 111 | if (flush) { |
Kevin Wolf | 602b57f | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 112 | ret = bdrv_flush(s->bs); |
Kevin Wolf | 602b57f | 2016-11-14 14:56:32 +0100 | [diff] [blame] | 113 | if (ret < 0) { |
| 114 | goto out; |
| 115 | } |
| 116 | } |
| 117 | |
| 118 | ret = 0; |
| 119 | out: |
| 120 | qemu_vfree(new_table); |
Kevin Wolf | 453e53e | 2016-11-15 11:14:01 +0100 | [diff] [blame] | 121 | return ret; |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 122 | } |
| 123 | |
| 124 | int qed_read_l1_table_sync(BDRVQEDState *s) |
| 125 | { |
Kevin Wolf | f651352 | 2016-11-14 16:08:44 +0100 | [diff] [blame] | 126 | return qed_read_table(s, s->header.l1_table_offset, s->l1_table); |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 127 | } |
| 128 | |
Paolo Bonzini | 2fd6163 | 2018-03-01 17:36:19 +0100 | [diff] [blame] | 129 | /* Called with table_lock held. */ |
Kevin Wolf | 453e53e | 2016-11-15 11:14:01 +0100 | [diff] [blame] | 130 | int qed_write_l1_table(BDRVQEDState *s, unsigned int index, unsigned int n) |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 131 | { |
| 132 | BLKDBG_EVENT(s->bs->file, BLKDBG_L1_UPDATE); |
Kevin Wolf | 453e53e | 2016-11-15 11:14:01 +0100 | [diff] [blame] | 133 | return qed_write_table(s, s->header.l1_table_offset, |
| 134 | s->l1_table, index, n, false); |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 135 | } |
| 136 | |
| 137 | int qed_write_l1_table_sync(BDRVQEDState *s, unsigned int index, |
| 138 | unsigned int n) |
| 139 | { |
Kevin Wolf | 453e53e | 2016-11-15 11:14:01 +0100 | [diff] [blame] | 140 | return qed_write_l1_table(s, index, n); |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 141 | } |
| 142 | |
Paolo Bonzini | 2fd6163 | 2018-03-01 17:36:19 +0100 | [diff] [blame] | 143 | /* Called with table_lock held. */ |
Kevin Wolf | a8165d2 | 2016-11-14 16:26:14 +0100 | [diff] [blame] | 144 | int qed_read_l2_table(BDRVQEDState *s, QEDRequest *request, uint64_t offset) |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 145 | { |
Kevin Wolf | f651352 | 2016-11-14 16:08:44 +0100 | [diff] [blame] | 146 | int ret; |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 147 | |
| 148 | qed_unref_l2_cache_entry(request->l2_table); |
| 149 | |
| 150 | /* Check for cached L2 entry */ |
| 151 | request->l2_table = qed_find_l2_cache_entry(&s->l2_cache, offset); |
| 152 | if (request->l2_table) { |
Kevin Wolf | a8165d2 | 2016-11-14 16:26:14 +0100 | [diff] [blame] | 153 | return 0; |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 154 | } |
| 155 | |
| 156 | request->l2_table = qed_alloc_l2_cache_entry(&s->l2_cache); |
| 157 | request->l2_table->table = qed_alloc_table(s); |
| 158 | |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 159 | BLKDBG_EVENT(s->bs->file, BLKDBG_L2_LOAD); |
Kevin Wolf | f651352 | 2016-11-14 16:08:44 +0100 | [diff] [blame] | 160 | ret = qed_read_table(s, offset, request->l2_table->table); |
| 161 | |
Kevin Wolf | f651352 | 2016-11-14 16:08:44 +0100 | [diff] [blame] | 162 | if (ret) { |
| 163 | /* can't trust loaded L2 table anymore */ |
| 164 | qed_unref_l2_cache_entry(request->l2_table); |
| 165 | request->l2_table = NULL; |
| 166 | } else { |
| 167 | request->l2_table->offset = offset; |
| 168 | |
| 169 | qed_commit_l2_cache_entry(&s->l2_cache, request->l2_table); |
| 170 | |
| 171 | /* This is guaranteed to succeed because we just committed the entry |
| 172 | * to the cache. |
| 173 | */ |
| 174 | request->l2_table = qed_find_l2_cache_entry(&s->l2_cache, offset); |
| 175 | assert(request->l2_table != NULL); |
| 176 | } |
Kevin Wolf | f651352 | 2016-11-14 16:08:44 +0100 | [diff] [blame] | 177 | |
Kevin Wolf | a8165d2 | 2016-11-14 16:26:14 +0100 | [diff] [blame] | 178 | return ret; |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 179 | } |
| 180 | |
| 181 | int qed_read_l2_table_sync(BDRVQEDState *s, QEDRequest *request, uint64_t offset) |
| 182 | { |
Kevin Wolf | a8165d2 | 2016-11-14 16:26:14 +0100 | [diff] [blame] | 183 | return qed_read_l2_table(s, request, offset); |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 184 | } |
| 185 | |
Paolo Bonzini | 2fd6163 | 2018-03-01 17:36:19 +0100 | [diff] [blame] | 186 | /* Called with table_lock held. */ |
Kevin Wolf | 453e53e | 2016-11-15 11:14:01 +0100 | [diff] [blame] | 187 | int qed_write_l2_table(BDRVQEDState *s, QEDRequest *request, |
| 188 | unsigned int index, unsigned int n, bool flush) |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 189 | { |
| 190 | BLKDBG_EVENT(s->bs->file, BLKDBG_L2_UPDATE); |
Kevin Wolf | 453e53e | 2016-11-15 11:14:01 +0100 | [diff] [blame] | 191 | return qed_write_table(s, request->l2_table->offset, |
| 192 | request->l2_table->table, index, n, flush); |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 193 | } |
| 194 | |
| 195 | int qed_write_l2_table_sync(BDRVQEDState *s, QEDRequest *request, |
| 196 | unsigned int index, unsigned int n, bool flush) |
| 197 | { |
Kevin Wolf | 453e53e | 2016-11-15 11:14:01 +0100 | [diff] [blame] | 198 | return qed_write_l2_table(s, request, index, n, flush); |
Stefan Hajnoczi | 298800c | 2010-12-06 16:08:01 +0000 | [diff] [blame] | 199 | } |