14
#include "qemu/osdep.h"
15
#include "block/block-io.h"
20
BdrvCheckResult *result;
24
uint32_t *used_clusters;
29
static bool qed_test_bit(uint32_t *bitmap, uint64_t n) {
30
return !!(bitmap[n / 32] & (1 << (n % 32)));
33
static void qed_set_bit(uint32_t *bitmap, uint64_t n) {
34
bitmap[n / 32] |= 1 << (n % 32);
44
static bool qed_set_used_clusters(QEDCheck *check, uint64_t offset,
47
uint64_t cluster = qed_bytes_to_clusters(check->s, offset);
48
unsigned int corruptions = 0;
52
if (qed_test_bit(check->used_clusters, cluster)) {
56
qed_set_bit(check->used_clusters, cluster);
60
check->result->corruptions += corruptions;
61
return corruptions == 0;
69
static unsigned int qed_check_l2_table(QEDCheck *check, QEDTable *table)
71
BDRVQEDState *s = check->s;
72
unsigned int i, num_invalid = 0;
73
uint64_t last_offset = 0;
75
for (i = 0; i < s->table_nelems; i++) {
76
uint64_t offset = table->offsets[i];
78
if (qed_offset_is_unalloc_cluster(offset) ||
79
qed_offset_is_zero_cluster(offset)) {
82
check->result->bfi.allocated_clusters++;
83
if (last_offset && (last_offset + s->header.cluster_size != offset)) {
84
check->result->bfi.fragmented_clusters++;
89
if (!qed_check_cluster_offset(s, offset)) {
91
table->offsets[i] = 0;
92
check->result->corruptions_fixed++;
94
check->result->corruptions++;
101
qed_set_used_clusters(check, offset, 1);
110
static int coroutine_fn GRAPH_RDLOCK
111
qed_check_l1_table(QEDCheck *check, QEDTable *table)
113
BDRVQEDState *s = check->s;
114
unsigned int i, num_invalid_l1 = 0;
115
int ret, last_error = 0;
118
qed_set_used_clusters(check, s->header.l1_table_offset,
119
s->header.table_size);
121
for (i = 0; i < s->table_nelems; i++) {
122
unsigned int num_invalid_l2;
123
uint64_t offset = table->offsets[i];
125
if (qed_offset_is_unalloc_cluster(offset)) {
130
if (!qed_check_table_offset(s, offset)) {
133
table->offsets[i] = 0;
134
check->result->corruptions_fixed++;
136
check->result->corruptions++;
143
if (!qed_set_used_clusters(check, offset, s->header.table_size)) {
147
ret = qed_read_l2_table_sync(s, &check->request, offset);
149
check->result->check_errors++;
154
num_invalid_l2 = qed_check_l2_table(check,
155
check->request.l2_table->table);
158
if (num_invalid_l2 > 0 && check->fix) {
159
ret = qed_write_l2_table_sync(s, &check->request, 0,
160
s->table_nelems, false);
162
check->result->check_errors++;
170
qed_unref_l2_cache_entry(check->request.l2_table);
171
check->request.l2_table = NULL;
174
if (num_invalid_l1 > 0 && check->fix) {
175
ret = qed_write_l1_table_sync(s, 0, s->table_nelems);
177
check->result->check_errors++;
188
static void qed_check_for_leaks(QEDCheck *check)
190
BDRVQEDState *s = check->s;
193
for (i = s->header.header_size; i < check->nclusters; i++) {
194
if (!qed_test_bit(check->used_clusters, i)) {
195
check->result->leaks++;
203
static void coroutine_fn GRAPH_RDLOCK
204
qed_check_mark_clean(BDRVQEDState *s, BdrvCheckResult *result)
207
if (result->corruptions > 0 || result->check_errors > 0) {
212
if (!(s->header.features & QED_F_NEED_CHECK)) {
217
bdrv_co_flush(s->bs);
219
s->header.features &= ~QED_F_NEED_CHECK;
220
qed_write_header_sync(s);
224
int coroutine_fn qed_check(BDRVQEDState *s, BdrvCheckResult *result, bool fix)
229
.nclusters = qed_bytes_to_clusters(s, s->file_size),
230
.request = { .l2_table = NULL },
235
check.used_clusters = g_try_new0(uint32_t, (check.nclusters + 31) / 32);
236
if (check.nclusters && check.used_clusters == NULL) {
240
check.result->bfi.total_clusters =
241
DIV_ROUND_UP(s->header.image_size, s->header.cluster_size);
242
ret = qed_check_l1_table(&check, s->l1_table);
245
qed_check_for_leaks(&check);
248
qed_check_mark_clean(s, result);
252
g_free(check.used_clusters);