1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
|
// SPDX-License-Identifier: GPL-2.0-only
/*
* Copyright 2023 Red Hat
*/
/*
* This file contains the main entry points for normal operations on a vdo as well as functions for
* constructing and destroying vdo instances (in memory).
*/
/**
* DOC:
*
* A read_only_notifier has a single completion which is used to perform read-only notifications,
* however, vdo_enter_read_only_mode() may be called from any thread. A pair of fields, protected
* by a spinlock, are used to control the read-only mode entry process. The first field holds the
* read-only error. The second is the state field, which may hold any of the four special values
* enumerated here.
*
* When vdo_enter_read_only_mode() is called from some vdo thread, if the read_only_error field
* already contains an error (i.e. its value is not VDO_SUCCESS), then some other error has already
* initiated the read-only process, and nothing more is done. Otherwise, the new error is stored in
* the read_only_error field, and the state field is consulted. If the state is MAY_NOTIFY, it is
* set to NOTIFYING, and the notification process begins. If the state is MAY_NOT_NOTIFY, then
* notifications are currently disallowed, generally due to the vdo being suspended. In this case,
* the nothing more will be done until the vdo is resumed, at which point the notification will be
* performed. In any other case, the vdo is already read-only, and there is nothing more to do.
*/
#include "vdo.h"
#include <linux/completion.h>
#include <linux/device-mapper.h>
#include <linux/kernel.h>
#include <linux/lz4.h>
#include <linux/module.h>
#include <linux/mutex.h>
#include <linux/spinlock.h>
#include <linux/types.h>
#include "logger.h"
#include "memory-alloc.h"
#include "permassert.h"
#include "string-utils.h"
#include "block-map.h"
#include "completion.h"
#include "data-vio.h"
#include "dedupe.h"
#include "encodings.h"
#include "funnel-workqueue.h"
#include "io-submitter.h"
#include "logical-zone.h"
#include "packer.h"
#include "physical-zone.h"
#include "recovery-journal.h"
#include "slab-depot.h"
#include "statistics.h"
#include "status-codes.h"
#include "vio.h"
#define PARANOID_THREAD_CONSISTENCY_CHECKS 0
struct sync_completion {
struct vdo_completion vdo_completion;
struct completion completion;
};
/* A linked list is adequate for the small number of entries we expect. */
struct device_registry {
struct list_head links;
/* TODO: Convert to rcu per kernel recommendation. */
rwlock_t lock;
};
static struct device_registry registry;
/**
* vdo_initialize_device_registry_once() - Initialize the necessary structures for the device
* registry.
*/
void vdo_initialize_device_registry_once(void)
{
INIT_LIST_HEAD(®istry.links);
rwlock_init(®istry.lock);
}
/** vdo_is_equal() - Implements vdo_filter_fn. */
static bool vdo_is_equal(struct vdo *vdo, const void *context)
{
return (vdo == context);
}
/**
* filter_vdos_locked() - Find a vdo in the registry if it exists there.
* @filter: The filter function to apply to devices.
* @context: A bit of context to provide the filter.
*
* Context: Must be called holding the lock.
*
* Return: the vdo object found, if any.
*/
static struct vdo * __must_check filter_vdos_locked(vdo_filter_fn filter,
const void *context)
{
struct vdo *vdo;
list_for_each_entry(vdo, ®istry.links, registration) {
if (filter(vdo, context))
return vdo;
}
return NULL;
}
/**
* vdo_find_matching() - Find and return the first (if any) vdo matching a given filter function.
* @filter: The filter function to apply to vdos.
* @context: A bit of context to provide the filter.
*/
struct vdo *vdo_find_matching(vdo_filter_fn filter, const void *context)
{
struct vdo *vdo;
read_lock(®istry.lock);
vdo = filter_vdos_locked(filter, context);
read_unlock(®istry.lock);
return vdo;
}
static void start_vdo_request_queue(void *ptr)
{
struct vdo_thread *thread = vdo_get_work_queue_owner(vdo_get_current_work_queue());
vdo_register_allocating_thread(&thread->allocating_thread,
&thread->vdo->allocations_allowed);
}
static void finish_vdo_request_queue(void *ptr)
{
vdo_unregister_allocating_thread();
}
#ifdef MODULE
#define MODULE_NAME THIS_MODULE->name
#else
#define MODULE_NAME "dm-vdo"
#endif /* MODULE */
static const struct vdo_work_queue_type default_queue_type = {
.start = start_vdo_request_queue,
.finish = finish_vdo_request_queue,
.max_priority = VDO_DEFAULT_Q_MAX_PRIORITY,
.default_priority = VDO_DEFAULT_Q_COMPLETION_PRIORITY,
};
static const struct vdo_work_queue_type bio_ack_q_type = {
.start = NULL,
.finish = NULL,
.max_priority = BIO_ACK_Q_MAX_PRIORITY,
.default_priority = BIO_ACK_Q_ACK_PRIORITY,
};
static const struct vdo_work_queue_type cpu_q_type = {
.start = NULL,
.finish = NULL,
.max_priority = CPU_Q_MAX_PRIORITY,
.default_priority = CPU_Q_MAX_PRIORITY,
};
static void uninitialize_thread_config(struct thread_config *config)
{
vdo_free(vdo_forget(config->logical_threads));
vdo_free(vdo_forget(config->physical_threads));
vdo_free(vdo_forget(config->hash_zone_threads));
vdo_free(vdo_forget(config->bio_threads));
memset(config, 0, sizeof(struct thread_config));
}
static void assign_thread_ids(struct thread_config *config,
thread_id_t thread_ids[], zone_count_t count)
{
zone_count_t zone;
for (zone = 0; zone < count; zone++)
thread_ids[zone] = config->thread_count++;
}
/**
* initialize_thread_config() - Initialize the thread mapping
*
* If the logical, physical, and hash zone counts are all 0, a single thread will be shared by all
* three plus the packer and recovery journal. Otherwise, there must be at least one of each type,
* and each will have its own thread, as will the packer and recovery journal.
*
* Return: VDO_SUCCESS or an error.
*/
static int __must_check initialize_thread_config(struct thread_count_config counts,
struct thread_config *config)
{
int result;
bool single = ((counts.logical_zones + counts.physical_zones + counts.hash_zones) == 0);
config->bio_thread_count = counts.bio_threads;
if (single) {
config->logical_zone_count = 1;
config->physical_zone_count = 1;
config->hash_zone_count = 1;
} else {
config->logical_zone_count = counts.logical_zones;
config->physical_zone_count = counts.physical_zones;
config->hash_zone_count = counts.hash_zones;
}
result = vdo_allocate(config->logical_zone_count, thread_id_t,
"logical thread array", &config->logical_threads);
if (result != VDO_SUCCESS) {
uninitialize_thread_config(config);
return result;
}
result = vdo_allocate(config->physical_zone_count, thread_id_t,
"physical thread array", &config->physical_threads);
if (result != VDO_SUCCESS) {
uninitialize_thread_config(config);
return result;
}
result = vdo_allocate(config->hash_zone_count, thread_id_t,
"hash thread array", &config->hash_zone_threads);
if (result != VDO_SUCCESS) {
uninitialize_thread_config(config);
return result;
}
result = vdo_allocate(config->bio_thread_count, thread_id_t,
"bio thread array", &config->bio_threads);
if (result != VDO_SUCCESS) {
uninitialize_thread_config(config);
return result;
}
if (single) {
config->logical_threads[0] = config->thread_count;
config->physical_threads[0] = config->thread_count;
config->hash_zone_threads[0] = config->thread_count++;
} else {
config->admin_thread = config->thread_count;
config->journal_thread = config->thread_count++;
config->packer_thread = config->thread_count++;
assign_thread_ids(config, config->logical_threads, counts.logical_zones);
assign_thread_ids(config, config->physical_threads, counts.physical_zones);
assign_thread_ids(config, config->hash_zone_threads, counts.hash_zones);
}
config->dedupe_thread = config->thread_count++;
config->bio_ack_thread =
((counts.bio_ack_threads > 0) ? config->thread_count++ : VDO_INVALID_THREAD_ID);
config->cpu_thread = config->thread_count++;
assign_thread_ids(config, config->bio_threads, counts.bio_threads);
return VDO_SUCCESS;
}
/**
* read_geometry_block() - Synchronously read the geometry block from a vdo's underlying block
* device.
* @vdo: The vdo whose geometry is to be read.
*
* Return: VDO_SUCCESS or an error code.
*/
static int __must_check read_geometry_block(struct vdo *vdo)
{
struct vio *vio;
char *block;
int result;
result = vdo_allocate(VDO_BLOCK_SIZE, u8, __func__, &block);
if (result != VDO_SUCCESS)
return result;
result = create_metadata_vio(vdo, VIO_TYPE_GEOMETRY, VIO_PRIORITY_HIGH, NULL,
block, &vio);
if (result != VDO_SUCCESS) {
vdo_free(block);
return result;
}
/*
* This is only safe because, having not already loaded the geometry, the vdo's geometry's
* bio_offset field is 0, so the fact that vio_reset_bio() will subtract that offset from
* the supplied pbn is not a problem.
*/
result = vio_reset_bio(vio, block, NULL, REQ_OP_READ,
VDO_GEOMETRY_BLOCK_LOCATION);
if (result != VDO_SUCCESS) {
free_vio(vdo_forget(vio));
vdo_free(block);
return result;
}
bio_set_dev(vio->bio, vdo_get_backing_device(vdo));
submit_bio_wait(vio->bio);
result = blk_status_to_errno(vio->bio->bi_status);
free_vio(vdo_forget(vio));
if (result != 0) {
vdo_log_error_strerror(result, "synchronous read failed");
vdo_free(block);
return -EIO;
}
result = vdo_parse_geometry_block((u8 *) block, &vdo->geometry);
vdo_free(block);
return result;
}
static bool get_zone_thread_name(const thread_id_t thread_ids[], zone_count_t count,
thread_id_t id, const char *prefix,
char *buffer, size_t buffer_length)
{
if (id >= thread_ids[0]) {
thread_id_t index = id - thread_ids[0];
if (index < count) {
snprintf(buffer, buffer_length, "%s%d", prefix, index);
return true;
}
}
return false;
}
/**
* get_thread_name() - Format the name of the worker thread desired to support a given work queue.
* @thread_config: The thread configuration.
* @thread_id: The thread id.
* @buffer: Where to put the formatted name.
* @buffer_length: Size of the output buffer.
*
* The physical layer may add a prefix identifying the product; the output from this function
* should just identify the thread.
*/
static void get_thread_name(const struct thread_config *thread_config,
thread_id_t thread_id, char *buffer, size_t buffer_length)
{
if (thread_id == thread_config->journal_thread) {
if (thread_config->packer_thread == thread_id) {
/*
* This is the "single thread" config where one thread is used for the
* journal, packer, logical, physical, and hash zones. In that case, it is
* known as the "request queue."
*/
snprintf(buffer, buffer_length, "reqQ");
return;
}
snprintf(buffer, buffer_length, "journalQ");
return;
} else if (thread_id == thread_config->admin_thread) {
/* Theoretically this could be different from the journal thread. */
snprintf(buffer, buffer_length, "adminQ");
return;
} else if (thread_id == thread_config->packer_thread) {
snprintf(buffer, buffer_length, "packerQ");
return;
} else if (thread_id == thread_config->dedupe_thread) {
snprintf(buffer, buffer_length, "dedupeQ");
return;
} else if (thread_id == thread_config->bio_ack_thread) {
snprintf(buffer, buffer_length, "ackQ");
return;
} else if (thread_id == thread_config->cpu_thread) {
snprintf(buffer, buffer_length, "cpuQ");
return;
}
if (get_zone_thread_name(thread_config->logical_threads,
thread_config->logical_zone_count,
thread_id, "logQ", buffer, buffer_length))
return;
if (get_zone_thread_name(thread_config->physical_threads,
thread_config->physical_zone_count,
thread_id, "physQ", buffer, buffer_length))
return;
if (get_zone_thread_name(thread_config->hash_zone_threads,
thread_config->hash_zone_count,
thread_id, "hashQ", buffer, buffer_length))
return;
if (get_zone_thread_name(thread_config->bio_threads,
thread_config->bio_thread_count,
thread_id, "bioQ", buffer, buffer_length))
return;
/* Some sort of misconfiguration? */
snprintf(buffer, buffer_length, "reqQ%d", thread_id);
}
/**
* vdo_make_thread() - Construct a single vdo work_queue and its associated thread (or threads for
* round-robin queues).
* @vdo: The vdo which owns the thread.
* @thread_id: The id of the thread to create (as determined by the thread_config).
* @type: The description of the work queue for this thread.
* @queue_count: The number of actual threads/queues contained in the "thread".
* @contexts: An array of queue_count contexts, one for each individual queue; may be NULL.
*
* Each "thread" constructed by this method is represented by a unique thread id in the thread
* config, and completions can be enqueued to the queue and run on the threads comprising this
* entity.
*
* Return: VDO_SUCCESS or an error.
*/
int vdo_make_thread(struct vdo *vdo, thread_id_t thread_id,
const struct vdo_work_queue_type *type,
unsigned int queue_count, void *contexts[])
{
struct vdo_thread *thread = &vdo->threads[thread_id];
char queue_name[MAX_VDO_WORK_QUEUE_NAME_LEN];
if (type == NULL)
type = &default_queue_type;
if (thread->queue != NULL) {
return VDO_ASSERT(vdo_work_queue_type_is(thread->queue, type),
"already constructed vdo thread %u is of the correct type",
thread_id);
}
thread->vdo = vdo;
thread->thread_id = thread_id;
get_thread_name(&vdo->thread_config, thread_id, queue_name, sizeof(queue_name));
return vdo_make_work_queue(vdo->thread_name_prefix, queue_name, thread,
type, queue_count, contexts, &thread->queue);
}
/**
* register_vdo() - Register a VDO; it must not already be registered.
* @vdo: The vdo to register.
*
* Return: VDO_SUCCESS or an error.
*/
static int register_vdo(struct vdo *vdo)
{
int result;
write_lock(®istry.lock);
result = VDO_ASSERT(filter_vdos_locked(vdo_is_equal, vdo) == NULL,
"VDO not already registered");
if (result == VDO_SUCCESS) {
INIT_LIST_HEAD(&vdo->registration);
list_add_tail(&vdo->registration, ®istry.links);
}
write_unlock(®istry.lock);
return result;
}
/**
* initialize_vdo() - Do the portion of initializing a vdo which will clean up after itself on
* error.
* @vdo: The vdo being initialized
* @config: The configuration of the vdo
* @instance: The instance number of the vdo
* @reason: The buffer to hold the failure reason on error
*/
static int initialize_vdo(struct vdo *vdo, struct device_config *config,
unsigned int instance, char **reason)
{
int result;
zone_count_t i;
vdo->device_config = config;
vdo->starting_sector_offset = config->owning_target->begin;
vdo->instance = instance;
vdo->allocations_allowed = true;
vdo_set_admin_state_code(&vdo->admin.state, VDO_ADMIN_STATE_NEW);
INIT_LIST_HEAD(&vdo->device_config_list);
vdo_initialize_completion(&vdo->admin.completion, vdo, VDO_ADMIN_COMPLETION);
init_completion(&vdo->admin.callback_sync);
mutex_init(&vdo->stats_mutex);
result = read_geometry_block(vdo);
if (result != VDO_SUCCESS) {
*reason = "Could not load geometry block";
return result;
}
result = initialize_thread_config(config->thread_counts, &vdo->thread_config);
if (result != VDO_SUCCESS) {
*reason = "Cannot create thread configuration";
return result;
}
vdo_log_info("zones: %d logical, %d physical, %d hash; total threads: %d",
config->thread_counts.logical_zones,
config->thread_counts.physical_zones,
config->thread_counts.hash_zones, vdo->thread_config.thread_count);
/* Compression context storage */
result = vdo_allocate(config->thread_counts.cpu_threads, char *, "LZ4 context",
&vdo->compression_context);
if (result != VDO_SUCCESS) {
*reason = "cannot allocate LZ4 context";
return result;
}
for (i = 0; i < config->thread_counts.cpu_threads; i++) {
result = vdo_allocate(LZ4_MEM_COMPRESS, char, "LZ4 context",
&vdo->compression_context[i]);
if (result != VDO_SUCCESS) {
*reason = "cannot allocate LZ4 context";
return result;
}
}
result = register_vdo(vdo);
if (result != VDO_SUCCESS) {
*reason = "Cannot add VDO to device registry";
return result;
}
vdo_set_admin_state_code(&vdo->admin.state, VDO_ADMIN_STATE_INITIALIZED);
return result;
}
/**
* vdo_make() - Allocate and initialize a vdo.
* @instance: Device instantiation counter.
* @config: The device configuration.
* @reason: The reason for any failure during this call.
* @vdo_ptr: A pointer to hold the created vdo.
*
* Return: VDO_SUCCESS or an error.
*/
int vdo_make(unsigned int instance, struct device_config *config, char **reason,
struct vdo **vdo_ptr)
{
int result;
struct vdo *vdo;
/* Initialize with a generic failure reason to prevent returning garbage. */
*reason = "Unspecified error";
result = vdo_allocate(1, struct vdo, __func__, &vdo);
if (result != VDO_SUCCESS) {
*reason = "Cannot allocate VDO";
return result;
}
result = initialize_vdo(vdo, config, instance, reason);
if (result != VDO_SUCCESS) {
vdo_destroy(vdo);
return result;
}
/* From here on, the caller will clean up if there is an error. */
*vdo_ptr = vdo;
snprintf(vdo->thread_name_prefix, sizeof(vdo->thread_name_prefix),
"%s%u", MODULE_NAME, instance);
BUG_ON(vdo->thread_name_prefix[0] == '\0');
result = vdo_allocate(vdo->thread_config.thread_count,
struct vdo_thread, __func__, &vdo->threads);
if (result != VDO_SUCCESS) {
*reason = "Cannot allocate thread structures";
return result;
}
result = vdo_make_thread(vdo, vdo->thread_config.admin_thread,
&default_queue_type, 1, NULL);
if (result != VDO_SUCCESS) {
*reason = "Cannot make admin thread";
return result;
}
result = vdo_make_flusher(vdo);
if (result != VDO_SUCCESS) {
*reason = "Cannot make flusher zones";
return result;
}
result = vdo_make_packer(vdo, DEFAULT_PACKER_BINS, &vdo->packer);
if (result != VDO_SUCCESS) {
*reason = "Cannot make packer zones";
return result;
}
BUG_ON(vdo->device_config->logical_block_size <= 0);
BUG_ON(vdo->device_config->owned_device == NULL);
result = make_data_vio_pool(vdo, MAXIMUM_VDO_USER_VIOS,
MAXIMUM_VDO_USER_VIOS * 3 / 4,
&vdo->data_vio_pool);
if (result != VDO_SUCCESS) {
*reason = "Cannot allocate data_vio pool";
return result;
}
result = vdo_make_io_submitter(config->thread_counts.bio_threads,
config->thread_counts.bio_rotation_interval,
get_data_vio_pool_request_limit(vdo->data_vio_pool),
vdo, &vdo->io_submitter);
if (result != VDO_SUCCESS) {
*reason = "bio submission initialization failed";
return result;
}
if (vdo_uses_bio_ack_queue(vdo)) {
result = vdo_make_thread(vdo, vdo->thread_config.bio_ack_thread,
&bio_ack_q_type,
config->thread_counts.bio_ack_threads, NULL);
if (result != VDO_SUCCESS) {
*reason = "bio ack queue initialization failed";
return result;
}
}
result = vdo_make_thread(vdo, vdo->thread_config.cpu_thread, &cpu_q_type,
config->thread_counts.cpu_threads,
(void **) vdo->compression_context);
if (result != VDO_SUCCESS) {
*reason = "CPU queue initialization failed";
return result;
}
return VDO_SUCCESS;
}
static void finish_vdo(struct vdo *vdo)
{
int i;
if (vdo->threads == NULL)
return;
vdo_cleanup_io_submitter(vdo->io_submitter);
vdo_finish_dedupe_index(vdo->hash_zones);
for (i = 0; i < vdo->thread_config.thread_count; i++)
vdo_finish_work_queue(vdo->threads[i].queue);
}
/**
* free_listeners() - Free the list of read-only listeners associated with a thread.
* @thread_data: The thread holding the list to free.
*/
static void free_listeners(struct vdo_thread *thread)
{
struct read_only_listener *listener, *next;
for (listener = vdo_forget(thread->listeners); listener != NULL; listener = next) {
next = vdo_forget(listener->next);
vdo_free(listener);
}
}
static void uninitialize_super_block(struct vdo_super_block *super_block)
{
free_vio_components(&super_block->vio);
vdo_free(super_block->buffer);
}
/**
* unregister_vdo() - Remove a vdo from the device registry.
* @vdo: The vdo to remove.
*/
static void unregister_vdo(struct vdo *vdo)
{
write_lock(®istry.lock);
if (filter_vdos_locked(vdo_is_equal, vdo) == vdo)
list_del_init(&vdo->registration);
write_unlock(®istry.lock);
}
/**
* vdo_destroy() - Destroy a vdo instance.
* @vdo: The vdo to destroy (may be NULL).
*/
void vdo_destroy(struct vdo *vdo)
{
unsigned int i;
if (vdo == NULL)
return;
/* A running VDO should never be destroyed without suspending first. */
BUG_ON(vdo_get_admin_state(vdo)->normal);
vdo->allocations_allowed = true;
finish_vdo(vdo);
unregister_vdo(vdo);
free_data_vio_pool(vdo->data_vio_pool);
vdo_free_io_submitter(vdo_forget(vdo->io_submitter));
vdo_free_flusher(vdo_forget(vdo->flusher));
vdo_free_packer(vdo_forget(vdo->packer));
vdo_free_recovery_journal(vdo_forget(vdo->recovery_journal));
vdo_free_slab_depot(vdo_forget(vdo->depot));
vdo_uninitialize_layout(&vdo->layout);
vdo_uninitialize_layout(&vdo->next_layout);
if (vdo->partition_copier)
dm_kcopyd_client_destroy(vdo_forget(vdo->partition_copier));
uninitialize_super_block(&vdo->super_block);
vdo_free_block_map(vdo_forget(vdo->block_map));
vdo_free_hash_zones(vdo_forget(vdo->hash_zones));
vdo_free_physical_zones(vdo_forget(vdo->physical_zones));
vdo_free_logical_zones(vdo_forget(vdo->logical_zones));
if (vdo->threads != NULL) {
for (i = 0; i < vdo->thread_config.thread_count; i++) {
free_listeners(&vdo->threads[i]);
vdo_free_work_queue(vdo_forget(vdo->threads[i].queue));
}
vdo_free(vdo_forget(vdo->threads));
}
uninitialize_thread_config(&vdo->thread_config);
if (vdo->compression_context != NULL) {
for (i = 0; i < vdo->device_config->thread_counts.cpu_threads; i++)
vdo_free(vdo_forget(vdo->compression_context[i]));
vdo_free(vdo_forget(vdo->compression_context));
}
vdo_free(vdo);
}
static int initialize_super_block(struct vdo *vdo, struct vdo_super_block *super_block)
{
int result;
result = vdo_allocate(VDO_BLOCK_SIZE, char, "encoded super block",
(char **) &vdo->super_block.buffer);
if (result != VDO_SUCCESS)
return result;
return allocate_vio_components(vdo, VIO_TYPE_SUPER_BLOCK,
VIO_PRIORITY_METADATA, NULL, 1,
(char *) super_block->buffer,
&vdo->super_block.vio);
}
/**
* finish_reading_super_block() - Continue after loading the super block.
* @completion: The super block vio.
*
* This callback is registered in vdo_load_super_block().
*/
static void finish_reading_super_block(struct vdo_completion *completion)
{
struct vdo_super_block *super_block =
container_of(as_vio(completion), struct vdo_super_block, vio);
vdo_continue_completion(vdo_forget(completion->parent),
vdo_decode_super_block(super_block->buffer));
}
/**
* handle_super_block_read_error() - Handle an error reading the super block.
* @completion: The super block vio.
*
* This error handler is registered in vdo_load_super_block().
*/
static void handle_super_block_read_error(struct vdo_completion *completion)
{
vio_record_metadata_io_error(as_vio(completion));
finish_reading_super_block(completion);
}
static void read_super_block_endio(struct bio *bio)
{
struct vio *vio = bio->bi_private;
struct vdo_completion *parent = vio->completion.parent;
continue_vio_after_io(vio, finish_reading_super_block,
parent->callback_thread_id);
}
/**
* vdo_load_super_block() - Allocate a super block and read its contents from storage.
* @vdo: The vdo containing the super block on disk.
* @parent: The completion to notify after loading the super block.
*/
void vdo_load_super_block(struct vdo *vdo, struct vdo_completion *parent)
{
int result;
result = initialize_super_block(vdo, &vdo->super_block);
if (result != VDO_SUCCESS) {
vdo_continue_completion(parent, result);
return;
}
vdo->super_block.vio.completion.parent = parent;
vdo_submit_metadata_vio(&vdo->super_block.vio,
vdo_get_data_region_start(vdo->geometry),
read_super_block_endio,
handle_super_block_read_error,
REQ_OP_READ);
}
/**
* vdo_get_backing_device() - Get the block device object underlying a vdo.
* @vdo: The vdo.
*
* Return: The vdo's current block device.
*/
struct block_device *vdo_get_backing_device(const struct vdo *vdo)
{
return vdo->device_config->owned_device->bdev;
}
/**
* vdo_get_device_name() - Get the device name associated with the vdo target.
* @target: The target device interface.
*
* Return: The block device name.
*/
const char *vdo_get_device_name(const struct dm_target *target)
{
return dm_device_name(dm_table_get_md(target->table));
}
/**
* vdo_synchronous_flush() - Issue a flush request and wait for it to complete.
* @vdo: The vdo.
*
* Return: VDO_SUCCESS or an error.
*/
int vdo_synchronous_flush(struct vdo *vdo)
{
int result;
struct bio bio;
bio_init(&bio, vdo_get_backing_device(vdo), NULL, 0,
REQ_OP_WRITE | REQ_PREFLUSH);
submit_bio_wait(&bio);
result = blk_status_to_errno(bio.bi_status);
atomic64_inc(&vdo->stats.flush_out);
if (result != 0) {
vdo_log_error_strerror(result, "synchronous flush failed");
result = -EIO;
}
bio_uninit(&bio);
return result;
}
/**
* vdo_get_state() - Get the current state of the vdo.
* @vdo: The vdo.
* Context: This method may be called from any thread.
*
* Return: The current state of the vdo.
*/
enum vdo_state vdo_get_state(const struct vdo *vdo)
{
enum vdo_state state = atomic_read(&vdo->state);
/* pairs with barriers where state field is changed */
smp_rmb();
return state;
}
/**
* vdo_set_state() - Set the current state of the vdo.
* @vdo: The vdo whose state is to be set.
* @state: The new state of the vdo.
*
* Context: This method may be called from any thread.
*/
void vdo_set_state(struct vdo *vdo, enum vdo_state state)
{
/* pairs with barrier in vdo_get_state */
smp_wmb();
atomic_set(&vdo->state, state);
}
/**
* vdo_get_admin_state() - Get the admin state of the vdo.
* @vdo: The vdo.
*
* Return: The code for the vdo's current admin state.
*/
const struct admin_state_code *vdo_get_admin_state(const struct vdo *vdo)
{
return vdo_get_admin_state_code(&vdo->admin.state);
}
/**
* record_vdo() - Record the state of the VDO for encoding in the super block.
*/
static void record_vdo(struct vdo *vdo)
{
/* This is for backwards compatibility. */
vdo->states.unused = vdo->geometry.unused;
vdo->states.vdo.state = vdo_get_state(vdo);
vdo->states.block_map = vdo_record_block_map(vdo->block_map);
vdo->states.recovery_journal = vdo_record_recovery_journal(vdo->recovery_journal);
vdo->states.slab_depot = vdo_record_slab_depot(vdo->depot);
vdo->states.layout = vdo->layout;
}
/**
* continue_super_block_parent() - Continue the parent of a super block save operation.
* @completion: The super block vio.
*
* This callback is registered in vdo_save_components().
*/
static void continue_super_block_parent(struct vdo_completion *completion)
{
vdo_continue_completion(vdo_forget(completion->parent), completion->result);
}
/**
* handle_save_error() - Log a super block save error.
* @completion: The super block vio.
*
* This error handler is registered in vdo_save_components().
*/
static void handle_save_error(struct vdo_completion *completion)
{
struct vdo_super_block *super_block =
container_of(as_vio(completion), struct vdo_super_block, vio);
vio_record_metadata_io_error(&super_block->vio);
vdo_log_error_strerror(completion->result, "super block save failed");
/*
* Mark the super block as unwritable so that we won't attempt to write it again. This
* avoids the case where a growth attempt fails writing the super block with the new size,
* but the subsequent attempt to write out the read-only state succeeds. In this case,
* writes which happened just before the suspend would not be visible if the VDO is
* restarted without rebuilding, but, after a read-only rebuild, the effects of those
* writes would reappear.
*/
super_block->unwritable = true;
completion->callback(completion);
}
static void super_block_write_endio(struct bio *bio)
{
struct vio *vio = bio->bi_private;
struct vdo_completion *parent = vio->completion.parent;
continue_vio_after_io(vio, continue_super_block_parent,
parent->callback_thread_id);
}
/**
* vdo_save_components() - Encode the vdo and save the super block asynchronously.
* @vdo: The vdo whose state is being saved.
* @parent: The completion to notify when the save is complete.
*/
void vdo_save_components(struct vdo *vdo, struct vdo_completion *parent)
{
struct vdo_super_block *super_block = &vdo->super_block;
if (super_block->unwritable) {
vdo_continue_completion(parent, VDO_READ_ONLY);
return;
}
if (super_block->vio.completion.parent != NULL) {
vdo_continue_completion(parent, VDO_COMPONENT_BUSY);
return;
}
record_vdo(vdo);
vdo_encode_super_block(super_block->buffer, &vdo->states);
super_block->vio.completion.parent = parent;
super_block->vio.completion.callback_thread_id = parent->callback_thread_id;
vdo_submit_metadata_vio(&super_block->vio,
vdo_get_data_region_start(vdo->geometry),
super_block_write_endio, handle_save_error,
REQ_OP_WRITE | REQ_PREFLUSH | REQ_FUA);
}
/**
* vdo_register_read_only_listener() - Register a listener to be notified when the VDO goes
* read-only.
* @vdo: The vdo to register with.
* @listener: The object to notify.
* @notification: The function to call to send the notification.
* @thread_id: The id of the thread on which to send the notification.
*
* Return: VDO_SUCCESS or an error.
*/
int vdo_register_read_only_listener(struct vdo *vdo, void *listener,
vdo_read_only_notification_fn notification,
thread_id_t thread_id)
{
struct vdo_thread *thread = &vdo->threads[thread_id];
struct read_only_listener *read_only_listener;
int result;
result = VDO_ASSERT(thread_id != vdo->thread_config.dedupe_thread,
"read only listener not registered on dedupe thread");
if (result != VDO_SUCCESS)
return result;
result = vdo_allocate(1, struct read_only_listener, __func__,
&read_only_listener);
if (result != VDO_SUCCESS)
return result;
*read_only_listener = (struct read_only_listener) {
.listener = listener,
.notify = notification,
.next = thread->listeners,
};
thread->listeners = read_only_listener;
return VDO_SUCCESS;
}
/**
* notify_vdo_of_read_only_mode() - Notify a vdo that it is going read-only.
* @listener: The vdo.
* @parent: The completion to notify in order to acknowledge the notification.
*
* This will save the read-only state to the super block.
*
* Implements vdo_read_only_notification_fn.
*/
static void notify_vdo_of_read_only_mode(void *listener, struct vdo_completion *parent)
{
struct vdo *vdo = listener;
if (vdo_in_read_only_mode(vdo))
vdo_finish_completion(parent);
vdo_set_state(vdo, VDO_READ_ONLY_MODE);
vdo_save_components(vdo, parent);
}
/**
* vdo_enable_read_only_entry() - Enable a vdo to enter read-only mode on errors.
* @vdo: The vdo to enable.
*
* Return: VDO_SUCCESS or an error.
*/
int vdo_enable_read_only_entry(struct vdo *vdo)
{
thread_id_t id;
bool is_read_only = vdo_in_read_only_mode(vdo);
struct read_only_notifier *notifier = &vdo->read_only_notifier;
if (is_read_only) {
notifier->read_only_error = VDO_READ_ONLY;
notifier->state = NOTIFIED;
} else {
notifier->state = MAY_NOT_NOTIFY;
}
spin_lock_init(¬ifier->lock);
vdo_initialize_completion(¬ifier->completion, vdo,
VDO_READ_ONLY_MODE_COMPLETION);
for (id = 0; id < vdo->thread_config.thread_count; id++)
vdo->threads[id].is_read_only = is_read_only;
return vdo_register_read_only_listener(vdo, vdo, notify_vdo_of_read_only_mode,
vdo->thread_config.admin_thread);
}
/**
* vdo_wait_until_not_entering_read_only_mode() - Wait until no read-only notifications are in
* progress and prevent any subsequent
* notifications.
* @parent: The completion to notify when no threads are entering read-only mode.
*
* Notifications may be re-enabled by calling vdo_allow_read_only_mode_entry().
*/
void vdo_wait_until_not_entering_read_only_mode(struct vdo_completion *parent)
{
struct vdo *vdo = parent->vdo;
struct read_only_notifier *notifier = &vdo->read_only_notifier;
vdo_assert_on_admin_thread(vdo, __func__);
if (notifier->waiter != NULL) {
vdo_continue_completion(parent, VDO_COMPONENT_BUSY);
return;
}
spin_lock(¬ifier->lock);
if (notifier->state == NOTIFYING)
notifier->waiter = parent;
else if (notifier->state == MAY_NOTIFY)
notifier->state = MAY_NOT_NOTIFY;
spin_unlock(¬ifier->lock);
if (notifier->waiter == NULL) {
/*
* A notification was not in progress, and now they are
* disallowed.
*/
vdo_launch_completion(parent);
return;
}
}
/**
* as_notifier() - Convert a generic vdo_completion to a read_only_notifier.
* @completion: The completion to convert.
*
* Return: The completion as a read_only_notifier.
*/
static inline struct read_only_notifier *as_notifier(struct vdo_completion *completion)
{
vdo_assert_completion_type(completion, VDO_READ_ONLY_MODE_COMPLETION);
return container_of(completion, struct read_only_notifier, completion);
}
/**
* finish_entering_read_only_mode() - Complete the process of entering read only mode.
* @completion: The read-only mode completion.
*/
static void finish_entering_read_only_mode(struct vdo_completion *completion)
{
struct read_only_notifier *notifier = as_notifier(completion);
vdo_assert_on_admin_thread(completion->vdo, __func__);
spin_lock(¬ifier->lock);
notifier->state = NOTIFIED;
spin_unlock(¬ifier->lock);
if (notifier->waiter != NULL)
vdo_continue_completion(vdo_forget(notifier->waiter),
completion->result);
}
/**
* make_thread_read_only() - Inform each thread that the VDO is in read-only mode.
* @completion: The read-only mode completion.
*/
static void make_thread_read_only(struct vdo_completion *completion)
{
struct vdo *vdo = completion->vdo;
thread_id_t thread_id = completion->callback_thread_id;
struct read_only_notifier *notifier = as_notifier(completion);
struct read_only_listener *listener = completion->parent;
if (listener == NULL) {
/* This is the first call on this thread */
struct vdo_thread *thread = &vdo->threads[thread_id];
thread->is_read_only = true;
listener = thread->listeners;
if (thread_id == 0)
vdo_log_error_strerror(READ_ONCE(notifier->read_only_error),
"Unrecoverable error, entering read-only mode");
} else {
/* We've just finished notifying a listener */
listener = listener->next;
}
if (listener != NULL) {
/* We have a listener to notify */
vdo_prepare_completion(completion, make_thread_read_only,
make_thread_read_only, thread_id,
listener);
listener->notify(listener->listener, completion);
return;
}
/* We're done with this thread */
if (++thread_id == vdo->thread_config.dedupe_thread) {
/*
* We don't want to notify the dedupe thread since it may be
* blocked rebuilding the index.
*/
thread_id++;
}
if (thread_id >= vdo->thread_config.thread_count) {
/* There are no more threads */
vdo_prepare_completion(completion, finish_entering_read_only_mode,
finish_entering_read_only_mode,
vdo->thread_config.admin_thread, NULL);
} else {
vdo_prepare_completion(completion, make_thread_read_only,
make_thread_read_only, thread_id, NULL);
}
vdo_launch_completion(completion);
}
/**
* vdo_allow_read_only_mode_entry() - Allow the notifier to put the VDO into read-only mode,
* reversing the effects of
* vdo_wait_until_not_entering_read_only_mode().
* @parent: The object to notify once the operation is complete.
*
* If some thread tried to put the vdo into read-only mode while notifications were disallowed, it
* will be done when this method is called. If that happens, the parent will not be notified until
* the vdo has actually entered read-only mode and attempted to save the super block.
*
* Context: This method may only be called from the admin thread.
*/
void vdo_allow_read_only_mode_entry(struct vdo_completion *parent)
{
struct vdo *vdo = parent->vdo;
struct read_only_notifier *notifier = &vdo->read_only_notifier;
vdo_assert_on_admin_thread(vdo, __func__);
if (notifier->waiter != NULL) {
vdo_continue_completion(parent, VDO_COMPONENT_BUSY);
return;
}
spin_lock(¬ifier->lock);
if (notifier->state == MAY_NOT_NOTIFY) {
if (notifier->read_only_error == VDO_SUCCESS) {
notifier->state = MAY_NOTIFY;
} else {
notifier->state = NOTIFYING;
notifier->waiter = parent;
}
}
spin_unlock(¬ifier->lock);
if (notifier->waiter == NULL) {
/* We're done */
vdo_launch_completion(parent);
return;
}
/* Do the pending notification. */
make_thread_read_only(¬ifier->completion);
}
/**
* vdo_enter_read_only_mode() - Put a VDO into read-only mode and save the read-only state in the
* super block.
* @vdo: The vdo.
* @error_code: The error which caused the VDO to enter read-only mode.
*
* This method is a no-op if the VDO is already read-only.
*/
void vdo_enter_read_only_mode(struct vdo *vdo, int error_code)
{
bool notify = false;
thread_id_t thread_id = vdo_get_callback_thread_id();
struct read_only_notifier *notifier = &vdo->read_only_notifier;
struct vdo_thread *thread;
if (thread_id != VDO_INVALID_THREAD_ID) {
thread = &vdo->threads[thread_id];
if (thread->is_read_only) {
/* This thread has already gone read-only. */
return;
}
/* Record for this thread that the VDO is read-only. */
thread->is_read_only = true;
}
spin_lock(¬ifier->lock);
if (notifier->read_only_error == VDO_SUCCESS) {
WRITE_ONCE(notifier->read_only_error, error_code);
if (notifier->state == MAY_NOTIFY) {
notifier->state = NOTIFYING;
notify = true;
}
}
spin_unlock(¬ifier->lock);
if (!notify) {
/* The notifier is already aware of a read-only error */
return;
}
/* Initiate a notification starting on the lowest numbered thread. */
vdo_launch_completion_callback(¬ifier->completion, make_thread_read_only, 0);
}
/**
* vdo_is_read_only() - Check whether the VDO is read-only.
* @vdo: The vdo.
*
* Return: true if the vdo is read-only.
*
* This method may be called from any thread, as opposed to examining the VDO's state field which
* is only safe to check from the admin thread.
*/
bool vdo_is_read_only(struct vdo *vdo)
{
return vdo->threads[vdo_get_callback_thread_id()].is_read_only;
}
/**
* vdo_in_read_only_mode() - Check whether a vdo is in read-only mode.
* @vdo: The vdo to query.
*
* Return: true if the vdo is in read-only mode.
*/
bool vdo_in_read_only_mode(const struct vdo *vdo)
{
return (vdo_get_state(vdo) == VDO_READ_ONLY_MODE);
}
/**
* vdo_in_recovery_mode() - Check whether the vdo is in recovery mode.
* @vdo: The vdo to query.
*
* Return: true if the vdo is in recovery mode.
*/
bool vdo_in_recovery_mode(const struct vdo *vdo)
{
return (vdo_get_state(vdo) == VDO_RECOVERING);
}
/**
* vdo_enter_recovery_mode() - Put the vdo into recovery mode.
* @vdo: The vdo.
*/
void vdo_enter_recovery_mode(struct vdo *vdo)
{
vdo_assert_on_admin_thread(vdo, __func__);
if (vdo_in_read_only_mode(vdo))
return;
vdo_log_info("Entering recovery mode");
vdo_set_state(vdo, VDO_RECOVERING);
}
/**
* complete_synchronous_action() - Signal the waiting thread that a synchronous action is complete.
* @completion: The sync completion.
*/
static void complete_synchronous_action(struct vdo_completion *completion)
{
vdo_assert_completion_type(completion, VDO_SYNC_COMPLETION);
complete(&(container_of(completion, struct sync_completion,
vdo_completion)->completion));
}
/**
* perform_synchronous_action() - Launch an action on a VDO thread and wait for it to complete.
* @vdo: The vdo.
* @action: The callback to launch.
* @thread_id: The thread on which to run the action.
* @parent: The parent of the sync completion (may be NULL).
*/
static int perform_synchronous_action(struct vdo *vdo, vdo_action_fn action,
thread_id_t thread_id, void *parent)
{
struct sync_completion sync;
vdo_initialize_completion(&sync.vdo_completion, vdo, VDO_SYNC_COMPLETION);
init_completion(&sync.completion);
sync.vdo_completion.parent = parent;
vdo_launch_completion_callback(&sync.vdo_completion, action, thread_id);
wait_for_completion(&sync.completion);
return sync.vdo_completion.result;
}
/**
* set_compression_callback() - Callback to turn compression on or off.
* @completion: The completion.
*/
static void set_compression_callback(struct vdo_completion *completion)
{
struct vdo *vdo = completion->vdo;
bool *enable = completion->parent;
bool was_enabled = vdo_get_compressing(vdo);
if (*enable != was_enabled) {
WRITE_ONCE(vdo->compressing, *enable);
if (was_enabled) {
/* Signal the packer to flush since compression has been disabled. */
vdo_flush_packer(vdo->packer);
}
}
vdo_log_info("compression is %s", (*enable ? "enabled" : "disabled"));
*enable = was_enabled;
complete_synchronous_action(completion);
}
/**
* vdo_set_compressing() - Turn compression on or off.
* @vdo: The vdo.
* @enable: Whether to enable or disable compression.
*
* Return: Whether compression was previously on or off.
*/
bool vdo_set_compressing(struct vdo *vdo, bool enable)
{
perform_synchronous_action(vdo, set_compression_callback,
vdo->thread_config.packer_thread,
&enable);
return enable;
}
/**
* vdo_get_compressing() - Get whether compression is enabled in a vdo.
* @vdo: The vdo.
*
* Return: State of compression.
*/
bool vdo_get_compressing(struct vdo *vdo)
{
return READ_ONCE(vdo->compressing);
}
static size_t get_block_map_cache_size(const struct vdo *vdo)
{
return ((size_t) vdo->device_config->cache_size) * VDO_BLOCK_SIZE;
}
static struct error_statistics __must_check get_vdo_error_statistics(const struct vdo *vdo)
{
/*
* The error counts can be incremented from arbitrary threads and so must be incremented
* atomically, but they are just statistics with no semantics that could rely on memory
* order, so unfenced reads are sufficient.
*/
const struct atomic_statistics *atoms = &vdo->stats;
return (struct error_statistics) {
.invalid_advice_pbn_count = atomic64_read(&atoms->invalid_advice_pbn_count),
.no_space_error_count = atomic64_read(&atoms->no_space_error_count),
.read_only_error_count = atomic64_read(&atoms->read_only_error_count),
};
}
static void copy_bio_stat(struct bio_stats *b, const struct atomic_bio_stats *a)
{
b->read = atomic64_read(&a->read);
b->write = atomic64_read(&a->write);
b->discard = atomic64_read(&a->discard);
b->flush = atomic64_read(&a->flush);
b->empty_flush = atomic64_read(&a->empty_flush);
b->fua = atomic64_read(&a->fua);
}
static struct bio_stats subtract_bio_stats(struct bio_stats minuend,
struct bio_stats subtrahend)
{
return (struct bio_stats) {
.read = minuend.read - subtrahend.read,
.write = minuend.write - subtrahend.write,
.discard = minuend.discard - subtrahend.discard,
.flush = minuend.flush - subtrahend.flush,
.empty_flush = minuend.empty_flush - subtrahend.empty_flush,
.fua = minuend.fua - subtrahend.fua,
};
}
/**
* vdo_get_physical_blocks_allocated() - Get the number of physical blocks in use by user data.
* @vdo: The vdo.
*
* Return: The number of blocks allocated for user data.
*/
static block_count_t __must_check vdo_get_physical_blocks_allocated(const struct vdo *vdo)
{
return (vdo_get_slab_depot_allocated_blocks(vdo->depot) -
vdo_get_journal_block_map_data_blocks_used(vdo->recovery_journal));
}
/**
* vdo_get_physical_blocks_overhead() - Get the number of physical blocks used by vdo metadata.
* @vdo: The vdo.
*
* Return: The number of overhead blocks.
*/
static block_count_t __must_check vdo_get_physical_blocks_overhead(const struct vdo *vdo)
{
/*
* config.physical_blocks is mutated during resize and is in a packed structure,
* but resize runs on admin thread.
* TODO: Verify that this is always safe.
*/
return (vdo->states.vdo.config.physical_blocks -
vdo_get_slab_depot_data_blocks(vdo->depot) +
vdo_get_journal_block_map_data_blocks_used(vdo->recovery_journal));
}
static const char *vdo_describe_state(enum vdo_state state)
{
/* These strings should all fit in the 15 chars of VDOStatistics.mode. */
switch (state) {
case VDO_RECOVERING:
return "recovering";
case VDO_READ_ONLY_MODE:
return "read-only";
default:
return "normal";
}
}
/**
* get_vdo_statistics() - Populate a vdo_statistics structure on the admin thread.
* @vdo: The vdo.
* @stats: The statistics structure to populate.
*/
static void get_vdo_statistics(const struct vdo *vdo, struct vdo_statistics *stats)
{
struct recovery_journal *journal = vdo->recovery_journal;
enum vdo_state state = vdo_get_state(vdo);
vdo_assert_on_admin_thread(vdo, __func__);
/* start with a clean slate */
memset(stats, 0, sizeof(struct vdo_statistics));
/*
* These are immutable properties of the vdo object, so it is safe to query them from any
* thread.
*/
stats->version = STATISTICS_VERSION;
stats->logical_blocks = vdo->states.vdo.config.logical_blocks;
/*
* config.physical_blocks is mutated during resize and is in a packed structure, but resize
* runs on the admin thread.
* TODO: verify that this is always safe
*/
stats->physical_blocks = vdo->states.vdo.config.physical_blocks;
stats->block_size = VDO_BLOCK_SIZE;
stats->complete_recoveries = vdo->states.vdo.complete_recoveries;
stats->read_only_recoveries = vdo->states.vdo.read_only_recoveries;
stats->block_map_cache_size = get_block_map_cache_size(vdo);
/* The callees are responsible for thread-safety. */
stats->data_blocks_used = vdo_get_physical_blocks_allocated(vdo);
stats->overhead_blocks_used = vdo_get_physical_blocks_overhead(vdo);
stats->logical_blocks_used = vdo_get_recovery_journal_logical_blocks_used(journal);
vdo_get_slab_depot_statistics(vdo->depot, stats);
stats->journal = vdo_get_recovery_journal_statistics(journal);
stats->packer = vdo_get_packer_statistics(vdo->packer);
stats->block_map = vdo_get_block_map_statistics(vdo->block_map);
vdo_get_dedupe_statistics(vdo->hash_zones, stats);
stats->errors = get_vdo_error_statistics(vdo);
stats->in_recovery_mode = (state == VDO_RECOVERING);
snprintf(stats->mode, sizeof(stats->mode), "%s", vdo_describe_state(state));
stats->instance = vdo->instance;
stats->current_vios_in_progress = get_data_vio_pool_active_requests(vdo->data_vio_pool);
stats->max_vios = get_data_vio_pool_maximum_requests(vdo->data_vio_pool);
stats->flush_out = atomic64_read(&vdo->stats.flush_out);
stats->logical_block_size = vdo->device_config->logical_block_size;
copy_bio_stat(&stats->bios_in, &vdo->stats.bios_in);
copy_bio_stat(&stats->bios_in_partial, &vdo->stats.bios_in_partial);
copy_bio_stat(&stats->bios_out, &vdo->stats.bios_out);
copy_bio_stat(&stats->bios_meta, &vdo->stats.bios_meta);
copy_bio_stat(&stats->bios_journal, &vdo->stats.bios_journal);
copy_bio_stat(&stats->bios_page_cache, &vdo->stats.bios_page_cache);
copy_bio_stat(&stats->bios_out_completed, &vdo->stats.bios_out_completed);
copy_bio_stat(&stats->bios_meta_completed, &vdo->stats.bios_meta_completed);
copy_bio_stat(&stats->bios_journal_completed,
&vdo->stats.bios_journal_completed);
copy_bio_stat(&stats->bios_page_cache_completed,
&vdo->stats.bios_page_cache_completed);
copy_bio_stat(&stats->bios_acknowledged, &vdo->stats.bios_acknowledged);
copy_bio_stat(&stats->bios_acknowledged_partial, &vdo->stats.bios_acknowledged_partial);
stats->bios_in_progress =
subtract_bio_stats(stats->bios_in, stats->bios_acknowledged);
vdo_get_memory_stats(&stats->memory_usage.bytes_used,
&stats->memory_usage.peak_bytes_used);
}
/**
* vdo_fetch_statistics_callback() - Action to populate a vdo_statistics
* structure on the admin thread.
* @completion: The completion.
*
* This callback is registered in vdo_fetch_statistics().
*/
static void vdo_fetch_statistics_callback(struct vdo_completion *completion)
{
get_vdo_statistics(completion->vdo, completion->parent);
complete_synchronous_action(completion);
}
/**
* vdo_fetch_statistics() - Fetch statistics on the correct thread.
* @vdo: The vdo.
* @stats: The vdo statistics are returned here.
*/
void vdo_fetch_statistics(struct vdo *vdo, struct vdo_statistics *stats)
{
perform_synchronous_action(vdo, vdo_fetch_statistics_callback,
vdo->thread_config.admin_thread, stats);
}
/**
* vdo_get_callback_thread_id() - Get the id of the callback thread on which a completion is
* currently running.
*
* Return: The current thread ID, or -1 if no such thread.
*/
thread_id_t vdo_get_callback_thread_id(void)
{
struct vdo_work_queue *queue = vdo_get_current_work_queue();
struct vdo_thread *thread;
thread_id_t thread_id;
if (queue == NULL)
return VDO_INVALID_THREAD_ID;
thread = vdo_get_work_queue_owner(queue);
thread_id = thread->thread_id;
if (PARANOID_THREAD_CONSISTENCY_CHECKS) {
BUG_ON(thread_id >= thread->vdo->thread_config.thread_count);
BUG_ON(thread != &thread->vdo->threads[thread_id]);
}
return thread_id;
}
/**
* vdo_dump_status() - Dump status information about a vdo to the log for debugging.
* @vdo: The vdo to dump.
*/
void vdo_dump_status(const struct vdo *vdo)
{
zone_count_t zone;
vdo_dump_flusher(vdo->flusher);
vdo_dump_recovery_journal_statistics(vdo->recovery_journal);
vdo_dump_packer(vdo->packer);
vdo_dump_slab_depot(vdo->depot);
for (zone = 0; zone < vdo->thread_config.logical_zone_count; zone++)
vdo_dump_logical_zone(&vdo->logical_zones->zones[zone]);
for (zone = 0; zone < vdo->thread_config.physical_zone_count; zone++)
vdo_dump_physical_zone(&vdo->physical_zones->zones[zone]);
vdo_dump_hash_zones(vdo->hash_zones);
}
/**
* vdo_assert_on_admin_thread() - Assert that we are running on the admin thread.
* @vdo: The vdo.
* @name: The name of the function which should be running on the admin thread (for logging).
*/
void vdo_assert_on_admin_thread(const struct vdo *vdo, const char *name)
{
VDO_ASSERT_LOG_ONLY((vdo_get_callback_thread_id() == vdo->thread_config.admin_thread),
"%s called on admin thread", name);
}
/**
* vdo_assert_on_logical_zone_thread() - Assert that this function was called on the specified
* logical zone thread.
* @vdo: The vdo.
* @logical_zone: The number of the logical zone.
* @name: The name of the calling function.
*/
void vdo_assert_on_logical_zone_thread(const struct vdo *vdo, zone_count_t logical_zone,
const char *name)
{
VDO_ASSERT_LOG_ONLY((vdo_get_callback_thread_id() ==
vdo->thread_config.logical_threads[logical_zone]),
"%s called on logical thread", name);
}
/**
* vdo_assert_on_physical_zone_thread() - Assert that this function was called on the specified
* physical zone thread.
* @vdo: The vdo.
* @physical_zone: The number of the physical zone.
* @name: The name of the calling function.
*/
void vdo_assert_on_physical_zone_thread(const struct vdo *vdo,
zone_count_t physical_zone, const char *name)
{
VDO_ASSERT_LOG_ONLY((vdo_get_callback_thread_id() ==
vdo->thread_config.physical_threads[physical_zone]),
"%s called on physical thread", name);
}
/**
* vdo_get_physical_zone() - Get the physical zone responsible for a given physical block number.
* @vdo: The vdo containing the physical zones.
* @pbn: The PBN of the data block.
* @zone_ptr: A pointer to return the physical zone.
*
* Gets the physical zone responsible for a given physical block number of a data block in this vdo
* instance, or of the zero block (for which a NULL zone is returned). For any other block number
* that is not in the range of valid data block numbers in any slab, an error will be returned.
* This function is safe to call on invalid block numbers; it will not put the vdo into read-only
* mode.
*
* Return: VDO_SUCCESS or VDO_OUT_OF_RANGE if the block number is invalid or an error code for any
* other failure.
*/
int vdo_get_physical_zone(const struct vdo *vdo, physical_block_number_t pbn,
struct physical_zone **zone_ptr)
{
struct vdo_slab *slab;
int result;
if (pbn == VDO_ZERO_BLOCK) {
*zone_ptr = NULL;
return VDO_SUCCESS;
}
/*
* Used because it does a more restrictive bounds check than vdo_get_slab(), and done first
* because it won't trigger read-only mode on an invalid PBN.
*/
if (!vdo_is_physical_data_block(vdo->depot, pbn))
return VDO_OUT_OF_RANGE;
/* With the PBN already checked, we should always succeed in finding a slab. */
slab = vdo_get_slab(vdo->depot, pbn);
result = VDO_ASSERT(slab != NULL, "vdo_get_slab must succeed on all valid PBNs");
if (result != VDO_SUCCESS)
return result;
*zone_ptr = &vdo->physical_zones->zones[slab->allocator->zone_number];
return VDO_SUCCESS;
}
|