mapped_deviceRegular
4.4: ✅struct mapped_device {
struct srcu_struct io_barrier;
struct mutex suspend_lock;
atomic_t holders;
atomic_t open_count;
struct dm_table *map;
struct list_head table_devices;
struct mutex table_devices_lock;
long unsigned int flags;
struct request_queue *queue;
unsigned int type;
struct mutex type_lock;
struct target_type *immutable_target_type;
struct gendisk *disk;
char name[16];
void *interface_ptr;
atomic_t pending[2];
wait_queue_head_t wait;
struct work_struct work;
struct bio_list deferred;
spinlock_t deferred_lock;
struct workqueue_struct *wq;
mempool_t *io_pool;
mempool_t *rq_pool;
struct bio_set *bs;
atomic_t event_nr;
wait_queue_head_t eventq;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
struct super_block *frozen_sb;
struct block_device *bdev;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct bio flush_bio;
unsigned int internal_suspend_count;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set tag_set;
bool use_blk_mq;
};
4.8: ✅struct mapped_device {
struct srcu_struct io_barrier;
struct mutex suspend_lock;
void *map;
struct list_head table_devices;
struct mutex table_devices_lock;
long unsigned int flags;
struct request_queue *queue;
int numa_node_id;
unsigned int type;
struct mutex type_lock;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
struct gendisk *disk;
char name[16];
void *interface_ptr;
atomic_t pending[2];
wait_queue_head_t wait;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct workqueue_struct *wq;
mempool_t *io_pool;
mempool_t *rq_pool;
struct bio_set *bs;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct block_device *bdev;
struct dm_kobject_holder kobj_holder;
struct bio flush_bio;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set *tag_set;
bool use_blk_mq;
bool init_tio_pdu;
};
4.10: ✅struct mapped_device {
struct srcu_struct io_barrier;
struct mutex suspend_lock;
void *map;
struct list_head table_devices;
struct mutex table_devices_lock;
long unsigned int flags;
struct request_queue *queue;
int numa_node_id;
unsigned int type;
struct mutex type_lock;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
struct gendisk *disk;
char name[16];
void *interface_ptr;
atomic_t pending[2];
wait_queue_head_t wait;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct workqueue_struct *wq;
mempool_t *io_pool;
mempool_t *rq_pool;
struct bio_set *bs;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct block_device *bdev;
struct dm_kobject_holder kobj_holder;
struct bio flush_bio;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set *tag_set;
bool use_blk_mq;
bool init_tio_pdu;
};
4.13: ✅struct mapped_device {
struct srcu_struct io_barrier;
struct mutex suspend_lock;
void *map;
struct list_head table_devices;
struct mutex table_devices_lock;
long unsigned int flags;
struct request_queue *queue;
int numa_node_id;
enum dm_queue_mode type;
struct mutex type_lock;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
struct gendisk *disk;
struct dax_device *dax_dev;
char name[16];
void *interface_ptr;
atomic_t pending[2];
wait_queue_head_t wait;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct workqueue_struct *wq;
mempool_t *io_pool;
struct bio_set *bs;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct block_device *bdev;
struct dm_kobject_holder kobj_holder;
struct bio flush_bio;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set *tag_set;
bool use_blk_mq;
bool init_tio_pdu;
};
4.15: ✅struct mapped_device {
struct mutex suspend_lock;
void *map;
struct list_head table_devices;
struct mutex table_devices_lock;
long unsigned int flags;
struct request_queue *queue;
int numa_node_id;
enum dm_queue_mode type;
struct mutex type_lock;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
struct gendisk *disk;
struct dax_device *dax_dev;
char name[16];
void *interface_ptr;
atomic_t pending[2];
wait_queue_head_t wait;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct workqueue_struct *wq;
mempool_t *io_pool;
struct bio_set *bs;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct block_device *bdev;
struct dm_kobject_holder kobj_holder;
struct bio flush_bio;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set *tag_set;
bool use_blk_mq;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
4.18: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
atomic_t pending[2];
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct bio flush_bio;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set *tag_set;
bool use_blk_mq;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.0: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.3: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.4: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.8: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.11: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.13: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.15: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
unsigned int nr_zones;
unsigned int *zwp_offset;
struct dm_ima_measurements ima;
};
5.19: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
wait_queue_head_t wait;
long unsigned int *pending_io;
struct hd_geometry geometry;
struct workqueue_struct *wq;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
bool init_tio_pdu;
struct blk_mq_tag_set *tag_set;
struct dm_stats stats;
unsigned int internal_suspend_count;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_md_mempools *mempools;
struct dm_kobject_holder kobj_holder;
struct srcu_struct io_barrier;
unsigned int nr_zones;
unsigned int *zwp_offset;
struct dm_ima_measurements ima;
};
6.2: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
wait_queue_head_t wait;
long unsigned int *pending_io;
struct hd_geometry geometry;
struct workqueue_struct *wq;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
struct work_struct requeue_work;
struct dm_io *requeue_list;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
bool init_tio_pdu;
struct blk_mq_tag_set *tag_set;
struct dm_stats stats;
unsigned int internal_suspend_count;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_md_mempools *mempools;
struct dm_kobject_holder kobj_holder;
struct srcu_struct io_barrier;
unsigned int nr_zones;
unsigned int *zwp_offset;
struct dm_ima_measurements ima;
};
6.5: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
wait_queue_head_t wait;
long unsigned int *pending_io;
struct hd_geometry geometry;
struct workqueue_struct *wq;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
struct work_struct requeue_work;
struct dm_io *requeue_list;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
bool init_tio_pdu;
struct blk_mq_tag_set *tag_set;
struct dm_stats stats;
unsigned int internal_suspend_count;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_md_mempools *mempools;
struct dm_kobject_holder kobj_holder;
struct srcu_struct io_barrier;
unsigned int nr_zones;
unsigned int *zwp_offset;
struct dm_ima_measurements ima;
};
6.8: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
wait_queue_head_t wait;
long unsigned int *pending_io;
struct hd_geometry geometry;
struct workqueue_struct *wq;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
struct work_struct requeue_work;
struct dm_io *requeue_list;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
bool init_tio_pdu;
struct blk_mq_tag_set *tag_set;
struct dm_stats stats;
unsigned int internal_suspend_count;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_md_mempools *mempools;
struct dm_kobject_holder kobj_holder;
struct srcu_struct io_barrier;
unsigned int nr_zones;
unsigned int *zwp_offset;
struct dm_ima_measurements ima;
};
arm64: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
armhf: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
ppc64el: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
riscv64: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
aws: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
azure: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
gcp: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
lowlatency: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
Regular
4.4 and 4.8 ⚠️int numa_node_id
struct dm_target *immutable_target
bool init_tio_pdu
struct dm_table *map ➡️ void *map
struct blk_mq_tag_set tag_set ➡️ struct blk_mq_tag_set *tag_set
4.8 and 4.10 ✅
4.10 and 4.13 ⚠️struct dax_device *dax_dev
mempool_t *rq_pool
unsigned int type ➡️ enum dm_queue_mode type
4.13 and 4.15 ✅
4.15 and 4.18 ⚠️struct bio_set io_bs
mempool_t *io_pool
struct bio_set *bs ➡️ struct bio_set bs
4.18 and 5.0 ⚠️atomic_t pending[2]
struct bio flush_bio
struct kthread_worker kworker
struct task_struct *kworker_task
unsigned int seq_rq_merge_deadline_usecs
int last_rq_rw
sector_t last_rq_pos
ktime_t last_rq_start_time
bool use_blk_mq
5.0 and 5.3 ✅
5.3 and 5.4 ✅
5.4 and 5.8 ✅
5.8 and 5.11 ⚠️int swap_bios
struct semaphore swap_bios_semaphore
struct mutex swap_bios_lock
struct super_block *frozen_sb
struct block_device *bdev
5.11 and 5.13 ✅
5.13 and 5.15 ⚠️unsigned int nr_zones
unsigned int *zwp_offset
struct dm_ima_measurements ima
5.15 and 5.19 ⚠️long unsigned int *pending_io
struct dm_md_mempools *mempools
struct bio_set io_bs
struct bio_set bs
5.19 and 6.2 ⚠️struct work_struct requeue_work
struct dm_io *requeue_list
6.2 and 6.5 ✅
6.5 and 6.8 ✅
amd64 and arm64 ✅
amd64 and armhf ✅
amd64 and ppc64el ✅
amd64 and riscv64 ✅
generic and aws ✅
generic and azure ✅
generic and gcp ✅
generic and lowlatency ✅