mapped_device
Regular
4.4
: ✅struct mapped_device {
struct srcu_struct io_barrier;
struct mutex suspend_lock;
atomic_t holders;
atomic_t open_count;
struct dm_table *map;
struct list_head table_devices;
struct mutex table_devices_lock;
long unsigned int flags;
struct request_queue *queue;
unsigned int type;
struct mutex type_lock;
struct target_type *immutable_target_type;
struct gendisk *disk;
char name[16];
void *interface_ptr;
atomic_t pending[2];
wait_queue_head_t wait;
struct work_struct work;
struct bio_list deferred;
spinlock_t deferred_lock;
struct workqueue_struct *wq;
mempool_t *io_pool;
mempool_t *rq_pool;
struct bio_set *bs;
atomic_t event_nr;
wait_queue_head_t eventq;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
struct super_block *frozen_sb;
struct block_device *bdev;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct bio flush_bio;
unsigned int internal_suspend_count;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set tag_set;
bool use_blk_mq;
};
4.8
: ✅struct mapped_device {
struct srcu_struct io_barrier;
struct mutex suspend_lock;
void *map;
struct list_head table_devices;
struct mutex table_devices_lock;
long unsigned int flags;
struct request_queue *queue;
int numa_node_id;
unsigned int type;
struct mutex type_lock;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
struct gendisk *disk;
char name[16];
void *interface_ptr;
atomic_t pending[2];
wait_queue_head_t wait;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct workqueue_struct *wq;
mempool_t *io_pool;
mempool_t *rq_pool;
struct bio_set *bs;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct block_device *bdev;
struct dm_kobject_holder kobj_holder;
struct bio flush_bio;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set *tag_set;
bool use_blk_mq;
bool init_tio_pdu;
};
4.10
: ✅struct mapped_device {
struct srcu_struct io_barrier;
struct mutex suspend_lock;
void *map;
struct list_head table_devices;
struct mutex table_devices_lock;
long unsigned int flags;
struct request_queue *queue;
int numa_node_id;
unsigned int type;
struct mutex type_lock;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
struct gendisk *disk;
char name[16];
void *interface_ptr;
atomic_t pending[2];
wait_queue_head_t wait;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct workqueue_struct *wq;
mempool_t *io_pool;
mempool_t *rq_pool;
struct bio_set *bs;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct block_device *bdev;
struct dm_kobject_holder kobj_holder;
struct bio flush_bio;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set *tag_set;
bool use_blk_mq;
bool init_tio_pdu;
};
4.13
: ✅struct mapped_device {
struct srcu_struct io_barrier;
struct mutex suspend_lock;
void *map;
struct list_head table_devices;
struct mutex table_devices_lock;
long unsigned int flags;
struct request_queue *queue;
int numa_node_id;
enum dm_queue_mode type;
struct mutex type_lock;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
struct gendisk *disk;
struct dax_device *dax_dev;
char name[16];
void *interface_ptr;
atomic_t pending[2];
wait_queue_head_t wait;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct workqueue_struct *wq;
mempool_t *io_pool;
struct bio_set *bs;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct block_device *bdev;
struct dm_kobject_holder kobj_holder;
struct bio flush_bio;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set *tag_set;
bool use_blk_mq;
bool init_tio_pdu;
};
4.15
: ✅struct mapped_device {
struct mutex suspend_lock;
void *map;
struct list_head table_devices;
struct mutex table_devices_lock;
long unsigned int flags;
struct request_queue *queue;
int numa_node_id;
enum dm_queue_mode type;
struct mutex type_lock;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
struct gendisk *disk;
struct dax_device *dax_dev;
char name[16];
void *interface_ptr;
atomic_t pending[2];
wait_queue_head_t wait;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct workqueue_struct *wq;
mempool_t *io_pool;
struct bio_set *bs;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct block_device *bdev;
struct dm_kobject_holder kobj_holder;
struct bio flush_bio;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set *tag_set;
bool use_blk_mq;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
4.18
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
atomic_t pending[2];
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct bio flush_bio;
struct dm_stats stats;
struct kthread_worker kworker;
struct task_struct *kworker_task;
unsigned int seq_rq_merge_deadline_usecs;
int last_rq_rw;
sector_t last_rq_pos;
ktime_t last_rq_start_time;
struct blk_mq_tag_set *tag_set;
bool use_blk_mq;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.0
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.3
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.4
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.8
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.11
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.13
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
5.15
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
unsigned int nr_zones;
unsigned int *zwp_offset;
struct dm_ima_measurements ima;
};
5.19
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
wait_queue_head_t wait;
long unsigned int *pending_io;
struct hd_geometry geometry;
struct workqueue_struct *wq;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
bool init_tio_pdu;
struct blk_mq_tag_set *tag_set;
struct dm_stats stats;
unsigned int internal_suspend_count;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_md_mempools *mempools;
struct dm_kobject_holder kobj_holder;
struct srcu_struct io_barrier;
unsigned int nr_zones;
unsigned int *zwp_offset;
struct dm_ima_measurements ima;
};
6.2
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
wait_queue_head_t wait;
long unsigned int *pending_io;
struct hd_geometry geometry;
struct workqueue_struct *wq;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
struct work_struct requeue_work;
struct dm_io *requeue_list;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
bool init_tio_pdu;
struct blk_mq_tag_set *tag_set;
struct dm_stats stats;
unsigned int internal_suspend_count;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_md_mempools *mempools;
struct dm_kobject_holder kobj_holder;
struct srcu_struct io_barrier;
unsigned int nr_zones;
unsigned int *zwp_offset;
struct dm_ima_measurements ima;
};
6.5
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
wait_queue_head_t wait;
long unsigned int *pending_io;
struct hd_geometry geometry;
struct workqueue_struct *wq;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
struct work_struct requeue_work;
struct dm_io *requeue_list;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
bool init_tio_pdu;
struct blk_mq_tag_set *tag_set;
struct dm_stats stats;
unsigned int internal_suspend_count;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_md_mempools *mempools;
struct dm_kobject_holder kobj_holder;
struct srcu_struct io_barrier;
unsigned int nr_zones;
unsigned int *zwp_offset;
struct dm_ima_measurements ima;
};
6.8
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
wait_queue_head_t wait;
long unsigned int *pending_io;
struct hd_geometry geometry;
struct workqueue_struct *wq;
struct work_struct work;
spinlock_t deferred_lock;
struct bio_list deferred;
struct work_struct requeue_work;
struct dm_io *requeue_list;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
bool init_tio_pdu;
struct blk_mq_tag_set *tag_set;
struct dm_stats stats;
unsigned int internal_suspend_count;
int swap_bios;
struct semaphore swap_bios_semaphore;
struct mutex swap_bios_lock;
struct dm_md_mempools *mempools;
struct dm_kobject_holder kobj_holder;
struct srcu_struct io_barrier;
unsigned int nr_zones;
unsigned int *zwp_offset;
struct dm_ima_measurements ima;
};
arm64
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
armhf
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
ppc64el
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
riscv64
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
aws
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
azure
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
gcp
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
lowlatency
: ✅struct mapped_device {
struct mutex suspend_lock;
struct mutex table_devices_lock;
struct list_head table_devices;
void *map;
long unsigned int flags;
struct mutex type_lock;
enum dm_queue_mode type;
int numa_node_id;
struct request_queue *queue;
atomic_t holders;
atomic_t open_count;
struct dm_target *immutable_target;
struct target_type *immutable_target_type;
char name[16];
struct gendisk *disk;
struct dax_device *dax_dev;
struct work_struct work;
wait_queue_head_t wait;
spinlock_t deferred_lock;
struct bio_list deferred;
void *interface_ptr;
wait_queue_head_t eventq;
atomic_t event_nr;
atomic_t uevent_seq;
struct list_head uevent_list;
spinlock_t uevent_lock;
unsigned int internal_suspend_count;
struct bio_set io_bs;
struct bio_set bs;
struct workqueue_struct *wq;
struct super_block *frozen_sb;
struct hd_geometry geometry;
struct dm_kobject_holder kobj_holder;
struct block_device *bdev;
struct dm_stats stats;
struct blk_mq_tag_set *tag_set;
bool init_tio_pdu;
struct srcu_struct io_barrier;
};
Regular
4.4
and 4.8
⚠️int numa_node_id
struct dm_target *immutable_target
bool init_tio_pdu
struct dm_table *map
➡️ void *map
struct blk_mq_tag_set tag_set
➡️ struct blk_mq_tag_set *tag_set
4.8
and 4.10
✅
4.10
and 4.13
⚠️struct dax_device *dax_dev
mempool_t *rq_pool
unsigned int type
➡️ enum dm_queue_mode type
4.13
and 4.15
✅
4.15
and 4.18
⚠️struct bio_set io_bs
mempool_t *io_pool
struct bio_set *bs
➡️ struct bio_set bs
4.18
and 5.0
⚠️atomic_t pending[2]
struct bio flush_bio
struct kthread_worker kworker
struct task_struct *kworker_task
unsigned int seq_rq_merge_deadline_usecs
int last_rq_rw
sector_t last_rq_pos
ktime_t last_rq_start_time
bool use_blk_mq
5.0
and 5.3
✅
5.3
and 5.4
✅
5.4
and 5.8
✅
5.8
and 5.11
⚠️int swap_bios
struct semaphore swap_bios_semaphore
struct mutex swap_bios_lock
struct super_block *frozen_sb
struct block_device *bdev
5.11
and 5.13
✅
5.13
and 5.15
⚠️unsigned int nr_zones
unsigned int *zwp_offset
struct dm_ima_measurements ima
5.15
and 5.19
⚠️long unsigned int *pending_io
struct dm_md_mempools *mempools
struct bio_set io_bs
struct bio_set bs
5.19
and 6.2
⚠️struct work_struct requeue_work
struct dm_io *requeue_list
6.2
and 6.5
✅
6.5
and 6.8
✅
amd64
and arm64
✅
amd64
and armhf
✅
amd64
and ppc64el
✅
amd64
and riscv64
✅
generic
and aws
✅
generic
and azure
✅
generic
and gcp
✅
generic
and lowlatency
✅