Lines Matching refs:mapped_device

82 void dm_lock_md_type(struct mapped_device *md);
83 void dm_unlock_md_type(struct mapped_device *md);
84 void dm_set_md_type(struct mapped_device *md, unsigned type);
85 unsigned dm_get_md_type(struct mapped_device *md);
86 struct target_type *dm_get_immutable_target_type(struct mapped_device *md);
88 int dm_setup_md_queue(struct mapped_device *md);
127 int dm_deleting_md(struct mapped_device *md);
132 int dm_suspended_md(struct mapped_device *md);
137 int dm_suspended_internally_md(struct mapped_device *md);
138 void dm_internal_suspend_fast(struct mapped_device *md);
139 void dm_internal_resume_fast(struct mapped_device *md);
140 void dm_internal_suspend_noflush(struct mapped_device *md);
141 void dm_internal_resume(struct mapped_device *md);
146 int dm_test_deferred_remove_flag(struct mapped_device *md);
173 int dm_sysfs_init(struct mapped_device *md);
174 void dm_sysfs_exit(struct mapped_device *md);
175 struct kobject *dm_kobject(struct mapped_device *md);
176 struct mapped_device *dm_get_from_kobject(struct kobject *kobj);
195 void dm_destroy(struct mapped_device *md);
196 void dm_destroy_immediate(struct mapped_device *md);
197 int dm_open_count(struct mapped_device *md);
198 int dm_lock_for_deletion(struct mapped_device *md, bool mark_deferred, bool only_deferred);
199 int dm_cancel_deferred_remove(struct mapped_device *md);
200 int dm_request_based(struct mapped_device *md);
201 sector_t dm_get_size(struct mapped_device *md);
202 struct request_queue *dm_get_md_queue(struct mapped_device *md);
203 int dm_get_table_device(struct mapped_device *md, dev_t dev, fmode_t mode,
205 void dm_put_table_device(struct mapped_device *md, struct dm_dev *d);
206 struct dm_stats *dm_get_stats(struct mapped_device *md);
208 int dm_kobject_uevent(struct mapped_device *md, enum kobject_action action,
211 void dm_internal_suspend(struct mapped_device *md);
212 void dm_internal_resume(struct mapped_device *md);
214 bool dm_use_blk_mq(struct mapped_device *md);
225 struct dm_md_mempools *dm_alloc_md_mempools(struct mapped_device *md, unsigned type,
240 ssize_t dm_attr_rq_based_seq_io_merge_deadline_show(struct mapped_device *md, char *buf);
241 ssize_t dm_attr_rq_based_seq_io_merge_deadline_store(struct mapped_device *md,