Home
last modified time | relevance | path

Searched refs:md (Results 1 – 138 of 138) sorted by relevance

/linux-4.1.27/drivers/video/fbdev/omap/
Dlcd_mipid.c68 static void mipid_transfer(struct mipid_device *md, int cmd, const u8 *wbuf, in mipid_transfer() argument
76 BUG_ON(md->spi == NULL); in mipid_transfer()
117 r = spi_sync(md->spi, &m); in mipid_transfer()
119 dev_dbg(&md->spi->dev, "spi_sync %d\n", r); in mipid_transfer()
125 static inline void mipid_cmd(struct mipid_device *md, int cmd) in mipid_cmd() argument
127 mipid_transfer(md, cmd, NULL, 0, NULL, 0); in mipid_cmd()
130 static inline void mipid_write(struct mipid_device *md, in mipid_write() argument
133 mipid_transfer(md, reg, buf, len, NULL, 0); in mipid_write()
136 static inline void mipid_read(struct mipid_device *md, in mipid_read() argument
139 mipid_transfer(md, reg, NULL, 0, buf, len); in mipid_read()
[all …]
/linux-4.1.27/drivers/md/
Ddm.c69 struct mapped_device *md; member
83 struct mapped_device *md; member
240 bool dm_use_blk_mq(struct mapped_device *md) in dm_use_blk_mq() argument
242 return md->use_blk_mq; in dm_use_blk_mq()
437 int dm_deleting_md(struct mapped_device *md) in dm_deleting_md() argument
439 return test_bit(DMF_DELETING, &md->flags); in dm_deleting_md()
444 struct mapped_device *md; in dm_blk_open() local
448 md = bdev->bd_disk->private_data; in dm_blk_open()
449 if (!md) in dm_blk_open()
452 if (test_bit(DMF_FREEING, &md->flags) || in dm_blk_open()
[all …]
Ddm-era-target.c33 struct writeset_metadata md; member
73 ws->md.nr_bits = nr_blocks; in writeset_alloc()
74 ws->md.root = INVALID_WRITESET_ROOT; in writeset_alloc()
91 memset(ws->bits, 0, bitset_size(ws->md.nr_bits)); in writeset_init()
93 r = setup_on_disk_bitset(info, ws->md.nr_bits, &ws->md.root); in writeset_init()
137 r = dm_bitset_set_bit(info, ws->md.root, block, &ws->md.root); in writeset_test_and_set()
300 static int superblock_read_lock(struct era_metadata *md, in superblock_read_lock() argument
303 return dm_bm_read_lock(md->bm, SUPERBLOCK_LOCATION, in superblock_read_lock()
307 static int superblock_lock_zero(struct era_metadata *md, in superblock_lock_zero() argument
310 return dm_bm_write_lock_zero(md->bm, SUPERBLOCK_LOCATION, in superblock_lock_zero()
[all …]
Ddm.h82 void dm_lock_md_type(struct mapped_device *md);
83 void dm_unlock_md_type(struct mapped_device *md);
84 void dm_set_md_type(struct mapped_device *md, unsigned type);
85 unsigned dm_get_md_type(struct mapped_device *md);
86 struct target_type *dm_get_immutable_target_type(struct mapped_device *md);
88 int dm_setup_md_queue(struct mapped_device *md);
127 int dm_deleting_md(struct mapped_device *md);
132 int dm_suspended_md(struct mapped_device *md);
137 int dm_suspended_internally_md(struct mapped_device *md);
138 void dm_internal_suspend_fast(struct mapped_device *md);
[all …]
Ddm-ioctl.c35 struct mapped_device *md; member
118 dm_get(hc->md); in __get_name_cell()
132 dm_get(hc->md); in __get_uuid_cell()
141 struct mapped_device *md; in __get_dev_cell() local
144 md = dm_get_md(huge_decode_dev(dev)); in __get_dev_cell()
145 if (!md) in __get_dev_cell()
148 hc = dm_get_mdptr(md); in __get_dev_cell()
150 dm_put(md); in __get_dev_cell()
161 struct mapped_device *md) in alloc_cell() argument
189 hc->md = md; in alloc_cell()
[all …]
Ddm-sysfs.c25 struct mapped_device *md; in dm_attr_show() local
32 md = dm_get_from_kobject(kobj); in dm_attr_show()
33 if (!md) in dm_attr_show()
36 ret = dm_attr->show(md, page); in dm_attr_show()
37 dm_put(md); in dm_attr_show()
50 struct mapped_device *md; in dm_attr_store() local
57 md = dm_get_from_kobject(kobj); in dm_attr_store()
58 if (!md) in dm_attr_store()
61 ret = dm_attr->store(md, page, count); in dm_attr_store()
62 dm_put(md); in dm_attr_store()
[all …]
Ddm-raid.c68 struct mddev md; member
163 mddev_init(&rs->md); in context_alloc()
167 rs->md.raid_disks = raid_devs; in context_alloc()
168 rs->md.level = raid_type->level; in context_alloc()
169 rs->md.new_level = rs->md.level; in context_alloc()
170 rs->md.layout = raid_type->algorithm; in context_alloc()
171 rs->md.new_layout = rs->md.layout; in context_alloc()
172 rs->md.delta_disks = 0; in context_alloc()
173 rs->md.recovery_cp = 0; in context_alloc()
194 for (i = 0; i < rs->md.raid_disks; i++) { in context_free()
[all …]
Ddm-uevent.c44 struct mapped_device *md; member
57 static struct dm_uevent *dm_uevent_alloc(struct mapped_device *md) in dm_uevent_alloc() argument
66 event->md = md; in dm_uevent_alloc()
71 static struct dm_uevent *dm_build_path_uevent(struct mapped_device *md, in dm_build_path_uevent() argument
80 event = dm_uevent_alloc(md); in dm_build_path_uevent()
101 dm_next_uevent_seq(md))) { in dm_build_path_uevent()
146 if (dm_copy_name_and_uuid(event->md, event->name, in dm_send_uevents()
186 struct mapped_device *md = dm_table_get_md(ti->table); in dm_path_uevent() local
194 event = dm_build_path_uevent(md, ti, in dm_path_uevent()
201 dm_uevent_add(md, &event->elist); in dm_path_uevent()
Ddm-table.c32 struct mapped_device *md; member
184 unsigned num_targets, struct mapped_device *md) in dm_table_create() argument
210 t->md = md; in dm_table_create()
215 static void free_devices(struct list_head *devices, struct mapped_device *md) in free_devices() argument
223 dm_device_name(md), dd->dm_dev->name); in free_devices()
224 dm_put_table_device(md, dd->dm_dev); in free_devices()
253 free_devices(&t->devices, t->md); in dm_table_destroy()
298 dm_device_name(ti->table->md), bdevname(bdev, b), in device_area_is_invalid()
311 dm_device_name(ti->table->md), bdevname(bdev, b), in device_area_is_invalid()
324 dm_device_name(ti->table->md), in device_area_is_invalid()
[all …]
Ddm-stats.c233 struct mapped_device *md) in dm_stats_create() argument
310 suspend_callback(md); in dm_stats_create()
332 resume_callback(md); in dm_stats_create()
338 resume_callback(md); in dm_stats_create()
766 static int message_stats_create(struct mapped_device *md, in message_stats_create() argument
786 len = dm_get_size(md); in message_stats_create()
828 id = dm_stats_create(dm_get_stats(md), start, end, step, program_id, aux_data, in message_stats_create()
829 dm_internal_suspend_fast, dm_internal_resume_fast, md); in message_stats_create()
838 static int message_stats_delete(struct mapped_device *md, in message_stats_delete() argument
850 return dm_stats_delete(dm_get_stats(md), id); in message_stats_delete()
[all …]
DMakefile18 md-mod-y += md.o bitmap.o
33 obj-$(CONFIG_MD_CLUSTER) += md-cluster.o
35 obj-$(CONFIG_BLK_DEV_MD) += md-mod.o
Ddm-stats.h28 int dm_stats_message(struct mapped_device *md, unsigned argc, char **argv,
Ddm-verity.c207 struct mapped_device *md = dm_table_get_md(v->ti->table); in verity_handle_err() local
237 kobject_uevent_env(&disk_to_dev(dm_disk(md))->kobj, KOBJ_CHANGE, envp); in verity_handle_err()
DKconfig194 source "drivers/md/bcache/Kconfig"
249 source "drivers/md/persistent-data/Kconfig"
Ddm-thin.c425 static struct pool *__pool_table_lookup(struct mapped_device *md) in __pool_table_lookup() argument
432 if (tmp->pool_md == md) { in __pool_table_lookup()
/linux-4.1.27/arch/ia64/kernel/
Defi.c254 is_memory_available (efi_memory_desc_t *md) in is_memory_available() argument
256 if (!(md->attribute & EFI_MEMORY_WB)) in is_memory_available()
259 switch (md->type) { in is_memory_available()
278 #define efi_md_size(md) (md->num_pages << EFI_PAGE_SHIFT) argument
287 efi_md_end(efi_memory_desc_t *md) in efi_md_end() argument
289 return (md->phys_addr + efi_md_size(md)); in efi_md_end()
293 efi_wb(efi_memory_desc_t *md) in efi_wb() argument
295 return (md->attribute & EFI_MEMORY_WB); in efi_wb()
299 efi_uc(efi_memory_desc_t *md) in efi_uc() argument
301 return (md->attribute & EFI_MEMORY_UC); in efi_uc()
[all …]
/linux-4.1.27/drivers/staging/lustre/lnet/lnet/
Dlib-md.c47 lnet_md_unlink(lnet_libmd_t *md) in lnet_md_unlink() argument
49 if ((md->md_flags & LNET_MD_FLAG_ZOMBIE) == 0) { in lnet_md_unlink()
51 lnet_me_t *me = md->md_me; in lnet_md_unlink()
53 md->md_flags |= LNET_MD_FLAG_ZOMBIE; in lnet_md_unlink()
60 lnet_ptl_detach_md(me, md); in lnet_md_unlink()
66 lnet_res_lh_invalidate(&md->md_lh); in lnet_md_unlink()
69 if (md->md_refcount != 0) { in lnet_md_unlink()
70 CDEBUG(D_NET, "Queueing unlink of md %p\n", md); in lnet_md_unlink()
74 CDEBUG(D_NET, "Unlinking md %p\n", md); in lnet_md_unlink()
76 if (md->md_eq != NULL) { in lnet_md_unlink()
[all …]
Dlib-me.c232 lnet_libmd_t *md; in LNetMEUnlink() local
248 md = me->me_md; in LNetMEUnlink()
249 if (md != NULL) { in LNetMEUnlink()
250 md->md_flags |= LNET_MD_FLAG_ABORTED; in LNetMEUnlink()
251 if (md->md_eq != NULL && md->md_refcount == 0) { in LNetMEUnlink()
252 lnet_build_unlink_event(md, &ev); in LNetMEUnlink()
253 lnet_eq_enqueue_event(md->md_eq, &ev); in LNetMEUnlink()
271 lnet_libmd_t *md = me->me_md; in lnet_me_unlink() local
274 lnet_ptl_detach_md(me, md); in lnet_me_unlink()
275 lnet_md_unlink(md); in lnet_me_unlink()
[all …]
Dlib-ptl.c139 lnet_try_match_md(lnet_libmd_t *md, in lnet_try_match_md() argument
146 lnet_me_t *me = md->md_me; in lnet_try_match_md()
149 if (lnet_md_exhausted(md)) in lnet_try_match_md()
153 if ((md->md_options & info->mi_opc) == 0) in lnet_try_match_md()
171 if ((md->md_options & LNET_MD_MANAGE_REMOTE) == 0) in lnet_try_match_md()
172 offset = md->md_offset; in lnet_try_match_md()
176 if ((md->md_options & LNET_MD_MAX_SIZE) != 0) { in lnet_try_match_md()
177 mlength = md->md_max_size; in lnet_try_match_md()
178 LASSERT(md->md_offset + mlength <= md->md_length); in lnet_try_match_md()
180 mlength = md->md_length - offset; in lnet_try_match_md()
[all …]
Dlib-msg.c46 lnet_build_unlink_event(lnet_libmd_t *md, lnet_event_t *ev) in lnet_build_unlink_event() argument
53 lnet_md_deconstruct(md, &ev->md); in lnet_build_unlink_event()
54 lnet_md2handle(&ev->md_handle, md); in lnet_build_unlink_event()
308 lnet_msg_attach_md(lnet_msg_t *msg, lnet_libmd_t *md, in lnet_msg_attach_md() argument
318 msg->msg_md = md; in lnet_msg_attach_md()
324 md->md_refcount++; in lnet_msg_attach_md()
325 if (md->md_threshold != LNET_MD_THRESH_INF) { in lnet_msg_attach_md()
326 LASSERT(md->md_threshold > 0); in lnet_msg_attach_md()
327 md->md_threshold--; in lnet_msg_attach_md()
331 lnet_md2handle(&msg->msg_ev.md_handle, md); in lnet_msg_attach_md()
[all …]
Dlib-move.c605 lnet_libmd_t *md = msg->msg_md; in lnet_setpayloadbuffer() local
609 LASSERT(md != NULL); in lnet_setpayloadbuffer()
614 msg->msg_niov = md->md_niov; in lnet_setpayloadbuffer()
615 if ((md->md_options & LNET_MD_KIOV) != 0) in lnet_setpayloadbuffer()
616 msg->msg_kiov = md->md_iov.kiov; in lnet_setpayloadbuffer()
618 msg->msg_iov = md->md_iov.iov; in lnet_setpayloadbuffer()
1503 lnet_libmd_t *md; in lnet_parse_reply() local
1515 md = lnet_wire_handle2md(&hdr->msg.reply.dst_wmd); in lnet_parse_reply()
1516 if (md == NULL || md->md_threshold == 0 || md->md_me != NULL) { in lnet_parse_reply()
1519 (md == NULL) ? "invalid" : "inactive", in lnet_parse_reply()
[all …]
Dapi-ni.c1657 lnet_md_t md = { NULL }; in lnet_ping_target_init() local
1692 md.start = the_lnet.ln_ping_info; in lnet_ping_target_init()
1693 md.length = infosz; in lnet_ping_target_init()
1694 md.threshold = LNET_MD_THRESH_INF; in lnet_ping_target_init()
1695 md.max_size = 0; in lnet_ping_target_init()
1696 md.options = LNET_MD_OP_GET | LNET_MD_TRUNCATE | in lnet_ping_target_init()
1698 md.user_ptr = NULL; in lnet_ping_target_init()
1699 md.eq_handle = the_lnet.ln_ping_target_eq; in lnet_ping_target_init()
1701 rc = LNetMDAttach(meh, md, in lnet_ping_target_init()
1765 lnet_md_t md = { NULL }; in lnet_ping() local
[all …]
DMakefile4 lib-md.o lib-ptl.o lib-move.o module.o lo.o router.o \
Drouter.c702 lnet_rc_data_t *rcd = event->md.user_ptr; in lnet_router_checker_event()
/linux-4.1.27/arch/mips/pci/
Dmsi-xlp.c134 struct xlp_msi_data *md = irq_data_get_irq_handler_data(d); in xlp_msi_enable() local
139 spin_lock_irqsave(&md->msi_lock, flags); in xlp_msi_enable()
140 md->msi_enabled_mask |= 1u << vec; in xlp_msi_enable()
142 nlm_write_reg(md->lnkbase, PCIE_9XX_MSI_EN, in xlp_msi_enable()
143 md->msi_enabled_mask); in xlp_msi_enable()
145 nlm_write_reg(md->lnkbase, PCIE_MSI_EN, md->msi_enabled_mask); in xlp_msi_enable()
146 spin_unlock_irqrestore(&md->msi_lock, flags); in xlp_msi_enable()
151 struct xlp_msi_data *md = irq_data_get_irq_handler_data(d); in xlp_msi_disable() local
156 spin_lock_irqsave(&md->msi_lock, flags); in xlp_msi_disable()
157 md->msi_enabled_mask &= ~(1u << vec); in xlp_msi_disable()
[all …]
/linux-4.1.27/arch/x86/platform/efi/
Defi.c131 efi_memory_desc_t *md = p; in do_add_efi_memmap() local
132 unsigned long long start = md->phys_addr; in do_add_efi_memmap()
133 unsigned long long size = md->num_pages << EFI_PAGE_SHIFT; in do_add_efi_memmap()
136 switch (md->type) { in do_add_efi_memmap()
142 if (md->attribute & EFI_MEMORY_WB) in do_add_efi_memmap()
204 efi_memory_desc_t *md; in print_efi_memmap() local
213 md = p; in print_efi_memmap()
215 i, efi_md_typeattr_format(buf, sizeof(buf), md), in print_efi_memmap()
216 md->phys_addr, in print_efi_memmap()
217 md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT), in print_efi_memmap()
[all …]
Dquirks.c146 efi_memory_desc_t *md = p; in efi_reserve_boot_services() local
147 u64 start = md->phys_addr; in efi_reserve_boot_services()
148 u64 size = md->num_pages << EFI_PAGE_SHIFT; in efi_reserve_boot_services()
150 if (md->type != EFI_BOOT_SERVICES_CODE && in efi_reserve_boot_services()
151 md->type != EFI_BOOT_SERVICES_DATA) in efi_reserve_boot_services()
164 md->num_pages = 0; in efi_reserve_boot_services()
177 efi_memory_desc_t *md = p; in efi_free_boot_services() local
178 unsigned long long start = md->phys_addr; in efi_free_boot_services()
179 unsigned long long size = md->num_pages << EFI_PAGE_SHIFT; in efi_free_boot_services()
181 if (md->type != EFI_BOOT_SERVICES_CODE && in efi_free_boot_services()
[all …]
Defi_64.c63 efi_memory_desc_t *md; in early_code_mapping_set_exec() local
71 md = p; in early_code_mapping_set_exec()
72 if (md->type == EFI_RUNTIME_SERVICES_CODE || in early_code_mapping_set_exec()
73 md->type == EFI_BOOT_SERVICES_CODE) in early_code_mapping_set_exec()
74 efi_set_executable(md, executable); in early_code_mapping_set_exec()
204 static void __init __map_region(efi_memory_desc_t *md, u64 va) in __map_region() argument
209 if (!(md->attribute & EFI_MEMORY_WB)) in __map_region()
212 if (kernel_map_pages_in_pgd(pgd, md->phys_addr, va, md->num_pages, pf)) in __map_region()
214 md->phys_addr, va); in __map_region()
217 void __init efi_map_region(efi_memory_desc_t *md) in efi_map_region() argument
[all …]
Defi_32.c51 void __init efi_map_region(efi_memory_desc_t *md) in efi_map_region() argument
53 old_map_region(md); in efi_map_region()
56 void __init efi_map_region_fixed(efi_memory_desc_t *md) {} in efi_map_region_fixed() argument
/linux-4.1.27/arch/arm64/kernel/
Defi.c63 static int __init is_normal_ram(efi_memory_desc_t *md) in is_normal_ram() argument
65 if (md->attribute & EFI_MEMORY_WB) in is_normal_ram()
77 efi_memory_desc_t *md; in efi_to_phys() local
79 for_each_efi_memory_desc(&memmap, md) { in efi_to_phys()
80 if (!(md->attribute & EFI_MEMORY_RUNTIME)) in efi_to_phys()
82 if (md->virt_addr == 0) in efi_to_phys()
85 if (md->virt_addr <= addr && in efi_to_phys()
86 (addr - md->virt_addr) < (md->num_pages << EFI_PAGE_SHIFT)) in efi_to_phys()
87 return md->phys_addr + addr - md->virt_addr; in efi_to_phys()
153 static __init int is_reserve_region(efi_memory_desc_t *md) in is_reserve_region() argument
[all …]
/linux-4.1.27/drivers/mmc/card/
Dblock.c139 struct mmc_blk_data *md);
157 struct mmc_blk_data *md; in mmc_blk_get() local
160 md = disk->private_data; in mmc_blk_get()
161 if (md && md->usage == 0) in mmc_blk_get()
162 md = NULL; in mmc_blk_get()
163 if (md) in mmc_blk_get()
164 md->usage++; in mmc_blk_get()
167 return md; in mmc_blk_get()
180 static void mmc_blk_put(struct mmc_blk_data *md) in mmc_blk_put() argument
183 md->usage--; in mmc_blk_put()
[all …]
/linux-4.1.27/drivers/video/fbdev/matrox/
Dmatroxfb_maven.c134 static int* get_ctrl_ptr(struct maven_data* md, int idx) { in get_ctrl_ptr() argument
135 return (int*)((char*)(md->primary_head) + maven_controls[idx].control); in get_ctrl_ptr()
339 static unsigned char maven_compute_deflicker (const struct maven_data* md) { in maven_compute_deflicker() argument
342 df = (md->version == MGATVO_B?0x40:0x00); in maven_compute_deflicker()
343 switch (md->primary_head->altout.tvo_params.deflicker) { in maven_compute_deflicker()
357 static void maven_compute_bwlevel (const struct maven_data* md, in maven_compute_bwlevel() argument
359 const int b = md->primary_head->altout.tvo_params.brightness + BLMIN; in maven_compute_bwlevel()
360 const int c = md->primary_head->altout.tvo_params.contrast; in maven_compute_bwlevel()
366 static const struct maven_gamma* maven_compute_gamma (const struct maven_data* md) { in maven_compute_gamma() argument
367 return maven_gamma + md->primary_head->altout.tvo_params.gamma; in maven_compute_gamma()
[all …]
Dmatroxfb_g450.c141 static int g450_query_ctrl(void* md, struct v4l2_queryctrl *p) { in g450_query_ctrl() argument
162 static int g450_set_ctrl(void* md, struct v4l2_control *p) { in g450_set_ctrl() argument
164 struct matrox_fb_info *minfo = md; in g450_set_ctrl()
216 static int g450_get_ctrl(void* md, struct v4l2_control *p) { in g450_get_ctrl() argument
218 struct matrox_fb_info *minfo = md; in g450_get_ctrl()
520 static int matroxfb_g450_compute(void* md, struct my_timming* mt) { in matroxfb_g450_compute() argument
521 struct matrox_fb_info *minfo = md; in matroxfb_g450_compute()
558 static int matroxfb_g450_program(void* md) { in matroxfb_g450_program() argument
559 struct matrox_fb_info *minfo = md; in matroxfb_g450_program()
567 static int matroxfb_g450_verify_mode(void* md, u_int32_t arg) { in matroxfb_g450_verify_mode() argument
[all …]
/linux-4.1.27/drivers/net/ethernet/tile/
Dtilegx.c360 struct mpipe_data *md = &mpipe_data[instance]; in tile_net_provide_buffer() local
382 gxio_mpipe_push_buffer(&md->context, md->first_buffer_stack + kind, in tile_net_provide_buffer()
410 struct mpipe_data *md = &mpipe_data[instance]; in tile_net_pop_all_buffers() local
414 (tile_io_addr_t)gxio_mpipe_pop_buffer(&md->context, in tile_net_pop_all_buffers()
462 struct mpipe_data *md = &mpipe_data[instance]; in tile_tx_timestamp() local
467 gxio_mpipe_get_timestamp(&md->context, &ts); in tile_tx_timestamp()
588 struct mpipe_data *md = &mpipe_data[instance]; in tile_net_handle_packet() local
589 struct net_device *dev = md->tile_net_devs_for_channel[idesc->channel]; in tile_net_handle_packet()
657 struct mpipe_data *md; in tile_net_poll() local
682 md = &mpipe_data[instance]; in tile_net_poll()
[all …]
/linux-4.1.27/drivers/w1/
Dw1.c84 struct w1_master *md = dev_to_w1_master(dev); in w1_master_release() local
86 dev_dbg(dev, "%s: Releasing %s.\n", __func__, md->name); in w1_master_release()
87 memset(md, 0, sizeof(struct w1_master) + sizeof(struct w1_bus_master)); in w1_master_release()
88 kfree(md); in w1_master_release()
223 struct w1_master *md = dev_to_w1_master(dev); in w1_master_attribute_show_name() local
226 mutex_lock(&md->mutex); in w1_master_attribute_show_name()
227 count = sprintf(buf, "%s\n", md->name); in w1_master_attribute_show_name()
228 mutex_unlock(&md->mutex); in w1_master_attribute_show_name()
238 struct w1_master *md = dev_to_w1_master(dev); in w1_master_attribute_store_search() local
245 mutex_lock(&md->mutex); in w1_master_attribute_store_search()
[all …]
/linux-4.1.27/drivers/staging/lustre/include/linux/lnet/
Dlib-lnet.h87 static inline int lnet_md_exhausted(lnet_libmd_t *md) in lnet_md_exhausted() argument
89 return (md->md_threshold == 0 || in lnet_md_exhausted()
90 ((md->md_options & LNET_MD_MAX_SIZE) != 0 && in lnet_md_exhausted()
91 md->md_offset + md->md_max_size > md->md_length)); in lnet_md_exhausted()
94 static inline int lnet_md_unlinkable(lnet_libmd_t *md) in lnet_md_unlinkable() argument
101 if (md->md_refcount != 0) in lnet_md_unlinkable()
104 if ((md->md_flags & LNET_MD_FLAG_ZOMBIE) != 0) in lnet_md_unlinkable()
107 return ((md->md_flags & LNET_MD_FLAG_AUTO_UNLINK) != 0 && in lnet_md_unlinkable()
108 lnet_md_exhausted(md)); in lnet_md_unlinkable()
254 lnet_libmd_t *md; in lnet_md_alloc() local
[all …]
Dtypes.h426 lnet_md_t md; member
/linux-4.1.27/drivers/staging/lustre/lustre/ptlrpc/
Dpers.c46 void ptlrpc_fill_bulk_md(lnet_md_t *md, struct ptlrpc_bulk_desc *desc, in ptlrpc_fill_bulk_md() argument
53 LASSERT(!(md->options & (LNET_MD_IOVEC | LNET_MD_KIOV | in ptlrpc_fill_bulk_md()
56 md->options |= LNET_MD_KIOV; in ptlrpc_fill_bulk_md()
57 md->length = max(0, desc->bd_iov_count - mdidx * LNET_MAX_IOV); in ptlrpc_fill_bulk_md()
58 md->length = min_t(unsigned int, LNET_MAX_IOV, md->length); in ptlrpc_fill_bulk_md()
60 md->start = &desc->bd_enc_iov[mdidx * LNET_MAX_IOV]; in ptlrpc_fill_bulk_md()
62 md->start = &desc->bd_iov[mdidx * LNET_MAX_IOV]; in ptlrpc_fill_bulk_md()
Dniobuf.c56 lnet_md_t md; in ptl_send_buf() local
61 md.start = base; in ptl_send_buf()
62 md.length = len; in ptl_send_buf()
63 md.threshold = (ack == LNET_ACK_REQ) ? 2 : 1; in ptl_send_buf()
64 md.options = PTLRPC_MD_OPTIONS; in ptl_send_buf()
65 md.user_ptr = cbid; in ptl_send_buf()
66 md.eq_handle = ptlrpc_eq_h; in ptl_send_buf()
75 rc = LNetMDBind(md, LNET_UNLINK, mdh); in ptl_send_buf()
124 lnet_md_t md; in ptlrpc_register_bulk() local
169 md.user_ptr = &desc->bd_cbid; in ptlrpc_register_bulk()
[all …]
Devents.c56 struct ptlrpc_cb_id *cbid = ev->md.user_ptr; in request_out_callback()
89 struct ptlrpc_cb_id *cbid = ev->md.user_ptr; in reply_in_callback()
95 LASSERT(ev->md.start == req->rq_repbuf); in reply_in_callback()
99 LASSERT((ev->md.options & LNET_MD_MANAGE_REMOTE) != 0); in reply_in_callback()
175 struct ptlrpc_cb_id *cbid = ev->md.user_ptr; in client_bulk_callback()
284 struct ptlrpc_cb_id *cbid = ev->md.user_ptr; in request_in_callback()
292 LASSERT((char *)ev->md.start >= rqbd->rqbd_buffer); in request_in_callback()
293 LASSERT((char *)ev->md.start + ev->offset + ev->mlength <= in request_in_callback()
327 req->rq_reqbuf = ev->md.start + ev->offset; in request_in_callback()
383 struct ptlrpc_cb_id *cbid = ev->md.user_ptr; in reply_out_callback()
[all …]
Dptlrpc_internal.h233 void ptlrpc_fill_bulk_md(lnet_md_t *md, struct ptlrpc_bulk_desc *desc,
/linux-4.1.27/include/linux/
Ddevice-mapper.h371 int dm_create(int minor, struct mapped_device **md);
377 void dm_get(struct mapped_device *md);
378 int dm_hold(struct mapped_device *md);
379 void dm_put(struct mapped_device *md);
384 void dm_set_mdptr(struct mapped_device *md, void *ptr);
385 void *dm_get_mdptr(struct mapped_device *md);
390 int dm_suspend(struct mapped_device *md, unsigned suspend_flags);
391 int dm_resume(struct mapped_device *md);
396 uint32_t dm_get_event_nr(struct mapped_device *md);
397 int dm_wait_event(struct mapped_device *md, int event_nr);
[all …]
Defi.h896 #define for_each_efi_memory_desc(m, md) \ argument
897 for ((md) = (m)->map; \
898 (md) <= (efi_memory_desc_t *)((m)->map_end - (m)->desc_size); \
899 (md) = (void *)(md) + (m)->desc_size)
906 const efi_memory_desc_t *md);
/linux-4.1.27/drivers/input/touchscreen/
Dcyttsp4_core.c719 static void cyttsp4_report_slot_liftoff(struct cyttsp4_mt_data *md, in cyttsp4_report_slot_liftoff() argument
724 if (md->num_prv_tch == 0) in cyttsp4_report_slot_liftoff()
728 input_mt_slot(md->input, t); in cyttsp4_report_slot_liftoff()
729 input_mt_report_slot_state(md->input, in cyttsp4_report_slot_liftoff()
734 static void cyttsp4_lift_all(struct cyttsp4_mt_data *md) in cyttsp4_lift_all() argument
736 if (!md->si) in cyttsp4_lift_all()
739 if (md->num_prv_tch != 0) { in cyttsp4_lift_all()
740 cyttsp4_report_slot_liftoff(md, in cyttsp4_lift_all()
741 md->si->si_ofs.tch_abs[CY_TCH_T].max); in cyttsp4_lift_all()
742 input_sync(md->input); in cyttsp4_lift_all()
[all …]
Dcyttsp4_core.h360 struct cyttsp4_mt_data md; member
/linux-4.1.27/drivers/mtd/nand/
Dnand_bbt.c392 struct nand_bbt_descr *td, struct nand_bbt_descr *md) in read_abs_bbts() argument
406 if (md && (md->options & NAND_BBT_VERSION)) { in read_abs_bbts()
407 scan_read(mtd, buf, (loff_t)md->pages[0] << this->page_shift, in read_abs_bbts()
408 mtd->writesize, md); in read_abs_bbts()
409 md->version[0] = buf[bbt_get_ver_offs(mtd, md)]; in read_abs_bbts()
411 md->pages[0], md->version[0]); in read_abs_bbts()
600 struct nand_bbt_descr *md) in search_read_bbts() argument
606 if (md) in search_read_bbts()
607 search_bbt(mtd, buf, md); in search_read_bbts()
621 struct nand_bbt_descr *td, struct nand_bbt_descr *md, in write_bbt() argument
[all …]
/linux-4.1.27/arch/unicore32/mm/
Dmmu.c208 static void __init create_mapping(struct map_desc *md) in create_mapping() argument
214 if (md->virtual != vectors_base() && md->virtual < TASK_SIZE) { in create_mapping()
217 __pfn_to_phys((u64)md->pfn), md->virtual); in create_mapping()
221 if ((md->type == MT_DEVICE || md->type == MT_ROM) && in create_mapping()
222 md->virtual >= PAGE_OFFSET && md->virtual < VMALLOC_END) { in create_mapping()
225 __pfn_to_phys((u64)md->pfn), md->virtual); in create_mapping()
228 type = &mem_types[md->type]; in create_mapping()
230 addr = md->virtual & PAGE_MASK; in create_mapping()
231 phys = (unsigned long)__pfn_to_phys(md->pfn); in create_mapping()
232 length = PAGE_ALIGN(md->length + (md->virtual & ~PAGE_MASK)); in create_mapping()
[all …]
/linux-4.1.27/drivers/clk/qcom/
Dclk-rcg.c114 static u32 md_to_m(struct mn *mn, u32 md) in md_to_m() argument
116 md >>= mn->m_val_shift; in md_to_m()
117 md &= BIT(mn->width) - 1; in md_to_m()
118 return md; in md_to_m()
140 static u32 mn_to_md(struct mn *mn, u32 m, u32 n, u32 md) in mn_to_md() argument
146 md &= ~mask; in mn_to_md()
150 md |= m; in mn_to_md()
151 md |= ~n & mask_w; in mn_to_md()
154 return md; in mn_to_md()
208 u32 ns, md, reg; in configure_bank() local
[all …]
/linux-4.1.27/arch/arm/mm/
Dmmu.c779 static void __init create_36bit_mapping(struct map_desc *md, in create_36bit_mapping() argument
786 addr = md->virtual; in create_36bit_mapping()
787 phys = __pfn_to_phys(md->pfn); in create_36bit_mapping()
788 length = PAGE_ALIGN(md->length); in create_36bit_mapping()
792 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping()
804 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping()
808 if ((addr | length | __pfn_to_phys(md->pfn)) & ~SUPERSECTION_MASK) { in create_36bit_mapping()
810 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping()
818 phys |= (((md->pfn >> (32 - PAGE_SHIFT)) & 0xF) << 20); in create_36bit_mapping()
844 static void __init create_mapping(struct map_desc *md) in create_mapping() argument
[all …]
/linux-4.1.27/drivers/firmware/efi/
Defi.c254 efi_memory_desc_t *md = p; in efi_lookup_mapped_addr() local
255 u64 size = md->num_pages << EFI_PAGE_SHIFT; in efi_lookup_mapped_addr()
256 u64 end = md->phys_addr + size; in efi_lookup_mapped_addr()
257 if (!(md->attribute & EFI_MEMORY_RUNTIME) && in efi_lookup_mapped_addr()
258 md->type != EFI_BOOT_SERVICES_CODE && in efi_lookup_mapped_addr()
259 md->type != EFI_BOOT_SERVICES_DATA) in efi_lookup_mapped_addr()
261 if (!md->virt_addr) in efi_lookup_mapped_addr()
263 if (phys_addr >= md->phys_addr && phys_addr < end) { in efi_lookup_mapped_addr()
264 phys_addr += md->virt_addr - md->phys_addr; in efi_lookup_mapped_addr()
485 const efi_memory_desc_t *md) in efi_md_typeattr_format() argument
[all …]
Druntime-map.c22 efi_memory_desc_t md; member
40 return snprintf(buf, PAGE_SIZE, "0x%x\n", entry->md.type); in type_show()
43 #define EFI_RUNTIME_FIELD(var) entry->md.var
127 memcpy(&entry->md, efi_runtime_map + nr * efi_memdesc_size, in add_sysfs_runtime_map_entry()
/linux-4.1.27/fs/proc/
Dtask_mmu.c1359 struct numa_maps md; member
1362 static void gather_stats(struct page *page, struct numa_maps *md, int pte_dirty, in gather_stats() argument
1367 md->pages += nr_pages; in gather_stats()
1369 md->dirty += nr_pages; in gather_stats()
1372 md->swapcache += nr_pages; in gather_stats()
1375 md->active += nr_pages; in gather_stats()
1378 md->writeback += nr_pages; in gather_stats()
1381 md->anon += nr_pages; in gather_stats()
1383 if (count > md->mapcount_max) in gather_stats()
1384 md->mapcount_max = count; in gather_stats()
[all …]
/linux-4.1.27/drivers/s390/char/
Dtape_3590.c993 sense->fmt.f70.md); in tape_3590_print_mim_msg_f0()
1004 "procedure %i", sense->fmt.f70.md); in tape_3590_print_mim_msg_f0()
1047 "interface 0x%02x", sense->fmt.f71.md[0]); in tape_3590_print_io_sim_msg_f1()
1051 "0x%02x", sense->fmt.f71.md[0]); in tape_3590_print_io_sim_msg_f1()
1055 "0x%02x", sense->fmt.f71.md[0]); in tape_3590_print_io_sim_msg_f1()
1059 sense->fmt.f71.md[0]); in tape_3590_print_io_sim_msg_f1()
1063 "0x%02x", sense->fmt.f71.md[0]); in tape_3590_print_io_sim_msg_f1()
1081 "0x%x on CU", sense->fmt.f71.md[1]); in tape_3590_print_io_sim_msg_f1()
1084 "nodes (0x%x-0x%x) on CU", sense->fmt.f71.md[1], in tape_3590_print_io_sim_msg_f1()
1085 sense->fmt.f71.md[2]); in tape_3590_print_io_sim_msg_f1()
[all …]
Dtape_3590.h96 unsigned int md:8; member
109 unsigned char md[3]; member
/linux-4.1.27/scripts/mod/
Dsumversion.c225 static inline void add_char(unsigned char c, struct md4_ctx *md) in add_char() argument
227 md4_update(md, &c, 1); in add_char()
231 struct md4_ctx *md) in parse_string() argument
235 add_char(file[0], md); in parse_string()
237 add_char(file[i], md); in parse_string()
256 static int parse_file(const char *fname, struct md4_ctx *md) in parse_file() argument
278 i += parse_string(file+i, len - i, md); in parse_file()
288 add_char(file[i], md); in parse_file()
305 static int parse_source_files(const char *objfile, struct md4_ctx *md) in parse_source_files() argument
351 if (!parse_file(p, md)) { in parse_source_files()
[all …]
/linux-4.1.27/block/partitions/
Dmac.c41 struct mac_driver_desc *md; in mac_partition() local
44 md = read_part_sector(state, 0, &sect); in mac_partition()
45 if (!md) in mac_partition()
47 if (be16_to_cpu(md->signature) != MAC_DRIVER_MAGIC) { in mac_partition()
51 secsize = be16_to_cpu(md->block_size); in mac_partition()
/linux-4.1.27/arch/sparc/mm/
Dinit_64.c929 static int scan_pio_for_cfg_handle(struct mdesc_handle *md, u64 pio, in scan_pio_for_cfg_handle() argument
934 mdesc_for_each_arc(arc, md, pio, MDESC_ARC_TYPE_FWD) { in scan_pio_for_cfg_handle()
935 u64 target = mdesc_arc_target(md, arc); in scan_pio_for_cfg_handle()
938 val = mdesc_get_property(md, target, in scan_pio_for_cfg_handle()
946 static int scan_arcs_for_cfg_handle(struct mdesc_handle *md, u64 grp, in scan_arcs_for_cfg_handle() argument
952 mdesc_for_each_arc(arc, md, grp, MDESC_ARC_TYPE_FWD) { in scan_arcs_for_cfg_handle()
953 u64 target = mdesc_arc_target(md, arc); in scan_arcs_for_cfg_handle()
954 const char *name = mdesc_node_name(md, target); in scan_arcs_for_cfg_handle()
960 val = mdesc_get_property(md, target, "latency", NULL); in scan_arcs_for_cfg_handle()
973 return scan_pio_for_cfg_handle(md, candidate, cfg_handle); in scan_arcs_for_cfg_handle()
[all …]
/linux-4.1.27/fs/ncpfs/
Dinode.c493 struct ncp_mount_data* md = (struct ncp_mount_data*)raw_data; in ncp_fill_super() local
495 data.flags = md->flags; in ncp_fill_super()
497 data.mounted_uid = make_kuid(current_user_ns(), md->mounted_uid); in ncp_fill_super()
498 data.wdog_pid = find_get_pid(md->wdog_pid); in ncp_fill_super()
499 data.ncp_fd = md->ncp_fd; in ncp_fill_super()
500 data.time_out = md->time_out; in ncp_fill_super()
501 data.retry_count = md->retry_count; in ncp_fill_super()
502 data.uid = make_kuid(current_user_ns(), md->uid); in ncp_fill_super()
503 data.gid = make_kgid(current_user_ns(), md->gid); in ncp_fill_super()
504 data.file_mode = md->file_mode; in ncp_fill_super()
[all …]
/linux-4.1.27/net/openvswitch/
Dvport-vxlan.c65 struct vxlan_metadata *md) in vxlan_rcv() argument
72 .gbp = md->gbp, in vxlan_rcv()
79 if (vxlan_port->exts & VXLAN_F_GBP && md->gbp) in vxlan_rcv()
84 key = cpu_to_be64(ntohl(md->vni) >> 8); in vxlan_rcv()
228 struct vxlan_metadata md = {0}; in vxlan_tnl_send() local
254 md.vni = htonl(be64_to_cpu(tun_key->tun_id) << 8); in vxlan_tnl_send()
255 md.gbp = vxlan_ext_gbp(skb); in vxlan_tnl_send()
262 &md, false, vxflags); in vxlan_tnl_send()
/linux-4.1.27/arch/ia64/hp/sim/boot/
Dfw-emu.c238 efi_memory_desc_t *efi_memmap, *md; in sys_fw_init() local
247 md = efi_memmap + i++; \ in sys_fw_init()
248 md->type = typ; \ in sys_fw_init()
249 md->pad = 0; \ in sys_fw_init()
250 md->phys_addr = start; \ in sys_fw_init()
251 md->virt_addr = 0; \ in sys_fw_init()
252 md->num_pages = (end - start) >> 12; \ in sys_fw_init()
253 md->attribute = attr; \ in sys_fw_init()
/linux-4.1.27/drivers/media/platform/s3c-camif/
Dcamif-core.c310 struct media_device *md = &camif->media_dev; in camif_media_dev_register() local
315 memset(md, 0, sizeof(*md)); in camif_media_dev_register()
316 snprintf(md->model, sizeof(md->model), "SAMSUNG S3C%s CAMIF", in camif_media_dev_register()
318 strlcpy(md->bus_info, "platform", sizeof(md->bus_info)); in camif_media_dev_register()
319 md->hw_revision = ip_rev; in camif_media_dev_register()
320 md->driver_version = KERNEL_VERSION(1, 0, 0); in camif_media_dev_register()
322 md->dev = camif->dev; in camif_media_dev_register()
325 v4l2_dev->mdev = md; in camif_media_dev_register()
331 ret = media_device_register(md); in camif_media_dev_register()
/linux-4.1.27/drivers/staging/lustre/lustre/mdc/
Dmdc_request.c467 static int mdc_unpack_acl(struct ptlrpc_request *req, struct lustre_md *md) in mdc_unpack_acl() argument
470 struct mdt_body *body = md->body; in mdc_unpack_acl()
500 md->posix_acl = acl; in mdc_unpack_acl()
504 #define mdc_unpack_acl(req, md) 0 argument
509 struct lustre_md *md) in mdc_get_lustre_md() argument
514 LASSERT(md); in mdc_get_lustre_md()
515 memset(md, 0, sizeof(*md)); in mdc_get_lustre_md()
517 md->body = req_capsule_server_get(pill, &RMF_MDT_BODY); in mdc_get_lustre_md()
518 LASSERT(md->body != NULL); in mdc_get_lustre_md()
520 if (md->body->valid & OBD_MD_FLEASIZE) { in mdc_get_lustre_md()
[all …]
Dmdc_internal.h119 struct lustre_md *md);
121 int mdc_free_lustre_md(struct obd_export *exp, struct lustre_md *md);
/linux-4.1.27/drivers/clk/shmobile/
Dclk-r8a7779.c55 #define CPG_CLK_CONFIG_INDEX(md) (((md) & (BIT(2)|BIT(1))) >> 1) argument
83 #define CPG_PLLA_MULT_INDEX(md) (((md) & (BIT(12)|BIT(11))) >> 11) argument
Dclk-rcar-gen2.c267 #define CPG_PLL_CONFIG_INDEX(md) ((((md) & BIT(14)) >> 12) | \ argument
268 (((md) & BIT(13)) >> 12) | \
269 (((md) & BIT(19)) >> 19))
/linux-4.1.27/drivers/staging/lustre/lustre/llite/
Dllite_lib.c152 static int client_common_fill_super(struct super_block *sb, char *md, char *dt, in client_common_fill_super() argument
168 obd = class_name2obd(md); in client_common_fill_super()
170 CERROR("MD %s: not setup or attached\n", md); in client_common_fill_super()
186 dt, md); in client_common_fill_super()
256 md); in client_common_fill_super()
259 CERROR("cannot connect to %s: rc = %d\n", md, err); in client_common_fill_super()
916 char *dt = NULL, *md = NULL; in ll_fill_super() local
986 md = kasprintf(GFP_NOFS, "%s-%p", lprof->lp_md, cfg->cfg_instance); in ll_fill_super()
987 if (!md) { in ll_fill_super()
993 err = client_common_fill_super(sb, md, dt, mnt); in ll_fill_super()
[all …]
Dnamei.c61 struct lustre_md *md = opaque; in ll_test_inode() local
63 if (unlikely(!(md->body->valid & OBD_MD_FLID))) { in ll_test_inode()
68 if (!lu_fid_eq(&lli->lli_fid, &md->body->fid1)) in ll_test_inode()
109 struct lustre_md *md) in ll_iget() argument
114 inode = iget5_locked(sb, hash, ll_test_inode, ll_set_inode, md); in ll_iget()
120 ll_read_inode2(inode, md); in ll_iget()
125 ll_get_fsname(sb, NULL, 0), md->lsm, in ll_iget()
127 rc = cl_file_inode_init(inode, md); in ll_iget()
137 ll_update_inode(inode, md); in ll_iget()
139 inode, PFID(&md->body->fid1)); in ll_iget()
Dfile.c3408 struct lustre_md md = { NULL }; in ll_layout_lock_set() local
3451 rc = obd_unpackmd(sbi->ll_dt_exp, &md.lsm, in ll_layout_lock_set()
3455 if (md.lsm != NULL) in ll_layout_lock_set()
3456 *gen = md.lsm->lsm_layout_gen; in ll_layout_lock_set()
3473 conf.u.coc_md = &md; in ll_layout_lock_set()
3476 if (md.lsm != NULL) in ll_layout_lock_set()
3477 obd_free_memmd(sbi->ll_dt_exp, &md.lsm); in ll_layout_lock_set()
Dllite_internal.h807 void ll_update_inode(struct inode *inode, struct lustre_md *md);
/linux-4.1.27/drivers/block/drbd/
Ddrbd_main.c850 spin_lock_irq(&device->ldev->md.uuid_lock); in _drbd_send_uuids()
852 p->uuid[i] = cpu_to_be64(device->ldev->md.uuid[i]); in _drbd_send_uuids()
853 spin_unlock_irq(&device->ldev->md.uuid_lock); in _drbd_send_uuids()
881 u64 *uuid = device->ldev->md.uuid; in drbd_print_uuids()
905 uuid = device->ldev->md.uuid[UI_BITMAP]; in drbd_gen_and_send_sync_uuid()
3065 buffer->uuid[i] = cpu_to_be64(device->ldev->md.uuid[i]); in drbd_md_write()
3066 buffer->flags = cpu_to_be32(device->ldev->md.flags); in drbd_md_write()
3069 buffer->md_size_sect = cpu_to_be32(device->ldev->md.md_size_sect); in drbd_md_write()
3070 buffer->al_offset = cpu_to_be32(device->ldev->md.al_offset); in drbd_md_write()
3073 buffer->device_uuid = cpu_to_be64(device->ldev->md.device_uuid); in drbd_md_write()
[all …]
Ddrbd_nl.c667 device->ldev->md.uuid[UI_CURRENT] &= ~(u64)1; in drbd_set_role()
680 && device->ldev->md.uuid[UI_BITMAP] == 0) || forced) in drbd_set_role()
683 device->ldev->md.uuid[UI_CURRENT] |= (u64)1; in drbd_set_role()
776 unsigned int al_size_sect = bdev->md.al_size_4k * 8; in drbd_md_set_sector_offsets()
778 bdev->md.md_offset = drbd_md_ss(bdev); in drbd_md_set_sector_offsets()
780 switch (bdev->md.meta_dev_idx) { in drbd_md_set_sector_offsets()
783 bdev->md.md_size_sect = MD_128MB_SECT; in drbd_md_set_sector_offsets()
784 bdev->md.al_offset = MD_4kB_SECT; in drbd_md_set_sector_offsets()
785 bdev->md.bm_offset = MD_4kB_SECT + al_size_sect; in drbd_md_set_sector_offsets()
789 bdev->md.md_size_sect = drbd_get_capacity(bdev->md_bdev); in drbd_md_set_sector_offsets()
[all …]
Ddrbd_actlog.c443 const unsigned int stripes = device->ldev->md.al_stripes; in al_tr_number_to_on_disk_sector()
444 const unsigned int stripe_size_4kB = device->ldev->md.al_stripe_size_4k; in al_tr_number_to_on_disk_sector()
447 unsigned int t = device->al_tr_number % (device->ldev->md.al_size_4k); in al_tr_number_to_on_disk_sector()
456 return device->ldev->md.md_offset + device->ldev->md.al_offset + t; in al_tr_number_to_on_disk_sector()
612 struct drbd_md *md = &device->ldev->md; in drbd_initialize_al() local
613 sector_t al_base = md->md_offset + md->al_offset; in drbd_initialize_al()
614 int al_size_4k = md->al_stripes * md->al_stripe_size_4k; in drbd_initialize_al()
Ddrbd_int.h611 struct drbd_md md; member
1815 switch (bdev->md.meta_dev_idx) { in drbd_md_first_sector()
1818 return bdev->md.md_offset + bdev->md.bm_offset; in drbd_md_first_sector()
1821 return bdev->md.md_offset; in drbd_md_first_sector()
1831 switch (bdev->md.meta_dev_idx) { in drbd_md_last_sector()
1834 return bdev->md.md_offset + MD_4kB_SECT -1; in drbd_md_last_sector()
1837 return bdev->md.md_offset + bdev->md.md_size_sect -1; in drbd_md_last_sector()
1860 switch (bdev->md.meta_dev_idx) { in drbd_get_max_capacity()
1873 BM_EXT_TO_SECT(bdev->md.md_size_sect in drbd_get_max_capacity()
1874 - bdev->md.bm_offset)); in drbd_get_max_capacity()
[all …]
Ddrbd_debugfs.c766 struct drbd_md *md; in device_data_gen_id_show() local
772 md = &device->ldev->md; in device_data_gen_id_show()
773 spin_lock_irq(&md->uuid_lock); in device_data_gen_id_show()
775 seq_printf(m, "0x%016llX\n", md->uuid[idx]); in device_data_gen_id_show()
777 spin_unlock_irq(&md->uuid_lock); in device_data_gen_id_show()
Ddrbd_receiver.c2726 self = device->ldev->md.uuid[UI_BITMAP] & 1; in drbd_asb_recover_0p()
2940 self = device->ldev->md.uuid[UI_CURRENT] & ~((u64)1); in drbd_uuid_compare()
2960 if (device->p_uuid[UI_BITMAP] == (u64)0 && device->ldev->md.uuid[UI_BITMAP] != (u64)0) { in drbd_uuid_compare()
2965 …if ((device->ldev->md.uuid[UI_BITMAP] & ~((u64)1)) == (device->p_uuid[UI_HISTORY_START] & ~((u64)1… in drbd_uuid_compare()
2966 …(device->ldev->md.uuid[UI_HISTORY_START] & ~((u64)1)) == (device->p_uuid[UI_HISTORY_START + 1] & ~… in drbd_uuid_compare()
2969 device->ldev->md.uuid[UI_HISTORY_START] = device->ldev->md.uuid[UI_BITMAP]; in drbd_uuid_compare()
2970 device->ldev->md.uuid[UI_BITMAP] = 0; in drbd_uuid_compare()
2972 drbd_uuid_dump(device, "self", device->ldev->md.uuid, in drbd_uuid_compare()
2983 if (device->ldev->md.uuid[UI_BITMAP] == (u64)0 && device->p_uuid[UI_BITMAP] != (u64)0) { in drbd_uuid_compare()
2988 …if ((device->ldev->md.uuid[UI_HISTORY_START] & ~((u64)1)) == (device->p_uuid[UI_BITMAP] & ~((u64)1… in drbd_uuid_compare()
[all …]
Ddrbd_bitmap.c620 if (ldev->md.al_offset == 8) in drbd_md_on_disk_bits()
621 bitmap_sectors = ldev->md.md_size_sect - ldev->md.bm_offset; in drbd_md_on_disk_bits()
623 bitmap_sectors = ldev->md.al_offset - ldev->md.bm_offset; in drbd_md_on_disk_bits()
1003 device->ldev->md.md_offset + device->ldev->md.bm_offset; in bm_page_io_async()
Ddrbd_state.c809 if (device->ed_uuid == device->ldev->md.uuid[UI_CURRENT]) { in sanitize_state()
1121 u32 mdf = device->ldev->md.flags & ~(MDF_CONSISTENT|MDF_PRIMARY_IND| in __drbd_set_state()
1139 if (mdf != device->ldev->md.flags) { in __drbd_set_state()
1140 device->ldev->md.flags = mdf; in __drbd_set_state()
1144 drbd_set_ed_uuid(device, device->ldev->md.uuid[UI_CURRENT]); in __drbd_set_state()
1367 device->ldev->md.uuid[UI_BITMAP] == 0 && ns.disk >= D_UP_TO_DATE) { in after_state_ch()
1381 device->ldev->md.uuid[UI_BITMAP] == 0 && ns.disk >= D_UP_TO_DATE) { in after_state_ch()
Ddrbd_worker.c953 drbd_uuid_set(device, UI_BITMAP, device->ldev->md.uuid[UI_CURRENT]); in drbd_resync_finished()
970 device->p_uuid[i] = device->ldev->md.uuid[i]; in drbd_resync_finished()
/linux-4.1.27/arch/x86/include/asm/
Defi.h92 extern void __init efi_set_executable(efi_memory_desc_t *md, bool executable);
98 extern void __init efi_map_region(efi_memory_desc_t *md);
99 extern void __init efi_map_region_fixed(efi_memory_desc_t *md);
103 extern void __init old_map_region(efi_memory_desc_t *md);
/linux-4.1.27/Documentation/
Dmd.txt1 Tools that manage md devices can be found at
8 You can boot with your md device with the following kernel command
12 md=<md device no.>,<raid level>,<chunk size factor>,<fault level>,dev0,dev1,...,devn
15 md=<md device no.>,dev0,dev1,...,devn
17 md=d<md device no.>,dev0,dev1,...,devn
19 md device no. = the number of the md device ...
39 e:\loadlin\loadlin e:\zimage root=/dev/md0 md=0,0,4,0,/dev/hdb2,/dev/hdc3 ro
45 When md is compiled into the kernel (not as module), partitions of
63 For this reason, md will normally refuse to start such an array. This
70 array, md supports a module parameter "start_dirty_degraded" which,
[all …]
Dmd-cluster.txt12 | idle | md super | bm super [0] + bits |
53 updated, and the node must re-read the md superblock. This is performed
136 The resync process, is the regular md resync. However, in a clustered
D00-INDEX304 md.txt
Dkernel-parameters.txt1958 md= [HW] RAID subsystems devices and level
1959 See Documentation/md.txt.
2974 See Documentation/md.txt.
/linux-4.1.27/tools/perf/util/
Devlist.c636 struct perf_mmap *md = &evlist->mmap[idx]; in perf_evlist__mmap_read() local
637 u64 head = perf_mmap__read_head(md); in perf_evlist__mmap_read()
638 u64 old = md->prev; in perf_evlist__mmap_read()
639 unsigned char *data = md->base + page_size; in perf_evlist__mmap_read()
652 if (diff > md->mask / 2 || diff < 0) { in perf_evlist__mmap_read()
665 event = (union perf_event *)&data[old & md->mask]; in perf_evlist__mmap_read()
672 if ((old & md->mask) + size != ((old + size) & md->mask)) { in perf_evlist__mmap_read()
675 void *dst = md->event_copy; in perf_evlist__mmap_read()
678 cpy = min(md->mask + 1 - (offset & md->mask), len); in perf_evlist__mmap_read()
679 memcpy(dst, &data[offset & md->mask], cpy); in perf_evlist__mmap_read()
[all …]
Dsymbol.c1080 struct kcore_mapfn_data *md = data; in kcore_mapfn() local
1083 map = map__new2(start, md->dso, md->type); in kcore_mapfn()
1090 list_add(&map->node, &md->maps); in kcore_mapfn()
1100 struct kcore_mapfn_data md; in dso__load_kcore() local
1124 md.dso = dso; in dso__load_kcore()
1125 md.type = map->type; in dso__load_kcore()
1126 INIT_LIST_HEAD(&md.maps); in dso__load_kcore()
1133 err = file__read_maps(fd, md.type == MAP__FUNCTION, kcore_mapfn, &md, in dso__load_kcore()
1139 if (list_empty(&md.maps)) { in dso__load_kcore()
1156 list_for_each_entry(new_map, &md.maps, node) { in dso__load_kcore()
[all …]
Devlist.h200 static inline void perf_mmap__write_tail(struct perf_mmap *md, u64 tail) in perf_mmap__write_tail() argument
202 struct perf_event_mmap_page *pc = md->base; in perf_mmap__write_tail()
/linux-4.1.27/arch/cris/include/arch-v32/arch/hwregs/iop/
Diop_dmc_out_defs.h153 unsigned int md : 16; member
177 unsigned int md : 16; member
184 unsigned int md : 16; member
210 unsigned int md : 16; member
Diop_dmc_in_defs.h163 unsigned int md : 16; member
187 unsigned int md : 16; member
194 unsigned int md : 16; member
220 unsigned int md : 16; member
/linux-4.1.27/tools/perf/
Dbuiltin-record.c72 struct perf_mmap *md = &rec->evlist->mmap[idx]; in record__mmap_read() local
73 u64 head = perf_mmap__read_head(md); in record__mmap_read()
74 u64 old = md->prev; in record__mmap_read()
75 unsigned char *data = md->base + page_size; in record__mmap_read()
87 if ((old & md->mask) + size != (head & md->mask)) { in record__mmap_read()
88 buf = &data[old & md->mask]; in record__mmap_read()
89 size = md->mask + 1 - (old & md->mask); in record__mmap_read()
98 buf = &data[old & md->mask]; in record__mmap_read()
107 md->prev = old; in record__mmap_read()
/linux-4.1.27/drivers/staging/lustre/lnet/selftest/
Drpc.c361 lnet_md_t md; in srpc_post_passive_rdma() local
372 md.threshold = 1; in srpc_post_passive_rdma()
373 md.user_ptr = ev; in srpc_post_passive_rdma()
374 md.start = buf; in srpc_post_passive_rdma()
375 md.length = len; in srpc_post_passive_rdma()
376 md.options = options; in srpc_post_passive_rdma()
377 md.eq_handle = srpc_data.rpc_lnet_eq; in srpc_post_passive_rdma()
379 rc = LNetMDAttach(meh, md, LNET_UNLINK, mdh); in srpc_post_passive_rdma()
401 lnet_md_t md; in srpc_post_active_rdma() local
403 md.user_ptr = ev; in srpc_post_active_rdma()
[all …]
/linux-4.1.27/net/ipv4/
Dtcp_highspeed.c17 unsigned int md; member
156 return max(tp->snd_cwnd - ((tp->snd_cwnd * hstcp_aimd_vals[ca->ai].md) >> 8), 2U); in hstcp_ssthresh()
/linux-4.1.27/arch/cris/include/arch-v32/arch/hwregs/
Ddma.h27 unsigned md : 16; member
68 unsigned md : 16; member
Dsser_defs.h182 unsigned int md : 1; member
191 unsigned int md : 1; member
Ddma_defs.h125 unsigned int md : 16; member
277 unsigned int md : 16; member
/linux-4.1.27/Documentation/device-mapper/
Ddm-log.txt21 disk drivers/md/dm-log.c
22 core drivers/md/dm-log.c
23 userspace drivers/md/dm-log-userspace* include/linux/dm-log-userspace.h
Dpersistent-data.txt25 under drivers/md/persistent-data.
/linux-4.1.27/drivers/firmware/efi/libstub/
Defi-stub-helper.c119 efi_memory_desc_t *md; in get_dram_base() local
128 for_each_efi_memory_desc(&map, md) in get_dram_base()
129 if (md->attribute & EFI_MEMORY_WB) in get_dram_base()
130 if (membase > md->phys_addr) in get_dram_base()
131 membase = md->phys_addr; in get_dram_base()
/linux-4.1.27/include/net/
Dvxlan.h104 struct vxlan_metadata *md);
150 __be16 src_port, __be16 dst_port, struct vxlan_metadata *md,
Dcfg80211.h2107 u16 md; member
/linux-4.1.27/drivers/net/
Dvxlan.c1170 struct vxlan_metadata md = {0}; in vxlan_udp_encap_recv() local
1212 md.gbp = ntohs(gbp->policy_id); in vxlan_udp_encap_recv()
1215 md.gbp |= VXLAN_GBP_DONT_LEARN; in vxlan_udp_encap_recv()
1218 md.gbp |= VXLAN_GBP_POLICY_APPLIED; in vxlan_udp_encap_recv()
1236 md.vni = vxh->vx_vni; in vxlan_udp_encap_recv()
1237 vs->rcv(vs, skb, &md); in vxlan_udp_encap_recv()
1255 struct vxlan_metadata *md) in vxlan_rcv() argument
1266 vni = ntohl(md->vni) >> 8; in vxlan_rcv()
1300 skb->mark = md->gbp; in vxlan_rcv()
1651 struct vxlan_metadata *md) in vxlan_build_gbp_hdr() argument
[all …]
/linux-4.1.27/drivers/usb/gadget/function/
Df_uvc.c795 struct uvc_color_matching_descriptor *md; in uvc_alloc_inst() local
842 md = &opts->uvc_color_matching; in uvc_alloc_inst()
843 md->bLength = UVC_DT_COLOR_MATCHING_SIZE; in uvc_alloc_inst()
844 md->bDescriptorType = USB_DT_CS_INTERFACE; in uvc_alloc_inst()
845 md->bDescriptorSubType = UVC_VS_COLORFORMAT; in uvc_alloc_inst()
846 md->bColorPrimaries = 1; in uvc_alloc_inst()
847 md->bTransferCharacteristics = 1; in uvc_alloc_inst()
848 md->bMatrixCoefficients = 4; in uvc_alloc_inst()
/linux-4.1.27/arch/m68k/ifpsp060/src/
Dilsp.S548 mov.l %d1,%d4 # md in d4
550 swap %d4 # hi(md) in lo d4
553 mulu.w %d1,%d0 # [1] lo(mr) * lo(md)
554 mulu.w %d3,%d1 # [2] hi(mr) * lo(md)
555 mulu.w %d4,%d2 # [3] lo(mr) * hi(md)
556 mulu.w %d4,%d3 # [4] hi(mr) * hi(md)
674 mov.l %d1,%d4 # md in d4
676 swap %d4 # hi(md) in lo d4
679 mulu.w %d1,%d0 # [1] lo(mr) * lo(md)
680 mulu.w %d3,%d1 # [2] hi(mr) * lo(md)
[all …]
/linux-4.1.27/Documentation/video4linux/
Dfimc.txt101 API the driver creates a sysfs entry associated with "s5p-fimc-md" platform
102 device. The entry path is: /sys/platform/devices/s5p-fimc-md/subdev_conf_mode.
118 # echo "sub-dev" > /sys/platform/devices/s5p-fimc-md/subdev_conf_mode
122 # echo "vid-dev" > /sys/platform/devices/s5p-fimc-md/subdev_conf_mode
/linux-4.1.27/drivers/scsi/
Dmesh.c305 volatile struct dbdma_regs __iomem *md = ms->dma; in mesh_dump_regs() local
310 ms, mr, md); in mesh_dump_regs()
320 in_le32(&md->status), in_le32(&md->cmdptr)); in mesh_dump_regs()
360 volatile struct dbdma_regs __iomem *md = ms->dma; in mesh_init() local
366 out_le32(&md->control, (RUN|PAUSE|FLUSH|WAKE) << 16); /* stop dma */ in mesh_init()
682 volatile struct dbdma_regs __iomem *md = ms->dma; in start_phase() local
786 out_le32(&md->cmdptr, virt_to_phys(ms->dma_cmds)); in start_phase()
787 out_le32(&md->control, (RUN << 16) | RUN); in start_phase()
1320 volatile struct dbdma_regs __iomem *md = ms->dma; in halt_dma() local
1329 && (in_le32(&md->status) & ACTIVE) != 0) { in halt_dma()
[all …]
Dsg.c1670 struct rq_map_data *md, map_data; in sg_start_req() local
1724 md = NULL; in sg_start_req()
1726 md = &map_data; in sg_start_req()
1728 if (md) { in sg_start_req()
1737 md->pages = req_schp->pages; in sg_start_req()
1738 md->page_order = req_schp->page_order; in sg_start_req()
1739 md->nr_entries = req_schp->k_use_sg; in sg_start_req()
1740 md->offset = 0; in sg_start_req()
1741 md->null_mapped = hp->dxferp ? 0 : 1; in sg_start_req()
1743 md->from_user = 1; in sg_start_req()
[all …]
/linux-4.1.27/drivers/video/fbdev/core/
Dmodedb.c1181 const struct fb_videomode *m, *m1 = NULL, *md = NULL, *best = NULL; in fb_find_best_display() local
1198 md = m; in fb_find_best_display()
1205 best = md; in fb_find_best_display()
1224 if (md) { in fb_find_best_display()
1225 best = md; in fb_find_best_display()
/linux-4.1.27/drivers/crypto/
Dhifn_795x.c1196 u16 md = 0; in hifn_setup_cmd_desc() local
1199 md |= HIFN_CRYPT_CMD_NEW_KEY; in hifn_setup_cmd_desc()
1201 md |= HIFN_CRYPT_CMD_NEW_IV; in hifn_setup_cmd_desc()
1205 md |= HIFN_CRYPT_CMD_MODE_ECB; in hifn_setup_cmd_desc()
1208 md |= HIFN_CRYPT_CMD_MODE_CBC; in hifn_setup_cmd_desc()
1211 md |= HIFN_CRYPT_CMD_MODE_CFB; in hifn_setup_cmd_desc()
1214 md |= HIFN_CRYPT_CMD_MODE_OFB; in hifn_setup_cmd_desc()
1224 md |= HIFN_CRYPT_CMD_KSZ_128 | in hifn_setup_cmd_desc()
1230 md |= HIFN_CRYPT_CMD_KSZ_192 | in hifn_setup_cmd_desc()
1236 md |= HIFN_CRYPT_CMD_KSZ_256 | in hifn_setup_cmd_desc()
[all …]
/linux-4.1.27/drivers/staging/lustre/lustre/lov/
Dlov_obd.c2190 void lov_stripe_lock(struct lov_stripe_md *md) in lov_stripe_lock() argument
2191 __acquires(&md->lsm_lock) in lov_stripe_lock()
2193 LASSERT(md->lsm_lock_owner != current_pid()); in lov_stripe_lock()
2194 spin_lock(&md->lsm_lock); in lov_stripe_lock()
2195 LASSERT(md->lsm_lock_owner == 0); in lov_stripe_lock()
2196 md->lsm_lock_owner = current_pid(); in lov_stripe_lock()
2200 void lov_stripe_unlock(struct lov_stripe_md *md) in lov_stripe_unlock() argument
2201 __releases(&md->lsm_lock) in lov_stripe_unlock()
2203 LASSERT(md->lsm_lock_owner == current_pid()); in lov_stripe_unlock()
2204 md->lsm_lock_owner = 0; in lov_stripe_unlock()
[all …]
/linux-4.1.27/drivers/iio/imu/
Dadis16480.c156 u16 md, year; in adis16480_show_firmware_date() local
165 ret = adis_read_reg_16(&adis16480->adis, ADIS16480_REG_FIRM_DM, &md); in adis16480_show_firmware_date()
170 md >> 8, md & 0xff, year); in adis16480_show_firmware_date()
/linux-4.1.27/lib/
Dinflate.c601 unsigned ml, md; /* masks for bl and bd bits */ in inflate_codes() local
613 md = mask_bits[bd]; in inflate_codes()
649 if ((e = (t = td + ((unsigned)b & md))->e) > 16) in inflate_codes()
/linux-4.1.27/drivers/clk/st/
Dclkgen-fsyn.c698 u32 md; member
728 CLKGEN_WRITE(fs, mdiv[fs->chan], fs->md); in quadfs_fsynth_program_rate()
882 fs->md = params->mdiv; in quadfs_fsynt_get_hw_value_for_recalc()
975 fs->md = params->mdiv; in quadfs_program_and_enable()
/linux-4.1.27/drivers/staging/lustre/lustre/lclient/
Dlcommon_cl.c1089 int cl_file_inode_init(struct inode *inode, struct lustre_md *md) in cl_file_inode_init() argument
1099 .coc_md = md in cl_file_inode_init()
1105 LASSERT(md->body->valid & OBD_MD_FLID); in cl_file_inode_init()
1132 lli->lli_has_smd = lsm_has_objects(md->lsm); in cl_file_inode_init()
/linux-4.1.27/arch/cris/arch-v32/drivers/
Dcryptocop.c428 cdesc->dma_descr->md = REG_TYPE_CONV(unsigned short int, struct strcop_meta_out, mo); in create_pad_descriptor()
487 key_desc->dma_descr->md = REG_TYPE_CONV(unsigned short int, struct strcop_meta_out, mo); in setup_key_dl_desc()
518 iv_desc->dma_descr->md = REG_TYPE_CONV(unsigned short int, struct strcop_meta_out, mo); in setup_cipher_iv_desc()
580 outdesc->dma_descr->md = REG_TYPE_CONV(unsigned short int, struct strcop_meta_in, mi); in create_input_descriptors()
596 outdesc->dma_descr->md = REG_TYPE_CONV(unsigned short int, struct strcop_meta_in, mi); in create_input_descriptors()
642 cdesc->dma_descr->md = REG_TYPE_CONV(unsigned short int, struct strcop_meta_out, (*meta_out)); in create_output_descriptors()
683 … (*current_in_cdesc)->dma_descr->md = REG_TYPE_CONV(unsigned short int, struct strcop_meta_in, mi); in append_input_descriptors()
687 … (*current_in_cdesc)->dma_descr->md = REG_TYPE_CONV(unsigned short int, struct strcop_meta_in, mi); in append_input_descriptors()
1242 ed->dma_descr->md = REG_TYPE_CONV(unsigned short int, struct strcop_meta_out, ed_mo); in cryptocop_setup_dma_list()
3193 td->md, in print_dma_descriptors()
[all …]
/linux-4.1.27/drivers/
DMakefile111 obj-$(CONFIG_MD) += md/
DKconfig31 source "drivers/md/Kconfig"
/linux-4.1.27/sound/oss/dmasound/
Ddmasound_core.c926 static int shared_resources_are_mine(fmode_t md) in shared_resources_are_mine() argument
929 return (shared_resource_owner & md) != 0; in shared_resources_are_mine()
931 shared_resource_owner = md ; in shared_resources_are_mine()
/linux-4.1.27/drivers/staging/lustre/lustre/obdecho/
Decho_client.c970 struct lustre_md *md; in cl_echo_object_find() local
972 md = &info->eti_md; in cl_echo_object_find()
973 memset(md, 0, sizeof(*md)); in cl_echo_object_find()
974 md->lsm = lsm; in cl_echo_object_find()
975 conf->eoc_cl.u.coc_md = md; in cl_echo_object_find()
/linux-4.1.27/drivers/media/tuners/
Dmt2063.c723 u32 ma, mb, mc, md, me, mf; in IsSpurInBand() local
747 md = (n * ((f_LO1 + hgds) / gd_Scale) - in IsSpurInBand()
751 if (md >= pAS_Info->maxH1) in IsSpurInBand()
758 if (md == ma) in IsSpurInBand()
763 if (mc != md) { in IsSpurInBand()
/linux-4.1.27/drivers/staging/rtl8723au/hal/
Drtl8723au_recv.c175 pattrib->mdata = (u8)prxreport->md; in update_recvframe_attrib()
/linux-4.1.27/Documentation/crypto/
Dasync-tx-api.txt37 xor-parity-calculations of the md-raid5 driver using the offload engines
164 ops_run_* and ops_complete_* routines in drivers/md/raid5.c for more
/linux-4.1.27/drivers/net/wireless/
Dmac80211_hwsim.c899 struct mac80211_hwsim_addr_match_data *md = data; in mac80211_hwsim_addr_iter() local
901 if (memcmp(mac, md->addr, ETH_ALEN) == 0) in mac80211_hwsim_addr_iter()
902 md->ret = true; in mac80211_hwsim_addr_iter()
908 struct mac80211_hwsim_addr_match_data md = { in mac80211_hwsim_addr_match() local
915 memcpy(md.addr, addr, ETH_ALEN); in mac80211_hwsim_addr_match()
920 &md); in mac80211_hwsim_addr_match()
922 return md.ret; in mac80211_hwsim_addr_match()
/linux-4.1.27/drivers/staging/rtl8723au/include/
Drtl8723a_hal.h435 u32 md:1; member
/linux-4.1.27/drivers/staging/lustre/lustre/include/
Dlclient.h379 int cl_file_inode_init(struct inode *inode, struct lustre_md *md);
Dobd_class.h1690 struct lustre_md *md) in md_get_lustre_md() argument
1694 return MDP(exp->exp_obd, get_lustre_md)(exp, req, dt_exp, md_exp, md); in md_get_lustre_md()
1698 struct lustre_md *md) in md_free_lustre_md() argument
1702 return MDP(exp->exp_obd, free_lustre_md)(exp, md); in md_free_lustre_md()
Dobd.h240 void lov_stripe_lock(struct lov_stripe_md *md);
241 void lov_stripe_unlock(struct lov_stripe_md *md);
/linux-4.1.27/drivers/staging/lustre/lustre/lmv/
Dlmv_obd.c2601 struct lustre_md *md) in lmv_get_lustre_md() argument
2605 return md_get_lustre_md(lmv->tgts[0]->ltd_exp, req, dt_exp, md_exp, md); in lmv_get_lustre_md()
2608 static int lmv_free_lustre_md(struct obd_export *exp, struct lustre_md *md) in lmv_free_lustre_md() argument
2613 if (md->mea) in lmv_free_lustre_md()
2614 obd_free_memmd(exp, (void *)&md->mea); in lmv_free_lustre_md()
2615 return md_free_lustre_md(lmv->tgts[0]->ltd_exp, md); in lmv_free_lustre_md()
/linux-4.1.27/Documentation/block/
Dbiovecs.txt82 occasionally in stacking block drivers and various code (e.g. md and
Dbiodoc.txt403 through layered drivers (lvm, md, evms), with minimal overhead.
638 in lvm or md.
/linux-4.1.27/Documentation/filesystems/
Dntfs.txt427 raid0run /dev/md0 to start a particular md device, in this case /dev/md0.
432 It is advisable to do the mount read-only to see if the md volume has been
440 Using the md driver will not work properly if any of your NTFS partitions have
Dcoda.txt921 44..1133.. rrmmddiirr
/linux-4.1.27/Documentation/fault-injection/
Dfault-injection.txt4 See also drivers/md/faulty.c and "every_nth" module option for scsi_debug.
/linux-4.1.27/net/wireless/
Dtrace.h1877 __field(u16, md)
1883 __entry->md = ftie->md;
1887 WIPHY_PR_ARG, NETDEV_PR_ARG, __entry->md)
Dnl80211.c9821 ft_params.md = nla_get_u16(info->attrs[NL80211_ATTR_MDID]); in nl80211_update_ft_ies()
/linux-4.1.27/drivers/scsi/ibmvscsi/
Dibmvscsi.c665 struct srp_direct_buf *md) in map_sg_list() argument
672 struct srp_direct_buf *descr = md + i; in map_sg_list()
Dibmvfc.c1307 struct srp_direct_buf *md) in ibmvfc_map_sg_list() argument
1313 md[i].va = cpu_to_be64(sg_dma_address(sg)); in ibmvfc_map_sg_list()
1314 md[i].len = cpu_to_be32(sg_dma_len(sg)); in ibmvfc_map_sg_list()
1315 md[i].key = 0; in ibmvfc_map_sg_list()
/linux-4.1.27/drivers/dma/
Dimx-sdma.c210 u32 md; member
/linux-4.1.27/
DMAINTAINERS1974 F: drivers/md/bcache/
2864 M: Andy Walls <awalls@md.metrocast.net>
3156 F: drivers/md/dm*
3157 F: drivers/md/persistent-data/
5497 M: Andy Walls <awalls@md.metrocast.net>
9202 F: drivers/md/
DCREDITS1624 D: md boot support