vn 158 drivers/gpu/drm/i915/gem/i915_gem_object.c struct i915_vma *vma, *vn; vn 164 drivers/gpu/drm/i915/gem/i915_gem_object.c list_for_each_entry_safe(vma, vn, &obj->vma.list, obj_link) { vn 509 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_vma *vma, *vn; vn 511 drivers/gpu/drm/i915/i915_gem_gtt.c list_for_each_entry_safe(vma, vn, *phase, vm_link) vn 2752 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_vma *vma, *vn; vn 2761 drivers/gpu/drm/i915/i915_gem_gtt.c list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link) vn 3304 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_vma *vma, *vn; vn 3316 drivers/gpu/drm/i915/i915_gem_gtt.c list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link) { vn 149 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c struct nvif_mem_ram_vn vn; vn 186 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c if ( (ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c struct gf100_mem_map_vn vn; vn 50 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 73 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c struct gf100_mem_vn vn; vn 82 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 35 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c struct nv04_mem_map_vn vn; vn 41 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) vn 55 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c struct nv04_mem_vn vn; vn 59 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv04.c if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) vn 39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c struct nv50_mem_map_vn vn; vn 52 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 70 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c struct nv50_mem_vn vn; vn 80 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 245 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c struct gf100_vmm_map_vn vn; vn 263 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 150 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c struct gm200_vmm_vn vn; vn 163 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 318 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c struct gp100_vmm_map_vn vn; vn 336 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 429 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c struct gp100_vmm_fault_replay_vn vn; vn 433 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 514 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c struct gp100_vmm_vn vn; vn 523 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 84 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c struct nv04_vmm_map_vn vn; vn 87 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) vn 108 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c struct nv04_vmm_vn vn; vn 117 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c return nvif_unvers(-ENOSYS, &argv, &argc, args->vn); vn 232 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c struct nv50_vmm_map_vn vn; vn 251 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { vn 79 drivers/gpu/drm/sun4i/sun8i_vi_layer.c u32 vn = 0, vm = 0; vn 156 drivers/gpu/drm/sun4i/sun8i_vi_layer.c vn = (u32)ability * dst_h / 100; vn 157 drivers/gpu/drm/sun4i/sun8i_vi_layer.c src_h = vn; vn 192 drivers/gpu/drm/sun4i/sun8i_vi_layer.c SUN8I_MIXER_CHAN_VI_DS_N(vn) | vn 196 drivers/gpu/drm/sun4i/sun8i_vi_layer.c SUN8I_MIXER_CHAN_VI_DS_N(vn) | vn 1367 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define BP_FW_MB_IDX_VN(bp, vn) (BP_PORT(bp) +\ vn 1368 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h (vn) * ((CHIP_IS_E1x(bp) || (CHIP_MODE_IS_4_PORT(bp))) ? 2 : 1)) vn 906 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h static inline int func_by_vn(struct bnx2x *bp, int vn) vn 908 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h return 2 * vn + BP_PORT(bp); vn 1329 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h int vn; vn 1332 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) { vn 1333 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h if (vn == BP_VN(bp)) vn 1336 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h func = func_by_vn(bp, vn); vn 2446 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int vn; vn 2448 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) { vn 2449 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c u32 vn_cfg = bp->mf_config[vn]; vn 2462 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c input->vnic_min_rate[vn] = vn_min_rate; vn 2480 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c static void bnx2x_calc_vn_max(struct bnx2x *bp, int vn, vn 2484 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c u32 vn_cfg = bp->mf_config[vn]; vn 2499 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c DP(NETIF_MSG_IFUP, "vn %d: vn_max_rate %d\n", vn, vn_max_rate); vn 2501 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c input->vnic_max_rate[vn] = vn_max_rate; vn 2516 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int vn, n = (CHIP_MODE_IS_4_PORT(bp) ? 2 : 1); vn 2532 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) { vn 2533 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int /*abs*/func = n * (2 * vn + BP_PORT(bp)) + BP_PATH(bp); vn 2538 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->mf_config[vn] = vn 2558 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int vn; vn 2569 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) vn 2570 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bnx2x_calc_vn_max(bp, vn, &input); vn 2589 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int vn; vn 2597 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) { vn 2598 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int func = func_by_vn(bp, vn); vn 2604 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c (u32 *)&cmng->vnic.vnic_max_rate[vn]); vn 2610 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c (u32 *)&cmng->vnic.vnic_min_rate[vn]); vn 11200 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int vn = BP_VN(bp); vn 11202 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->igu_base_sb = (CHIP_MODE_IS_4_PORT(bp) ? pfid : vn) * vn 11206 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c (CHIP_MODE_IS_4_PORT(bp) ? pfid : vn); vn 11989 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int vn; vn 12077 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c vn = BP_VN(bp); vn 12115 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->mf_config[vn] = MF_CFG_RD(bp, vn 12129 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->mf_config[vn] = MF_CFG_RD(bp, vn 12137 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->mf_config[vn] = vn 12161 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->mf_config[vn] = vn 12166 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->mf_config[vn] = 0; vn 12180 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->mf_config[vn] = 0; vn 12187 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->mf_config[vn] = 0; vn 12226 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (vn) { vn 12229 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c vn); vn 167 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(net, vxlan_net_id); vn 169 drivers/net/vxlan.c return &vn->sock_list[hash_32(ntohs(port), PORT_HASH_BITS)]; vn 1344 drivers/net/vxlan.c static bool vxlan_group_used(struct vxlan_net *vn, struct vxlan_dev *dev) vn 1366 drivers/net/vxlan.c list_for_each_entry(vxlan, &vn->vxlan_list, next) { vn 1395 drivers/net/vxlan.c struct vxlan_net *vn; vn 1402 drivers/net/vxlan.c vn = net_generic(sock_net(vs->sock->sk), vxlan_net_id); vn 1403 drivers/net/vxlan.c spin_lock(&vn->sock_lock); vn 1409 drivers/net/vxlan.c spin_unlock(&vn->sock_lock); vn 2761 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); vn 2763 drivers/net/vxlan.c spin_lock(&vn->sock_lock); vn 2768 drivers/net/vxlan.c spin_unlock(&vn->sock_lock); vn 2774 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); vn 2778 drivers/net/vxlan.c spin_lock(&vn->sock_lock); vn 2780 drivers/net/vxlan.c spin_unlock(&vn->sock_lock); vn 2877 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); vn 2881 drivers/net/vxlan.c !vxlan_group_used(vn, vxlan)) vn 3004 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(net, vxlan_net_id); vn 3007 drivers/net/vxlan.c spin_lock(&vn->sock_lock); vn 3009 drivers/net/vxlan.c hlist_for_each_entry_rcu(vs, &vn->sock_list[i], hlist) { vn 3023 drivers/net/vxlan.c spin_unlock(&vn->sock_lock); vn 3227 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(net, vxlan_net_id); vn 3250 drivers/net/vxlan.c spin_lock(&vn->sock_lock); vn 3256 drivers/net/vxlan.c spin_unlock(&vn->sock_lock); vn 3275 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); vn 3285 drivers/net/vxlan.c spin_lock(&vn->sock_lock); vn 3290 drivers/net/vxlan.c spin_unlock(&vn->sock_lock); vn 3293 drivers/net/vxlan.c spin_unlock(&vn->sock_lock); vn 3343 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(src_net, vxlan_net_id); vn 3476 drivers/net/vxlan.c list_for_each_entry(tmp, &vn->vxlan_list, next) { vn 3578 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(net, vxlan_net_id); vn 3640 drivers/net/vxlan.c list_add(&vxlan->next, &vn->vxlan_list); vn 4226 drivers/net/vxlan.c static void vxlan_handle_lowerdev_unregister(struct vxlan_net *vn, vn 4232 drivers/net/vxlan.c list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) { vn 4252 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(dev_net(dev), vxlan_net_id); vn 4256 drivers/net/vxlan.c vxlan_handle_lowerdev_unregister(vn, dev); vn 4399 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(net, vxlan_net_id); vn 4402 drivers/net/vxlan.c INIT_LIST_HEAD(&vn->vxlan_list); vn 4403 drivers/net/vxlan.c spin_lock_init(&vn->sock_lock); vn 4406 drivers/net/vxlan.c INIT_HLIST_HEAD(&vn->sock_list[h]); vn 4413 drivers/net/vxlan.c struct vxlan_net *vn = net_generic(net, vxlan_net_id); vn 4422 drivers/net/vxlan.c list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) { vn 4431 drivers/net/vxlan.c WARN_ON_ONCE(!hlist_empty(&vn->sock_list[h])); vn 538 drivers/scsi/fcoe/fcoe_ctlr.c struct fip_vn_desc *vn; vn 548 drivers/scsi/fcoe/fcoe_ctlr.c len = sizeof(*kal) + ports * sizeof(*vn); vn 563 drivers/scsi/fcoe/fcoe_ctlr.c ports * sizeof(*vn)) / FIP_BPW); vn 572 drivers/scsi/fcoe/fcoe_ctlr.c vn = (struct fip_vn_desc *)(kal + 1); vn 573 drivers/scsi/fcoe/fcoe_ctlr.c vn->fd_desc.fip_dtype = FIP_DT_VN_ID; vn 574 drivers/scsi/fcoe/fcoe_ctlr.c vn->fd_desc.fip_dlen = sizeof(*vn) / FIP_BPW; vn 575 drivers/scsi/fcoe/fcoe_ctlr.c memcpy(vn->fd_mac, fip->get_src_addr(lport), ETH_ALEN); vn 576 drivers/scsi/fcoe/fcoe_ctlr.c hton24(vn->fd_fc_id, lport->port_id); vn 577 drivers/scsi/fcoe/fcoe_ctlr.c put_unaligned_be64(lport->wwpn, &vn->fd_wwpn); vn 2028 drivers/scsi/fcoe/fcoe_ctlr.c struct fip_vn_desc vn; vn 2043 drivers/scsi/fcoe/fcoe_ctlr.c dlen += sizeof(frame->mac) + sizeof(frame->wwnn) + sizeof(frame->vn); vn 2075 drivers/scsi/fcoe/fcoe_ctlr.c frame->vn.fd_desc.fip_dtype = FIP_DT_VN_ID; vn 2076 drivers/scsi/fcoe/fcoe_ctlr.c frame->vn.fd_desc.fip_dlen = sizeof(frame->vn) / FIP_BPW; vn 2077 drivers/scsi/fcoe/fcoe_ctlr.c hton24(frame->vn.fd_mac, FIP_VN_FC_MAP); vn 2078 drivers/scsi/fcoe/fcoe_ctlr.c hton24(frame->vn.fd_mac + 3, fip->port_id); vn 2079 drivers/scsi/fcoe/fcoe_ctlr.c hton24(frame->vn.fd_fc_id, fip->port_id); vn 2080 drivers/scsi/fcoe/fcoe_ctlr.c put_unaligned_be64(fip->lp->wwpn, &frame->vn.fd_wwpn); vn 2278 drivers/scsi/fcoe/fcoe_ctlr.c struct fip_vn_desc *vn = NULL; vn 2354 drivers/scsi/fcoe/fcoe_ctlr.c vn = (struct fip_vn_desc *)desc; vn 2355 drivers/scsi/fcoe/fcoe_ctlr.c memcpy(frport->vn_mac, vn->fd_mac, ETH_ALEN); vn 2356 drivers/scsi/fcoe/fcoe_ctlr.c frport->rdata.ids.port_id = ntoh24(vn->fd_fc_id); vn 2358 drivers/scsi/fcoe/fcoe_ctlr.c get_unaligned_be64(&vn->fd_wwpn); vn 2849 drivers/usb/gadget/function/f_mass_storage.c void fsg_common_set_inquiry_string(struct fsg_common *common, const char *vn, vn 2857 drivers/usb/gadget/function/f_mass_storage.c "%-8s%-16s%04x", vn ?: "Linux", vn 137 drivers/usb/gadget/function/f_mass_storage.h void fsg_common_set_inquiry_string(struct fsg_common *common, const char *vn, vn 54 fs/reiserfs/fix_node.c struct virtual_node *vn = tb->tb_vn; vn 61 fs/reiserfs/fix_node.c vn->vn_size = vn 66 fs/reiserfs/fix_node.c vn->vn_nr_item = (vn->vn_size - DC_SIZE) / (DC_SIZE + KEY_SIZE); vn 71 fs/reiserfs/fix_node.c vn->vn_nr_item = vn 72 fs/reiserfs/fix_node.c B_NR_ITEMS(Sh) + ((vn->vn_mode == M_INSERT) ? 1 : 0) - vn 73 fs/reiserfs/fix_node.c ((vn->vn_mode == M_DELETE) ? 1 : 0); vn 76 fs/reiserfs/fix_node.c vn->vn_vi = (struct virtual_item *)(tb->tb_vn + 1); vn 77 fs/reiserfs/fix_node.c memset(vn->vn_vi, 0, vn->vn_nr_item * sizeof(struct virtual_item)); vn 78 fs/reiserfs/fix_node.c vn->vn_free_ptr += vn->vn_nr_item * sizeof(struct virtual_item); vn 85 fs/reiserfs/fix_node.c && (vn->vn_mode != M_DELETE || vn->vn_affected_item_num)) vn 86 fs/reiserfs/fix_node.c vn->vn_vi[0].vi_type |= VI_TYPE_LEFT_MERGEABLE; vn 92 fs/reiserfs/fix_node.c for (new_num = 0; new_num < vn->vn_nr_item; new_num++) { vn 94 fs/reiserfs/fix_node.c struct virtual_item *vi = vn->vn_vi + new_num; vn 96 fs/reiserfs/fix_node.c ((new_num != vn->vn_affected_item_num) ? 0 : 1); vn 98 fs/reiserfs/fix_node.c if (is_affected && vn->vn_mode == M_INSERT) vn 102 fs/reiserfs/fix_node.c j = old_item_num(new_num, vn->vn_affected_item_num, vn 103 fs/reiserfs/fix_node.c vn->vn_mode); vn 108 fs/reiserfs/fix_node.c vi->vi_uarea = vn->vn_free_ptr; vn 114 fs/reiserfs/fix_node.c vn->vn_free_ptr += vn 115 fs/reiserfs/fix_node.c op_create_vi(vn, vi, is_affected, tb->insert_size[0]); vn 116 fs/reiserfs/fix_node.c if (tb->vn_buf + tb->vn_buf_size < vn->vn_free_ptr) vn 124 fs/reiserfs/fix_node.c if (vn->vn_mode == M_PASTE || vn->vn_mode == M_CUT) { vn 125 fs/reiserfs/fix_node.c vn->vn_vi[new_num].vi_item_len += tb->insert_size[0]; vn 127 fs/reiserfs/fix_node.c vi->vi_new_data = vn->vn_data; vn 132 fs/reiserfs/fix_node.c if (vn->vn_mode == M_INSERT) { vn 133 fs/reiserfs/fix_node.c struct virtual_item *vi = vn->vn_vi + vn->vn_affected_item_num; vn 135 fs/reiserfs/fix_node.c RFALSE(vn->vn_ins_ih == NULL, vn 138 fs/reiserfs/fix_node.c vi->vi_ih = vn->vn_ins_ih; vn 139 fs/reiserfs/fix_node.c vi->vi_item = vn->vn_data; vn 140 fs/reiserfs/fix_node.c vi->vi_uarea = vn->vn_free_ptr; vn 142 fs/reiserfs/fix_node.c op_create_vi(vn, vi, 0 /*not pasted or cut */ , vn 155 fs/reiserfs/fix_node.c && (vn->vn_mode != M_DELETE vn 156 fs/reiserfs/fix_node.c || vn->vn_affected_item_num != B_NR_ITEMS(Sh) - 1)) vn 157 fs/reiserfs/fix_node.c vn->vn_vi[vn->vn_nr_item - 1].vi_type |= vn 162 fs/reiserfs/fix_node.c !(vn->vn_mode != M_DELETE vn 163 fs/reiserfs/fix_node.c || vn->vn_affected_item_num != B_NR_ITEMS(Sh) - 1)) { vn 181 fs/reiserfs/fix_node.c key, vn->vn_affected_item_num, vn 182 fs/reiserfs/fix_node.c vn->vn_mode, M_DELETE); vn 197 fs/reiserfs/fix_node.c struct virtual_node *vn = tb->tb_vn; vn 211 fs/reiserfs/fix_node.c if (!cur_free || !vn->vn_nr_item) { vn 221 fs/reiserfs/fix_node.c vi = vn->vn_vi; vn 223 fs/reiserfs/fix_node.c (vn->vn_size - vn 227 fs/reiserfs/fix_node.c RFALSE(vn->vn_mode == M_INSERT || vn->vn_mode == M_PASTE, vn 230 fs/reiserfs/fix_node.c tb->lnum[0] = vn->vn_nr_item; vn 242 fs/reiserfs/fix_node.c for (i = 0; i < vn->vn_nr_item; vn 283 fs/reiserfs/fix_node.c struct virtual_node *vn = tb->tb_vn; vn 297 fs/reiserfs/fix_node.c if (!cur_free || !vn->vn_nr_item) { vn 307 fs/reiserfs/fix_node.c vi = vn->vn_vi + vn->vn_nr_item - 1; vn 309 fs/reiserfs/fix_node.c (vn->vn_size - vn 313 fs/reiserfs/fix_node.c RFALSE(vn->vn_mode == M_INSERT || vn->vn_mode == M_PASTE, vn 316 fs/reiserfs/fix_node.c tb->rnum[h] = vn->vn_nr_item; vn 328 fs/reiserfs/fix_node.c for (i = vn->vn_nr_item - 1; i >= 0; vn 380 fs/reiserfs/fix_node.c struct virtual_node *vn = tb->tb_vn; vn 447 fs/reiserfs/fix_node.c end_item = vn->vn_nr_item - to - 1; vn 458 fs/reiserfs/fix_node.c struct virtual_item *vi = vn->vn_vi + i; vn 566 fs/reiserfs/fix_node.c op_unit_num(&vn->vn_vi[split_item_num]) - snum012[4] - vn 569 fs/reiserfs/fix_node.c if (vn->vn_vi[split_item_num].vi_index != TYPE_DIRENTRY && vn 570 fs/reiserfs/fix_node.c vn->vn_vi[split_item_num].vi_index != TYPE_INDIRECT) vn 594 fs/reiserfs/fix_node.c op_unit_num(&vn->vn_vi[split_item_num]) - snum012[3] - vn 654 fs/reiserfs/fix_node.c struct virtual_node *vn = tb->tb_vn; vn 665 fs/reiserfs/fix_node.c remain_items = vn->vn_nr_item; vn 672 fs/reiserfs/fix_node.c set_parameters(tb, 0, to_left, vn->vn_nr_item - to_left, 0, vn 684 fs/reiserfs/fix_node.c size = op_unit_num(&vn->vn_vi[to_left]); vn 698 fs/reiserfs/fix_node.c struct virtual_node *vn = tb->tb_vn; vn 705 fs/reiserfs/fix_node.c if (vn->vn_nr_item) { vn 706 fs/reiserfs/fix_node.c if (vn->vn_vi[0].vi_type & VI_TYPE_LEFT_MERGEABLE) vn 709 fs/reiserfs/fix_node.c if (vn->vn_vi[vn->vn_nr_item - 1]. vn 748 fs/reiserfs/fix_node.c if (MAX_CHILD_SIZE(S0) + vn->vn_size <= rfree + lfree + ih_size) { vn 763 fs/reiserfs/fix_node.c to_l = (MAX_NR_KEY(Sh)+1 - lpar + vn->vn_nr_item + 1) / 2 -\ vn 783 fs/reiserfs/fix_node.c to_r = (MAX_NR_KEY(Sh)+1 - rpar + vn->vn_nr_item + 1) / 2 - (MAX_NR_KEY(Sh) + 1 - rpar);\ vn 1326 fs/reiserfs/fix_node.c struct virtual_node *vn = tb->tb_vn; vn 1408 fs/reiserfs/fix_node.c if (can_node_be_removed(vn->vn_mode, lfree, sfree, rfree, tb, h) == vn 1434 fs/reiserfs/fix_node.c if (h && (tb->rnum[h] + tb->lnum[h] >= vn->vn_nr_item + 1)) { vn 1447 fs/reiserfs/fix_node.c vn->vn_nr_item + 1) / 2 - (MAX_NR_KEY(Sh) + 1 - vn 1449 fs/reiserfs/fix_node.c set_parameters(tb, h, vn->vn_nr_item + 1 - to_r, to_r, 0, NULL, vn 1459 fs/reiserfs/fix_node.c (tb->lnum[h] >= vn->vn_nr_item + 1 || vn 1460 fs/reiserfs/fix_node.c tb->rnum[h] >= vn->vn_nr_item + 1), vn 1462 fs/reiserfs/fix_node.c RFALSE(!h && ((tb->lnum[h] >= vn->vn_nr_item && (tb->lbytes == -1)) || vn 1463 fs/reiserfs/fix_node.c (tb->rnum[h] >= vn->vn_nr_item && (tb->rbytes == -1))), vn 1483 fs/reiserfs/fix_node.c tb->s0num = vn->vn_nr_item; vn 1523 fs/reiserfs/fix_node.c nver = get_num_ver(vn->vn_mode, tb, h, vn 1524 fs/reiserfs/fix_node.c 0, -1, h ? vn->vn_nr_item : 0, -1, vn 1534 fs/reiserfs/fix_node.c nver1 = get_num_ver(vn->vn_mode, tb, h, vn 1550 fs/reiserfs/fix_node.c lnver = get_num_ver(vn->vn_mode, tb, h, vn 1552 fs/reiserfs/fix_node.c -1, h ? vn->vn_nr_item : 0, -1, vn 1557 fs/reiserfs/fix_node.c lnver1 = get_num_ver(vn->vn_mode, tb, h, vn 1575 fs/reiserfs/fix_node.c rnver = get_num_ver(vn->vn_mode, tb, h, vn 1577 fs/reiserfs/fix_node.c h ? (vn->vn_nr_item - rpar) : (rpar - vn 1586 fs/reiserfs/fix_node.c rnver1 = get_num_ver(vn->vn_mode, tb, h, vn 1605 fs/reiserfs/fix_node.c lrnver = get_num_ver(vn->vn_mode, tb, h, vn 1608 fs/reiserfs/fix_node.c h ? (vn->vn_nr_item - rpar) : (rpar - vn 1617 fs/reiserfs/fix_node.c lrnver1 = get_num_ver(vn->vn_mode, tb, h, vn 1728 fs/reiserfs/fix_node.c struct virtual_node *vn = tb->tb_vn; vn 1751 fs/reiserfs/fix_node.c if (vn->vn_nr_item > 0) { vn 1779 fs/reiserfs/fix_node.c if (vn->vn_nr_item >= MIN_NR_KEY(Sh)) { vn 1784 fs/reiserfs/fix_node.c if (vn->vn_nr_item == MIN_NR_KEY(Sh)) { vn 1786 fs/reiserfs/fix_node.c if (tb->lnum[h] >= vn->vn_nr_item + 1) { vn 1803 fs/reiserfs/fix_node.c if (tb->rnum[h] >= vn->vn_nr_item + 1) { vn 1824 fs/reiserfs/fix_node.c if (tb->rnum[h] + tb->lnum[h] >= vn->vn_nr_item + 1) { vn 1829 fs/reiserfs/fix_node.c tb->rnum[h] + vn->vn_nr_item + 1) / 2 - vn 1831 fs/reiserfs/fix_node.c set_parameters(tb, h, vn->vn_nr_item + 1 - to_r, to_r, vn 1846 fs/reiserfs/fix_node.c if (tb->lnum[h] >= vn->vn_nr_item + 1) vn 1848 fs/reiserfs/fix_node.c || tb->rnum[h] < vn->vn_nr_item + 1 || !tb->FR[h]) { vn 1864 fs/reiserfs/fix_node.c if (tb->rnum[h] >= vn->vn_nr_item + 1) { vn 1879 fs/reiserfs/fix_node.c if (tb->rnum[h] + tb->lnum[h] >= vn->vn_nr_item + 1) { vn 1884 fs/reiserfs/fix_node.c vn->vn_nr_item + 1) / 2 - (MAX_NR_KEY(Sh) + 1 - vn 1886 fs/reiserfs/fix_node.c set_parameters(tb, h, vn->vn_nr_item + 1 - to_r, to_r, 0, NULL, vn 1899 fs/reiserfs/fix_node.c (MAX_NR_KEY(Sh) + 1 - tb->lnum[h] + vn->vn_nr_item + vn 1900 fs/reiserfs/fix_node.c 1) / 2 - (vn->vn_nr_item + 1); vn 1906 fs/reiserfs/fix_node.c -((MAX_NR_KEY(Sh) + 1 - tb->rnum[h] + vn->vn_nr_item + vn 1907 fs/reiserfs/fix_node.c 1) / 2 - (vn->vn_nr_item + 1)), 1, NULL, -1, -1); vn 1927 fs/reiserfs/fix_node.c struct virtual_node *vn = tb->tb_vn; vn 1986 fs/reiserfs/fix_node.c if (tb->lnum[0] >= vn->vn_nr_item && tb->lbytes == -1) vn 1987 fs/reiserfs/fix_node.c if (is_left_neighbor_in_cache(tb, h) || ((tb->rnum[0] - ((tb->rbytes == -1) ? 0 : 1)) < vn->vn_nr_item) || /* S can not be merged with R */ vn 1999 fs/reiserfs/fix_node.c if (tb->rnum[0] >= vn->vn_nr_item && tb->rbytes == -1) { vn 2012 fs/reiserfs/fix_node.c tb->s0num = vn->vn_nr_item; vn 2068 fs/reiserfs/fix_node.c struct virtual_node *vn; vn 2070 fs/reiserfs/fix_node.c vn = tb->tb_vn = (struct virtual_node *)(tb->vn_buf); vn 2071 fs/reiserfs/fix_node.c vn->vn_free_ptr = (char *)(tb->tb_vn + 1); vn 2072 fs/reiserfs/fix_node.c vn->vn_mode = mode; vn 2073 fs/reiserfs/fix_node.c vn->vn_affected_item_num = inum; vn 2074 fs/reiserfs/fix_node.c vn->vn_pos_in_item = pos_in_item; vn 2075 fs/reiserfs/fix_node.c vn->vn_ins_ih = ins_ih; vn 2076 fs/reiserfs/fix_node.c vn->vn_data = data; vn 2078 fs/reiserfs/fix_node.c RFALSE(mode == M_INSERT && !vn->vn_ins_ih, vn 60 fs/reiserfs/item_ops.c static int sd_create_vi(struct virtual_node *vn, vn 150 fs/reiserfs/item_ops.c static int direct_create_vi(struct virtual_node *vn, vn 288 fs/reiserfs/item_ops.c static int indirect_create_vi(struct virtual_node *vn, vn 458 fs/reiserfs/item_ops.c static int direntry_create_vi(struct virtual_node *vn, vn 479 fs/reiserfs/item_ops.c ((is_affected) ? ((vn->vn_mode == M_CUT) ? -1 : vn 480 fs/reiserfs/item_ops.c (vn->vn_mode == M_PASTE ? 1 : 0)) : 0); vn 483 fs/reiserfs/item_ops.c j = old_entry_num(is_affected, i, vn->vn_pos_in_item, vn 484 fs/reiserfs/item_ops.c vn->vn_mode); vn 493 fs/reiserfs/item_ops.c if (is_affected && vn->vn_mode == M_PASTE) vn 494 fs/reiserfs/item_ops.c dir_u->entry_sizes[vn->vn_pos_in_item] = insert_size; vn 507 fs/reiserfs/item_ops.c && (vn->vn_mode == M_PASTE vn 508 fs/reiserfs/item_ops.c || vn->vn_mode == M_CUT)) ? insert_size : 0)) { vn 512 fs/reiserfs/item_ops.c vn->vn_mode, insert_size); vn 669 fs/reiserfs/item_ops.c static int errcatch_create_vi(struct virtual_node *vn, vn 2636 fs/reiserfs/reiserfs.h int (*create_vi) (struct virtual_node * vn, struct virtual_item * vi, vn 2652 fs/reiserfs/reiserfs.h #define op_create_vi(vn,vi,is_affected,insert_size) item_ops[le_ih_k_type ((vi)->vi_ih)]->create_vi (vn,vi,is_affected,insert_size) vn 218 net/8021q/vlan.c struct vlan_net *vn = net_generic(net, vlan_net_id); vn 231 net/8021q/vlan.c switch (vn->name_type) { vn 608 net/8021q/vlan.c struct vlan_net *vn; vn 610 net/8021q/vlan.c vn = net_generic(net, vlan_net_id); vn 611 net/8021q/vlan.c vn->name_type = args.u.name_type; vn 660 net/8021q/vlan.c struct vlan_net *vn = net_generic(net, vlan_net_id); vn 663 net/8021q/vlan.c vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD; vn 94 net/8021q/vlanproc.c struct vlan_net *vn = net_generic(net, vlan_net_id); vn 96 net/8021q/vlanproc.c if (vn->proc_vlan_conf) vn 97 net/8021q/vlanproc.c remove_proc_entry(name_conf, vn->proc_vlan_dir); vn 99 net/8021q/vlanproc.c if (vn->proc_vlan_dir) vn 113 net/8021q/vlanproc.c struct vlan_net *vn = net_generic(net, vlan_net_id); vn 115 net/8021q/vlanproc.c vn->proc_vlan_dir = proc_net_mkdir(net, name_root, net->proc_net); vn 116 net/8021q/vlanproc.c if (!vn->proc_vlan_dir) vn 119 net/8021q/vlanproc.c vn->proc_vlan_conf = proc_create_net(name_conf, S_IFREG | 0600, vn 120 net/8021q/vlanproc.c vn->proc_vlan_dir, &vlan_seq_ops, vn 122 net/8021q/vlanproc.c if (!vn->proc_vlan_conf) vn 139 net/8021q/vlanproc.c struct vlan_net *vn = net_generic(dev_net(vlandev), vlan_net_id); vn 144 net/8021q/vlanproc.c vn->proc_vlan_dir, vlandev_seq_show, vlandev); vn 219 net/8021q/vlanproc.c struct vlan_net *vn = net_generic(net, vlan_net_id); vn 226 net/8021q/vlanproc.c if (vn->name_type < ARRAY_SIZE(vlan_name_type_str)) vn 227 net/8021q/vlanproc.c nmtype = vlan_name_type_str[vn->name_type]; vn 249 net/batman-adv/netlink.c (u8)atomic_read(&bat_priv->tt.vn))) vn 776 net/batman-adv/routing.c orig_ttvn = (u8)atomic_read(&bat_priv->tt.vn); vn 854 net/batman-adv/routing.c curr_ttvn = (u8)atomic_read(&bat_priv->tt.vn); vn 796 net/batman-adv/soft-interface.c atomic_set(&bat_priv->tt.vn, 0); vn 729 net/batman-adv/translation-table.c (u8)atomic_read(&bat_priv->tt.vn)); vn 966 net/batman-adv/translation-table.c (*tt_data)->ttvn = atomic_read(&bat_priv->tt.vn); vn 1096 net/batman-adv/translation-table.c net_dev->name, (u8)atomic_read(&bat_priv->tt.vn)); vn 3381 net/batman-adv/translation-table.c my_ttvn = (u8)atomic_read(&bat_priv->tt.vn); vn 3420 net/batman-adv/translation-table.c req_ttvn = (u8)atomic_read(&bat_priv->tt.vn); vn 3941 net/batman-adv/translation-table.c atomic_inc(&bat_priv->tt.vn); vn 3944 net/batman-adv/translation-table.c (u8)atomic_read(&bat_priv->tt.vn)); vn 959 net/batman-adv/types.h atomic_t vn;