tn 211 arch/arm/vfp/vfpdouble.c int tn, tm = 0; tn 213 arch/arm/vfp/vfpdouble.c tn = vfp_double_type(vdn); tn 229 arch/arm/vfp/vfpdouble.c if (tn == VFP_SNAN || (tm != VFP_SNAN && tn == VFP_QNAN)) tn 244 arch/arm/vfp/vfpdouble.c return tn == VFP_SNAN || tm == VFP_SNAN ? FPSCR_IOC : VFP_NAN_FLAG; tn 688 arch/arm/vfp/vfpdouble.c int tn, tm; tn 690 arch/arm/vfp/vfpdouble.c tn = vfp_double_type(vdn); tn 693 arch/arm/vfp/vfpdouble.c if (tn & tm & VFP_INFINITY) { tn 709 arch/arm/vfp/vfpdouble.c } else if (tn & VFP_INFINITY && tm & VFP_NUMBER) { tn 1011 arch/arm/vfp/vfpdouble.c int tm, tn; tn 1018 arch/arm/vfp/vfpdouble.c tn = vfp_double_type(&vdn); tn 1024 arch/arm/vfp/vfpdouble.c if (tn & VFP_NAN) tn 1037 arch/arm/vfp/vfpdouble.c if (tm & tn & (VFP_INFINITY|VFP_ZERO)) tn 1043 arch/arm/vfp/vfpdouble.c if (tn & VFP_INFINITY) tn 1055 arch/arm/vfp/vfpdouble.c if (tm & VFP_INFINITY || tn & VFP_ZERO) tn 1058 arch/arm/vfp/vfpdouble.c if (tn & VFP_DENORMAL) tn 219 arch/arm/vfp/vfpsingle.c int tn, tm = 0; tn 221 arch/arm/vfp/vfpsingle.c tn = vfp_single_type(vsn); tn 237 arch/arm/vfp/vfpsingle.c if (tn == VFP_SNAN || (tm != VFP_SNAN && tn == VFP_QNAN)) tn 252 arch/arm/vfp/vfpsingle.c return tn == VFP_SNAN || tm == VFP_SNAN ? FPSCR_IOC : VFP_NAN_FLAG; tn 734 arch/arm/vfp/vfpsingle.c int tn, tm; tn 736 arch/arm/vfp/vfpsingle.c tn = vfp_single_type(vsn); tn 739 arch/arm/vfp/vfpsingle.c if (tn & tm & VFP_INFINITY) { tn 755 arch/arm/vfp/vfpsingle.c } else if (tn & VFP_INFINITY && tm & VFP_NUMBER) { tn 1057 arch/arm/vfp/vfpsingle.c int tm, tn; tn 1066 arch/arm/vfp/vfpsingle.c tn = vfp_single_type(&vsn); tn 1072 arch/arm/vfp/vfpsingle.c if (tn & VFP_NAN) tn 1085 arch/arm/vfp/vfpsingle.c if (tm & tn & (VFP_INFINITY|VFP_ZERO)) tn 1091 arch/arm/vfp/vfpsingle.c if (tn & VFP_INFINITY) tn 1103 arch/arm/vfp/vfpsingle.c if (tm & VFP_INFINITY || tn & VFP_ZERO) tn 1106 arch/arm/vfp/vfpsingle.c if (tn & VFP_DENORMAL) tn 87 drivers/ata/pata_mpc52xx.c u8 tn; tn 91 drivers/ata/pata_mpc52xx.c { .t0M = 32, .td = 15, .th = 2, .tj = 2, .tkw = 15, .tm = 4, .tn = 1 }, tn 92 drivers/ata/pata_mpc52xx.c { .t0M = 10, .td = 6, .th = 1, .tj = 1, .tkw = 4, .tm = 2, .tn = 1 }, tn 93 drivers/ata/pata_mpc52xx.c { .t0M = 8, .td = 5, .th = 1, .tj = 1, .tkw = 2, .tm = 2, .tn = 1 }, tn 97 drivers/ata/pata_mpc52xx.c { .t0M = 64, .td = 29, .th = 3, .tj = 3, .tkw = 29, .tm = 7, .tn = 2 }, tn 98 drivers/ata/pata_mpc52xx.c { .t0M = 20, .td = 11, .th = 2, .tj = 1, .tkw = 7, .tm = 4, .tn = 1 }, tn 99 drivers/ata/pata_mpc52xx.c { .t0M = 16, .td = 10, .th = 2, .tj = 1, .tkw = 4, .tm = 4, .tn = 1 }, tn 303 drivers/ata/pata_mpc52xx.c t->mdma2 = ((u32)s->th << 24) | ((u32)s->tj << 16) | ((u32)s->tn << 8); tn 451 drivers/bluetooth/bt3c_cs.c unsigned int iobase, tmp, tn; tn 493 drivers/bluetooth/bt3c_cs.c if (kstrtouint(b, 16, &tn)) tn 495 drivers/bluetooth/bt3c_cs.c tmp += tn; tn 1520 drivers/gpu/drm/i915/i915_request.c struct intel_timeline *tl, *tn; tn 1525 drivers/gpu/drm/i915/i915_request.c list_for_each_entry_safe(tl, tn, &timelines->active_list, link) { tn 1539 drivers/gpu/drm/i915/i915_request.c list_safe_reset_next(tl, tn, link); tn 1553 drivers/gpu/drm/i915/i915_request.c list_for_each_entry_safe(tl, tn, &free, link) tn 750 drivers/infiniband/ulp/ipoib/ipoib_main.c struct ipoib_neigh *neigh, *tn; tn 806 drivers/infiniband/ulp/ipoib/ipoib_main.c list_for_each_entry_safe(neigh, tn, &path->neigh_list, list) { tn 1907 drivers/scsi/ncr53c8xx.c static void ncr_alloc_ccb (struct ncb *np, u_char tn, u_char ln); tn 1912 drivers/scsi/ncr53c8xx.c static void ncr_init_tcb (struct ncb *np, u_char tn); tn 1913 drivers/scsi/ncr53c8xx.c static struct lcb * ncr_alloc_lcb (struct ncb *np, u_char tn, u_char ln); tn 5625 drivers/scsi/ncr53c8xx.c unsigned char tn = sdev->id, ln = sdev->lun; tn 5626 drivers/scsi/ncr53c8xx.c struct tcb *tp = &np->target[tn]; tn 7137 drivers/scsi/ncr53c8xx.c u_char tn = cmd->device->id; tn 7139 drivers/scsi/ncr53c8xx.c struct tcb *tp = &np->target[tn]; tn 7159 drivers/scsi/ncr53c8xx.c ncr_alloc_ccb(np, tn, ln); tn 7227 drivers/scsi/ncr53c8xx.c cp->target = tn; tn 7347 drivers/scsi/ncr53c8xx.c static void ncr_alloc_ccb(struct ncb *np, u_char tn, u_char ln) tn 7349 drivers/scsi/ncr53c8xx.c struct tcb *tp = &np->target[tn]; tn 7396 drivers/scsi/ncr53c8xx.c static void ncr_init_tcb (struct ncb *np, u_char tn) tn 7398 drivers/scsi/ncr53c8xx.c struct tcb *tp = &np->target[tn]; tn 7400 drivers/scsi/ncr53c8xx.c int th = tn & 3; tn 7408 drivers/scsi/ncr53c8xx.c cpu_to_scr((SCR_JUMP ^ IFFALSE (DATA (0x80 + tn)))); tn 7483 drivers/scsi/ncr53c8xx.c static struct lcb *ncr_alloc_lcb (struct ncb *np, u_char tn, u_char ln) tn 7485 drivers/scsi/ncr53c8xx.c struct tcb *tp = &np->target[tn]; tn 7509 drivers/scsi/ncr53c8xx.c ncr_init_tcb(np, tn); tn 7575 drivers/scsi/ncr53c8xx.c unsigned char tn = sdev->id, ln = sdev->lun; tn 7576 drivers/scsi/ncr53c8xx.c struct tcb *tp = &np->target[tn]; tn 7580 drivers/scsi/ncr53c8xx.c if (!lp && !(lp = ncr_alloc_lcb(np, tn, ln))) tn 45 drivers/scsi/sym53c8xx_2/sym_hipd.c static void sym_alloc_lcb_tags (struct sym_hcb *np, u_char tn, u_char ln); tn 4647 drivers/scsi/sym53c8xx_2/sym_hipd.c u_char tn = cmd->device->id; tn 4649 drivers/scsi/sym53c8xx_2/sym_hipd.c struct sym_tcb *tp = &np->target[tn]; tn 4681 drivers/scsi/sym53c8xx_2/sym_hipd.c sym_alloc_lcb_tags(np, tn, ln); tn 4755 drivers/scsi/sym53c8xx_2/sym_hipd.c cp->target = tn; tn 4952 drivers/scsi/sym53c8xx_2/sym_hipd.c static void sym_init_tcb (struct sym_hcb *np, u_char tn) tn 4968 drivers/scsi/sym53c8xx_2/sym_hipd.c struct sym_lcb *sym_alloc_lcb (struct sym_hcb *np, u_char tn, u_char ln) tn 4970 drivers/scsi/sym53c8xx_2/sym_hipd.c struct sym_tcb *tp = &np->target[tn]; tn 4976 drivers/scsi/sym53c8xx_2/sym_hipd.c sym_init_tcb (np, tn); tn 5049 drivers/scsi/sym53c8xx_2/sym_hipd.c static void sym_alloc_lcb_tags (struct sym_hcb *np, u_char tn, u_char ln) tn 5051 drivers/scsi/sym53c8xx_2/sym_hipd.c struct sym_tcb *tp = &np->target[tn]; tn 5095 drivers/scsi/sym53c8xx_2/sym_hipd.c int sym_free_lcb(struct sym_hcb *np, u_char tn, u_char ln) tn 5097 drivers/scsi/sym53c8xx_2/sym_hipd.c struct sym_tcb *tp = &np->target[tn]; tn 1055 drivers/scsi/sym53c8xx_2/sym_hipd.h struct sym_lcb *sym_alloc_lcb(struct sym_hcb *np, u_char tn, u_char ln); tn 1056 drivers/scsi/sym53c8xx_2/sym_hipd.h int sym_free_lcb(struct sym_hcb *np, u_char tn, u_char ln); tn 84 drivers/scsi/sym53c8xx_2/sym_nvram.c Symbios_target *tn = &nvram->target[target]; tn 86 drivers/scsi/sym53c8xx_2/sym_nvram.c if (!(tn->flags & SYMBIOS_QUEUE_TAGS_ENABLED)) tn 88 drivers/scsi/sym53c8xx_2/sym_nvram.c if (!(tn->flags & SYMBIOS_DISCONNECT_ENABLE)) tn 90 drivers/scsi/sym53c8xx_2/sym_nvram.c if (!(tn->flags & SYMBIOS_SCAN_AT_BOOT_TIME)) tn 92 drivers/scsi/sym53c8xx_2/sym_nvram.c if (!(tn->flags & SYMBIOS_SCAN_LUNS)) tn 94 drivers/scsi/sym53c8xx_2/sym_nvram.c tp->usr_period = (tn->sync_period + 3) / 4; tn 95 drivers/scsi/sym53c8xx_2/sym_nvram.c tp->usr_width = (tn->bus_width == 0x8) ? 0 : 1; tn 108 drivers/scsi/sym53c8xx_2/sym_nvram.c struct Tekram_target *tn = &nvram->target[target]; tn 110 drivers/scsi/sym53c8xx_2/sym_nvram.c if (tn->flags & TEKRAM_TAGGED_COMMANDS) { tn 114 drivers/scsi/sym53c8xx_2/sym_nvram.c if (tn->flags & TEKRAM_DISCONNECT_ENABLE) tn 117 drivers/scsi/sym53c8xx_2/sym_nvram.c if (tn->flags & TEKRAM_SYNC_NEGO) tn 118 drivers/scsi/sym53c8xx_2/sym_nvram.c tp->usr_period = Tekram_sync[tn->sync_index & 0xf]; tn 119 drivers/scsi/sym53c8xx_2/sym_nvram.c tp->usr_width = (tn->flags & TEKRAM_WIDE_NEGO) ? 1 : 0; tn 159 drivers/scsi/sym53c8xx_2/sym_nvram.c struct Symbios_target *tn = &nvram->target[i]; tn 162 drivers/scsi/sym53c8xx_2/sym_nvram.c (tn->flags & SYMBIOS_DISCONNECT_ENABLE) ? " DISC" : "", tn 163 drivers/scsi/sym53c8xx_2/sym_nvram.c (tn->flags & SYMBIOS_SCAN_AT_BOOT_TIME) ? " SCAN_BOOT" : "", tn 164 drivers/scsi/sym53c8xx_2/sym_nvram.c (tn->flags & SYMBIOS_SCAN_LUNS) ? " SCAN_LUNS" : "", tn 165 drivers/scsi/sym53c8xx_2/sym_nvram.c (tn->flags & SYMBIOS_QUEUE_TAGS_ENABLED)? " TCQ" : "", tn 166 drivers/scsi/sym53c8xx_2/sym_nvram.c tn->bus_width, tn 167 drivers/scsi/sym53c8xx_2/sym_nvram.c tn->sync_period / 4, tn 168 drivers/scsi/sym53c8xx_2/sym_nvram.c tn->timeout); tn 207 drivers/scsi/sym53c8xx_2/sym_nvram.c struct Tekram_target *tn = &nvram->target[i]; tn 208 drivers/scsi/sym53c8xx_2/sym_nvram.c j = tn->sync_index & 0xf; tn 212 drivers/scsi/sym53c8xx_2/sym_nvram.c (tn->flags & TEKRAM_PARITY_CHECK) ? " PARITY" : "", tn 213 drivers/scsi/sym53c8xx_2/sym_nvram.c (tn->flags & TEKRAM_SYNC_NEGO) ? " SYNC" : "", tn 214 drivers/scsi/sym53c8xx_2/sym_nvram.c (tn->flags & TEKRAM_DISCONNECT_ENABLE) ? " DISC" : "", tn 215 drivers/scsi/sym53c8xx_2/sym_nvram.c (tn->flags & TEKRAM_START_CMD) ? " START" : "", tn 216 drivers/scsi/sym53c8xx_2/sym_nvram.c (tn->flags & TEKRAM_TAGGED_COMMANDS) ? " TCQ" : "", tn 217 drivers/scsi/sym53c8xx_2/sym_nvram.c (tn->flags & TEKRAM_WIDE_NEGO) ? " WIDE" : "", tn 112 fs/cifs/cifs_debug.h const char *tn = ""; \ tn 114 fs/cifs/cifs_debug.h tn = tcon->treeName; \ tn 117 fs/cifs/cifs_debug.h __FILE__, tn, ##__VA_ARGS__); \ tn 120 fs/cifs/cifs_debug.h tn, ##__VA_ARGS__); \ tn 123 fs/cifs/cifs_debug.h tn, ##__VA_ARGS__); \ tn 354 fs/jffs2/nodelist.h #define tn_next(tn) rb_entry(rb_next(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) tn 355 fs/jffs2/nodelist.h #define tn_prev(tn) rb_entry(rb_prev(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) tn 356 fs/jffs2/nodelist.h #define tn_parent(tn) rb_entry(rb_parent(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) tn 357 fs/jffs2/nodelist.h #define tn_left(tn) rb_entry((tn)->rb.rb_left, struct jffs2_tmp_dnode_info, rb) tn 358 fs/jffs2/nodelist.h #define tn_right(tn) rb_entry((tn)->rb.rb_right, struct jffs2_tmp_dnode_info, rb) tn 359 fs/jffs2/nodelist.h #define tn_erase(tn, list) rb_erase(&tn->rb, list); tn 31 fs/jffs2/readinode.c static int check_node_data(struct jffs2_sb_info *c, struct jffs2_tmp_dnode_info *tn) tn 33 fs/jffs2/readinode.c struct jffs2_raw_node_ref *ref = tn->fn->raw; tn 40 fs/jffs2/readinode.c BUG_ON(tn->csize == 0); tn 44 fs/jffs2/readinode.c len = tn->csize; tn 51 fs/jffs2/readinode.c if (adj >= tn->csize) { tn 53 fs/jffs2/readinode.c ref_offset(ref), tn->csize, ofs); tn 62 fs/jffs2/readinode.c ref_offset(ref), tn->csize, tn->partial_crc, tn->data_crc, ofs - len, ofs, len); tn 69 fs/jffs2/readinode.c JFFS2_WARNING("MTD point returned len too short: %zu instead of %u.\n", retlen, tn->csize); tn 99 fs/jffs2/readinode.c crc = crc32(tn->partial_crc, buffer, len); tn 107 fs/jffs2/readinode.c if (crc != tn->data_crc) { tn 109 fs/jffs2/readinode.c ref_offset(ref), tn->data_crc, crc); tn 150 fs/jffs2/readinode.c static int check_tn_node(struct jffs2_sb_info *c, struct jffs2_tmp_dnode_info *tn) tn 154 fs/jffs2/readinode.c BUG_ON(ref_obsolete(tn->fn->raw)); tn 157 fs/jffs2/readinode.c if (ref_flags(tn->fn->raw) != REF_UNCHECKED) tn 161 fs/jffs2/readinode.c tn->fn->ofs, tn->fn->ofs + tn->fn->size, ref_offset(tn->fn->raw)); tn 163 fs/jffs2/readinode.c ret = check_node_data(c, tn); tn 169 fs/jffs2/readinode.c jffs2_mark_node_obsolete(c, tn->fn->raw); tn 178 fs/jffs2/readinode.c struct jffs2_tmp_dnode_info *tn = NULL; tn 185 fs/jffs2/readinode.c tn = rb_entry(next, struct jffs2_tmp_dnode_info, rb); tn 187 fs/jffs2/readinode.c if (tn->fn->ofs < offset) tn 188 fs/jffs2/readinode.c next = tn->rb.rb_right; tn 189 fs/jffs2/readinode.c else if (tn->fn->ofs >= offset) tn 190 fs/jffs2/readinode.c next = tn->rb.rb_left; tn 195 fs/jffs2/readinode.c return tn; tn 199 fs/jffs2/readinode.c static void jffs2_kill_tn(struct jffs2_sb_info *c, struct jffs2_tmp_dnode_info *tn) tn 201 fs/jffs2/readinode.c jffs2_mark_node_obsolete(c, tn->fn->raw); tn 202 fs/jffs2/readinode.c jffs2_free_full_dnode(tn->fn); tn 203 fs/jffs2/readinode.c jffs2_free_tmp_dnode_info(tn); tn 220 fs/jffs2/readinode.c struct jffs2_tmp_dnode_info *tn) tn 222 fs/jffs2/readinode.c uint32_t fn_end = tn->fn->ofs + tn->fn->size; tn 225 fs/jffs2/readinode.c dbg_readinode("insert fragment %#04x-%#04x, ver %u at %08x\n", tn->fn->ofs, fn_end, tn->version, ref_offset(tn->fn->raw)); tn 231 fs/jffs2/readinode.c if (!tn->fn->size) { tn 233 fs/jffs2/readinode.c if (rii->mdata_tn->version < tn->version) { tn 239 fs/jffs2/readinode.c tn->version, rii->mdata_tn->version); tn 240 fs/jffs2/readinode.c jffs2_kill_tn(c, tn); tn 244 fs/jffs2/readinode.c rii->mdata_tn = tn; tn 245 fs/jffs2/readinode.c dbg_readinode("keep new mdata with ver %d\n", tn->version); tn 250 fs/jffs2/readinode.c this = jffs2_lookup_tn(&rii->tn_root, tn->fn->ofs); tn 275 fs/jffs2/readinode.c if (this->version == tn->version) { tn 281 fs/jffs2/readinode.c jffs2_kill_tn(c, tn); tn 286 fs/jffs2/readinode.c rb_replace_node(&this->rb, &tn->rb, &rii->tn_root); tn 292 fs/jffs2/readinode.c if (this->version < tn->version && tn 293 fs/jffs2/readinode.c this->fn->ofs >= tn->fn->ofs && tn 296 fs/jffs2/readinode.c if (check_tn_node(c, tn)) { tn 298 fs/jffs2/readinode.c jffs2_kill_tn(c, tn); tn 304 fs/jffs2/readinode.c if (this->version < tn->version) { tn 316 fs/jffs2/readinode.c if (this->version > tn->version && tn 317 fs/jffs2/readinode.c this->fn->ofs <= tn->fn->ofs && tn 322 fs/jffs2/readinode.c jffs2_kill_tn(c, tn); tn 345 fs/jffs2/readinode.c if (tn->fn->ofs > insert_point->fn->ofs) tn 347 fs/jffs2/readinode.c else if (tn->fn->ofs < insert_point->fn->ofs || tn 348 fs/jffs2/readinode.c tn->fn->size < insert_point->fn->size) tn 353 fs/jffs2/readinode.c rb_link_node(&tn->rb, &insert_point->rb, link); tn 354 fs/jffs2/readinode.c rb_insert_color(&tn->rb, &rii->tn_root); tn 358 fs/jffs2/readinode.c this = tn_prev(tn); tn 361 fs/jffs2/readinode.c if (this->fn->ofs + this->fn->size > tn->fn->ofs) { tn 365 fs/jffs2/readinode.c tn->overlapped = 1; tn 385 fs/jffs2/readinode.c this = tn_next(tn); tn 424 fs/jffs2/readinode.c static void ver_insert(struct rb_root *ver_root, struct jffs2_tmp_dnode_info *tn) tn 434 fs/jffs2/readinode.c if (tn->version > this_tn->version) tn 440 fs/jffs2/readinode.c rb_link_node(&tn->rb, parent, link); tn 441 fs/jffs2/readinode.c rb_insert_color(&tn->rb, ver_root); tn 546 fs/jffs2/readinode.c struct jffs2_tmp_dnode_info *tn, *next; tn 548 fs/jffs2/readinode.c rbtree_postorder_for_each_entry_safe(tn, next, list, rb) { tn 549 fs/jffs2/readinode.c jffs2_free_full_dnode(tn->fn); tn 550 fs/jffs2/readinode.c jffs2_free_tmp_dnode_info(tn); tn 701 fs/jffs2/readinode.c struct jffs2_tmp_dnode_info *tn; tn 717 fs/jffs2/readinode.c tn = jffs2_alloc_tmp_dnode_info(); tn 718 fs/jffs2/readinode.c if (!tn) { tn 719 fs/jffs2/readinode.c JFFS2_ERROR("failed to allocate tn (%zu bytes).\n", sizeof(*tn)); tn 723 fs/jffs2/readinode.c tn->partial_crc = 0; tn 780 fs/jffs2/readinode.c tn->partial_crc = crc32(0, buf, len); tn 782 fs/jffs2/readinode.c dbg_readinode("Calculates CRC (%#08x) for %d bytes, csize %d\n", tn->partial_crc, len, csize); tn 786 fs/jffs2/readinode.c if (len >= csize && unlikely(tn->partial_crc != je32_to_cpu(rd->data_crc))) { tn 788 fs/jffs2/readinode.c ref_offset(ref), tn->partial_crc, je32_to_cpu(rd->data_crc)); tn 816 fs/jffs2/readinode.c tn->fn = jffs2_alloc_full_dnode(); tn 817 fs/jffs2/readinode.c if (!tn->fn) { tn 823 fs/jffs2/readinode.c tn->version = je32_to_cpu(rd->version); tn 824 fs/jffs2/readinode.c tn->fn->ofs = je32_to_cpu(rd->offset); tn 825 fs/jffs2/readinode.c tn->data_crc = je32_to_cpu(rd->data_crc); tn 826 fs/jffs2/readinode.c tn->csize = csize; tn 827 fs/jffs2/readinode.c tn->fn->raw = ref; tn 828 fs/jffs2/readinode.c tn->overlapped = 0; tn 830 fs/jffs2/readinode.c if (tn->version > rii->highest_version) tn 831 fs/jffs2/readinode.c rii->highest_version = tn->version; tn 836 fs/jffs2/readinode.c tn->fn->size = csize; tn 838 fs/jffs2/readinode.c tn->fn->size = je32_to_cpu(rd->dsize); tn 844 fs/jffs2/readinode.c ret = jffs2_add_tn_to_tree(c, rii, tn); tn 847 fs/jffs2/readinode.c jffs2_free_full_dnode(tn->fn); tn 849 fs/jffs2/readinode.c jffs2_free_tmp_dnode_info(tn); tn 854 fs/jffs2/readinode.c tn = tn_first(&rii->tn_root); tn 855 fs/jffs2/readinode.c while (tn) { tn 857 fs/jffs2/readinode.c tn, tn->version, tn->fn->ofs, tn 858 fs/jffs2/readinode.c tn->fn->ofs+tn->fn->size, tn->overlapped); tn 859 fs/jffs2/readinode.c tn = tn_next(tn); tn 118 include/net/act_api.h int tc_action_net_init(struct net *net, struct tc_action_net *tn, tn 123 include/net/act_api.h tn->idrinfo = kmalloc(sizeof(*tn->idrinfo), GFP_KERNEL); tn 124 include/net/act_api.h if (!tn->idrinfo) tn 126 include/net/act_api.h tn->ops = ops; tn 127 include/net/act_api.h tn->idrinfo->net = net; tn 128 include/net/act_api.h mutex_init(&tn->idrinfo->lock); tn 129 include/net/act_api.h idr_init(&tn->idrinfo->action_idr); tn 143 include/net/act_api.h struct tc_action_net *tn = net_generic(net, id); tn 145 include/net/act_api.h tcf_idrinfo_destroy(tn->ops, tn->idrinfo); tn 146 include/net/act_api.h kfree(tn->idrinfo); tn 151 include/net/act_api.h int tcf_generic_walker(struct tc_action_net *tn, struct sk_buff *skb, tn 155 include/net/act_api.h int tcf_idr_search(struct tc_action_net *tn, struct tc_action **a, u32 index); tn 156 include/net/act_api.h int tcf_idr_create(struct tc_action_net *tn, u32 index, struct nlattr *est, tn 159 include/net/act_api.h void tcf_idr_insert(struct tc_action_net *tn, struct tc_action *a); tn 161 include/net/act_api.h void tcf_idr_cleanup(struct tc_action_net *tn, u32 index); tn 162 include/net/act_api.h int tcf_idr_check_alloc(struct tc_action_net *tn, u32 *index, tn 1207 kernel/trace/ftrace.c struct hlist_node *tn; tn 1217 kernel/trace/ftrace.c hlist_for_each_entry_safe(entry, tn, hhd, hlist) tn 1377 kernel/trace/ftrace.c struct hlist_node *tn; tn 1409 kernel/trace/ftrace.c hlist_for_each_entry_safe(entry, tn, hhd, hlist) { tn 173 net/ipv4/fib_trie.c static struct key_vector *resize(struct trie *t, struct key_vector *tn); tn 194 net/ipv4/fib_trie.c #define node_parent(tn) rtnl_dereference(tn_info(tn)->parent) tn 195 net/ipv4/fib_trie.c #define get_child(tn, i) rtnl_dereference((tn)->tnode[i]) tn 198 net/ipv4/fib_trie.c #define node_parent_rcu(tn) rcu_dereference_rtnl(tn_info(tn)->parent) tn 199 net/ipv4/fib_trie.c #define get_child_rcu(tn, i) rcu_dereference_rtnl((tn)->tnode[i]) tn 213 net/ipv4/fib_trie.c static inline unsigned long child_length(const struct key_vector *tn) tn 215 net/ipv4/fib_trie.c return (1ul << tn->bits) & ~(1ul); tn 381 net/ipv4/fib_trie.c struct key_vector *tn; tn 399 net/ipv4/fib_trie.c tn = tnode->kv; tn 400 net/ipv4/fib_trie.c tn->key = (shift < KEYLENGTH) ? (key >> shift) << shift : 0; tn 401 net/ipv4/fib_trie.c tn->pos = pos; tn 402 net/ipv4/fib_trie.c tn->bits = bits; tn 403 net/ipv4/fib_trie.c tn->slen = pos; tn 405 net/ipv4/fib_trie.c return tn; tn 411 net/ipv4/fib_trie.c static inline int tnode_full(struct key_vector *tn, struct key_vector *n) tn 413 net/ipv4/fib_trie.c return n && ((n->pos + n->bits) == tn->pos) && IS_TNODE(n); tn 419 net/ipv4/fib_trie.c static void put_child(struct key_vector *tn, unsigned long i, tn 422 net/ipv4/fib_trie.c struct key_vector *chi = get_child(tn, i); tn 425 net/ipv4/fib_trie.c BUG_ON(i >= child_length(tn)); tn 429 net/ipv4/fib_trie.c empty_child_inc(tn); tn 431 net/ipv4/fib_trie.c empty_child_dec(tn); tn 434 net/ipv4/fib_trie.c wasfull = tnode_full(tn, chi); tn 435 net/ipv4/fib_trie.c isfull = tnode_full(tn, n); tn 438 net/ipv4/fib_trie.c tn_info(tn)->full_children--; tn 440 net/ipv4/fib_trie.c tn_info(tn)->full_children++; tn 442 net/ipv4/fib_trie.c if (n && (tn->slen < n->slen)) tn 443 net/ipv4/fib_trie.c tn->slen = n->slen; tn 445 net/ipv4/fib_trie.c rcu_assign_pointer(tn->tnode[i], n); tn 448 net/ipv4/fib_trie.c static void update_children(struct key_vector *tn) tn 453 net/ipv4/fib_trie.c for (i = child_length(tn); i;) { tn 454 net/ipv4/fib_trie.c struct key_vector *inode = get_child(tn, --i); tn 463 net/ipv4/fib_trie.c if (node_parent(inode) == tn) tn 466 net/ipv4/fib_trie.c node_set_parent(inode, tn); tn 479 net/ipv4/fib_trie.c static inline void tnode_free_init(struct key_vector *tn) tn 481 net/ipv4/fib_trie.c tn_info(tn)->rcu.next = NULL; tn 484 net/ipv4/fib_trie.c static inline void tnode_free_append(struct key_vector *tn, tn 487 net/ipv4/fib_trie.c tn_info(n)->rcu.next = tn_info(tn)->rcu.next; tn 488 net/ipv4/fib_trie.c tn_info(tn)->rcu.next = &tn_info(n)->rcu; tn 491 net/ipv4/fib_trie.c static void tnode_free(struct key_vector *tn) tn 493 net/ipv4/fib_trie.c struct callback_head *head = &tn_info(tn)->rcu; tn 497 net/ipv4/fib_trie.c tnode_free_size += TNODE_SIZE(1ul << tn->bits); tn 498 net/ipv4/fib_trie.c node_free(tn); tn 500 net/ipv4/fib_trie.c tn = container_of(head, struct tnode, rcu)->kv; tn 511 net/ipv4/fib_trie.c struct key_vector *tn) tn 517 net/ipv4/fib_trie.c NODE_INIT_PARENT(tn, tp); tn 518 net/ipv4/fib_trie.c put_child_root(tp, tn->key, tn); tn 521 net/ipv4/fib_trie.c update_children(tn); tn 527 net/ipv4/fib_trie.c for (i = child_length(tn); i;) { tn 528 net/ipv4/fib_trie.c struct key_vector *inode = get_child(tn, --i); tn 531 net/ipv4/fib_trie.c if (tnode_full(tn, inode)) tn 532 net/ipv4/fib_trie.c tn = resize(t, inode); tn 541 net/ipv4/fib_trie.c struct key_vector *tn; tn 547 net/ipv4/fib_trie.c tn = tnode_new(oldtnode->key, oldtnode->pos - 1, oldtnode->bits + 1); tn 548 net/ipv4/fib_trie.c if (!tn) tn 559 net/ipv4/fib_trie.c for (i = child_length(oldtnode), m = 1u << tn->pos; i;) { tn 570 net/ipv4/fib_trie.c put_child(tn, get_index(inode->key, tn), inode); tn 579 net/ipv4/fib_trie.c put_child(tn, 2 * i + 1, get_child(inode, 1)); tn 580 net/ipv4/fib_trie.c put_child(tn, 2 * i, get_child(inode, 0)); tn 603 net/ipv4/fib_trie.c tnode_free_append(tn, node1); tn 606 net/ipv4/fib_trie.c tnode_free_append(tn, node0); tn 617 net/ipv4/fib_trie.c NODE_INIT_PARENT(node1, tn); tn 618 net/ipv4/fib_trie.c NODE_INIT_PARENT(node0, tn); tn 621 net/ipv4/fib_trie.c put_child(tn, 2 * i + 1, node1); tn 622 net/ipv4/fib_trie.c put_child(tn, 2 * i, node0); tn 626 net/ipv4/fib_trie.c return replace(t, oldtnode, tn); tn 629 net/ipv4/fib_trie.c tnode_free(tn); tn 637 net/ipv4/fib_trie.c struct key_vector *tn; tn 642 net/ipv4/fib_trie.c tn = tnode_new(oldtnode->key, oldtnode->pos + 1, oldtnode->bits - 1); tn 643 net/ipv4/fib_trie.c if (!tn) tn 661 net/ipv4/fib_trie.c put_child(tn, i / 2, node1 ? : node0); tn 669 net/ipv4/fib_trie.c tnode_free_append(tn, inode); tn 674 net/ipv4/fib_trie.c NODE_INIT_PARENT(inode, tn); tn 677 net/ipv4/fib_trie.c put_child(tn, i / 2, inode); tn 681 net/ipv4/fib_trie.c return replace(t, oldtnode, tn); tn 684 net/ipv4/fib_trie.c tnode_free(tn); tn 710 net/ipv4/fib_trie.c static unsigned char update_suffix(struct key_vector *tn) tn 712 net/ipv4/fib_trie.c unsigned char slen = tn->pos; tn 720 net/ipv4/fib_trie.c slen_max = min_t(unsigned char, tn->pos + tn->bits - 1, tn->slen); tn 727 net/ipv4/fib_trie.c for (i = 0, stride = 0x2ul ; i < child_length(tn); i += stride) { tn 728 net/ipv4/fib_trie.c struct key_vector *n = get_child(tn, i); tn 743 net/ipv4/fib_trie.c tn->slen = slen; tn 805 net/ipv4/fib_trie.c static inline bool should_inflate(struct key_vector *tp, struct key_vector *tn) tn 807 net/ipv4/fib_trie.c unsigned long used = child_length(tn); tn 812 net/ipv4/fib_trie.c used -= tn_info(tn)->empty_children; tn 813 net/ipv4/fib_trie.c used += tn_info(tn)->full_children; tn 817 net/ipv4/fib_trie.c return (used > 1) && tn->pos && ((50 * used) >= threshold); tn 820 net/ipv4/fib_trie.c static inline bool should_halve(struct key_vector *tp, struct key_vector *tn) tn 822 net/ipv4/fib_trie.c unsigned long used = child_length(tn); tn 827 net/ipv4/fib_trie.c used -= tn_info(tn)->empty_children; tn 831 net/ipv4/fib_trie.c return (used > 1) && (tn->bits > 1) && ((100 * used) < threshold); tn 834 net/ipv4/fib_trie.c static inline bool should_collapse(struct key_vector *tn) tn 836 net/ipv4/fib_trie.c unsigned long used = child_length(tn); tn 838 net/ipv4/fib_trie.c used -= tn_info(tn)->empty_children; tn 841 net/ipv4/fib_trie.c if ((tn->bits == KEYLENGTH) && tn_info(tn)->full_children) tn 849 net/ipv4/fib_trie.c static struct key_vector *resize(struct trie *t, struct key_vector *tn) tn 854 net/ipv4/fib_trie.c struct key_vector *tp = node_parent(tn); tn 855 net/ipv4/fib_trie.c unsigned long cindex = get_index(tn->key, tp); tn 859 net/ipv4/fib_trie.c tn, inflate_threshold, halve_threshold); tn 865 net/ipv4/fib_trie.c BUG_ON(tn != get_child(tp, cindex)); tn 870 net/ipv4/fib_trie.c while (should_inflate(tp, tn) && max_work) { tn 871 net/ipv4/fib_trie.c tp = inflate(t, tn); tn 880 net/ipv4/fib_trie.c tn = get_child(tp, cindex); tn 884 net/ipv4/fib_trie.c tp = node_parent(tn); tn 893 net/ipv4/fib_trie.c while (should_halve(tp, tn) && max_work) { tn 894 net/ipv4/fib_trie.c tp = halve(t, tn); tn 903 net/ipv4/fib_trie.c tn = get_child(tp, cindex); tn 907 net/ipv4/fib_trie.c if (should_collapse(tn)) tn 908 net/ipv4/fib_trie.c return collapse(t, tn); tn 911 net/ipv4/fib_trie.c return node_parent(tn); tn 914 net/ipv4/fib_trie.c static void node_pull_suffix(struct key_vector *tn, unsigned char slen) tn 916 net/ipv4/fib_trie.c unsigned char node_slen = tn->slen; tn 918 net/ipv4/fib_trie.c while ((node_slen > tn->pos) && (node_slen > slen)) { tn 919 net/ipv4/fib_trie.c slen = update_suffix(tn); tn 923 net/ipv4/fib_trie.c tn = node_parent(tn); tn 924 net/ipv4/fib_trie.c node_slen = tn->slen; tn 928 net/ipv4/fib_trie.c static void node_push_suffix(struct key_vector *tn, unsigned char slen) tn 930 net/ipv4/fib_trie.c while (tn->slen < slen) { tn 931 net/ipv4/fib_trie.c tn->slen = slen; tn 932 net/ipv4/fib_trie.c tn = node_parent(tn); tn 1008 net/ipv4/fib_trie.c static void trie_rebalance(struct trie *t, struct key_vector *tn) tn 1010 net/ipv4/fib_trie.c while (!IS_TRIE(tn)) tn 1011 net/ipv4/fib_trie.c tn = resize(t, tn); tn 1033 net/ipv4/fib_trie.c struct key_vector *tn; tn 1035 net/ipv4/fib_trie.c tn = tnode_new(key, __fls(key ^ n->key), 1); tn 1036 net/ipv4/fib_trie.c if (!tn) tn 1040 net/ipv4/fib_trie.c NODE_INIT_PARENT(tn, tp); tn 1041 net/ipv4/fib_trie.c put_child(tn, get_index(key, tn) ^ 1, n); tn 1044 net/ipv4/fib_trie.c put_child_root(tp, key, tn); tn 1045 net/ipv4/fib_trie.c node_set_parent(n, tn); tn 1048 net/ipv4/fib_trie.c tp = tn; tn 1618 net/ipv4/fib_trie.c static struct key_vector *leaf_walk_rcu(struct key_vector **tn, t_key key) tn 1620 net/ipv4/fib_trie.c struct key_vector *pn, *n = *tn; tn 1667 net/ipv4/fib_trie.c *tn = pn; tn 1671 net/ipv4/fib_trie.c *tn = pn; tn 470 net/netfilter/nf_conntrack_proto_tcp.c struct nf_tcp_net *tn = nf_tcp_pernet(net); tn 675 net/netfilter/nf_conntrack_proto_tcp.c tn->tcp_be_liberal) tn 764 net/netfilter/nf_conntrack_proto_tcp.c const struct nf_tcp_net *tn = nf_tcp_pernet(net); tn 790 net/netfilter/nf_conntrack_proto_tcp.c } else if (tn->tcp_loose == 0) { tn 844 net/netfilter/nf_conntrack_proto_tcp.c struct nf_tcp_net *tn = nf_tcp_pernet(net); tn 1115 net/netfilter/nf_conntrack_proto_tcp.c timeouts = tn->timeouts; tn 1117 net/netfilter/nf_conntrack_proto_tcp.c if (ct->proto.tcp.retrans >= tn->tcp_max_retrans && tn 1310 net/netfilter/nf_conntrack_proto_tcp.c struct nf_tcp_net *tn = nf_tcp_pernet(net); tn 1315 net/netfilter/nf_conntrack_proto_tcp.c timeouts = tn->timeouts; tn 1318 net/netfilter/nf_conntrack_proto_tcp.c timeouts[i] = tn->timeouts[i]; tn 1421 net/netfilter/nf_conntrack_proto_tcp.c struct nf_tcp_net *tn = nf_tcp_pernet(net); tn 1425 net/netfilter/nf_conntrack_proto_tcp.c tn->timeouts[i] = tcp_timeouts[i]; tn 1430 net/netfilter/nf_conntrack_proto_tcp.c tn->timeouts[0] = tcp_timeouts[TCP_CONNTRACK_SYN_SENT]; tn 1431 net/netfilter/nf_conntrack_proto_tcp.c tn->tcp_loose = nf_ct_tcp_loose; tn 1432 net/netfilter/nf_conntrack_proto_tcp.c tn->tcp_be_liberal = nf_ct_tcp_be_liberal; tn 1433 net/netfilter/nf_conntrack_proto_tcp.c tn->tcp_max_retrans = nf_ct_tcp_max_retrans; tn 940 net/netfilter/nf_conntrack_standalone.c struct nf_tcp_net *tn = nf_tcp_pernet(net); tn 942 net/netfilter/nf_conntrack_standalone.c #define XASSIGN(XNAME, tn) \ tn 944 net/netfilter/nf_conntrack_standalone.c &(tn)->timeouts[TCP_CONNTRACK_ ## XNAME] tn 946 net/netfilter/nf_conntrack_standalone.c XASSIGN(SYN_SENT, tn); tn 947 net/netfilter/nf_conntrack_standalone.c XASSIGN(SYN_RECV, tn); tn 948 net/netfilter/nf_conntrack_standalone.c XASSIGN(ESTABLISHED, tn); tn 949 net/netfilter/nf_conntrack_standalone.c XASSIGN(FIN_WAIT, tn); tn 950 net/netfilter/nf_conntrack_standalone.c XASSIGN(CLOSE_WAIT, tn); tn 951 net/netfilter/nf_conntrack_standalone.c XASSIGN(LAST_ACK, tn); tn 952 net/netfilter/nf_conntrack_standalone.c XASSIGN(TIME_WAIT, tn); tn 953 net/netfilter/nf_conntrack_standalone.c XASSIGN(CLOSE, tn); tn 954 net/netfilter/nf_conntrack_standalone.c XASSIGN(RETRANS, tn); tn 955 net/netfilter/nf_conntrack_standalone.c XASSIGN(UNACK, tn); tn 960 net/netfilter/nf_conntrack_standalone.c XASSIGN(LOOSE, &tn->tcp_loose); tn 961 net/netfilter/nf_conntrack_standalone.c XASSIGN(LIBERAL, &tn->tcp_be_liberal); tn 962 net/netfilter/nf_conntrack_standalone.c XASSIGN(MAX_RETRANS, &tn->tcp_max_retrans); tn 65 net/netfilter/xt_TEE.c struct tee_net *tn = net_generic(net, tee_net_id); tn 68 net/netfilter/xt_TEE.c mutex_lock(&tn->lock); tn 69 net/netfilter/xt_TEE.c list_for_each_entry(priv, &tn->priv_list, list) { tn 87 net/netfilter/xt_TEE.c mutex_unlock(&tn->lock); tn 94 net/netfilter/xt_TEE.c struct tee_net *tn = net_generic(par->net, tee_net_id); tn 122 net/netfilter/xt_TEE.c mutex_lock(&tn->lock); tn 123 net/netfilter/xt_TEE.c list_add(&priv->list, &tn->priv_list); tn 124 net/netfilter/xt_TEE.c mutex_unlock(&tn->lock); tn 134 net/netfilter/xt_TEE.c struct tee_net *tn = net_generic(par->net, tee_net_id); tn 138 net/netfilter/xt_TEE.c mutex_lock(&tn->lock); tn 140 net/netfilter/xt_TEE.c mutex_unlock(&tn->lock); tn 175 net/netfilter/xt_TEE.c struct tee_net *tn = net_generic(net, tee_net_id); tn 177 net/netfilter/xt_TEE.c INIT_LIST_HEAD(&tn->priv_list); tn 178 net/netfilter/xt_TEE.c mutex_init(&tn->lock); tn 327 net/sched/act_api.c int tcf_generic_walker(struct tc_action_net *tn, struct sk_buff *skb, tn 332 net/sched/act_api.c struct tcf_idrinfo *idrinfo = tn->idrinfo; tn 346 net/sched/act_api.c int tcf_idr_search(struct tc_action_net *tn, struct tc_action **a, u32 index) tn 348 net/sched/act_api.c struct tcf_idrinfo *idrinfo = tn->idrinfo; tn 400 net/sched/act_api.c int tcf_idr_create(struct tc_action_net *tn, u32 index, struct nlattr *est, tn 405 net/sched/act_api.c struct tcf_idrinfo *idrinfo = tn->idrinfo; tn 454 net/sched/act_api.c void tcf_idr_insert(struct tc_action_net *tn, struct tc_action *a) tn 456 net/sched/act_api.c struct tcf_idrinfo *idrinfo = tn->idrinfo; tn 467 net/sched/act_api.c void tcf_idr_cleanup(struct tc_action_net *tn, u32 index) tn 469 net/sched/act_api.c struct tcf_idrinfo *idrinfo = tn->idrinfo; tn 484 net/sched/act_api.c int tcf_idr_check_alloc(struct tc_action_net *tn, u32 *index, tn 487 net/sched/act_api.c struct tcf_idrinfo *idrinfo = tn->idrinfo; tn 280 net/sched/act_bpf.c struct tc_action_net *tn = net_generic(net, bpf_net_id); tn 303 net/sched/act_bpf.c ret = tcf_idr_check_alloc(tn, &index, act, bind); tn 305 net/sched/act_bpf.c ret = tcf_idr_create(tn, index, est, act, tn 308 net/sched/act_bpf.c tcf_idr_cleanup(tn, index); tn 365 net/sched/act_bpf.c tcf_idr_insert(tn, *act); tn 396 net/sched/act_bpf.c struct tc_action_net *tn = net_generic(net, bpf_net_id); tn 398 net/sched/act_bpf.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 403 net/sched/act_bpf.c struct tc_action_net *tn = net_generic(net, bpf_net_id); tn 405 net/sched/act_bpf.c return tcf_idr_search(tn, a, index); tn 423 net/sched/act_bpf.c struct tc_action_net *tn = net_generic(net, bpf_net_id); tn 425 net/sched/act_bpf.c return tc_action_net_init(net, tn, &act_bpf_ops); tn 100 net/sched/act_connmark.c struct tc_action_net *tn = net_generic(net, connmark_net_id); tn 121 net/sched/act_connmark.c ret = tcf_idr_check_alloc(tn, &index, a, bind); tn 123 net/sched/act_connmark.c ret = tcf_idr_create(tn, index, est, a, tn 126 net/sched/act_connmark.c tcf_idr_cleanup(tn, index); tn 139 net/sched/act_connmark.c tcf_idr_insert(tn, *a); tn 206 net/sched/act_connmark.c struct tc_action_net *tn = net_generic(net, connmark_net_id); tn 208 net/sched/act_connmark.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 213 net/sched/act_connmark.c struct tc_action_net *tn = net_generic(net, connmark_net_id); tn 215 net/sched/act_connmark.c return tcf_idr_search(tn, a, index); tn 232 net/sched/act_connmark.c struct tc_action_net *tn = net_generic(net, connmark_net_id); tn 234 net/sched/act_connmark.c return tc_action_net_init(net, tn, &act_connmark_ops); tn 48 net/sched/act_csum.c struct tc_action_net *tn = net_generic(net, csum_net_id); tn 69 net/sched/act_csum.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 71 net/sched/act_csum.c ret = tcf_idr_create(tn, index, est, a, tn 74 net/sched/act_csum.c tcf_idr_cleanup(tn, index); tn 114 net/sched/act_csum.c tcf_idr_insert(tn, *a); tn 682 net/sched/act_csum.c struct tc_action_net *tn = net_generic(net, csum_net_id); tn 684 net/sched/act_csum.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 689 net/sched/act_csum.c struct tc_action_net *tn = net_generic(net, csum_net_id); tn 691 net/sched/act_csum.c return tcf_idr_search(tn, a, index); tn 715 net/sched/act_csum.c struct tc_action_net *tn = net_generic(net, csum_net_id); tn 717 net/sched/act_csum.c return tc_action_net_init(net, tn, &act_csum_ops); tn 38 net/sched/act_ct.c struct tc_action_net tn; /* Must be first */ tn 593 net/sched/act_ct.c struct tc_ct_action_net *tn = net_generic(net, ct_net_id); tn 629 net/sched/act_ct.c if (!tn->labels) { tn 673 net/sched/act_ct.c struct tc_action_net *tn = net_generic(net, ct_net_id); tn 697 net/sched/act_ct.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 702 net/sched/act_ct.c err = tcf_idr_create(tn, index, est, a, tn 705 net/sched/act_ct.c tcf_idr_cleanup(tn, index); tn 744 net/sched/act_ct.c tcf_idr_insert(tn, *a); tn 902 net/sched/act_ct.c struct tc_action_net *tn = net_generic(net, ct_net_id); tn 904 net/sched/act_ct.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 909 net/sched/act_ct.c struct tc_action_net *tn = net_generic(net, ct_net_id); tn 911 net/sched/act_ct.c return tcf_idr_search(tn, a, index); tn 944 net/sched/act_ct.c struct tc_ct_action_net *tn = net_generic(net, ct_net_id); tn 947 net/sched/act_ct.c tn->labels = false; tn 950 net/sched/act_ct.c tn->labels = true; tn 953 net/sched/act_ct.c return tc_action_net_init(net, &tn->tn, &act_ct_ops); tn 962 net/sched/act_ct.c struct tc_ct_action_net *tn = net_generic(net, ct_net_id); tn 964 net/sched/act_ct.c if (tn->labels) tn 159 net/sched/act_ctinfo.c struct tc_action_net *tn = net_generic(net, ctinfo_net_id); tn 210 net/sched/act_ctinfo.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 212 net/sched/act_ctinfo.c ret = tcf_idr_create(tn, index, est, a, tn 215 net/sched/act_ctinfo.c tcf_idr_cleanup(tn, index); tn 270 net/sched/act_ctinfo.c tcf_idr_insert(tn, *a); tn 351 net/sched/act_ctinfo.c struct tc_action_net *tn = net_generic(net, ctinfo_net_id); tn 353 net/sched/act_ctinfo.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 358 net/sched/act_ctinfo.c struct tc_action_net *tn = net_generic(net, ctinfo_net_id); tn 360 net/sched/act_ctinfo.c return tcf_idr_search(tn, a, index); tn 388 net/sched/act_ctinfo.c struct tc_action_net *tn = net_generic(net, ctinfo_net_id); tn 390 net/sched/act_ctinfo.c return tc_action_net_init(net, tn, &act_ctinfo_ops); tn 58 net/sched/act_gact.c struct tc_action_net *tn = net_generic(net, gact_net_id); tn 99 net/sched/act_gact.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 101 net/sched/act_gact.c ret = tcf_idr_create(tn, index, est, a, tn 104 net/sched/act_gact.c tcf_idr_cleanup(tn, index); tn 143 net/sched/act_gact.c tcf_idr_insert(tn, *a); tn 238 net/sched/act_gact.c struct tc_action_net *tn = net_generic(net, gact_net_id); tn 240 net/sched/act_gact.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 245 net/sched/act_gact.c struct tc_action_net *tn = net_generic(net, gact_net_id); tn 247 net/sched/act_gact.c return tcf_idr_search(tn, a, index); tn 279 net/sched/act_gact.c struct tc_action_net *tn = net_generic(net, gact_net_id); tn 281 net/sched/act_gact.c return tc_action_net_init(net, tn, &act_gact_ops); tn 470 net/sched/act_ife.c struct tc_action_net *tn = net_generic(net, ife_net_id); tn 512 net/sched/act_ife.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 524 net/sched/act_ife.c ret = tcf_idr_create(tn, index, est, a, &act_ife_ops, tn 527 net/sched/act_ife.c tcf_idr_cleanup(tn, index); tn 606 net/sched/act_ife.c tcf_idr_insert(tn, *a); tn 863 net/sched/act_ife.c struct tc_action_net *tn = net_generic(net, ife_net_id); tn 865 net/sched/act_ife.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 870 net/sched/act_ife.c struct tc_action_net *tn = net_generic(net, ife_net_id); tn 872 net/sched/act_ife.c return tcf_idr_search(tn, a, index); tn 890 net/sched/act_ife.c struct tc_action_net *tn = net_generic(net, ife_net_id); tn 892 net/sched/act_ife.c return tc_action_net_init(net, tn, &act_ife_ops); tn 100 net/sched/act_ipt.c struct tc_action_net *tn = net_generic(net, id); tn 121 net/sched/act_ipt.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 132 net/sched/act_ipt.c tcf_idr_cleanup(tn, index); tn 141 net/sched/act_ipt.c tcf_idr_cleanup(tn, index); tn 146 net/sched/act_ipt.c ret = tcf_idr_create(tn, index, est, a, ops, bind, tn 149 net/sched/act_ipt.c tcf_idr_cleanup(tn, index); tn 193 net/sched/act_ipt.c tcf_idr_insert(tn, *a); tn 325 net/sched/act_ipt.c struct tc_action_net *tn = net_generic(net, ipt_net_id); tn 327 net/sched/act_ipt.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 332 net/sched/act_ipt.c struct tc_action_net *tn = net_generic(net, ipt_net_id); tn 334 net/sched/act_ipt.c return tcf_idr_search(tn, a, index); tn 352 net/sched/act_ipt.c struct tc_action_net *tn = net_generic(net, ipt_net_id); tn 354 net/sched/act_ipt.c return tc_action_net_init(net, tn, &act_ipt_ops); tn 374 net/sched/act_ipt.c struct tc_action_net *tn = net_generic(net, xt_net_id); tn 376 net/sched/act_ipt.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 381 net/sched/act_ipt.c struct tc_action_net *tn = net_generic(net, xt_net_id); tn 383 net/sched/act_ipt.c return tcf_idr_search(tn, a, index); tn 401 net/sched/act_ipt.c struct tc_action_net *tn = net_generic(net, xt_net_id); tn 403 net/sched/act_ipt.c return tc_action_net_init(net, tn, &act_xt_ops); tn 98 net/sched/act_mirred.c struct tc_action_net *tn = net_generic(net, mirred_net_id); tn 123 net/sched/act_mirred.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 140 net/sched/act_mirred.c tcf_idr_cleanup(tn, index); tn 147 net/sched/act_mirred.c tcf_idr_cleanup(tn, index); tn 151 net/sched/act_mirred.c ret = tcf_idr_create(tn, index, est, a, tn 154 net/sched/act_mirred.c tcf_idr_cleanup(tn, index); tn 198 net/sched/act_mirred.c tcf_idr_insert(tn, *a); tn 369 net/sched/act_mirred.c struct tc_action_net *tn = net_generic(net, mirred_net_id); tn 371 net/sched/act_mirred.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 376 net/sched/act_mirred.c struct tc_action_net *tn = net_generic(net, mirred_net_id); tn 378 net/sched/act_mirred.c return tcf_idr_search(tn, a, index); tn 459 net/sched/act_mirred.c struct tc_action_net *tn = net_generic(net, mirred_net_id); tn 461 net/sched/act_mirred.c return tc_action_net_init(net, tn, &act_mirred_ops); tn 139 net/sched/act_mpls.c struct tc_action_net *tn = net_generic(net, mpls_net_id); tn 221 net/sched/act_mpls.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 229 net/sched/act_mpls.c ret = tcf_idr_create(tn, index, est, a, tn 232 net/sched/act_mpls.c tcf_idr_cleanup(tn, index); tn 277 net/sched/act_mpls.c tcf_idr_insert(tn, *a); tn 356 net/sched/act_mpls.c struct tc_action_net *tn = net_generic(net, mpls_net_id); tn 358 net/sched/act_mpls.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 363 net/sched/act_mpls.c struct tc_action_net *tn = net_generic(net, mpls_net_id); tn 365 net/sched/act_mpls.c return tcf_idr_search(tn, a, index); tn 383 net/sched/act_mpls.c struct tc_action_net *tn = net_generic(net, mpls_net_id); tn 385 net/sched/act_mpls.c return tc_action_net_init(net, tn, &act_mpls_ops); tn 41 net/sched/act_nat.c struct tc_action_net *tn = net_generic(net, nat_net_id); tn 61 net/sched/act_nat.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 63 net/sched/act_nat.c ret = tcf_idr_create(tn, index, est, a, tn 66 net/sched/act_nat.c tcf_idr_cleanup(tn, index); tn 97 net/sched/act_nat.c tcf_idr_insert(tn, *a); tn 302 net/sched/act_nat.c struct tc_action_net *tn = net_generic(net, nat_net_id); tn 304 net/sched/act_nat.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 309 net/sched/act_nat.c struct tc_action_net *tn = net_generic(net, nat_net_id); tn 311 net/sched/act_nat.c return tcf_idr_search(tn, a, index); tn 328 net/sched/act_nat.c struct tc_action_net *tn = net_generic(net, nat_net_id); tn 330 net/sched/act_nat.c return tc_action_net_init(net, tn, &act_nat_ops); tn 142 net/sched/act_pedit.c struct tc_action_net *tn = net_generic(net, pedit_net_id); tn 188 net/sched/act_pedit.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 190 net/sched/act_pedit.c ret = tcf_idr_create(tn, index, est, a, tn 193 net/sched/act_pedit.c tcf_idr_cleanup(tn, index); tn 241 net/sched/act_pedit.c tcf_idr_insert(tn, *a); tn 470 net/sched/act_pedit.c struct tc_action_net *tn = net_generic(net, pedit_net_id); tn 472 net/sched/act_pedit.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 477 net/sched/act_pedit.c struct tc_action_net *tn = net_generic(net, pedit_net_id); tn 479 net/sched/act_pedit.c return tcf_idr_search(tn, a, index); tn 497 net/sched/act_pedit.c struct tc_action_net *tn = net_generic(net, pedit_net_id); tn 499 net/sched/act_pedit.c return tc_action_net_init(net, tn, &act_pedit_ops); tn 33 net/sched/act_police.c struct tc_action_net *tn = net_generic(net, police_net_id); tn 35 net/sched/act_police.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 59 net/sched/act_police.c struct tc_action_net *tn = net_generic(net, police_net_id); tn 81 net/sched/act_police.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 89 net/sched/act_police.c ret = tcf_idr_create(tn, index, NULL, a, tn 92 net/sched/act_police.c tcf_idr_cleanup(tn, index); tn 205 net/sched/act_police.c tcf_idr_insert(tn, *a); tn 366 net/sched/act_police.c struct tc_action_net *tn = net_generic(net, police_net_id); tn 368 net/sched/act_police.c return tcf_idr_search(tn, a, index); tn 391 net/sched/act_police.c struct tc_action_net *tn = net_generic(net, police_net_id); tn 393 net/sched/act_police.c return tc_action_net_init(net, tn, &act_police_ops); tn 41 net/sched/act_sample.c struct tc_action_net *tn = net_generic(net, sample_net_id); tn 63 net/sched/act_sample.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 71 net/sched/act_sample.c ret = tcf_idr_create(tn, index, est, a, tn 74 net/sched/act_sample.c tcf_idr_cleanup(tn, index); tn 120 net/sched/act_sample.c tcf_idr_insert(tn, *a); tn 244 net/sched/act_sample.c struct tc_action_net *tn = net_generic(net, sample_net_id); tn 246 net/sched/act_sample.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 251 net/sched/act_sample.c struct tc_action_net *tn = net_generic(net, sample_net_id); tn 253 net/sched/act_sample.c return tcf_idr_search(tn, a, index); tn 298 net/sched/act_sample.c struct tc_action_net *tn = net_generic(net, sample_net_id); tn 300 net/sched/act_sample.c return tc_action_net_init(net, tn, &act_sample_ops); tn 91 net/sched/act_simple.c struct tc_action_net *tn = net_generic(net, simp_net_id); tn 113 net/sched/act_simple.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 124 net/sched/act_simple.c tcf_idr_cleanup(tn, index); tn 129 net/sched/act_simple.c ret = tcf_idr_create(tn, index, est, a, tn 132 net/sched/act_simple.c tcf_idr_cleanup(tn, index); tn 160 net/sched/act_simple.c tcf_idr_insert(tn, *a); tn 206 net/sched/act_simple.c struct tc_action_net *tn = net_generic(net, simp_net_id); tn 208 net/sched/act_simple.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 213 net/sched/act_simple.c struct tc_action_net *tn = net_generic(net, simp_net_id); tn 215 net/sched/act_simple.c return tcf_idr_search(tn, a, index); tn 233 net/sched/act_simple.c struct tc_action_net *tn = net_generic(net, simp_net_id); tn 235 net/sched/act_simple.c return tc_action_net_init(net, tn, &act_simp_ops); tn 92 net/sched/act_skbedit.c struct tc_action_net *tn = net_generic(net, skbedit_net_id); tn 151 net/sched/act_skbedit.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 162 net/sched/act_skbedit.c tcf_idr_cleanup(tn, index); tn 167 net/sched/act_skbedit.c ret = tcf_idr_create(tn, index, est, a, tn 170 net/sched/act_skbedit.c tcf_idr_cleanup(tn, index); tn 218 net/sched/act_skbedit.c tcf_idr_insert(tn, *a); tn 298 net/sched/act_skbedit.c struct tc_action_net *tn = net_generic(net, skbedit_net_id); tn 300 net/sched/act_skbedit.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 305 net/sched/act_skbedit.c struct tc_action_net *tn = net_generic(net, skbedit_net_id); tn 307 net/sched/act_skbedit.c return tcf_idr_search(tn, a, index); tn 337 net/sched/act_skbedit.c struct tc_action_net *tn = net_generic(net, skbedit_net_id); tn 339 net/sched/act_skbedit.c return tc_action_net_init(net, tn, &act_skbedit_ops); tn 85 net/sched/act_skbmod.c struct tc_action_net *tn = net_generic(net, skbmod_net_id); tn 129 net/sched/act_skbmod.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 140 net/sched/act_skbmod.c tcf_idr_cleanup(tn, index); tn 145 net/sched/act_skbmod.c ret = tcf_idr_create(tn, index, est, a, tn 148 net/sched/act_skbmod.c tcf_idr_cleanup(tn, index); tn 194 net/sched/act_skbmod.c tcf_idr_insert(tn, *a); tn 261 net/sched/act_skbmod.c struct tc_action_net *tn = net_generic(net, skbmod_net_id); tn 263 net/sched/act_skbmod.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 268 net/sched/act_skbmod.c struct tc_action_net *tn = net_generic(net, skbmod_net_id); tn 270 net/sched/act_skbmod.c return tcf_idr_search(tn, a, index); tn 288 net/sched/act_skbmod.c struct tc_action_net *tn = net_generic(net, skbmod_net_id); tn 290 net/sched/act_skbmod.c return tc_action_net_init(net, tn, &act_skbmod_ops); tn 218 net/sched/act_tunnel_key.c struct tc_action_net *tn = net_generic(net, tunnel_key_net_id); tn 254 net/sched/act_tunnel_key.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 354 net/sched/act_tunnel_key.c ret = tcf_idr_create(tn, index, est, a, tn 396 net/sched/act_tunnel_key.c tcf_idr_insert(tn, *a); tn 412 net/sched/act_tunnel_key.c tcf_idr_cleanup(tn, index); tn 578 net/sched/act_tunnel_key.c struct tc_action_net *tn = net_generic(net, tunnel_key_net_id); tn 580 net/sched/act_tunnel_key.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 585 net/sched/act_tunnel_key.c struct tc_action_net *tn = net_generic(net, tunnel_key_net_id); tn 587 net/sched/act_tunnel_key.c return tcf_idr_search(tn, a, index); tn 605 net/sched/act_tunnel_key.c struct tc_action_net *tn = net_generic(net, tunnel_key_net_id); tn 607 net/sched/act_tunnel_key.c return tc_action_net_init(net, tn, &act_tunnel_key_ops); tn 107 net/sched/act_vlan.c struct tc_action_net *tn = net_generic(net, vlan_net_id); tn 133 net/sched/act_vlan.c err = tcf_idr_check_alloc(tn, &index, a, bind); tn 149 net/sched/act_vlan.c tcf_idr_cleanup(tn, index); tn 157 net/sched/act_vlan.c tcf_idr_cleanup(tn, index); tn 171 net/sched/act_vlan.c tcf_idr_cleanup(tn, index); tn 185 net/sched/act_vlan.c tcf_idr_cleanup(tn, index); tn 191 net/sched/act_vlan.c ret = tcf_idr_create(tn, index, est, a, tn 194 net/sched/act_vlan.c tcf_idr_cleanup(tn, index); tn 232 net/sched/act_vlan.c tcf_idr_insert(tn, *a); tn 299 net/sched/act_vlan.c struct tc_action_net *tn = net_generic(net, vlan_net_id); tn 301 net/sched/act_vlan.c return tcf_generic_walker(tn, skb, cb, type, ops, extack); tn 319 net/sched/act_vlan.c struct tc_action_net *tn = net_generic(net, vlan_net_id); tn 321 net/sched/act_vlan.c return tcf_idr_search(tn, a, index); tn 349 net/sched/act_vlan.c struct tc_action_net *tn = net_generic(net, vlan_net_id); tn 351 net/sched/act_vlan.c return tc_action_net_init(net, tn, &act_vlan_ops); tn 872 net/sched/cls_api.c struct tcf_net *tn = net_generic(net, tcf_net_id); tn 876 net/sched/cls_api.c spin_lock(&tn->idr_lock); tn 877 net/sched/cls_api.c err = idr_alloc_u32(&tn->idr, block, &block->index, block->index, tn 879 net/sched/cls_api.c spin_unlock(&tn->idr_lock); tn 887 net/sched/cls_api.c struct tcf_net *tn = net_generic(net, tcf_net_id); tn 889 net/sched/cls_api.c spin_lock(&tn->idr_lock); tn 890 net/sched/cls_api.c idr_remove(&tn->idr, block->index); tn 891 net/sched/cls_api.c spin_unlock(&tn->idr_lock); tn 925 net/sched/cls_api.c struct tcf_net *tn = net_generic(net, tcf_net_id); tn 927 net/sched/cls_api.c return idr_find(&tn->idr, block_index); tn 3606 net/sched/cls_api.c struct tcf_net *tn = net_generic(net, tcf_net_id); tn 3608 net/sched/cls_api.c spin_lock_init(&tn->idr_lock); tn 3609 net/sched/cls_api.c idr_init(&tn->idr); tn 3615 net/sched/cls_api.c struct tcf_net *tn = net_generic(net, tcf_net_id); tn 3617 net/sched/cls_api.c idr_destroy(&tn->idr); tn 57 net/tipc/addr.c struct tipc_net *tn = tipc_net(net); tn 60 net/tipc/addr.c memcpy(tn->node_id, id, NODE_ID_LEN); tn 61 net/tipc/addr.c tipc_nodeid2string(tn->node_id_string, id); tn 62 net/tipc/addr.c tn->trial_addr = tmp[0] ^ tmp[1] ^ tmp[2] ^ tmp[3]; tn 64 net/tipc/addr.c tipc_own_id_string(net), tn->net_id); tn 69 net/tipc/addr.c struct tipc_net *tn = tipc_net(net); tn 72 net/tipc/addr.c tn->node_addr = addr; tn 77 net/tipc/addr.c tn->trial_addr = addr; tn 78 net/tipc/addr.c tn->addr_trial_end = jiffies; tn 53 net/tipc/addr.h struct tipc_net *tn = tipc_net(net); tn 55 net/tipc/addr.h if (!strlen(tn->node_id_string)) tn 57 net/tipc/addr.h return tn->node_id; tn 679 net/tipc/bcast.c struct tipc_net *tn = tipc_net(net); tn 686 net/tipc/bcast.c tn->bcbase = bb; tn 699 net/tipc/bcast.c tn->bcl = l; tn 711 net/tipc/bcast.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 714 net/tipc/bcast.c kfree(tn->bcbase); tn 715 net/tipc/bcast.c kfree(tn->bcl); tn 63 net/tipc/bearer.c struct tipc_net *tn = tipc_net(net); tn 65 net/tipc/bearer.c return rcu_dereference(tn->bearer_list[bearer_id]); tn 174 net/tipc/bearer.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 179 net/tipc/bearer.c b = rtnl_dereference(tn->bearer_list[i]); tn 193 net/tipc/bearer.c struct tipc_net *tn = tipc_net(net); tn 199 net/tipc/bearer.c b = rtnl_dereference(tn->bearer_list[bearer_id]); tn 209 net/tipc/bearer.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 213 net/tipc/bearer.c b = rcu_dereference(tn->bearer_list[bearer_id]); tn 221 net/tipc/bearer.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 225 net/tipc/bearer.c b = rcu_dereference(tn->bearer_list[bearer_id]); tn 238 net/tipc/bearer.c struct tipc_net *tn = tipc_net(net); tn 269 net/tipc/bearer.c b = rtnl_dereference(tn->bearer_list[bearer_id]); tn 326 net/tipc/bearer.c rcu_assign_pointer(tn->bearer_list[bearer_id], b); tn 361 net/tipc/bearer.c struct tipc_net *tn = tipc_net(net); tn 371 net/tipc/bearer.c RCU_INIT_POINTER(tn->bearer_list[bearer_id], NULL); tn 545 net/tipc/bearer.c struct tipc_net *tn = tipc_net(net); tn 546 net/tipc/bearer.c int net_id = tn->net_id; tn 669 net/tipc/bearer.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 674 net/tipc/bearer.c b = rtnl_dereference(tn->bearer_list[i]); tn 677 net/tipc/bearer.c tn->bearer_list[i] = NULL; tn 720 net/tipc/bearer.c struct tipc_net *tn = tipc_net(net); tn 726 net/tipc/bearer.c tn->loopback_pt.dev = dev; tn 727 net/tipc/bearer.c tn->loopback_pt.type = htons(ETH_P_TIPC); tn 728 net/tipc/bearer.c tn->loopback_pt.func = tipc_loopback_rcv_pkt; tn 729 net/tipc/bearer.c dev_add_pack(&tn->loopback_pt); tn 735 net/tipc/bearer.c struct tipc_net *tn = tipc_net(net); tn 737 net/tipc/bearer.c dev_remove_pack(&tn->loopback_pt); tn 805 net/tipc/bearer.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 816 net/tipc/bearer.c bearer = rtnl_dereference(tn->bearer_list[i]); tn 54 net/tipc/core.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 57 net/tipc/core.c tn->net_id = 4711; tn 58 net/tipc/core.c tn->node_addr = 0; tn 59 net/tipc/core.c tn->trial_addr = 0; tn 60 net/tipc/core.c tn->addr_trial_end = 0; tn 61 net/tipc/core.c tn->capabilities = TIPC_NODE_CAPABILITIES; tn 62 net/tipc/core.c memset(tn->node_id, 0, sizeof(tn->node_id)); tn 63 net/tipc/core.c memset(tn->node_id_string, 0, sizeof(tn->node_id_string)); tn 64 net/tipc/core.c tn->mon_threshold = TIPC_DEF_MON_THRESHOLD; tn 65 net/tipc/core.c get_random_bytes(&tn->random, sizeof(int)); tn 66 net/tipc/core.c INIT_LIST_HEAD(&tn->node_list); tn 67 net/tipc/core.c spin_lock_init(&tn->node_list_lock); tn 77 net/tipc/core.c INIT_LIST_HEAD(&tn->dist_queue); tn 83 net/tipc/discover.c struct tipc_net *tn = tipc_net(net); tn 88 net/tipc/discover.c tipc_msg_init(tn->trial_addr, hdr, LINK_CONFIG, mtyp, tn 92 net/tipc/discover.c msg_set_node_sig(hdr, tn->random); tn 95 net/tipc/discover.c msg_set_bc_netid(hdr, tn->net_id); tn 148 net/tipc/discover.c struct tipc_net *tn = tipc_net(net); tn 149 net/tipc/discover.c bool trial = time_before(jiffies, tn->addr_trial_end); tn 157 net/tipc/discover.c if (dst != tn->trial_addr) tn 161 net/tipc/discover.c tn->trial_addr = sugg_addr; tn 163 net/tipc/discover.c tn->addr_trial_end = jiffies + msecs_to_jiffies(1000); tn 169 net/tipc/discover.c tipc_sched_net_finalize(net, tn->trial_addr); tn 170 net/tipc/discover.c msg_set_prevnode(buf_msg(d->skb), tn->trial_addr); tn 194 net/tipc/discover.c struct tipc_net *tn = tipc_net(net); tn 197 net/tipc/discover.c bool legacy = tn->legacy_addr_format; tn 228 net/tipc/discover.c if (net_id != tn->net_id) tn 290 net/tipc/discover.c struct tipc_net *tn = tipc_net(d->net); tn 305 net/tipc/discover.c if (!time_before(jiffies, tn->addr_trial_end) && !tipc_own_addr(net)) { tn 308 net/tipc/discover.c tipc_sched_net_finalize(net, tn->trial_addr); tn 313 net/tipc/discover.c if (time_before(jiffies, tn->addr_trial_end)) { tn 322 net/tipc/discover.c msg_set_prevnode(buf_msg(d->skb), tn->trial_addr); tn 347 net/tipc/discover.c struct tipc_net *tn = tipc_net(net); tn 362 net/tipc/discover.c tn->addr_trial_end = jiffies + msecs_to_jiffies(1000); tn 2560 net/tipc/link.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 2563 net/tipc/link.c struct tipc_link *bcl = tn->bcl; tn 174 net/tipc/monitor.c struct tipc_net *tn = tipc_net(net); tn 176 net/tipc/monitor.c return mon->peer_cnt > tn->mon_threshold; tn 608 net/tipc/monitor.c struct tipc_net *tn = tipc_net(net); tn 613 net/tipc/monitor.c if (tn->monitors[bearer_id]) tn 625 net/tipc/monitor.c tn->monitors[bearer_id] = mon; tn 636 net/tipc/monitor.c mon->timer_intv = msecs_to_jiffies(MON_TIMEOUT + (tn->random & 0xffff)); tn 643 net/tipc/monitor.c struct tipc_net *tn = tipc_net(net); tn 653 net/tipc/monitor.c tn->monitors[bearer_id] = NULL; tn 685 net/tipc/monitor.c struct tipc_net *tn = tipc_net(net); tn 690 net/tipc/monitor.c tn->mon_threshold = cluster_size; tn 697 net/tipc/monitor.c struct tipc_net *tn = tipc_net(net); tn 699 net/tipc/monitor.c return tn->mon_threshold; tn 209 net/tipc/name_distr.c struct tipc_net *tn = tipc_net(net); tn 212 net/tipc/name_distr.c spin_lock_bh(&tn->nametbl_lock); tn 217 net/tipc/name_distr.c spin_unlock_bh(&tn->nametbl_lock); tn 235 net/tipc/name_distr.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 238 net/tipc/name_distr.c spin_lock_bh(&tn->nametbl_lock); tn 239 net/tipc/name_distr.c list_for_each_entry_safe(e, tmp, &tn->dist_queue, next) { tn 245 net/tipc/name_distr.c spin_unlock_bh(&tn->nametbl_lock); tn 303 net/tipc/name_distr.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 311 net/tipc/name_distr.c spin_lock_bh(&tn->nametbl_lock); tn 325 net/tipc/name_distr.c spin_unlock_bh(&tn->nametbl_lock); tn 338 net/tipc/name_distr.c struct tipc_net *tn = tipc_net(net); tn 342 net/tipc/name_distr.c spin_lock_bh(&tn->nametbl_lock); tn 349 net/tipc/name_distr.c spin_unlock_bh(&tn->nametbl_lock); tn 425 net/tipc/name_table.c struct tipc_net *tn = tipc_net(net); tn 426 net/tipc/name_table.c bool legacy = tn->legacy_addr_format; tn 615 net/tipc/name_table.c struct tipc_net *tn = tipc_net(net); tn 619 net/tipc/name_table.c spin_lock_bh(&tn->nametbl_lock); tn 633 net/tipc/name_table.c spin_unlock_bh(&tn->nametbl_lock); tn 647 net/tipc/name_table.c struct tipc_net *tn = tipc_net(net); tn 652 net/tipc/name_table.c spin_lock_bh(&tn->nametbl_lock); tn 664 net/tipc/name_table.c spin_unlock_bh(&tn->nametbl_lock); tn 679 net/tipc/name_table.c struct tipc_net *tn = tipc_net(sub->net); tn 685 net/tipc/name_table.c spin_lock_bh(&tn->nametbl_lock); tn 699 net/tipc/name_table.c spin_unlock_bh(&tn->nametbl_lock); tn 708 net/tipc/name_table.c struct tipc_net *tn = tipc_net(sub->net); tn 713 net/tipc/name_table.c spin_lock_bh(&tn->nametbl_lock); tn 729 net/tipc/name_table.c spin_unlock_bh(&tn->nametbl_lock); tn 734 net/tipc/name_table.c struct tipc_net *tn = tipc_net(net); tn 748 net/tipc/name_table.c tn->nametbl = nt; tn 749 net/tipc/name_table.c spin_lock_init(&tn->nametbl_lock); tn 778 net/tipc/name_table.c struct tipc_net *tn = tipc_net(net); tn 786 net/tipc/name_table.c spin_lock_bh(&tn->nametbl_lock); tn 795 net/tipc/name_table.c spin_unlock_bh(&tn->nametbl_lock); tn 898 net/tipc/name_table.c struct tipc_net *tn = tipc_net(net); tn 910 net/tipc/name_table.c head = &tn->nametbl->services[i]; tn 133 net/tipc/net.c struct tipc_net *tn = tipc_net(net); tn 135 net/tipc/net.c if (cmpxchg(&tn->node_addr, 0, addr)) tn 181 net/tipc/net.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 182 net/tipc/net.c u64 *w0 = (u64 *)&tn->node_id[0]; tn 183 net/tipc/net.c u64 *w1 = (u64 *)&tn->node_id[8]; tn 196 net/tipc/net.c if (nla_put_u32(msg->skb, TIPC_NLA_NET_ID, tn->net_id)) tn 244 net/tipc/net.c struct tipc_net *tn = tipc_net(net); tn 268 net/tipc/net.c tn->net_id = val; tn 277 net/tipc/net.c tn->legacy_addr_format = true; tn 261 net/tipc/node.c struct tipc_net *tn = tipc_net(net); tn 266 net/tipc/node.c hlist_for_each_entry_rcu(node, &tn->node_htable[thash], hash) { tn 283 net/tipc/node.c struct tipc_net *tn = tipc_net(net); tn 288 net/tipc/node.c list_for_each_entry_rcu(n, &tn->node_list, list) { tn 366 net/tipc/node.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 372 net/tipc/node.c spin_lock_bh(&tn->node_list_lock); tn 388 net/tipc/node.c tn->capabilities = TIPC_NODE_CAPABILITIES; tn 389 net/tipc/node.c list_for_each_entry_rcu(temp_node, &tn->node_list, list) { tn 390 net/tipc/node.c tn->capabilities &= temp_node->capabilities; tn 436 net/tipc/node.c hlist_add_head_rcu(&n->hash, &tn->node_htable[tipc_hashfn(addr)]); tn 437 net/tipc/node.c list_for_each_entry_rcu(temp_node, &tn->node_list, list) { tn 443 net/tipc/node.c tn->capabilities = TIPC_NODE_CAPABILITIES; tn 444 net/tipc/node.c list_for_each_entry_rcu(temp_node, &tn->node_list, list) { tn 445 net/tipc/node.c tn->capabilities &= temp_node->capabilities; tn 449 net/tipc/node.c spin_unlock_bh(&tn->node_list_lock); tn 484 net/tipc/node.c struct tipc_net *tn = tipc_net(net); tn 487 net/tipc/node.c spin_lock_bh(&tn->node_list_lock); tn 488 net/tipc/node.c list_for_each_entry_safe(node, t_node, &tn->node_list, list) tn 490 net/tipc/node.c spin_unlock_bh(&tn->node_list_lock); tn 604 net/tipc/node.c struct tipc_net *tn = tipc_net(peer->net); tn 608 net/tipc/node.c if (!spin_trylock_bh(&tn->node_list_lock)) tn 621 net/tipc/node.c tn->capabilities = TIPC_NODE_CAPABILITIES; tn 622 net/tipc/node.c list_for_each_entry_rcu(temp_node, &tn->node_list, list) { tn 623 net/tipc/node.c tn->capabilities &= temp_node->capabilities; tn 626 net/tipc/node.c spin_unlock_bh(&tn->node_list_lock); tn 951 net/tipc/node.c struct tipc_net *tn = tipc_net(net); tn 974 net/tipc/node.c if (tn->trial_addr == addr) tn 1119 net/tipc/node.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 1123 net/tipc/node.c list_for_each_entry_rcu(n, &tn->node_list, list) { tn 1897 net/tipc/node.c struct tipc_net *tn = tipc_net(net); tn 1907 net/tipc/node.c list_for_each_entry_rcu(n, &tn->node_list, list) { tn 1927 net/tipc/node.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 1951 net/tipc/node.c spin_lock_bh(&tn->node_list_lock); tn 1954 net/tipc/node.c spin_unlock_bh(&tn->node_list_lock); tn 1973 net/tipc/node.c spin_unlock_bh(&tn->node_list_lock); tn 1982 net/tipc/node.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 2013 net/tipc/node.c list_for_each_entry_rcu(node, &tn->node_list, list) { tn 2051 net/tipc/node.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 2059 net/tipc/node.c list_for_each_entry_rcu(n, &tn->node_list, list) { tn 2299 net/tipc/node.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 2329 net/tipc/node.c list_for_each_entry_continue_rcu(node, &tn->node_list, tn 2345 net/tipc/node.c list_for_each_entry_rcu(node, &tn->node_list, list) { tn 2780 net/tipc/socket.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 2785 net/tipc/socket.c rhashtable_walk_enter(&tn->sk_rht, &iter); tn 2810 net/tipc/socket.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 2814 net/tipc/socket.c tsk = rhashtable_lookup_fast(&tn->sk_rht, &portid, tsk_rht_params); tn 2826 net/tipc/socket.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 2836 net/tipc/socket.c if (!rhashtable_lookup_insert_fast(&tn->sk_rht, &tsk->node, tn 2848 net/tipc/socket.c struct tipc_net *tn = net_generic(sock_net(sk), tipc_net_id); tn 2850 net/tipc/socket.c if (!rhashtable_remove_fast(&tn->sk_rht, &tsk->node, tsk_rht_params)) { tn 2868 net/tipc/socket.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 2870 net/tipc/socket.c return rhashtable_init(&tn->sk_rht, &tsk_rht_params); tn 2875 net/tipc/socket.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 2880 net/tipc/socket.c rhashtable_destroy(&tn->sk_rht); tn 3410 net/tipc/socket.c struct tipc_net *tn = tipc_net(net); tn 3420 net/tipc/socket.c rhashtable_walk_enter(&tn->sk_rht, iter); tn 231 net/tipc/topsrv.c struct tipc_net *tn = tipc_net(con->server->net); tn 239 net/tipc/topsrv.c atomic_dec(&tn->subscription_count); tn 363 net/tipc/topsrv.c struct tipc_net *tn = tipc_net(srv->net); tn 371 net/tipc/topsrv.c if (atomic_read(&tn->subscription_count) >= TIPC_MAX_SUBSCR) { tn 378 net/tipc/topsrv.c atomic_inc(&tn->subscription_count); tn 643 net/tipc/topsrv.c struct tipc_net *tn = tipc_net(net); tn 656 net/tipc/topsrv.c tn->topsrv = srv; tn 657 net/tipc/topsrv.c atomic_set(&tn->subscription_count, 0); tn 490 net/tipc/udp_media.c struct tipc_net *tn = net_generic(net, tipc_net_id); tn 493 net/tipc/udp_media.c b = rtnl_dereference(tn->bearer_list[bid]);