tn                211 arch/arm/vfp/vfpdouble.c 	int tn, tm = 0;
tn                213 arch/arm/vfp/vfpdouble.c 	tn = vfp_double_type(vdn);
tn                229 arch/arm/vfp/vfpdouble.c 		if (tn == VFP_SNAN || (tm != VFP_SNAN && tn == VFP_QNAN))
tn                244 arch/arm/vfp/vfpdouble.c 	return tn == VFP_SNAN || tm == VFP_SNAN ? FPSCR_IOC : VFP_NAN_FLAG;
tn                688 arch/arm/vfp/vfpdouble.c 	int tn, tm;
tn                690 arch/arm/vfp/vfpdouble.c 	tn = vfp_double_type(vdn);
tn                693 arch/arm/vfp/vfpdouble.c 	if (tn & tm & VFP_INFINITY) {
tn                709 arch/arm/vfp/vfpdouble.c 	} else if (tn & VFP_INFINITY && tm & VFP_NUMBER) {
tn               1011 arch/arm/vfp/vfpdouble.c 	int tm, tn;
tn               1018 arch/arm/vfp/vfpdouble.c 	tn = vfp_double_type(&vdn);
tn               1024 arch/arm/vfp/vfpdouble.c 	if (tn & VFP_NAN)
tn               1037 arch/arm/vfp/vfpdouble.c 	if (tm & tn & (VFP_INFINITY|VFP_ZERO))
tn               1043 arch/arm/vfp/vfpdouble.c 	if (tn & VFP_INFINITY)
tn               1055 arch/arm/vfp/vfpdouble.c 	if (tm & VFP_INFINITY || tn & VFP_ZERO)
tn               1058 arch/arm/vfp/vfpdouble.c 	if (tn & VFP_DENORMAL)
tn                219 arch/arm/vfp/vfpsingle.c 	int tn, tm = 0;
tn                221 arch/arm/vfp/vfpsingle.c 	tn = vfp_single_type(vsn);
tn                237 arch/arm/vfp/vfpsingle.c 		if (tn == VFP_SNAN || (tm != VFP_SNAN && tn == VFP_QNAN))
tn                252 arch/arm/vfp/vfpsingle.c 	return tn == VFP_SNAN || tm == VFP_SNAN ? FPSCR_IOC : VFP_NAN_FLAG;
tn                734 arch/arm/vfp/vfpsingle.c 	int tn, tm;
tn                736 arch/arm/vfp/vfpsingle.c 	tn = vfp_single_type(vsn);
tn                739 arch/arm/vfp/vfpsingle.c 	if (tn & tm & VFP_INFINITY) {
tn                755 arch/arm/vfp/vfpsingle.c 	} else if (tn & VFP_INFINITY && tm & VFP_NUMBER) {
tn               1057 arch/arm/vfp/vfpsingle.c 	int tm, tn;
tn               1066 arch/arm/vfp/vfpsingle.c 	tn = vfp_single_type(&vsn);
tn               1072 arch/arm/vfp/vfpsingle.c 	if (tn & VFP_NAN)
tn               1085 arch/arm/vfp/vfpsingle.c 	if (tm & tn & (VFP_INFINITY|VFP_ZERO))
tn               1091 arch/arm/vfp/vfpsingle.c 	if (tn & VFP_INFINITY)
tn               1103 arch/arm/vfp/vfpsingle.c 	if (tm & VFP_INFINITY || tn & VFP_ZERO)
tn               1106 arch/arm/vfp/vfpsingle.c 	if (tn & VFP_DENORMAL)
tn                 87 drivers/ata/pata_mpc52xx.c 	u8 tn;
tn                 91 drivers/ata/pata_mpc52xx.c 	{ .t0M = 32, .td = 15, .th = 2, .tj = 2, .tkw = 15, .tm = 4, .tn = 1 },
tn                 92 drivers/ata/pata_mpc52xx.c 	{ .t0M = 10, .td = 6,  .th = 1, .tj = 1, .tkw = 4,  .tm = 2, .tn = 1 },
tn                 93 drivers/ata/pata_mpc52xx.c 	{ .t0M = 8,  .td = 5,  .th = 1, .tj = 1, .tkw = 2,  .tm = 2, .tn = 1 },
tn                 97 drivers/ata/pata_mpc52xx.c 	{ .t0M = 64, .td = 29, .th = 3, .tj = 3, .tkw = 29, .tm = 7, .tn = 2 },
tn                 98 drivers/ata/pata_mpc52xx.c 	{ .t0M = 20, .td = 11, .th = 2, .tj = 1, .tkw = 7,  .tm = 4, .tn = 1 },
tn                 99 drivers/ata/pata_mpc52xx.c 	{ .t0M = 16, .td = 10, .th = 2, .tj = 1, .tkw = 4,  .tm = 4, .tn = 1 },
tn                303 drivers/ata/pata_mpc52xx.c 	t->mdma2 = ((u32)s->th << 24) | ((u32)s->tj << 16) | ((u32)s->tn << 8);
tn                451 drivers/bluetooth/bt3c_cs.c 	unsigned int iobase, tmp, tn;
tn                493 drivers/bluetooth/bt3c_cs.c 			if (kstrtouint(b, 16, &tn))
tn                495 drivers/bluetooth/bt3c_cs.c 			tmp += tn;
tn               1520 drivers/gpu/drm/i915/i915_request.c 	struct intel_timeline *tl, *tn;
tn               1525 drivers/gpu/drm/i915/i915_request.c 	list_for_each_entry_safe(tl, tn, &timelines->active_list, link) {
tn               1539 drivers/gpu/drm/i915/i915_request.c 		list_safe_reset_next(tl, tn, link);
tn               1553 drivers/gpu/drm/i915/i915_request.c 	list_for_each_entry_safe(tl, tn, &free, link)
tn                750 drivers/infiniband/ulp/ipoib/ipoib_main.c 	struct ipoib_neigh *neigh, *tn;
tn                806 drivers/infiniband/ulp/ipoib/ipoib_main.c 		list_for_each_entry_safe(neigh, tn, &path->neigh_list, list) {
tn               1907 drivers/scsi/ncr53c8xx.c static	void	ncr_alloc_ccb	(struct ncb *np, u_char tn, u_char ln);
tn               1912 drivers/scsi/ncr53c8xx.c static	void	ncr_init_tcb	(struct ncb *np, u_char tn);
tn               1913 drivers/scsi/ncr53c8xx.c static	struct lcb *	ncr_alloc_lcb	(struct ncb *np, u_char tn, u_char ln);
tn               5625 drivers/scsi/ncr53c8xx.c 	unsigned char tn = sdev->id, ln = sdev->lun;
tn               5626 drivers/scsi/ncr53c8xx.c 	struct tcb *tp = &np->target[tn];
tn               7137 drivers/scsi/ncr53c8xx.c 	u_char tn = cmd->device->id;
tn               7139 drivers/scsi/ncr53c8xx.c 	struct tcb *tp = &np->target[tn];
tn               7159 drivers/scsi/ncr53c8xx.c 			ncr_alloc_ccb(np, tn, ln);
tn               7227 drivers/scsi/ncr53c8xx.c 	cp->target = tn;
tn               7347 drivers/scsi/ncr53c8xx.c static void ncr_alloc_ccb(struct ncb *np, u_char tn, u_char ln)
tn               7349 drivers/scsi/ncr53c8xx.c 	struct tcb *tp = &np->target[tn];
tn               7396 drivers/scsi/ncr53c8xx.c static void ncr_init_tcb (struct ncb *np, u_char tn)
tn               7398 drivers/scsi/ncr53c8xx.c 	struct tcb *tp = &np->target[tn];
tn               7400 drivers/scsi/ncr53c8xx.c 	int th = tn & 3;
tn               7408 drivers/scsi/ncr53c8xx.c 		cpu_to_scr((SCR_JUMP ^ IFFALSE (DATA (0x80 + tn))));
tn               7483 drivers/scsi/ncr53c8xx.c static struct lcb *ncr_alloc_lcb (struct ncb *np, u_char tn, u_char ln)
tn               7485 drivers/scsi/ncr53c8xx.c 	struct tcb *tp = &np->target[tn];
tn               7509 drivers/scsi/ncr53c8xx.c 		ncr_init_tcb(np, tn);
tn               7575 drivers/scsi/ncr53c8xx.c 	unsigned char tn = sdev->id, ln = sdev->lun;
tn               7576 drivers/scsi/ncr53c8xx.c 	struct tcb *tp = &np->target[tn];
tn               7580 drivers/scsi/ncr53c8xx.c 	if (!lp && !(lp = ncr_alloc_lcb(np, tn, ln)))
tn                 45 drivers/scsi/sym53c8xx_2/sym_hipd.c static void sym_alloc_lcb_tags (struct sym_hcb *np, u_char tn, u_char ln);
tn               4647 drivers/scsi/sym53c8xx_2/sym_hipd.c 	u_char tn = cmd->device->id;
tn               4649 drivers/scsi/sym53c8xx_2/sym_hipd.c 	struct sym_tcb *tp = &np->target[tn];
tn               4681 drivers/scsi/sym53c8xx_2/sym_hipd.c 				sym_alloc_lcb_tags(np, tn, ln);
tn               4755 drivers/scsi/sym53c8xx_2/sym_hipd.c 	cp->target = tn;
tn               4952 drivers/scsi/sym53c8xx_2/sym_hipd.c static void sym_init_tcb (struct sym_hcb *np, u_char tn)
tn               4968 drivers/scsi/sym53c8xx_2/sym_hipd.c struct sym_lcb *sym_alloc_lcb (struct sym_hcb *np, u_char tn, u_char ln)
tn               4970 drivers/scsi/sym53c8xx_2/sym_hipd.c 	struct sym_tcb *tp = &np->target[tn];
tn               4976 drivers/scsi/sym53c8xx_2/sym_hipd.c 	sym_init_tcb (np, tn);
tn               5049 drivers/scsi/sym53c8xx_2/sym_hipd.c static void sym_alloc_lcb_tags (struct sym_hcb *np, u_char tn, u_char ln)
tn               5051 drivers/scsi/sym53c8xx_2/sym_hipd.c 	struct sym_tcb *tp = &np->target[tn];
tn               5095 drivers/scsi/sym53c8xx_2/sym_hipd.c int sym_free_lcb(struct sym_hcb *np, u_char tn, u_char ln)
tn               5097 drivers/scsi/sym53c8xx_2/sym_hipd.c 	struct sym_tcb *tp = &np->target[tn];
tn               1055 drivers/scsi/sym53c8xx_2/sym_hipd.h struct sym_lcb *sym_alloc_lcb(struct sym_hcb *np, u_char tn, u_char ln);
tn               1056 drivers/scsi/sym53c8xx_2/sym_hipd.h int sym_free_lcb(struct sym_hcb *np, u_char tn, u_char ln);
tn                 84 drivers/scsi/sym53c8xx_2/sym_nvram.c 	Symbios_target *tn = &nvram->target[target];
tn                 86 drivers/scsi/sym53c8xx_2/sym_nvram.c 	if (!(tn->flags & SYMBIOS_QUEUE_TAGS_ENABLED))
tn                 88 drivers/scsi/sym53c8xx_2/sym_nvram.c 	if (!(tn->flags & SYMBIOS_DISCONNECT_ENABLE))
tn                 90 drivers/scsi/sym53c8xx_2/sym_nvram.c 	if (!(tn->flags & SYMBIOS_SCAN_AT_BOOT_TIME))
tn                 92 drivers/scsi/sym53c8xx_2/sym_nvram.c 	if (!(tn->flags & SYMBIOS_SCAN_LUNS))
tn                 94 drivers/scsi/sym53c8xx_2/sym_nvram.c 	tp->usr_period = (tn->sync_period + 3) / 4;
tn                 95 drivers/scsi/sym53c8xx_2/sym_nvram.c 	tp->usr_width = (tn->bus_width == 0x8) ? 0 : 1;
tn                108 drivers/scsi/sym53c8xx_2/sym_nvram.c 	struct Tekram_target *tn = &nvram->target[target];
tn                110 drivers/scsi/sym53c8xx_2/sym_nvram.c 	if (tn->flags & TEKRAM_TAGGED_COMMANDS) {
tn                114 drivers/scsi/sym53c8xx_2/sym_nvram.c 	if (tn->flags & TEKRAM_DISCONNECT_ENABLE)
tn                117 drivers/scsi/sym53c8xx_2/sym_nvram.c 	if (tn->flags & TEKRAM_SYNC_NEGO)
tn                118 drivers/scsi/sym53c8xx_2/sym_nvram.c 		tp->usr_period = Tekram_sync[tn->sync_index & 0xf];
tn                119 drivers/scsi/sym53c8xx_2/sym_nvram.c 	tp->usr_width = (tn->flags & TEKRAM_WIDE_NEGO) ? 1 : 0;
tn                159 drivers/scsi/sym53c8xx_2/sym_nvram.c 		struct Symbios_target *tn = &nvram->target[i];
tn                162 drivers/scsi/sym53c8xx_2/sym_nvram.c 		(tn->flags & SYMBIOS_DISCONNECT_ENABLE)	? " DISC"	: "",
tn                163 drivers/scsi/sym53c8xx_2/sym_nvram.c 		(tn->flags & SYMBIOS_SCAN_AT_BOOT_TIME)	? " SCAN_BOOT"	: "",
tn                164 drivers/scsi/sym53c8xx_2/sym_nvram.c 		(tn->flags & SYMBIOS_SCAN_LUNS)		? " SCAN_LUNS"	: "",
tn                165 drivers/scsi/sym53c8xx_2/sym_nvram.c 		(tn->flags & SYMBIOS_QUEUE_TAGS_ENABLED)? " TCQ"	: "",
tn                166 drivers/scsi/sym53c8xx_2/sym_nvram.c 		tn->bus_width,
tn                167 drivers/scsi/sym53c8xx_2/sym_nvram.c 		tn->sync_period / 4,
tn                168 drivers/scsi/sym53c8xx_2/sym_nvram.c 		tn->timeout);
tn                207 drivers/scsi/sym53c8xx_2/sym_nvram.c 		struct Tekram_target *tn = &nvram->target[i];
tn                208 drivers/scsi/sym53c8xx_2/sym_nvram.c 		j = tn->sync_index & 0xf;
tn                212 drivers/scsi/sym53c8xx_2/sym_nvram.c 		(tn->flags & TEKRAM_PARITY_CHECK)	? " PARITY"	: "",
tn                213 drivers/scsi/sym53c8xx_2/sym_nvram.c 		(tn->flags & TEKRAM_SYNC_NEGO)		? " SYNC"	: "",
tn                214 drivers/scsi/sym53c8xx_2/sym_nvram.c 		(tn->flags & TEKRAM_DISCONNECT_ENABLE)	? " DISC"	: "",
tn                215 drivers/scsi/sym53c8xx_2/sym_nvram.c 		(tn->flags & TEKRAM_START_CMD)		? " START"	: "",
tn                216 drivers/scsi/sym53c8xx_2/sym_nvram.c 		(tn->flags & TEKRAM_TAGGED_COMMANDS)	? " TCQ"	: "",
tn                217 drivers/scsi/sym53c8xx_2/sym_nvram.c 		(tn->flags & TEKRAM_WIDE_NEGO)		? " WIDE"	: "",
tn                112 fs/cifs/cifs_debug.h 	const char *tn = "";					\
tn                114 fs/cifs/cifs_debug.h 		tn = tcon->treeName;				\
tn                117 fs/cifs/cifs_debug.h 			__FILE__, tn, ##__VA_ARGS__);		\
tn                120 fs/cifs/cifs_debug.h 			tn, ##__VA_ARGS__);			\
tn                123 fs/cifs/cifs_debug.h 			tn, ##__VA_ARGS__);			\
tn                354 fs/jffs2/nodelist.h #define tn_next(tn) rb_entry(rb_next(&(tn)->rb), struct jffs2_tmp_dnode_info, rb)
tn                355 fs/jffs2/nodelist.h #define tn_prev(tn) rb_entry(rb_prev(&(tn)->rb), struct jffs2_tmp_dnode_info, rb)
tn                356 fs/jffs2/nodelist.h #define tn_parent(tn) rb_entry(rb_parent(&(tn)->rb), struct jffs2_tmp_dnode_info, rb)
tn                357 fs/jffs2/nodelist.h #define tn_left(tn) rb_entry((tn)->rb.rb_left, struct jffs2_tmp_dnode_info, rb)
tn                358 fs/jffs2/nodelist.h #define tn_right(tn) rb_entry((tn)->rb.rb_right, struct jffs2_tmp_dnode_info, rb)
tn                359 fs/jffs2/nodelist.h #define tn_erase(tn, list) rb_erase(&tn->rb, list);
tn                 31 fs/jffs2/readinode.c static int check_node_data(struct jffs2_sb_info *c, struct jffs2_tmp_dnode_info *tn)
tn                 33 fs/jffs2/readinode.c 	struct jffs2_raw_node_ref *ref = tn->fn->raw;
tn                 40 fs/jffs2/readinode.c 	BUG_ON(tn->csize == 0);
tn                 44 fs/jffs2/readinode.c 	len = tn->csize;
tn                 51 fs/jffs2/readinode.c 		if (adj >= tn->csize) {
tn                 53 fs/jffs2/readinode.c 				      ref_offset(ref), tn->csize, ofs);
tn                 62 fs/jffs2/readinode.c 		ref_offset(ref), tn->csize, tn->partial_crc, tn->data_crc, ofs - len, ofs, len);
tn                 69 fs/jffs2/readinode.c 		JFFS2_WARNING("MTD point returned len too short: %zu instead of %u.\n", retlen, tn->csize);
tn                 99 fs/jffs2/readinode.c 	crc = crc32(tn->partial_crc, buffer, len);
tn                107 fs/jffs2/readinode.c 	if (crc != tn->data_crc) {
tn                109 fs/jffs2/readinode.c 			     ref_offset(ref), tn->data_crc, crc);
tn                150 fs/jffs2/readinode.c static int check_tn_node(struct jffs2_sb_info *c, struct jffs2_tmp_dnode_info *tn)
tn                154 fs/jffs2/readinode.c 	BUG_ON(ref_obsolete(tn->fn->raw));
tn                157 fs/jffs2/readinode.c 	if (ref_flags(tn->fn->raw) != REF_UNCHECKED)
tn                161 fs/jffs2/readinode.c 		      tn->fn->ofs, tn->fn->ofs + tn->fn->size, ref_offset(tn->fn->raw));
tn                163 fs/jffs2/readinode.c 	ret = check_node_data(c, tn);
tn                169 fs/jffs2/readinode.c 		jffs2_mark_node_obsolete(c, tn->fn->raw);
tn                178 fs/jffs2/readinode.c 	struct jffs2_tmp_dnode_info *tn = NULL;
tn                185 fs/jffs2/readinode.c 		tn = rb_entry(next, struct jffs2_tmp_dnode_info, rb);
tn                187 fs/jffs2/readinode.c 		if (tn->fn->ofs < offset)
tn                188 fs/jffs2/readinode.c 			next = tn->rb.rb_right;
tn                189 fs/jffs2/readinode.c 		else if (tn->fn->ofs >= offset)
tn                190 fs/jffs2/readinode.c 			next = tn->rb.rb_left;
tn                195 fs/jffs2/readinode.c 	return tn;
tn                199 fs/jffs2/readinode.c static void jffs2_kill_tn(struct jffs2_sb_info *c, struct jffs2_tmp_dnode_info *tn)
tn                201 fs/jffs2/readinode.c 	jffs2_mark_node_obsolete(c, tn->fn->raw);
tn                202 fs/jffs2/readinode.c 	jffs2_free_full_dnode(tn->fn);
tn                203 fs/jffs2/readinode.c 	jffs2_free_tmp_dnode_info(tn);
tn                220 fs/jffs2/readinode.c 				struct jffs2_tmp_dnode_info *tn)
tn                222 fs/jffs2/readinode.c 	uint32_t fn_end = tn->fn->ofs + tn->fn->size;
tn                225 fs/jffs2/readinode.c 	dbg_readinode("insert fragment %#04x-%#04x, ver %u at %08x\n", tn->fn->ofs, fn_end, tn->version, ref_offset(tn->fn->raw));
tn                231 fs/jffs2/readinode.c 	if (!tn->fn->size) {
tn                233 fs/jffs2/readinode.c 			if (rii->mdata_tn->version < tn->version) {
tn                239 fs/jffs2/readinode.c 					      tn->version, rii->mdata_tn->version);
tn                240 fs/jffs2/readinode.c 				jffs2_kill_tn(c, tn);
tn                244 fs/jffs2/readinode.c 		rii->mdata_tn = tn;
tn                245 fs/jffs2/readinode.c 		dbg_readinode("keep new mdata with ver %d\n", tn->version);
tn                250 fs/jffs2/readinode.c 	this = jffs2_lookup_tn(&rii->tn_root, tn->fn->ofs);
tn                275 fs/jffs2/readinode.c 		if (this->version == tn->version) {
tn                281 fs/jffs2/readinode.c 				jffs2_kill_tn(c, tn);
tn                286 fs/jffs2/readinode.c 				rb_replace_node(&this->rb, &tn->rb, &rii->tn_root);
tn                292 fs/jffs2/readinode.c 		if (this->version < tn->version &&
tn                293 fs/jffs2/readinode.c 		    this->fn->ofs >= tn->fn->ofs &&
tn                296 fs/jffs2/readinode.c 			if (check_tn_node(c, tn)) {
tn                298 fs/jffs2/readinode.c 				jffs2_kill_tn(c, tn);
tn                304 fs/jffs2/readinode.c 				if (this->version < tn->version) {
tn                316 fs/jffs2/readinode.c 		if (this->version > tn->version &&
tn                317 fs/jffs2/readinode.c 		    this->fn->ofs <= tn->fn->ofs &&
tn                322 fs/jffs2/readinode.c 				jffs2_kill_tn(c, tn);
tn                345 fs/jffs2/readinode.c 			if (tn->fn->ofs > insert_point->fn->ofs)
tn                347 fs/jffs2/readinode.c 			else if (tn->fn->ofs < insert_point->fn->ofs ||
tn                348 fs/jffs2/readinode.c 				 tn->fn->size < insert_point->fn->size)
tn                353 fs/jffs2/readinode.c 		rb_link_node(&tn->rb, &insert_point->rb, link);
tn                354 fs/jffs2/readinode.c 		rb_insert_color(&tn->rb, &rii->tn_root);
tn                358 fs/jffs2/readinode.c 	this = tn_prev(tn);
tn                361 fs/jffs2/readinode.c 			if (this->fn->ofs + this->fn->size > tn->fn->ofs) {
tn                365 fs/jffs2/readinode.c 				tn->overlapped = 1;
tn                385 fs/jffs2/readinode.c 	this = tn_next(tn);
tn                424 fs/jffs2/readinode.c static void ver_insert(struct rb_root *ver_root, struct jffs2_tmp_dnode_info *tn)
tn                434 fs/jffs2/readinode.c 		if (tn->version > this_tn->version)
tn                440 fs/jffs2/readinode.c 	rb_link_node(&tn->rb, parent, link);
tn                441 fs/jffs2/readinode.c 	rb_insert_color(&tn->rb, ver_root);
tn                546 fs/jffs2/readinode.c 	struct jffs2_tmp_dnode_info *tn, *next;
tn                548 fs/jffs2/readinode.c 	rbtree_postorder_for_each_entry_safe(tn, next, list, rb) {
tn                549 fs/jffs2/readinode.c 			jffs2_free_full_dnode(tn->fn);
tn                550 fs/jffs2/readinode.c 			jffs2_free_tmp_dnode_info(tn);
tn                701 fs/jffs2/readinode.c 	struct jffs2_tmp_dnode_info *tn;
tn                717 fs/jffs2/readinode.c 	tn = jffs2_alloc_tmp_dnode_info();
tn                718 fs/jffs2/readinode.c 	if (!tn) {
tn                719 fs/jffs2/readinode.c 		JFFS2_ERROR("failed to allocate tn (%zu bytes).\n", sizeof(*tn));
tn                723 fs/jffs2/readinode.c 	tn->partial_crc = 0;
tn                780 fs/jffs2/readinode.c 			tn->partial_crc = crc32(0, buf, len);
tn                782 fs/jffs2/readinode.c 			dbg_readinode("Calculates CRC (%#08x) for %d bytes, csize %d\n", tn->partial_crc, len, csize);
tn                786 fs/jffs2/readinode.c 			if (len >= csize && unlikely(tn->partial_crc != je32_to_cpu(rd->data_crc))) {
tn                788 fs/jffs2/readinode.c 					ref_offset(ref), tn->partial_crc, je32_to_cpu(rd->data_crc));
tn                816 fs/jffs2/readinode.c 	tn->fn = jffs2_alloc_full_dnode();
tn                817 fs/jffs2/readinode.c 	if (!tn->fn) {
tn                823 fs/jffs2/readinode.c 	tn->version = je32_to_cpu(rd->version);
tn                824 fs/jffs2/readinode.c 	tn->fn->ofs = je32_to_cpu(rd->offset);
tn                825 fs/jffs2/readinode.c 	tn->data_crc = je32_to_cpu(rd->data_crc);
tn                826 fs/jffs2/readinode.c 	tn->csize = csize;
tn                827 fs/jffs2/readinode.c 	tn->fn->raw = ref;
tn                828 fs/jffs2/readinode.c 	tn->overlapped = 0;
tn                830 fs/jffs2/readinode.c 	if (tn->version > rii->highest_version)
tn                831 fs/jffs2/readinode.c 		rii->highest_version = tn->version;
tn                836 fs/jffs2/readinode.c 		tn->fn->size = csize;
tn                838 fs/jffs2/readinode.c 		tn->fn->size = je32_to_cpu(rd->dsize);
tn                844 fs/jffs2/readinode.c 	ret = jffs2_add_tn_to_tree(c, rii, tn);
tn                847 fs/jffs2/readinode.c 		jffs2_free_full_dnode(tn->fn);
tn                849 fs/jffs2/readinode.c 		jffs2_free_tmp_dnode_info(tn);
tn                854 fs/jffs2/readinode.c 	tn = tn_first(&rii->tn_root);
tn                855 fs/jffs2/readinode.c 	while (tn) {
tn                857 fs/jffs2/readinode.c 			       tn, tn->version, tn->fn->ofs,
tn                858 fs/jffs2/readinode.c 			       tn->fn->ofs+tn->fn->size, tn->overlapped);
tn                859 fs/jffs2/readinode.c 		tn = tn_next(tn);
tn                118 include/net/act_api.h int tc_action_net_init(struct net *net, struct tc_action_net *tn,
tn                123 include/net/act_api.h 	tn->idrinfo = kmalloc(sizeof(*tn->idrinfo), GFP_KERNEL);
tn                124 include/net/act_api.h 	if (!tn->idrinfo)
tn                126 include/net/act_api.h 	tn->ops = ops;
tn                127 include/net/act_api.h 	tn->idrinfo->net = net;
tn                128 include/net/act_api.h 	mutex_init(&tn->idrinfo->lock);
tn                129 include/net/act_api.h 	idr_init(&tn->idrinfo->action_idr);
tn                143 include/net/act_api.h 		struct tc_action_net *tn = net_generic(net, id);
tn                145 include/net/act_api.h 		tcf_idrinfo_destroy(tn->ops, tn->idrinfo);
tn                146 include/net/act_api.h 		kfree(tn->idrinfo);
tn                151 include/net/act_api.h int tcf_generic_walker(struct tc_action_net *tn, struct sk_buff *skb,
tn                155 include/net/act_api.h int tcf_idr_search(struct tc_action_net *tn, struct tc_action **a, u32 index);
tn                156 include/net/act_api.h int tcf_idr_create(struct tc_action_net *tn, u32 index, struct nlattr *est,
tn                159 include/net/act_api.h void tcf_idr_insert(struct tc_action_net *tn, struct tc_action *a);
tn                161 include/net/act_api.h void tcf_idr_cleanup(struct tc_action_net *tn, u32 index);
tn                162 include/net/act_api.h int tcf_idr_check_alloc(struct tc_action_net *tn, u32 *index,
tn               1207 kernel/trace/ftrace.c 	struct hlist_node *tn;
tn               1217 kernel/trace/ftrace.c 		hlist_for_each_entry_safe(entry, tn, hhd, hlist)
tn               1377 kernel/trace/ftrace.c 	struct hlist_node *tn;
tn               1409 kernel/trace/ftrace.c 		hlist_for_each_entry_safe(entry, tn, hhd, hlist) {
tn                173 net/ipv4/fib_trie.c static struct key_vector *resize(struct trie *t, struct key_vector *tn);
tn                194 net/ipv4/fib_trie.c #define node_parent(tn) rtnl_dereference(tn_info(tn)->parent)
tn                195 net/ipv4/fib_trie.c #define get_child(tn, i) rtnl_dereference((tn)->tnode[i])
tn                198 net/ipv4/fib_trie.c #define node_parent_rcu(tn) rcu_dereference_rtnl(tn_info(tn)->parent)
tn                199 net/ipv4/fib_trie.c #define get_child_rcu(tn, i) rcu_dereference_rtnl((tn)->tnode[i])
tn                213 net/ipv4/fib_trie.c static inline unsigned long child_length(const struct key_vector *tn)
tn                215 net/ipv4/fib_trie.c 	return (1ul << tn->bits) & ~(1ul);
tn                381 net/ipv4/fib_trie.c 	struct key_vector *tn;
tn                399 net/ipv4/fib_trie.c 	tn = tnode->kv;
tn                400 net/ipv4/fib_trie.c 	tn->key = (shift < KEYLENGTH) ? (key >> shift) << shift : 0;
tn                401 net/ipv4/fib_trie.c 	tn->pos = pos;
tn                402 net/ipv4/fib_trie.c 	tn->bits = bits;
tn                403 net/ipv4/fib_trie.c 	tn->slen = pos;
tn                405 net/ipv4/fib_trie.c 	return tn;
tn                411 net/ipv4/fib_trie.c static inline int tnode_full(struct key_vector *tn, struct key_vector *n)
tn                413 net/ipv4/fib_trie.c 	return n && ((n->pos + n->bits) == tn->pos) && IS_TNODE(n);
tn                419 net/ipv4/fib_trie.c static void put_child(struct key_vector *tn, unsigned long i,
tn                422 net/ipv4/fib_trie.c 	struct key_vector *chi = get_child(tn, i);
tn                425 net/ipv4/fib_trie.c 	BUG_ON(i >= child_length(tn));
tn                429 net/ipv4/fib_trie.c 		empty_child_inc(tn);
tn                431 net/ipv4/fib_trie.c 		empty_child_dec(tn);
tn                434 net/ipv4/fib_trie.c 	wasfull = tnode_full(tn, chi);
tn                435 net/ipv4/fib_trie.c 	isfull = tnode_full(tn, n);
tn                438 net/ipv4/fib_trie.c 		tn_info(tn)->full_children--;
tn                440 net/ipv4/fib_trie.c 		tn_info(tn)->full_children++;
tn                442 net/ipv4/fib_trie.c 	if (n && (tn->slen < n->slen))
tn                443 net/ipv4/fib_trie.c 		tn->slen = n->slen;
tn                445 net/ipv4/fib_trie.c 	rcu_assign_pointer(tn->tnode[i], n);
tn                448 net/ipv4/fib_trie.c static void update_children(struct key_vector *tn)
tn                453 net/ipv4/fib_trie.c 	for (i = child_length(tn); i;) {
tn                454 net/ipv4/fib_trie.c 		struct key_vector *inode = get_child(tn, --i);
tn                463 net/ipv4/fib_trie.c 		if (node_parent(inode) == tn)
tn                466 net/ipv4/fib_trie.c 			node_set_parent(inode, tn);
tn                479 net/ipv4/fib_trie.c static inline void tnode_free_init(struct key_vector *tn)
tn                481 net/ipv4/fib_trie.c 	tn_info(tn)->rcu.next = NULL;
tn                484 net/ipv4/fib_trie.c static inline void tnode_free_append(struct key_vector *tn,
tn                487 net/ipv4/fib_trie.c 	tn_info(n)->rcu.next = tn_info(tn)->rcu.next;
tn                488 net/ipv4/fib_trie.c 	tn_info(tn)->rcu.next = &tn_info(n)->rcu;
tn                491 net/ipv4/fib_trie.c static void tnode_free(struct key_vector *tn)
tn                493 net/ipv4/fib_trie.c 	struct callback_head *head = &tn_info(tn)->rcu;
tn                497 net/ipv4/fib_trie.c 		tnode_free_size += TNODE_SIZE(1ul << tn->bits);
tn                498 net/ipv4/fib_trie.c 		node_free(tn);
tn                500 net/ipv4/fib_trie.c 		tn = container_of(head, struct tnode, rcu)->kv;
tn                511 net/ipv4/fib_trie.c 				  struct key_vector *tn)
tn                517 net/ipv4/fib_trie.c 	NODE_INIT_PARENT(tn, tp);
tn                518 net/ipv4/fib_trie.c 	put_child_root(tp, tn->key, tn);
tn                521 net/ipv4/fib_trie.c 	update_children(tn);
tn                527 net/ipv4/fib_trie.c 	for (i = child_length(tn); i;) {
tn                528 net/ipv4/fib_trie.c 		struct key_vector *inode = get_child(tn, --i);
tn                531 net/ipv4/fib_trie.c 		if (tnode_full(tn, inode))
tn                532 net/ipv4/fib_trie.c 			tn = resize(t, inode);
tn                541 net/ipv4/fib_trie.c 	struct key_vector *tn;
tn                547 net/ipv4/fib_trie.c 	tn = tnode_new(oldtnode->key, oldtnode->pos - 1, oldtnode->bits + 1);
tn                548 net/ipv4/fib_trie.c 	if (!tn)
tn                559 net/ipv4/fib_trie.c 	for (i = child_length(oldtnode), m = 1u << tn->pos; i;) {
tn                570 net/ipv4/fib_trie.c 			put_child(tn, get_index(inode->key, tn), inode);
tn                579 net/ipv4/fib_trie.c 			put_child(tn, 2 * i + 1, get_child(inode, 1));
tn                580 net/ipv4/fib_trie.c 			put_child(tn, 2 * i, get_child(inode, 0));
tn                603 net/ipv4/fib_trie.c 		tnode_free_append(tn, node1);
tn                606 net/ipv4/fib_trie.c 		tnode_free_append(tn, node0);
tn                617 net/ipv4/fib_trie.c 		NODE_INIT_PARENT(node1, tn);
tn                618 net/ipv4/fib_trie.c 		NODE_INIT_PARENT(node0, tn);
tn                621 net/ipv4/fib_trie.c 		put_child(tn, 2 * i + 1, node1);
tn                622 net/ipv4/fib_trie.c 		put_child(tn, 2 * i, node0);
tn                626 net/ipv4/fib_trie.c 	return replace(t, oldtnode, tn);
tn                629 net/ipv4/fib_trie.c 	tnode_free(tn);
tn                637 net/ipv4/fib_trie.c 	struct key_vector *tn;
tn                642 net/ipv4/fib_trie.c 	tn = tnode_new(oldtnode->key, oldtnode->pos + 1, oldtnode->bits - 1);
tn                643 net/ipv4/fib_trie.c 	if (!tn)
tn                661 net/ipv4/fib_trie.c 			put_child(tn, i / 2, node1 ? : node0);
tn                669 net/ipv4/fib_trie.c 		tnode_free_append(tn, inode);
tn                674 net/ipv4/fib_trie.c 		NODE_INIT_PARENT(inode, tn);
tn                677 net/ipv4/fib_trie.c 		put_child(tn, i / 2, inode);
tn                681 net/ipv4/fib_trie.c 	return replace(t, oldtnode, tn);
tn                684 net/ipv4/fib_trie.c 	tnode_free(tn);
tn                710 net/ipv4/fib_trie.c static unsigned char update_suffix(struct key_vector *tn)
tn                712 net/ipv4/fib_trie.c 	unsigned char slen = tn->pos;
tn                720 net/ipv4/fib_trie.c 	slen_max = min_t(unsigned char, tn->pos + tn->bits - 1, tn->slen);
tn                727 net/ipv4/fib_trie.c 	for (i = 0, stride = 0x2ul ; i < child_length(tn); i += stride) {
tn                728 net/ipv4/fib_trie.c 		struct key_vector *n = get_child(tn, i);
tn                743 net/ipv4/fib_trie.c 	tn->slen = slen;
tn                805 net/ipv4/fib_trie.c static inline bool should_inflate(struct key_vector *tp, struct key_vector *tn)
tn                807 net/ipv4/fib_trie.c 	unsigned long used = child_length(tn);
tn                812 net/ipv4/fib_trie.c 	used -= tn_info(tn)->empty_children;
tn                813 net/ipv4/fib_trie.c 	used += tn_info(tn)->full_children;
tn                817 net/ipv4/fib_trie.c 	return (used > 1) && tn->pos && ((50 * used) >= threshold);
tn                820 net/ipv4/fib_trie.c static inline bool should_halve(struct key_vector *tp, struct key_vector *tn)
tn                822 net/ipv4/fib_trie.c 	unsigned long used = child_length(tn);
tn                827 net/ipv4/fib_trie.c 	used -= tn_info(tn)->empty_children;
tn                831 net/ipv4/fib_trie.c 	return (used > 1) && (tn->bits > 1) && ((100 * used) < threshold);
tn                834 net/ipv4/fib_trie.c static inline bool should_collapse(struct key_vector *tn)
tn                836 net/ipv4/fib_trie.c 	unsigned long used = child_length(tn);
tn                838 net/ipv4/fib_trie.c 	used -= tn_info(tn)->empty_children;
tn                841 net/ipv4/fib_trie.c 	if ((tn->bits == KEYLENGTH) && tn_info(tn)->full_children)
tn                849 net/ipv4/fib_trie.c static struct key_vector *resize(struct trie *t, struct key_vector *tn)
tn                854 net/ipv4/fib_trie.c 	struct key_vector *tp = node_parent(tn);
tn                855 net/ipv4/fib_trie.c 	unsigned long cindex = get_index(tn->key, tp);
tn                859 net/ipv4/fib_trie.c 		 tn, inflate_threshold, halve_threshold);
tn                865 net/ipv4/fib_trie.c 	BUG_ON(tn != get_child(tp, cindex));
tn                870 net/ipv4/fib_trie.c 	while (should_inflate(tp, tn) && max_work) {
tn                871 net/ipv4/fib_trie.c 		tp = inflate(t, tn);
tn                880 net/ipv4/fib_trie.c 		tn = get_child(tp, cindex);
tn                884 net/ipv4/fib_trie.c 	tp = node_parent(tn);
tn                893 net/ipv4/fib_trie.c 	while (should_halve(tp, tn) && max_work) {
tn                894 net/ipv4/fib_trie.c 		tp = halve(t, tn);
tn                903 net/ipv4/fib_trie.c 		tn = get_child(tp, cindex);
tn                907 net/ipv4/fib_trie.c 	if (should_collapse(tn))
tn                908 net/ipv4/fib_trie.c 		return collapse(t, tn);
tn                911 net/ipv4/fib_trie.c 	return node_parent(tn);
tn                914 net/ipv4/fib_trie.c static void node_pull_suffix(struct key_vector *tn, unsigned char slen)
tn                916 net/ipv4/fib_trie.c 	unsigned char node_slen = tn->slen;
tn                918 net/ipv4/fib_trie.c 	while ((node_slen > tn->pos) && (node_slen > slen)) {
tn                919 net/ipv4/fib_trie.c 		slen = update_suffix(tn);
tn                923 net/ipv4/fib_trie.c 		tn = node_parent(tn);
tn                924 net/ipv4/fib_trie.c 		node_slen = tn->slen;
tn                928 net/ipv4/fib_trie.c static void node_push_suffix(struct key_vector *tn, unsigned char slen)
tn                930 net/ipv4/fib_trie.c 	while (tn->slen < slen) {
tn                931 net/ipv4/fib_trie.c 		tn->slen = slen;
tn                932 net/ipv4/fib_trie.c 		tn = node_parent(tn);
tn               1008 net/ipv4/fib_trie.c static void trie_rebalance(struct trie *t, struct key_vector *tn)
tn               1010 net/ipv4/fib_trie.c 	while (!IS_TRIE(tn))
tn               1011 net/ipv4/fib_trie.c 		tn = resize(t, tn);
tn               1033 net/ipv4/fib_trie.c 		struct key_vector *tn;
tn               1035 net/ipv4/fib_trie.c 		tn = tnode_new(key, __fls(key ^ n->key), 1);
tn               1036 net/ipv4/fib_trie.c 		if (!tn)
tn               1040 net/ipv4/fib_trie.c 		NODE_INIT_PARENT(tn, tp);
tn               1041 net/ipv4/fib_trie.c 		put_child(tn, get_index(key, tn) ^ 1, n);
tn               1044 net/ipv4/fib_trie.c 		put_child_root(tp, key, tn);
tn               1045 net/ipv4/fib_trie.c 		node_set_parent(n, tn);
tn               1048 net/ipv4/fib_trie.c 		tp = tn;
tn               1618 net/ipv4/fib_trie.c static struct key_vector *leaf_walk_rcu(struct key_vector **tn, t_key key)
tn               1620 net/ipv4/fib_trie.c 	struct key_vector *pn, *n = *tn;
tn               1667 net/ipv4/fib_trie.c 	*tn = pn;
tn               1671 net/ipv4/fib_trie.c 	*tn = pn;
tn                470 net/netfilter/nf_conntrack_proto_tcp.c 	struct nf_tcp_net *tn = nf_tcp_pernet(net);
tn                675 net/netfilter/nf_conntrack_proto_tcp.c 		    tn->tcp_be_liberal)
tn                764 net/netfilter/nf_conntrack_proto_tcp.c 	const struct nf_tcp_net *tn = nf_tcp_pernet(net);
tn                790 net/netfilter/nf_conntrack_proto_tcp.c 	} else if (tn->tcp_loose == 0) {
tn                844 net/netfilter/nf_conntrack_proto_tcp.c 	struct nf_tcp_net *tn = nf_tcp_pernet(net);
tn               1115 net/netfilter/nf_conntrack_proto_tcp.c 		timeouts = tn->timeouts;
tn               1117 net/netfilter/nf_conntrack_proto_tcp.c 	if (ct->proto.tcp.retrans >= tn->tcp_max_retrans &&
tn               1310 net/netfilter/nf_conntrack_proto_tcp.c 	struct nf_tcp_net *tn = nf_tcp_pernet(net);
tn               1315 net/netfilter/nf_conntrack_proto_tcp.c 		timeouts = tn->timeouts;
tn               1318 net/netfilter/nf_conntrack_proto_tcp.c 		timeouts[i] = tn->timeouts[i];
tn               1421 net/netfilter/nf_conntrack_proto_tcp.c 	struct nf_tcp_net *tn = nf_tcp_pernet(net);
tn               1425 net/netfilter/nf_conntrack_proto_tcp.c 		tn->timeouts[i] = tcp_timeouts[i];
tn               1430 net/netfilter/nf_conntrack_proto_tcp.c 	tn->timeouts[0] = tcp_timeouts[TCP_CONNTRACK_SYN_SENT];
tn               1431 net/netfilter/nf_conntrack_proto_tcp.c 	tn->tcp_loose = nf_ct_tcp_loose;
tn               1432 net/netfilter/nf_conntrack_proto_tcp.c 	tn->tcp_be_liberal = nf_ct_tcp_be_liberal;
tn               1433 net/netfilter/nf_conntrack_proto_tcp.c 	tn->tcp_max_retrans = nf_ct_tcp_max_retrans;
tn                940 net/netfilter/nf_conntrack_standalone.c 	struct nf_tcp_net *tn = nf_tcp_pernet(net);
tn                942 net/netfilter/nf_conntrack_standalone.c #define XASSIGN(XNAME, tn) \
tn                944 net/netfilter/nf_conntrack_standalone.c 			&(tn)->timeouts[TCP_CONNTRACK_ ## XNAME]
tn                946 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(SYN_SENT, tn);
tn                947 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(SYN_RECV, tn);
tn                948 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(ESTABLISHED, tn);
tn                949 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(FIN_WAIT, tn);
tn                950 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(CLOSE_WAIT, tn);
tn                951 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(LAST_ACK, tn);
tn                952 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(TIME_WAIT, tn);
tn                953 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(CLOSE, tn);
tn                954 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(RETRANS, tn);
tn                955 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(UNACK, tn);
tn                960 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(LOOSE, &tn->tcp_loose);
tn                961 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(LIBERAL, &tn->tcp_be_liberal);
tn                962 net/netfilter/nf_conntrack_standalone.c 	XASSIGN(MAX_RETRANS, &tn->tcp_max_retrans);
tn                 65 net/netfilter/xt_TEE.c 	struct tee_net *tn = net_generic(net, tee_net_id);
tn                 68 net/netfilter/xt_TEE.c 	mutex_lock(&tn->lock);
tn                 69 net/netfilter/xt_TEE.c 	list_for_each_entry(priv, &tn->priv_list, list) {
tn                 87 net/netfilter/xt_TEE.c 	mutex_unlock(&tn->lock);
tn                 94 net/netfilter/xt_TEE.c 	struct tee_net *tn = net_generic(par->net, tee_net_id);
tn                122 net/netfilter/xt_TEE.c 		mutex_lock(&tn->lock);
tn                123 net/netfilter/xt_TEE.c 		list_add(&priv->list, &tn->priv_list);
tn                124 net/netfilter/xt_TEE.c 		mutex_unlock(&tn->lock);
tn                134 net/netfilter/xt_TEE.c 	struct tee_net *tn = net_generic(par->net, tee_net_id);
tn                138 net/netfilter/xt_TEE.c 		mutex_lock(&tn->lock);
tn                140 net/netfilter/xt_TEE.c 		mutex_unlock(&tn->lock);
tn                175 net/netfilter/xt_TEE.c 	struct tee_net *tn = net_generic(net, tee_net_id);
tn                177 net/netfilter/xt_TEE.c 	INIT_LIST_HEAD(&tn->priv_list);
tn                178 net/netfilter/xt_TEE.c 	mutex_init(&tn->lock);
tn                327 net/sched/act_api.c int tcf_generic_walker(struct tc_action_net *tn, struct sk_buff *skb,
tn                332 net/sched/act_api.c 	struct tcf_idrinfo *idrinfo = tn->idrinfo;
tn                346 net/sched/act_api.c int tcf_idr_search(struct tc_action_net *tn, struct tc_action **a, u32 index)
tn                348 net/sched/act_api.c 	struct tcf_idrinfo *idrinfo = tn->idrinfo;
tn                400 net/sched/act_api.c int tcf_idr_create(struct tc_action_net *tn, u32 index, struct nlattr *est,
tn                405 net/sched/act_api.c 	struct tcf_idrinfo *idrinfo = tn->idrinfo;
tn                454 net/sched/act_api.c void tcf_idr_insert(struct tc_action_net *tn, struct tc_action *a)
tn                456 net/sched/act_api.c 	struct tcf_idrinfo *idrinfo = tn->idrinfo;
tn                467 net/sched/act_api.c void tcf_idr_cleanup(struct tc_action_net *tn, u32 index)
tn                469 net/sched/act_api.c 	struct tcf_idrinfo *idrinfo = tn->idrinfo;
tn                484 net/sched/act_api.c int tcf_idr_check_alloc(struct tc_action_net *tn, u32 *index,
tn                487 net/sched/act_api.c 	struct tcf_idrinfo *idrinfo = tn->idrinfo;
tn                280 net/sched/act_bpf.c 	struct tc_action_net *tn = net_generic(net, bpf_net_id);
tn                303 net/sched/act_bpf.c 	ret = tcf_idr_check_alloc(tn, &index, act, bind);
tn                305 net/sched/act_bpf.c 		ret = tcf_idr_create(tn, index, est, act,
tn                308 net/sched/act_bpf.c 			tcf_idr_cleanup(tn, index);
tn                365 net/sched/act_bpf.c 		tcf_idr_insert(tn, *act);
tn                396 net/sched/act_bpf.c 	struct tc_action_net *tn = net_generic(net, bpf_net_id);
tn                398 net/sched/act_bpf.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                403 net/sched/act_bpf.c 	struct tc_action_net *tn = net_generic(net, bpf_net_id);
tn                405 net/sched/act_bpf.c 	return tcf_idr_search(tn, a, index);
tn                423 net/sched/act_bpf.c 	struct tc_action_net *tn = net_generic(net, bpf_net_id);
tn                425 net/sched/act_bpf.c 	return tc_action_net_init(net, tn, &act_bpf_ops);
tn                100 net/sched/act_connmark.c 	struct tc_action_net *tn = net_generic(net, connmark_net_id);
tn                121 net/sched/act_connmark.c 	ret = tcf_idr_check_alloc(tn, &index, a, bind);
tn                123 net/sched/act_connmark.c 		ret = tcf_idr_create(tn, index, est, a,
tn                126 net/sched/act_connmark.c 			tcf_idr_cleanup(tn, index);
tn                139 net/sched/act_connmark.c 		tcf_idr_insert(tn, *a);
tn                206 net/sched/act_connmark.c 	struct tc_action_net *tn = net_generic(net, connmark_net_id);
tn                208 net/sched/act_connmark.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                213 net/sched/act_connmark.c 	struct tc_action_net *tn = net_generic(net, connmark_net_id);
tn                215 net/sched/act_connmark.c 	return tcf_idr_search(tn, a, index);
tn                232 net/sched/act_connmark.c 	struct tc_action_net *tn = net_generic(net, connmark_net_id);
tn                234 net/sched/act_connmark.c 	return tc_action_net_init(net, tn, &act_connmark_ops);
tn                 48 net/sched/act_csum.c 	struct tc_action_net *tn = net_generic(net, csum_net_id);
tn                 69 net/sched/act_csum.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                 71 net/sched/act_csum.c 		ret = tcf_idr_create(tn, index, est, a,
tn                 74 net/sched/act_csum.c 			tcf_idr_cleanup(tn, index);
tn                114 net/sched/act_csum.c 		tcf_idr_insert(tn, *a);
tn                682 net/sched/act_csum.c 	struct tc_action_net *tn = net_generic(net, csum_net_id);
tn                684 net/sched/act_csum.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                689 net/sched/act_csum.c 	struct tc_action_net *tn = net_generic(net, csum_net_id);
tn                691 net/sched/act_csum.c 	return tcf_idr_search(tn, a, index);
tn                715 net/sched/act_csum.c 	struct tc_action_net *tn = net_generic(net, csum_net_id);
tn                717 net/sched/act_csum.c 	return tc_action_net_init(net, tn, &act_csum_ops);
tn                 38 net/sched/act_ct.c 	struct tc_action_net tn; /* Must be first */
tn                593 net/sched/act_ct.c 	struct tc_ct_action_net *tn = net_generic(net, ct_net_id);
tn                629 net/sched/act_ct.c 		if (!tn->labels) {
tn                673 net/sched/act_ct.c 	struct tc_action_net *tn = net_generic(net, ct_net_id);
tn                697 net/sched/act_ct.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                702 net/sched/act_ct.c 		err = tcf_idr_create(tn, index, est, a,
tn                705 net/sched/act_ct.c 			tcf_idr_cleanup(tn, index);
tn                744 net/sched/act_ct.c 		tcf_idr_insert(tn, *a);
tn                902 net/sched/act_ct.c 	struct tc_action_net *tn = net_generic(net, ct_net_id);
tn                904 net/sched/act_ct.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                909 net/sched/act_ct.c 	struct tc_action_net *tn = net_generic(net, ct_net_id);
tn                911 net/sched/act_ct.c 	return tcf_idr_search(tn, a, index);
tn                944 net/sched/act_ct.c 	struct tc_ct_action_net *tn = net_generic(net, ct_net_id);
tn                947 net/sched/act_ct.c 		tn->labels = false;
tn                950 net/sched/act_ct.c 		tn->labels = true;
tn                953 net/sched/act_ct.c 	return tc_action_net_init(net, &tn->tn, &act_ct_ops);
tn                962 net/sched/act_ct.c 		struct tc_ct_action_net *tn = net_generic(net, ct_net_id);
tn                964 net/sched/act_ct.c 		if (tn->labels)
tn                159 net/sched/act_ctinfo.c 	struct tc_action_net *tn = net_generic(net, ctinfo_net_id);
tn                210 net/sched/act_ctinfo.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                212 net/sched/act_ctinfo.c 		ret = tcf_idr_create(tn, index, est, a,
tn                215 net/sched/act_ctinfo.c 			tcf_idr_cleanup(tn, index);
tn                270 net/sched/act_ctinfo.c 		tcf_idr_insert(tn, *a);
tn                351 net/sched/act_ctinfo.c 	struct tc_action_net *tn = net_generic(net, ctinfo_net_id);
tn                353 net/sched/act_ctinfo.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                358 net/sched/act_ctinfo.c 	struct tc_action_net *tn = net_generic(net, ctinfo_net_id);
tn                360 net/sched/act_ctinfo.c 	return tcf_idr_search(tn, a, index);
tn                388 net/sched/act_ctinfo.c 	struct tc_action_net *tn = net_generic(net, ctinfo_net_id);
tn                390 net/sched/act_ctinfo.c 	return tc_action_net_init(net, tn, &act_ctinfo_ops);
tn                 58 net/sched/act_gact.c 	struct tc_action_net *tn = net_generic(net, gact_net_id);
tn                 99 net/sched/act_gact.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                101 net/sched/act_gact.c 		ret = tcf_idr_create(tn, index, est, a,
tn                104 net/sched/act_gact.c 			tcf_idr_cleanup(tn, index);
tn                143 net/sched/act_gact.c 		tcf_idr_insert(tn, *a);
tn                238 net/sched/act_gact.c 	struct tc_action_net *tn = net_generic(net, gact_net_id);
tn                240 net/sched/act_gact.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                245 net/sched/act_gact.c 	struct tc_action_net *tn = net_generic(net, gact_net_id);
tn                247 net/sched/act_gact.c 	return tcf_idr_search(tn, a, index);
tn                279 net/sched/act_gact.c 	struct tc_action_net *tn = net_generic(net, gact_net_id);
tn                281 net/sched/act_gact.c 	return tc_action_net_init(net, tn, &act_gact_ops);
tn                470 net/sched/act_ife.c 	struct tc_action_net *tn = net_generic(net, ife_net_id);
tn                512 net/sched/act_ife.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                524 net/sched/act_ife.c 		ret = tcf_idr_create(tn, index, est, a, &act_ife_ops,
tn                527 net/sched/act_ife.c 			tcf_idr_cleanup(tn, index);
tn                606 net/sched/act_ife.c 		tcf_idr_insert(tn, *a);
tn                863 net/sched/act_ife.c 	struct tc_action_net *tn = net_generic(net, ife_net_id);
tn                865 net/sched/act_ife.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                870 net/sched/act_ife.c 	struct tc_action_net *tn = net_generic(net, ife_net_id);
tn                872 net/sched/act_ife.c 	return tcf_idr_search(tn, a, index);
tn                890 net/sched/act_ife.c 	struct tc_action_net *tn = net_generic(net, ife_net_id);
tn                892 net/sched/act_ife.c 	return tc_action_net_init(net, tn, &act_ife_ops);
tn                100 net/sched/act_ipt.c 	struct tc_action_net *tn = net_generic(net, id);
tn                121 net/sched/act_ipt.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                132 net/sched/act_ipt.c 			tcf_idr_cleanup(tn, index);
tn                141 net/sched/act_ipt.c 			tcf_idr_cleanup(tn, index);
tn                146 net/sched/act_ipt.c 		ret = tcf_idr_create(tn, index, est, a, ops, bind,
tn                149 net/sched/act_ipt.c 			tcf_idr_cleanup(tn, index);
tn                193 net/sched/act_ipt.c 		tcf_idr_insert(tn, *a);
tn                325 net/sched/act_ipt.c 	struct tc_action_net *tn = net_generic(net, ipt_net_id);
tn                327 net/sched/act_ipt.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                332 net/sched/act_ipt.c 	struct tc_action_net *tn = net_generic(net, ipt_net_id);
tn                334 net/sched/act_ipt.c 	return tcf_idr_search(tn, a, index);
tn                352 net/sched/act_ipt.c 	struct tc_action_net *tn = net_generic(net, ipt_net_id);
tn                354 net/sched/act_ipt.c 	return tc_action_net_init(net, tn, &act_ipt_ops);
tn                374 net/sched/act_ipt.c 	struct tc_action_net *tn = net_generic(net, xt_net_id);
tn                376 net/sched/act_ipt.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                381 net/sched/act_ipt.c 	struct tc_action_net *tn = net_generic(net, xt_net_id);
tn                383 net/sched/act_ipt.c 	return tcf_idr_search(tn, a, index);
tn                401 net/sched/act_ipt.c 	struct tc_action_net *tn = net_generic(net, xt_net_id);
tn                403 net/sched/act_ipt.c 	return tc_action_net_init(net, tn, &act_xt_ops);
tn                 98 net/sched/act_mirred.c 	struct tc_action_net *tn = net_generic(net, mirred_net_id);
tn                123 net/sched/act_mirred.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                140 net/sched/act_mirred.c 			tcf_idr_cleanup(tn, index);
tn                147 net/sched/act_mirred.c 			tcf_idr_cleanup(tn, index);
tn                151 net/sched/act_mirred.c 		ret = tcf_idr_create(tn, index, est, a,
tn                154 net/sched/act_mirred.c 			tcf_idr_cleanup(tn, index);
tn                198 net/sched/act_mirred.c 		tcf_idr_insert(tn, *a);
tn                369 net/sched/act_mirred.c 	struct tc_action_net *tn = net_generic(net, mirred_net_id);
tn                371 net/sched/act_mirred.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                376 net/sched/act_mirred.c 	struct tc_action_net *tn = net_generic(net, mirred_net_id);
tn                378 net/sched/act_mirred.c 	return tcf_idr_search(tn, a, index);
tn                459 net/sched/act_mirred.c 	struct tc_action_net *tn = net_generic(net, mirred_net_id);
tn                461 net/sched/act_mirred.c 	return tc_action_net_init(net, tn, &act_mirred_ops);
tn                139 net/sched/act_mpls.c 	struct tc_action_net *tn = net_generic(net, mpls_net_id);
tn                221 net/sched/act_mpls.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                229 net/sched/act_mpls.c 		ret = tcf_idr_create(tn, index, est, a,
tn                232 net/sched/act_mpls.c 			tcf_idr_cleanup(tn, index);
tn                277 net/sched/act_mpls.c 		tcf_idr_insert(tn, *a);
tn                356 net/sched/act_mpls.c 	struct tc_action_net *tn = net_generic(net, mpls_net_id);
tn                358 net/sched/act_mpls.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                363 net/sched/act_mpls.c 	struct tc_action_net *tn = net_generic(net, mpls_net_id);
tn                365 net/sched/act_mpls.c 	return tcf_idr_search(tn, a, index);
tn                383 net/sched/act_mpls.c 	struct tc_action_net *tn = net_generic(net, mpls_net_id);
tn                385 net/sched/act_mpls.c 	return tc_action_net_init(net, tn, &act_mpls_ops);
tn                 41 net/sched/act_nat.c 	struct tc_action_net *tn = net_generic(net, nat_net_id);
tn                 61 net/sched/act_nat.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                 63 net/sched/act_nat.c 		ret = tcf_idr_create(tn, index, est, a,
tn                 66 net/sched/act_nat.c 			tcf_idr_cleanup(tn, index);
tn                 97 net/sched/act_nat.c 		tcf_idr_insert(tn, *a);
tn                302 net/sched/act_nat.c 	struct tc_action_net *tn = net_generic(net, nat_net_id);
tn                304 net/sched/act_nat.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                309 net/sched/act_nat.c 	struct tc_action_net *tn = net_generic(net, nat_net_id);
tn                311 net/sched/act_nat.c 	return tcf_idr_search(tn, a, index);
tn                328 net/sched/act_nat.c 	struct tc_action_net *tn = net_generic(net, nat_net_id);
tn                330 net/sched/act_nat.c 	return tc_action_net_init(net, tn, &act_nat_ops);
tn                142 net/sched/act_pedit.c 	struct tc_action_net *tn = net_generic(net, pedit_net_id);
tn                188 net/sched/act_pedit.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                190 net/sched/act_pedit.c 		ret = tcf_idr_create(tn, index, est, a,
tn                193 net/sched/act_pedit.c 			tcf_idr_cleanup(tn, index);
tn                241 net/sched/act_pedit.c 		tcf_idr_insert(tn, *a);
tn                470 net/sched/act_pedit.c 	struct tc_action_net *tn = net_generic(net, pedit_net_id);
tn                472 net/sched/act_pedit.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                477 net/sched/act_pedit.c 	struct tc_action_net *tn = net_generic(net, pedit_net_id);
tn                479 net/sched/act_pedit.c 	return tcf_idr_search(tn, a, index);
tn                497 net/sched/act_pedit.c 	struct tc_action_net *tn = net_generic(net, pedit_net_id);
tn                499 net/sched/act_pedit.c 	return tc_action_net_init(net, tn, &act_pedit_ops);
tn                 33 net/sched/act_police.c 	struct tc_action_net *tn = net_generic(net, police_net_id);
tn                 35 net/sched/act_police.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                 59 net/sched/act_police.c 	struct tc_action_net *tn = net_generic(net, police_net_id);
tn                 81 net/sched/act_police.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                 89 net/sched/act_police.c 		ret = tcf_idr_create(tn, index, NULL, a,
tn                 92 net/sched/act_police.c 			tcf_idr_cleanup(tn, index);
tn                205 net/sched/act_police.c 		tcf_idr_insert(tn, *a);
tn                366 net/sched/act_police.c 	struct tc_action_net *tn = net_generic(net, police_net_id);
tn                368 net/sched/act_police.c 	return tcf_idr_search(tn, a, index);
tn                391 net/sched/act_police.c 	struct tc_action_net *tn = net_generic(net, police_net_id);
tn                393 net/sched/act_police.c 	return tc_action_net_init(net, tn, &act_police_ops);
tn                 41 net/sched/act_sample.c 	struct tc_action_net *tn = net_generic(net, sample_net_id);
tn                 63 net/sched/act_sample.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                 71 net/sched/act_sample.c 		ret = tcf_idr_create(tn, index, est, a,
tn                 74 net/sched/act_sample.c 			tcf_idr_cleanup(tn, index);
tn                120 net/sched/act_sample.c 		tcf_idr_insert(tn, *a);
tn                244 net/sched/act_sample.c 	struct tc_action_net *tn = net_generic(net, sample_net_id);
tn                246 net/sched/act_sample.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                251 net/sched/act_sample.c 	struct tc_action_net *tn = net_generic(net, sample_net_id);
tn                253 net/sched/act_sample.c 	return tcf_idr_search(tn, a, index);
tn                298 net/sched/act_sample.c 	struct tc_action_net *tn = net_generic(net, sample_net_id);
tn                300 net/sched/act_sample.c 	return tc_action_net_init(net, tn, &act_sample_ops);
tn                 91 net/sched/act_simple.c 	struct tc_action_net *tn = net_generic(net, simp_net_id);
tn                113 net/sched/act_simple.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                124 net/sched/act_simple.c 			tcf_idr_cleanup(tn, index);
tn                129 net/sched/act_simple.c 		ret = tcf_idr_create(tn, index, est, a,
tn                132 net/sched/act_simple.c 			tcf_idr_cleanup(tn, index);
tn                160 net/sched/act_simple.c 		tcf_idr_insert(tn, *a);
tn                206 net/sched/act_simple.c 	struct tc_action_net *tn = net_generic(net, simp_net_id);
tn                208 net/sched/act_simple.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                213 net/sched/act_simple.c 	struct tc_action_net *tn = net_generic(net, simp_net_id);
tn                215 net/sched/act_simple.c 	return tcf_idr_search(tn, a, index);
tn                233 net/sched/act_simple.c 	struct tc_action_net *tn = net_generic(net, simp_net_id);
tn                235 net/sched/act_simple.c 	return tc_action_net_init(net, tn, &act_simp_ops);
tn                 92 net/sched/act_skbedit.c 	struct tc_action_net *tn = net_generic(net, skbedit_net_id);
tn                151 net/sched/act_skbedit.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                162 net/sched/act_skbedit.c 			tcf_idr_cleanup(tn, index);
tn                167 net/sched/act_skbedit.c 		ret = tcf_idr_create(tn, index, est, a,
tn                170 net/sched/act_skbedit.c 			tcf_idr_cleanup(tn, index);
tn                218 net/sched/act_skbedit.c 		tcf_idr_insert(tn, *a);
tn                298 net/sched/act_skbedit.c 	struct tc_action_net *tn = net_generic(net, skbedit_net_id);
tn                300 net/sched/act_skbedit.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                305 net/sched/act_skbedit.c 	struct tc_action_net *tn = net_generic(net, skbedit_net_id);
tn                307 net/sched/act_skbedit.c 	return tcf_idr_search(tn, a, index);
tn                337 net/sched/act_skbedit.c 	struct tc_action_net *tn = net_generic(net, skbedit_net_id);
tn                339 net/sched/act_skbedit.c 	return tc_action_net_init(net, tn, &act_skbedit_ops);
tn                 85 net/sched/act_skbmod.c 	struct tc_action_net *tn = net_generic(net, skbmod_net_id);
tn                129 net/sched/act_skbmod.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                140 net/sched/act_skbmod.c 			tcf_idr_cleanup(tn, index);
tn                145 net/sched/act_skbmod.c 		ret = tcf_idr_create(tn, index, est, a,
tn                148 net/sched/act_skbmod.c 			tcf_idr_cleanup(tn, index);
tn                194 net/sched/act_skbmod.c 		tcf_idr_insert(tn, *a);
tn                261 net/sched/act_skbmod.c 	struct tc_action_net *tn = net_generic(net, skbmod_net_id);
tn                263 net/sched/act_skbmod.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                268 net/sched/act_skbmod.c 	struct tc_action_net *tn = net_generic(net, skbmod_net_id);
tn                270 net/sched/act_skbmod.c 	return tcf_idr_search(tn, a, index);
tn                288 net/sched/act_skbmod.c 	struct tc_action_net *tn = net_generic(net, skbmod_net_id);
tn                290 net/sched/act_skbmod.c 	return tc_action_net_init(net, tn, &act_skbmod_ops);
tn                218 net/sched/act_tunnel_key.c 	struct tc_action_net *tn = net_generic(net, tunnel_key_net_id);
tn                254 net/sched/act_tunnel_key.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                354 net/sched/act_tunnel_key.c 		ret = tcf_idr_create(tn, index, est, a,
tn                396 net/sched/act_tunnel_key.c 		tcf_idr_insert(tn, *a);
tn                412 net/sched/act_tunnel_key.c 		tcf_idr_cleanup(tn, index);
tn                578 net/sched/act_tunnel_key.c 	struct tc_action_net *tn = net_generic(net, tunnel_key_net_id);
tn                580 net/sched/act_tunnel_key.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                585 net/sched/act_tunnel_key.c 	struct tc_action_net *tn = net_generic(net, tunnel_key_net_id);
tn                587 net/sched/act_tunnel_key.c 	return tcf_idr_search(tn, a, index);
tn                605 net/sched/act_tunnel_key.c 	struct tc_action_net *tn = net_generic(net, tunnel_key_net_id);
tn                607 net/sched/act_tunnel_key.c 	return tc_action_net_init(net, tn, &act_tunnel_key_ops);
tn                107 net/sched/act_vlan.c 	struct tc_action_net *tn = net_generic(net, vlan_net_id);
tn                133 net/sched/act_vlan.c 	err = tcf_idr_check_alloc(tn, &index, a, bind);
tn                149 net/sched/act_vlan.c 				tcf_idr_cleanup(tn, index);
tn                157 net/sched/act_vlan.c 				tcf_idr_cleanup(tn, index);
tn                171 net/sched/act_vlan.c 					tcf_idr_cleanup(tn, index);
tn                185 net/sched/act_vlan.c 			tcf_idr_cleanup(tn, index);
tn                191 net/sched/act_vlan.c 		ret = tcf_idr_create(tn, index, est, a,
tn                194 net/sched/act_vlan.c 			tcf_idr_cleanup(tn, index);
tn                232 net/sched/act_vlan.c 		tcf_idr_insert(tn, *a);
tn                299 net/sched/act_vlan.c 	struct tc_action_net *tn = net_generic(net, vlan_net_id);
tn                301 net/sched/act_vlan.c 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
tn                319 net/sched/act_vlan.c 	struct tc_action_net *tn = net_generic(net, vlan_net_id);
tn                321 net/sched/act_vlan.c 	return tcf_idr_search(tn, a, index);
tn                349 net/sched/act_vlan.c 	struct tc_action_net *tn = net_generic(net, vlan_net_id);
tn                351 net/sched/act_vlan.c 	return tc_action_net_init(net, tn, &act_vlan_ops);
tn                872 net/sched/cls_api.c 	struct tcf_net *tn = net_generic(net, tcf_net_id);
tn                876 net/sched/cls_api.c 	spin_lock(&tn->idr_lock);
tn                877 net/sched/cls_api.c 	err = idr_alloc_u32(&tn->idr, block, &block->index, block->index,
tn                879 net/sched/cls_api.c 	spin_unlock(&tn->idr_lock);
tn                887 net/sched/cls_api.c 	struct tcf_net *tn = net_generic(net, tcf_net_id);
tn                889 net/sched/cls_api.c 	spin_lock(&tn->idr_lock);
tn                890 net/sched/cls_api.c 	idr_remove(&tn->idr, block->index);
tn                891 net/sched/cls_api.c 	spin_unlock(&tn->idr_lock);
tn                925 net/sched/cls_api.c 	struct tcf_net *tn = net_generic(net, tcf_net_id);
tn                927 net/sched/cls_api.c 	return idr_find(&tn->idr, block_index);
tn               3606 net/sched/cls_api.c 	struct tcf_net *tn = net_generic(net, tcf_net_id);
tn               3608 net/sched/cls_api.c 	spin_lock_init(&tn->idr_lock);
tn               3609 net/sched/cls_api.c 	idr_init(&tn->idr);
tn               3615 net/sched/cls_api.c 	struct tcf_net *tn = net_generic(net, tcf_net_id);
tn               3617 net/sched/cls_api.c 	idr_destroy(&tn->idr);
tn                 57 net/tipc/addr.c 	struct tipc_net *tn = tipc_net(net);
tn                 60 net/tipc/addr.c 	memcpy(tn->node_id, id, NODE_ID_LEN);
tn                 61 net/tipc/addr.c 	tipc_nodeid2string(tn->node_id_string, id);
tn                 62 net/tipc/addr.c 	tn->trial_addr = tmp[0] ^ tmp[1] ^ tmp[2] ^ tmp[3];
tn                 64 net/tipc/addr.c 		tipc_own_id_string(net), tn->net_id);
tn                 69 net/tipc/addr.c 	struct tipc_net *tn = tipc_net(net);
tn                 72 net/tipc/addr.c 	tn->node_addr = addr;
tn                 77 net/tipc/addr.c 	tn->trial_addr = addr;
tn                 78 net/tipc/addr.c 	tn->addr_trial_end = jiffies;
tn                 53 net/tipc/addr.h 	struct tipc_net *tn = tipc_net(net);
tn                 55 net/tipc/addr.h 	if (!strlen(tn->node_id_string))
tn                 57 net/tipc/addr.h 	return tn->node_id;
tn                679 net/tipc/bcast.c 	struct tipc_net *tn = tipc_net(net);
tn                686 net/tipc/bcast.c 	tn->bcbase = bb;
tn                699 net/tipc/bcast.c 	tn->bcl = l;
tn                711 net/tipc/bcast.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                714 net/tipc/bcast.c 	kfree(tn->bcbase);
tn                715 net/tipc/bcast.c 	kfree(tn->bcl);
tn                 63 net/tipc/bearer.c 	struct tipc_net *tn = tipc_net(net);
tn                 65 net/tipc/bearer.c 	return rcu_dereference(tn->bearer_list[bearer_id]);
tn                174 net/tipc/bearer.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                179 net/tipc/bearer.c 		b = rtnl_dereference(tn->bearer_list[i]);
tn                193 net/tipc/bearer.c 	struct tipc_net *tn = tipc_net(net);
tn                199 net/tipc/bearer.c 	b = rtnl_dereference(tn->bearer_list[bearer_id]);
tn                209 net/tipc/bearer.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                213 net/tipc/bearer.c 	b = rcu_dereference(tn->bearer_list[bearer_id]);
tn                221 net/tipc/bearer.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                225 net/tipc/bearer.c 	b = rcu_dereference(tn->bearer_list[bearer_id]);
tn                238 net/tipc/bearer.c 	struct tipc_net *tn = tipc_net(net);
tn                269 net/tipc/bearer.c 		b = rtnl_dereference(tn->bearer_list[bearer_id]);
tn                326 net/tipc/bearer.c 	rcu_assign_pointer(tn->bearer_list[bearer_id], b);
tn                361 net/tipc/bearer.c 	struct tipc_net *tn = tipc_net(net);
tn                371 net/tipc/bearer.c 	RCU_INIT_POINTER(tn->bearer_list[bearer_id], NULL);
tn                545 net/tipc/bearer.c 	struct tipc_net *tn = tipc_net(net);
tn                546 net/tipc/bearer.c 	int net_id = tn->net_id;
tn                669 net/tipc/bearer.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                674 net/tipc/bearer.c 		b = rtnl_dereference(tn->bearer_list[i]);
tn                677 net/tipc/bearer.c 			tn->bearer_list[i] = NULL;
tn                720 net/tipc/bearer.c 	struct tipc_net *tn = tipc_net(net);
tn                726 net/tipc/bearer.c 	tn->loopback_pt.dev = dev;
tn                727 net/tipc/bearer.c 	tn->loopback_pt.type = htons(ETH_P_TIPC);
tn                728 net/tipc/bearer.c 	tn->loopback_pt.func = tipc_loopback_rcv_pkt;
tn                729 net/tipc/bearer.c 	dev_add_pack(&tn->loopback_pt);
tn                735 net/tipc/bearer.c 	struct tipc_net *tn = tipc_net(net);
tn                737 net/tipc/bearer.c 	dev_remove_pack(&tn->loopback_pt);
tn                805 net/tipc/bearer.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                816 net/tipc/bearer.c 		bearer = rtnl_dereference(tn->bearer_list[i]);
tn                 54 net/tipc/core.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                 57 net/tipc/core.c 	tn->net_id = 4711;
tn                 58 net/tipc/core.c 	tn->node_addr = 0;
tn                 59 net/tipc/core.c 	tn->trial_addr = 0;
tn                 60 net/tipc/core.c 	tn->addr_trial_end = 0;
tn                 61 net/tipc/core.c 	tn->capabilities = TIPC_NODE_CAPABILITIES;
tn                 62 net/tipc/core.c 	memset(tn->node_id, 0, sizeof(tn->node_id));
tn                 63 net/tipc/core.c 	memset(tn->node_id_string, 0, sizeof(tn->node_id_string));
tn                 64 net/tipc/core.c 	tn->mon_threshold = TIPC_DEF_MON_THRESHOLD;
tn                 65 net/tipc/core.c 	get_random_bytes(&tn->random, sizeof(int));
tn                 66 net/tipc/core.c 	INIT_LIST_HEAD(&tn->node_list);
tn                 67 net/tipc/core.c 	spin_lock_init(&tn->node_list_lock);
tn                 77 net/tipc/core.c 	INIT_LIST_HEAD(&tn->dist_queue);
tn                 83 net/tipc/discover.c 	struct tipc_net *tn = tipc_net(net);
tn                 88 net/tipc/discover.c 	tipc_msg_init(tn->trial_addr, hdr, LINK_CONFIG, mtyp,
tn                 92 net/tipc/discover.c 	msg_set_node_sig(hdr, tn->random);
tn                 95 net/tipc/discover.c 	msg_set_bc_netid(hdr, tn->net_id);
tn                148 net/tipc/discover.c 	struct tipc_net *tn = tipc_net(net);
tn                149 net/tipc/discover.c 	bool trial = time_before(jiffies, tn->addr_trial_end);
tn                157 net/tipc/discover.c 		if (dst != tn->trial_addr)
tn                161 net/tipc/discover.c 		tn->trial_addr = sugg_addr;
tn                163 net/tipc/discover.c 		tn->addr_trial_end = jiffies + msecs_to_jiffies(1000);
tn                169 net/tipc/discover.c 		tipc_sched_net_finalize(net, tn->trial_addr);
tn                170 net/tipc/discover.c 		msg_set_prevnode(buf_msg(d->skb), tn->trial_addr);
tn                194 net/tipc/discover.c 	struct tipc_net *tn = tipc_net(net);
tn                197 net/tipc/discover.c 	bool legacy = tn->legacy_addr_format;
tn                228 net/tipc/discover.c 	if (net_id != tn->net_id)
tn                290 net/tipc/discover.c 	struct tipc_net *tn = tipc_net(d->net);
tn                305 net/tipc/discover.c 	if (!time_before(jiffies, tn->addr_trial_end) && !tipc_own_addr(net)) {
tn                308 net/tipc/discover.c 		tipc_sched_net_finalize(net, tn->trial_addr);
tn                313 net/tipc/discover.c 	if (time_before(jiffies, tn->addr_trial_end)) {
tn                322 net/tipc/discover.c 		msg_set_prevnode(buf_msg(d->skb), tn->trial_addr);
tn                347 net/tipc/discover.c 	struct tipc_net *tn = tipc_net(net);
tn                362 net/tipc/discover.c 		tn->addr_trial_end = jiffies + msecs_to_jiffies(1000);
tn               2560 net/tipc/link.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               2563 net/tipc/link.c 	struct tipc_link *bcl = tn->bcl;
tn                174 net/tipc/monitor.c 	struct tipc_net *tn = tipc_net(net);
tn                176 net/tipc/monitor.c 	return mon->peer_cnt > tn->mon_threshold;
tn                608 net/tipc/monitor.c 	struct tipc_net *tn = tipc_net(net);
tn                613 net/tipc/monitor.c 	if (tn->monitors[bearer_id])
tn                625 net/tipc/monitor.c 	tn->monitors[bearer_id] = mon;
tn                636 net/tipc/monitor.c 	mon->timer_intv = msecs_to_jiffies(MON_TIMEOUT + (tn->random & 0xffff));
tn                643 net/tipc/monitor.c 	struct tipc_net *tn = tipc_net(net);
tn                653 net/tipc/monitor.c 	tn->monitors[bearer_id] = NULL;
tn                685 net/tipc/monitor.c 	struct tipc_net *tn = tipc_net(net);
tn                690 net/tipc/monitor.c 	tn->mon_threshold = cluster_size;
tn                697 net/tipc/monitor.c 	struct tipc_net *tn = tipc_net(net);
tn                699 net/tipc/monitor.c 	return tn->mon_threshold;
tn                209 net/tipc/name_distr.c 	struct tipc_net *tn = tipc_net(net);
tn                212 net/tipc/name_distr.c 	spin_lock_bh(&tn->nametbl_lock);
tn                217 net/tipc/name_distr.c 	spin_unlock_bh(&tn->nametbl_lock);
tn                235 net/tipc/name_distr.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                238 net/tipc/name_distr.c 	spin_lock_bh(&tn->nametbl_lock);
tn                239 net/tipc/name_distr.c 	list_for_each_entry_safe(e, tmp, &tn->dist_queue, next) {
tn                245 net/tipc/name_distr.c 	spin_unlock_bh(&tn->nametbl_lock);
tn                303 net/tipc/name_distr.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                311 net/tipc/name_distr.c 	spin_lock_bh(&tn->nametbl_lock);
tn                325 net/tipc/name_distr.c 	spin_unlock_bh(&tn->nametbl_lock);
tn                338 net/tipc/name_distr.c 	struct tipc_net *tn = tipc_net(net);
tn                342 net/tipc/name_distr.c 	spin_lock_bh(&tn->nametbl_lock);
tn                349 net/tipc/name_distr.c 	spin_unlock_bh(&tn->nametbl_lock);
tn                425 net/tipc/name_table.c 	struct tipc_net *tn = tipc_net(net);
tn                426 net/tipc/name_table.c 	bool legacy = tn->legacy_addr_format;
tn                615 net/tipc/name_table.c 	struct tipc_net *tn = tipc_net(net);
tn                619 net/tipc/name_table.c 	spin_lock_bh(&tn->nametbl_lock);
tn                633 net/tipc/name_table.c 	spin_unlock_bh(&tn->nametbl_lock);
tn                647 net/tipc/name_table.c 	struct tipc_net *tn = tipc_net(net);
tn                652 net/tipc/name_table.c 	spin_lock_bh(&tn->nametbl_lock);
tn                664 net/tipc/name_table.c 	spin_unlock_bh(&tn->nametbl_lock);
tn                679 net/tipc/name_table.c 	struct tipc_net *tn = tipc_net(sub->net);
tn                685 net/tipc/name_table.c 	spin_lock_bh(&tn->nametbl_lock);
tn                699 net/tipc/name_table.c 	spin_unlock_bh(&tn->nametbl_lock);
tn                708 net/tipc/name_table.c 	struct tipc_net *tn = tipc_net(sub->net);
tn                713 net/tipc/name_table.c 	spin_lock_bh(&tn->nametbl_lock);
tn                729 net/tipc/name_table.c 	spin_unlock_bh(&tn->nametbl_lock);
tn                734 net/tipc/name_table.c 	struct tipc_net *tn = tipc_net(net);
tn                748 net/tipc/name_table.c 	tn->nametbl = nt;
tn                749 net/tipc/name_table.c 	spin_lock_init(&tn->nametbl_lock);
tn                778 net/tipc/name_table.c 	struct tipc_net *tn = tipc_net(net);
tn                786 net/tipc/name_table.c 	spin_lock_bh(&tn->nametbl_lock);
tn                795 net/tipc/name_table.c 	spin_unlock_bh(&tn->nametbl_lock);
tn                898 net/tipc/name_table.c 	struct tipc_net *tn = tipc_net(net);
tn                910 net/tipc/name_table.c 		head = &tn->nametbl->services[i];
tn                133 net/tipc/net.c 	struct tipc_net *tn = tipc_net(net);
tn                135 net/tipc/net.c 	if (cmpxchg(&tn->node_addr, 0, addr))
tn                181 net/tipc/net.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                182 net/tipc/net.c 	u64 *w0 = (u64 *)&tn->node_id[0];
tn                183 net/tipc/net.c 	u64 *w1 = (u64 *)&tn->node_id[8];
tn                196 net/tipc/net.c 	if (nla_put_u32(msg->skb, TIPC_NLA_NET_ID, tn->net_id))
tn                244 net/tipc/net.c 	struct tipc_net *tn = tipc_net(net);
tn                268 net/tipc/net.c 		tn->net_id = val;
tn                277 net/tipc/net.c 		tn->legacy_addr_format = true;
tn                261 net/tipc/node.c 	struct tipc_net *tn = tipc_net(net);
tn                266 net/tipc/node.c 	hlist_for_each_entry_rcu(node, &tn->node_htable[thash], hash) {
tn                283 net/tipc/node.c 	struct tipc_net *tn = tipc_net(net);
tn                288 net/tipc/node.c 	list_for_each_entry_rcu(n, &tn->node_list, list) {
tn                366 net/tipc/node.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                372 net/tipc/node.c 	spin_lock_bh(&tn->node_list_lock);
tn                388 net/tipc/node.c 		tn->capabilities = TIPC_NODE_CAPABILITIES;
tn                389 net/tipc/node.c 		list_for_each_entry_rcu(temp_node, &tn->node_list, list) {
tn                390 net/tipc/node.c 			tn->capabilities &= temp_node->capabilities;
tn                436 net/tipc/node.c 	hlist_add_head_rcu(&n->hash, &tn->node_htable[tipc_hashfn(addr)]);
tn                437 net/tipc/node.c 	list_for_each_entry_rcu(temp_node, &tn->node_list, list) {
tn                443 net/tipc/node.c 	tn->capabilities = TIPC_NODE_CAPABILITIES;
tn                444 net/tipc/node.c 	list_for_each_entry_rcu(temp_node, &tn->node_list, list) {
tn                445 net/tipc/node.c 		tn->capabilities &= temp_node->capabilities;
tn                449 net/tipc/node.c 	spin_unlock_bh(&tn->node_list_lock);
tn                484 net/tipc/node.c 	struct tipc_net *tn = tipc_net(net);
tn                487 net/tipc/node.c 	spin_lock_bh(&tn->node_list_lock);
tn                488 net/tipc/node.c 	list_for_each_entry_safe(node, t_node, &tn->node_list, list)
tn                490 net/tipc/node.c 	spin_unlock_bh(&tn->node_list_lock);
tn                604 net/tipc/node.c 	struct tipc_net *tn = tipc_net(peer->net);
tn                608 net/tipc/node.c 	if (!spin_trylock_bh(&tn->node_list_lock))
tn                621 net/tipc/node.c 	tn->capabilities = TIPC_NODE_CAPABILITIES;
tn                622 net/tipc/node.c 	list_for_each_entry_rcu(temp_node, &tn->node_list, list) {
tn                623 net/tipc/node.c 		tn->capabilities &= temp_node->capabilities;
tn                626 net/tipc/node.c 	spin_unlock_bh(&tn->node_list_lock);
tn                951 net/tipc/node.c 	struct tipc_net *tn = tipc_net(net);
tn                974 net/tipc/node.c 	if (tn->trial_addr == addr)
tn               1119 net/tipc/node.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               1123 net/tipc/node.c 	list_for_each_entry_rcu(n, &tn->node_list, list) {
tn               1897 net/tipc/node.c 	struct tipc_net *tn = tipc_net(net);
tn               1907 net/tipc/node.c 	list_for_each_entry_rcu(n, &tn->node_list, list) {
tn               1927 net/tipc/node.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               1951 net/tipc/node.c 	spin_lock_bh(&tn->node_list_lock);
tn               1954 net/tipc/node.c 		spin_unlock_bh(&tn->node_list_lock);
tn               1973 net/tipc/node.c 	spin_unlock_bh(&tn->node_list_lock);
tn               1982 net/tipc/node.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               2013 net/tipc/node.c 	list_for_each_entry_rcu(node, &tn->node_list, list) {
tn               2051 net/tipc/node.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               2059 net/tipc/node.c 	list_for_each_entry_rcu(n, &tn->node_list, list) {
tn               2299 net/tipc/node.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               2329 net/tipc/node.c 		list_for_each_entry_continue_rcu(node, &tn->node_list,
tn               2345 net/tipc/node.c 		list_for_each_entry_rcu(node, &tn->node_list, list) {
tn               2780 net/tipc/socket.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               2785 net/tipc/socket.c 	rhashtable_walk_enter(&tn->sk_rht, &iter);
tn               2810 net/tipc/socket.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               2814 net/tipc/socket.c 	tsk = rhashtable_lookup_fast(&tn->sk_rht, &portid, tsk_rht_params);
tn               2826 net/tipc/socket.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               2836 net/tipc/socket.c 		if (!rhashtable_lookup_insert_fast(&tn->sk_rht, &tsk->node,
tn               2848 net/tipc/socket.c 	struct tipc_net *tn = net_generic(sock_net(sk), tipc_net_id);
tn               2850 net/tipc/socket.c 	if (!rhashtable_remove_fast(&tn->sk_rht, &tsk->node, tsk_rht_params)) {
tn               2868 net/tipc/socket.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               2870 net/tipc/socket.c 	return rhashtable_init(&tn->sk_rht, &tsk_rht_params);
tn               2875 net/tipc/socket.c 	struct tipc_net *tn = net_generic(net, tipc_net_id);
tn               2880 net/tipc/socket.c 	rhashtable_destroy(&tn->sk_rht);
tn               3410 net/tipc/socket.c 	struct tipc_net *tn = tipc_net(net);
tn               3420 net/tipc/socket.c 	rhashtable_walk_enter(&tn->sk_rht, iter);
tn                231 net/tipc/topsrv.c 	struct tipc_net *tn = tipc_net(con->server->net);
tn                239 net/tipc/topsrv.c 			atomic_dec(&tn->subscription_count);
tn                363 net/tipc/topsrv.c 	struct tipc_net *tn = tipc_net(srv->net);
tn                371 net/tipc/topsrv.c 	if (atomic_read(&tn->subscription_count) >= TIPC_MAX_SUBSCR) {
tn                378 net/tipc/topsrv.c 	atomic_inc(&tn->subscription_count);
tn                643 net/tipc/topsrv.c 	struct tipc_net *tn = tipc_net(net);
tn                656 net/tipc/topsrv.c 	tn->topsrv = srv;
tn                657 net/tipc/topsrv.c 	atomic_set(&tn->subscription_count, 0);
tn                490 net/tipc/udp_media.c 		struct tipc_net *tn = net_generic(net, tipc_net_id);
tn                493 net/tipc/udp_media.c 		b = rtnl_dereference(tn->bearer_list[bid]);