eq                 93 arch/arc/include/asm/tlb-mmu1.h xor.eq  r1,r1,1                 /* flip bottom bit of data index */
eq                183 arch/arm/include/asm/assembler.h 	asm_trace_hardirqs_on cond=eq
eq                192 arch/arm/include/asm/assembler.h 	.irp	c,,eq,ne,cs,cc,mi,pl,vs,vc,hi,ls,ge,lt,gt,le,hs,lo
eq                449 arch/arm/include/asm/assembler.h 	.irp	c,,eq,ne,cs,cc,mi,pl,vs,vc,hi,ls,ge,lt,gt,le,hs,lo
eq                117 arch/arm/include/asm/futex.h 	"2:	" TUSERCOND(str, eq) "	%3, [%4]\n"
eq                 39 arch/arm64/include/asm/fpsimdmacros.h 	b.eq	9999f
eq                204 arch/arm64/include/asm/fpsimdmacros.h 		b.eq		921f
eq                 68 arch/ia64/kernel/minstate.h 	cmp.eq pKStk,pUStk=r0,r17;		/* are we in kernel mode already? */		\
eq                128 arch/ia64/kernel/minstate.h 	cmp.eq pNonSys,pSys=r0,r0	/* initialize pSys=0, pNonSys=1 */			\
eq                 62 arch/unicore32/include/asm/assembler.h 	.ifc	\cond, eq
eq                408 arch/x86/include/asm/vmx.h #define DEBUG_REG_ACCESS_REG(eq)        (((eq) >> 8) & 0xf) /* 11:8, general purpose reg. */
eq               6426 block/bfq-iosched.c 	struct elevator_queue *eq;
eq               6428 block/bfq-iosched.c 	eq = elevator_alloc(q, e);
eq               6429 block/bfq-iosched.c 	if (!eq)
eq               6434 block/bfq-iosched.c 		kobject_put(&eq->kobj);
eq               6437 block/bfq-iosched.c 	eq->elevator_data = bfqd;
eq               6440 block/bfq-iosched.c 	q->elevator = eq;
eq               6551 block/bfq-iosched.c 	kobject_put(&eq->kobj);
eq                526 block/blk-mq-sched.c 	struct elevator_queue *eq;
eq                560 block/blk-mq-sched.c 				eq = q->elevator;
eq                562 block/blk-mq-sched.c 				blk_mq_exit_sched(q, eq);
eq                563 block/blk-mq-sched.c 				kobject_put(&eq->kobj);
eq                167 block/elevator.c 	struct elevator_queue *eq;
eq                169 block/elevator.c 	eq = kzalloc_node(sizeof(*eq), GFP_KERNEL, q->node);
eq                170 block/elevator.c 	if (unlikely(!eq))
eq                173 block/elevator.c 	eq->type = e;
eq                174 block/elevator.c 	kobject_init(&eq->kobj, &elv_ktype);
eq                175 block/elevator.c 	mutex_init(&eq->sysfs_lock);
eq                176 block/elevator.c 	hash_init(eq->hash);
eq                178 block/elevator.c 	return eq;
eq                419 block/kyber-iosched.c 	struct elevator_queue *eq;
eq                421 block/kyber-iosched.c 	eq = elevator_alloc(q, e);
eq                422 block/kyber-iosched.c 	if (!eq)
eq                427 block/kyber-iosched.c 		kobject_put(&eq->kobj);
eq                433 block/kyber-iosched.c 	eq->elevator_data = kqd;
eq                434 block/kyber-iosched.c 	q->elevator = eq;
eq                409 block/mq-deadline.c 	struct elevator_queue *eq;
eq                411 block/mq-deadline.c 	eq = elevator_alloc(q, e);
eq                412 block/mq-deadline.c 	if (!eq)
eq                417 block/mq-deadline.c 		kobject_put(&eq->kobj);
eq                420 block/mq-deadline.c 	eq->elevator_data = dd;
eq                435 block/mq-deadline.c 	q->elevator = eq;
eq               1175 drivers/block/drbd/drbd_worker.c 	int err, eq = 0;
eq               1201 drivers/block/drbd/drbd_worker.c 			eq = !memcmp(digest, di->digest, digest_size);
eq               1205 drivers/block/drbd/drbd_worker.c 		if (eq) {
eq               1298 drivers/block/drbd/drbd_worker.c 	int err, eq = 0;
eq               1323 drivers/block/drbd/drbd_worker.c 			eq = !memcmp(digest, di->digest, digest_size);
eq               1334 drivers/block/drbd/drbd_worker.c 	if (!eq)
eq               1340 drivers/block/drbd/drbd_worker.c 			       eq ? ID_IN_SYNC : ID_OUT_OF_SYNC);
eq                 49 drivers/clk/spear/clk-aux-synth.c 	u8 eq = rtbl[index].eq ? 1 : 2;
eq                 52 drivers/clk/spear/clk-aux-synth.c 			(rtbl[index].yscale * eq)) * 10000;
eq                115 drivers/clk/spear/clk-aux-synth.c 	val |= (rtbl[i].eq & aux->masks->eq_sel_mask) <<
eq                 46 drivers/clk/spear/clk.h 	u8 eq;
eq                255 drivers/clk/spear/spear1310_clock.c 	{.xscale = 10, .yscale = 204, .eq = 0}, /* 12.29 MHz */
eq                256 drivers/clk/spear/spear1310_clock.c 	{.xscale = 4, .yscale = 21, .eq = 0}, /* 48 MHz */
eq                257 drivers/clk/spear/spear1310_clock.c 	{.xscale = 2, .yscale = 6, .eq = 0}, /* 83 MHz */
eq                258 drivers/clk/spear/spear1310_clock.c 	{.xscale = 2, .yscale = 4, .eq = 0}, /* 125 MHz */
eq                259 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 3, .eq = 1}, /* 166 MHz */
eq                260 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 2, .eq = 1}, /* 250 MHz */
eq                266 drivers/clk/spear/spear1310_clock.c 	{.xscale = 2, .yscale = 6, .eq = 0}, /* divided by 6 */
eq                267 drivers/clk/spear/spear1310_clock.c 	{.xscale = 2, .yscale = 4, .eq = 0}, /* divided by 4 */
eq                268 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 3, .eq = 1}, /* divided by 3 */
eq                269 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 2, .eq = 1}, /* divided by 2 */
eq                314 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 12, .eq = 0}, /* 2.048 MHz, smp freq = 8Khz */
eq                315 drivers/clk/spear/spear1310_clock.c 	{.xscale = 11, .yscale = 96, .eq = 0}, /* 2.816 MHz, smp freq = 11Khz */
eq                316 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 6, .eq = 0}, /* 4.096 MHz, smp freq = 16Khz */
eq                317 drivers/clk/spear/spear1310_clock.c 	{.xscale = 11, .yscale = 48, .eq = 0}, /* 5.632 MHz, smp freq = 22Khz */
eq                323 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 3, .eq = 0},
eq                326 drivers/clk/spear/spear1310_clock.c 	{.xscale = 17, .yscale = 37, .eq = 0}, /* 11.289 MHz, smp freq = 44Khz*/
eq                328 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 2, .eq = 0}, /* 12.288 MHz */
eq                334 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 4, .eq = 0}, /* 1.53 MHz */
eq                335 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 2, .eq = 0}, /* 3.07 Mhz */
eq                342 drivers/clk/spear/spear1310_clock.c 	{.xscale = 1, .yscale = 31, .eq = 0}, /* 2.68 MHz */
eq                343 drivers/clk/spear/spear1310_clock.c 	{.xscale = 2, .yscale = 21, .eq = 0}, /* 7.94 MHz */
eq                344 drivers/clk/spear/spear1310_clock.c 	{.xscale = 4, .yscale = 21, .eq = 0}, /* 15.87 MHz */
eq                345 drivers/clk/spear/spear1310_clock.c 	{.xscale = 10, .yscale = 42, .eq = 0}, /* 19.84 MHz */
eq                267 drivers/clk/spear/spear1340_clock.c 	{.xscale = 5, .yscale = 122, .eq = 0},
eq                269 drivers/clk/spear/spear1340_clock.c 	{.xscale = 10, .yscale = 204, .eq = 0},
eq                271 drivers/clk/spear/spear1340_clock.c 	{.xscale = 4, .yscale = 25, .eq = 0},
eq                273 drivers/clk/spear/spear1340_clock.c 	{.xscale = 4, .yscale = 21, .eq = 0},
eq                275 drivers/clk/spear/spear1340_clock.c 	{.xscale = 5, .yscale = 18, .eq = 0},
eq                277 drivers/clk/spear/spear1340_clock.c 	{.xscale = 2, .yscale = 6, .eq = 0},
eq                279 drivers/clk/spear/spear1340_clock.c 	{.xscale = 5, .yscale = 12, .eq = 0},
eq                281 drivers/clk/spear/spear1340_clock.c 	{.xscale = 2, .yscale = 4, .eq = 0},
eq                283 drivers/clk/spear/spear1340_clock.c 	{.xscale = 5, .yscale = 18, .eq = 1},
eq                285 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 3, .eq = 1},
eq                287 drivers/clk/spear/spear1340_clock.c 	{.xscale = 5, .yscale = 12, .eq = 1},
eq                289 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 2, .eq = 1},
eq                295 drivers/clk/spear/spear1340_clock.c 	{.xscale = 2, .yscale = 6, .eq = 0}, /* divided by 6 */
eq                296 drivers/clk/spear/spear1340_clock.c 	{.xscale = 2, .yscale = 4, .eq = 0}, /* divided by 4 */
eq                297 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 3, .eq = 1}, /* divided by 3 */
eq                298 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 2, .eq = 1}, /* divided by 2 */
eq                353 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 12, .eq = 0}, /* 2.048 MHz, smp freq = 8Khz */
eq                354 drivers/clk/spear/spear1340_clock.c 	{.xscale = 11, .yscale = 96, .eq = 0}, /* 2.816 MHz, smp freq = 11Khz */
eq                355 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 6, .eq = 0}, /* 4.096 MHz, smp freq = 16Khz */
eq                356 drivers/clk/spear/spear1340_clock.c 	{.xscale = 11, .yscale = 48, .eq = 0}, /* 5.632 MHz, smp freq = 22Khz */
eq                362 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 3, .eq = 0},
eq                365 drivers/clk/spear/spear1340_clock.c 	{.xscale = 17, .yscale = 37, .eq = 0}, /* 11.289 MHz, smp freq = 44Khz*/
eq                366 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 2, .eq = 0}, /* 12.288 MHz, smp freq = 48Khz*/
eq                372 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 4, .eq = 0},
eq                373 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 2, .eq = 0},
eq                380 drivers/clk/spear/spear1340_clock.c 	{.xscale = 1, .yscale = 31, .eq = 0}, /* 2.68 MHz */
eq                381 drivers/clk/spear/spear1340_clock.c 	{.xscale = 2, .yscale = 21, .eq = 0}, /* 7.94 MHz */
eq                382 drivers/clk/spear/spear1340_clock.c 	{.xscale = 4, .yscale = 21, .eq = 0}, /* 15.87 MHz */
eq                383 drivers/clk/spear/spear1340_clock.c 	{.xscale = 10, .yscale = 42, .eq = 0}, /* 19.84 MHz */
eq                109 drivers/clk/spear/spear3xx_clock.c 	{.xscale = 1, .yscale = 81, .eq = 0}, /* 2.049 MHz */
eq                110 drivers/clk/spear/spear3xx_clock.c 	{.xscale = 1, .yscale = 59, .eq = 0}, /* 2.822 MHz */
eq                111 drivers/clk/spear/spear3xx_clock.c 	{.xscale = 2, .yscale = 81, .eq = 0}, /* 4.098 MHz */
eq                112 drivers/clk/spear/spear3xx_clock.c 	{.xscale = 3, .yscale = 89, .eq = 0}, /* 5.644 MHz */
eq                113 drivers/clk/spear/spear3xx_clock.c 	{.xscale = 4, .yscale = 81, .eq = 0}, /* 8.197 MHz */
eq                114 drivers/clk/spear/spear3xx_clock.c 	{.xscale = 4, .yscale = 59, .eq = 0}, /* 11.254 MHz */
eq                115 drivers/clk/spear/spear3xx_clock.c 	{.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */
eq                116 drivers/clk/spear/spear3xx_clock.c 	{.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */
eq                117 drivers/clk/spear/spear3xx_clock.c 	{.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */
eq                118 drivers/clk/spear/spear3xx_clock.c 	{.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */
eq                 93 drivers/clk/spear/spear6xx_clock.c 	{.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */
eq                 94 drivers/clk/spear/spear6xx_clock.c 	{.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */
eq                 95 drivers/clk/spear/spear6xx_clock.c 	{.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */
eq                 96 drivers/clk/spear/spear6xx_clock.c 	{.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */
eq                167 drivers/firmware/broadcom/bcm47xx_nvram.c 	char *var, *value, *end, *eq;
eq                183 drivers/firmware/broadcom/bcm47xx_nvram.c 		eq = strchr(var, '=');
eq                184 drivers/firmware/broadcom/bcm47xx_nvram.c 		if (!eq)
eq                186 drivers/firmware/broadcom/bcm47xx_nvram.c 		value = eq + 1;
eq                187 drivers/firmware/broadcom/bcm47xx_nvram.c 		if (eq - var == strlen(name) &&
eq                188 drivers/firmware/broadcom/bcm47xx_nvram.c 		    strncmp(var, name, eq - var) == 0)
eq                766 drivers/infiniband/hw/hfi1/pcie.c static int load_eq_table(struct hfi1_devdata *dd, const u8 eq[11][3], u8 fs,
eq                780 drivers/infiniband/hw/hfi1/pcie.c 		c_minus1 = eq[i][PREC] / div;
eq                781 drivers/infiniband/hw/hfi1/pcie.c 		c0 = fs - (eq[i][PREC] / div) - (eq[i][POST] / div);
eq                782 drivers/infiniband/hw/hfi1/pcie.c 		c_plus1 = eq[i][POST] / div;
eq                802 drivers/infiniband/hw/hfi1/pcie.c 				   i, (u32)eq[i][0], (u32)eq[i][1],
eq                803 drivers/infiniband/hw/hfi1/pcie.c 				   (u32)eq[i][2]);
eq                998 drivers/infiniband/hw/hfi1/pcie.c 	const u8 (*eq)[3];
eq               1136 drivers/infiniband/hw/hfi1/pcie.c 		eq = discrete_preliminary_eq;
eq               1146 drivers/infiniband/hw/hfi1/pcie.c 		eq = integrated_preliminary_eq;
eq               1157 drivers/infiniband/hw/hfi1/pcie.c 	ret = load_eq_table(dd, eq, fs, div);
eq                196 drivers/infiniband/hw/hns/hns_roce_cq.c 	synchronize_irq(hr_dev->eq_table.eq[hr_cq->vector].irq);
eq                791 drivers/infiniband/hw/hns/hns_roce_device.h 	struct hns_roce_eq	*eq;
eq               3694 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static void set_eq_cons_index_v1(struct hns_roce_eq *eq, int req_not)
eq               3696 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	roce_raw_write((eq->cons_index & HNS_ROCE_V1_CONS_IDX_M) |
eq               3697 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		      (req_not << eq->log_entries), eq->doorbell);
eq               3862 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_aeqe *get_aeqe_v1(struct hns_roce_eq *eq, u32 entry)
eq               3864 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	unsigned long off = (entry & (eq->entries - 1)) *
eq               3868 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		(eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
eq               3872 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_aeqe *next_aeqe_sw_v1(struct hns_roce_eq *eq)
eq               3874 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	struct hns_roce_aeqe *aeqe = get_aeqe_v1(eq, eq->cons_index);
eq               3877 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		!!(eq->cons_index & eq->entries)) ? aeqe : NULL;
eq               3881 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			       struct hns_roce_eq *eq)
eq               3888 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	while ((aeqe = next_aeqe_sw_v1(eq))) {
eq               3952 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				 event_type, eq->eqn, eq->cons_index);
eq               3956 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		eq->cons_index++;
eq               3959 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		if (eq->cons_index > 2 * hr_dev->caps.aeqe_depth - 1) {
eq               3961 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->cons_index = 0;
eq               3965 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	set_eq_cons_index_v1(eq, 0);
eq               3970 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_ceqe *get_ceqe_v1(struct hns_roce_eq *eq, u32 entry)
eq               3972 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	unsigned long off = (entry & (eq->entries - 1)) *
eq               3976 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			(eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
eq               3980 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_ceqe *next_ceqe_sw_v1(struct hns_roce_eq *eq)
eq               3982 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	struct hns_roce_ceqe *ceqe = get_ceqe_v1(eq, eq->cons_index);
eq               3986 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		(!!(eq->cons_index & eq->entries)) ? ceqe : NULL;
eq               3990 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			       struct hns_roce_eq *eq)
eq               3996 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	while ((ceqe = next_ceqe_sw_v1(eq))) {
eq               4008 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		++eq->cons_index;
eq               4011 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		if (eq->cons_index >
eq               4013 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			dev_warn(&eq->hr_dev->pdev->dev,
eq               4015 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->cons_index = 0;
eq               4019 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	set_eq_cons_index_v1(eq, 0);
eq               4026 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	struct hns_roce_eq  *eq  = eq_ptr;
eq               4027 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	struct hns_roce_dev *hr_dev = eq->hr_dev;
eq               4030 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	if (eq->type_flag == HNS_ROCE_CEQ)
eq               4032 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		int_work = hns_roce_v1_ceq_int(hr_dev, eq);
eq               4035 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		int_work = hns_roce_v1_aeq_int(hr_dev, eq);
eq               4179 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				struct hns_roce_eq *eq)
eq               4181 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	int npages = (PAGE_ALIGN(eq->eqe_size * eq->entries) +
eq               4185 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	if (!eq->buf_list)
eq               4190 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				  eq->buf_list[i].buf, eq->buf_list[i].map);
eq               4192 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	kfree(eq->buf_list);
eq               4221 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				 struct hns_roce_eq *eq)
eq               4223 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	void __iomem *eqc = hr_dev->eq_table.eqc_base[eq->eqn];
eq               4236 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	num_bas = (PAGE_ALIGN(eq->entries * eq->eqe_size) +
eq               4239 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	if ((eq->entries * eq->eqe_size) > HNS_ROCE_BA_SIZE) {
eq               4241 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			(eq->entries * eq->eqe_size), HNS_ROCE_BA_SIZE,
eq               4246 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	eq->buf_list = kcalloc(num_bas, sizeof(*eq->buf_list), GFP_KERNEL);
eq               4247 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	if (!eq->buf_list)
eq               4251 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		eq->buf_list[i].buf = dma_alloc_coherent(dev, HNS_ROCE_BA_SIZE,
eq               4254 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		if (!eq->buf_list[i].buf) {
eq               4259 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		eq->buf_list[i].map = tmp_dma_addr;
eq               4261 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	eq->cons_index = 0;
eq               4267 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		       eq->log_entries);
eq               4272 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	writel((u32)(eq->buf_list[0].map >> 12), eqc + 4);
eq               4282 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		       eq->buf_list[0].map >> 44);
eq               4298 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		dma_free_coherent(dev, HNS_ROCE_BA_SIZE, eq->buf_list[i].buf,
eq               4299 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				  eq->buf_list[i].map);
eq               4301 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	kfree(eq->buf_list);
eq               4309 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	struct hns_roce_eq *eq;
eq               4318 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	eq_table->eq = kcalloc(eq_num, sizeof(*eq_table->eq), GFP_KERNEL);
eq               4319 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	if (!eq_table->eq)
eq               4330 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		eq = &eq_table->eq[i];
eq               4331 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		eq->hr_dev = hr_dev;
eq               4332 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		eq->eqn = i;
eq               4333 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		eq->irq = hr_dev->irq[i];
eq               4334 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		eq->log_page_size = PAGE_SHIFT;
eq               4341 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->type_flag = HNS_ROCE_CEQ;
eq               4342 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->doorbell = hr_dev->reg_base +
eq               4345 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->entries = hr_dev->caps.ceqe_depth;
eq               4346 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->log_entries = ilog2(eq->entries);
eq               4347 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->eqe_size = HNS_ROCE_CEQ_ENTRY_SIZE;
eq               4352 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->type_flag = HNS_ROCE_AEQ;
eq               4353 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->doorbell = hr_dev->reg_base +
eq               4355 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->entries = hr_dev->caps.aeqe_depth;
eq               4356 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->log_entries = ilog2(eq->entries);
eq               4357 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			eq->eqe_size = HNS_ROCE_AEQ_ENTRY_SIZE;
eq               4373 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		ret = hns_roce_v1_create_eq(hr_dev, &eq_table->eq[i]);
eq               4385 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 					  &eq_table->eq[j]);
eq               4404 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		free_irq(hr_dev->irq[j], &eq_table->eq[j]);
eq               4408 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		hns_roce_v1_free_eq(hr_dev, &eq_table->eq[i]);
eq               4413 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	kfree(eq_table->eq);
eq               4431 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		free_irq(hr_dev->irq[i], &eq_table->eq[i]);
eq               4433 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		hns_roce_v1_free_eq(hr_dev, &eq_table->eq[i]);
eq               4439 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	kfree(eq_table->eq);
eq               4931 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				      struct hns_roce_eq *eq,
eq               4944 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	irq_work->event_type = eq->event_type;
eq               4945 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	irq_work->sub_type = eq->sub_type;
eq               4949 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static void set_eq_cons_index_v2(struct hns_roce_eq *eq)
eq               4951 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_dev *hr_dev = eq->hr_dev;
eq               4957 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (eq->type_flag == HNS_ROCE_AEQ) {
eq               4960 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			       eq->arm_st == HNS_ROCE_V2_EQ_ALWAYS_ARMED ?
eq               4965 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			       HNS_ROCE_V2_EQ_DB_TAG_S, eq->eqn);
eq               4969 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			       eq->arm_st == HNS_ROCE_V2_EQ_ALWAYS_ARMED ?
eq               4976 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       (eq->cons_index & HNS_ROCE_V2_CONS_IDX_M));
eq               4978 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	hns_roce_write64(hr_dev, doorbell, eq->doorbell);
eq               4981 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_aeqe *get_aeqe_v2(struct hns_roce_eq *eq, u32 entry)
eq               4986 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	buf_chk_sz = 1 << (eq->eqe_buf_pg_sz + PAGE_SHIFT);
eq               4987 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	off = (entry & (eq->entries - 1)) * HNS_ROCE_AEQ_ENTRY_SIZE;
eq               4989 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	return (struct hns_roce_aeqe *)((char *)(eq->buf_list->buf) +
eq               4993 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_aeqe *mhop_get_aeqe(struct hns_roce_eq *eq, u32 entry)
eq               4998 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	buf_chk_sz = 1 << (eq->eqe_buf_pg_sz + PAGE_SHIFT);
eq               5000 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	off = (entry & (eq->entries - 1)) * HNS_ROCE_AEQ_ENTRY_SIZE;
eq               5002 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (eq->hop_num == HNS_ROCE_HOP_NUM_0)
eq               5003 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		return (struct hns_roce_aeqe *)((u8 *)(eq->bt_l0) +
eq               5007 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			(eq->buf[off / buf_chk_sz]) + off % buf_chk_sz);
eq               5010 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_aeqe *next_aeqe_sw_v2(struct hns_roce_eq *eq)
eq               5014 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (!eq->hop_num)
eq               5015 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		aeqe = get_aeqe_v2(eq, eq->cons_index);
eq               5017 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		aeqe = mhop_get_aeqe(eq, eq->cons_index);
eq               5020 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		!!(eq->cons_index & eq->entries)) ? aeqe : NULL;
eq               5024 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			       struct hns_roce_eq *eq)
eq               5027 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_aeqe *aeqe = next_aeqe_sw_v2(eq);
eq               5090 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				event_type, eq->eqn, eq->cons_index);
eq               5094 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->event_type = event_type;
eq               5095 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->sub_type = sub_type;
eq               5096 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		++eq->cons_index;
eq               5099 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (eq->cons_index > (2 * eq->entries - 1))
eq               5100 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->cons_index = 0;
eq               5102 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		hns_roce_v2_init_irq_work(hr_dev, eq, qpn, cqn);
eq               5104 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		aeqe = next_aeqe_sw_v2(eq);
eq               5107 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	set_eq_cons_index_v2(eq);
eq               5111 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_ceqe *get_ceqe_v2(struct hns_roce_eq *eq, u32 entry)
eq               5116 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	buf_chk_sz = 1 << (eq->eqe_buf_pg_sz + PAGE_SHIFT);
eq               5117 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	off = (entry & (eq->entries - 1)) * HNS_ROCE_CEQ_ENTRY_SIZE;
eq               5119 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	return (struct hns_roce_ceqe *)((char *)(eq->buf_list->buf) +
eq               5123 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_ceqe *mhop_get_ceqe(struct hns_roce_eq *eq, u32 entry)
eq               5128 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	buf_chk_sz = 1 << (eq->eqe_buf_pg_sz + PAGE_SHIFT);
eq               5130 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	off = (entry & (eq->entries - 1)) * HNS_ROCE_CEQ_ENTRY_SIZE;
eq               5132 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (eq->hop_num == HNS_ROCE_HOP_NUM_0)
eq               5133 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		return (struct hns_roce_ceqe *)((u8 *)(eq->bt_l0) +
eq               5136 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		return (struct hns_roce_ceqe *)((u8 *)(eq->buf[off /
eq               5140 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_ceqe *next_ceqe_sw_v2(struct hns_roce_eq *eq)
eq               5144 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (!eq->hop_num)
eq               5145 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		ceqe = get_ceqe_v2(eq, eq->cons_index);
eq               5147 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		ceqe = mhop_get_ceqe(eq, eq->cons_index);
eq               5150 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		(!!(eq->cons_index & eq->entries)) ? ceqe : NULL;
eq               5154 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			       struct hns_roce_eq *eq)
eq               5157 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_ceqe *ceqe = next_ceqe_sw_v2(eq);
eq               5173 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		++eq->cons_index;
eq               5176 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (eq->cons_index > (EQ_DEPTH_COEFF * eq->entries - 1)) {
eq               5178 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->cons_index = 0;
eq               5181 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		ceqe = next_ceqe_sw_v2(eq);
eq               5184 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	set_eq_cons_index_v2(eq);
eq               5191 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_eq *eq = eq_ptr;
eq               5192 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_dev *hr_dev = eq->hr_dev;
eq               5195 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (eq->type_flag == HNS_ROCE_CEQ)
eq               5197 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		int_work = hns_roce_v2_ceq_int(hr_dev, eq);
eq               5200 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		int_work = hns_roce_v2_aeq_int(hr_dev, eq);
eq               5310 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				  struct hns_roce_eq *eq)
eq               5327 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		dma_free_coherent(dev, (unsigned int)(eq->entries *
eq               5328 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				  eq->eqe_size), eq->bt_l0, eq->l0_dma);
eq               5332 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	dma_free_coherent(dev, bt_chk_sz, eq->bt_l0, eq->l0_dma);
eq               5334 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		for (i = 0; i < eq->l0_last_num; i++) {
eq               5335 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			if (i == eq->l0_last_num - 1) {
eq               5336 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				eqe_alloc = i * (buf_chk_sz / eq->eqe_size);
eq               5337 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				size = (eq->entries - eqe_alloc) * eq->eqe_size;
eq               5338 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				dma_free_coherent(dev, size, eq->buf[i],
eq               5339 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 						  eq->buf_dma[i]);
eq               5342 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			dma_free_coherent(dev, buf_chk_sz, eq->buf[i],
eq               5343 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					  eq->buf_dma[i]);
eq               5346 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		for (i = 0; i < eq->l0_last_num; i++) {
eq               5347 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			dma_free_coherent(dev, bt_chk_sz, eq->bt_l1[i],
eq               5348 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					  eq->l1_dma[i]);
eq               5352 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				if ((i == eq->l0_last_num - 1)
eq               5353 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				     && j == eq->l1_last_num - 1) {
eq               5354 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					eqe_alloc = (buf_chk_sz / eq->eqe_size)
eq               5356 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					size = (eq->entries - eqe_alloc)
eq               5357 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 						* eq->eqe_size;
eq               5359 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 							  eq->buf[idx],
eq               5360 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 							  eq->buf_dma[idx]);
eq               5363 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				dma_free_coherent(dev, buf_chk_sz, eq->buf[idx],
eq               5364 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 						  eq->buf_dma[idx]);
eq               5368 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq->buf_dma);
eq               5369 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq->buf);
eq               5370 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq->l1_dma);
eq               5371 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq->bt_l1);
eq               5372 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->buf_dma = NULL;
eq               5373 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->buf = NULL;
eq               5374 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->l1_dma = NULL;
eq               5375 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->bt_l1 = NULL;
eq               5379 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				struct hns_roce_eq *eq)
eq               5383 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	buf_chk_sz = 1 << (eq->eqe_buf_pg_sz + PAGE_SHIFT);
eq               5386 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		hns_roce_mhop_free_eq(hr_dev, eq);
eq               5390 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	dma_free_coherent(hr_dev->dev, buf_chk_sz, eq->buf_list->buf,
eq               5391 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			  eq->buf_list->map);
eq               5392 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq->buf_list);
eq               5396 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				struct hns_roce_eq *eq,
eq               5405 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->doorbell = hr_dev->reg_base + ROCEE_VF_EQ_DB_CFG0_REG;
eq               5406 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->hop_num = hr_dev->caps.eqe_hop_num;
eq               5407 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->cons_index = 0;
eq               5408 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->over_ignore = HNS_ROCE_V2_EQ_OVER_IGNORE_0;
eq               5409 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->coalesce = HNS_ROCE_V2_EQ_COALESCE_0;
eq               5410 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->arm_st = HNS_ROCE_V2_EQ_ALWAYS_ARMED;
eq               5411 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->eqe_ba_pg_sz = hr_dev->caps.eqe_ba_pg_sz;
eq               5412 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->eqe_buf_pg_sz = hr_dev->caps.eqe_buf_pg_sz;
eq               5413 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->shift = ilog2((unsigned int)eq->entries);
eq               5415 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (!eq->hop_num)
eq               5416 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->eqe_ba = eq->buf_list->map;
eq               5418 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->eqe_ba = eq->l0_dma;
eq               5429 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_HOP_NUM_S, eq->hop_num);
eq               5434 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_OVER_IGNORE_S, eq->over_ignore);
eq               5439 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_COALESCE_S, eq->coalesce);
eq               5444 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_ARM_ST_S, eq->arm_st);
eq               5449 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_EQN_S, eq->eqn);
eq               5461 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       eq->eqe_ba_pg_sz + PG_SHIFT_OFFSET);
eq               5467 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       eq->eqe_buf_pg_sz + PG_SHIFT_OFFSET);
eq               5478 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_MAX_CNT_S, eq->eq_max_cnt);
eq               5483 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_PERIOD_S, eq->eq_period);
eq               5494 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_EQE_BA_L_S, eq->eqe_ba >> 3);
eq               5499 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_EQE_BA_H_S, eq->eqe_ba >> 35);
eq               5504 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_SHIFT_S, eq->shift);
eq               5515 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_CUR_EQE_BA_L_S, eq->cur_eqe_ba >> 12);
eq               5520 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_CUR_EQE_BA_M_S, eq->cur_eqe_ba >> 28);
eq               5525 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_CUR_EQE_BA_H_S, eq->cur_eqe_ba >> 60);
eq               5536 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_NXT_EQE_BA_L_S, eq->nxt_eqe_ba >> 12);
eq               5541 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       HNS_ROCE_EQC_NXT_EQE_BA_H_S, eq->nxt_eqe_ba >> 44);
eq               5545 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				  struct hns_roce_eq *eq)
eq               5567 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	ba_num = DIV_ROUND_UP(PAGE_ALIGN(eq->entries * eq->eqe_size),
eq               5572 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (eq->entries > buf_chk_sz / eq->eqe_size) {
eq               5574 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				eq->entries);
eq               5577 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->bt_l0 = dma_alloc_coherent(dev, eq->entries * eq->eqe_size,
eq               5578 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					       &(eq->l0_dma), GFP_KERNEL);
eq               5579 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (!eq->bt_l0)
eq               5582 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->cur_eqe_ba = eq->l0_dma;
eq               5583 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->nxt_eqe_ba = 0;
eq               5588 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->buf_dma = kcalloc(ba_num, sizeof(*eq->buf_dma), GFP_KERNEL);
eq               5589 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (!eq->buf_dma)
eq               5591 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->buf = kcalloc(ba_num, sizeof(*eq->buf), GFP_KERNEL);
eq               5592 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (!eq->buf)
eq               5596 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->l1_dma = kcalloc(bt_num, sizeof(*eq->l1_dma), GFP_KERNEL);
eq               5597 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (!eq->l1_dma)
eq               5600 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->bt_l1 = kcalloc(bt_num, sizeof(*eq->bt_l1), GFP_KERNEL);
eq               5601 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (!eq->bt_l1)
eq               5606 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->bt_l0 = dma_alloc_coherent(dev, bt_chk_sz, &eq->l0_dma, GFP_KERNEL);
eq               5607 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (!eq->bt_l0)
eq               5620 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				eqe_alloc = i * (buf_chk_sz / eq->eqe_size);
eq               5621 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				size = (eq->entries - eqe_alloc) * eq->eqe_size;
eq               5623 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->buf[i] = dma_alloc_coherent(dev, size,
eq               5624 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 							&(eq->buf_dma[i]),
eq               5626 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			if (!eq->buf[i])
eq               5629 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*(eq->bt_l0 + i) = eq->buf_dma[i];
eq               5635 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->cur_eqe_ba = eq->buf_dma[0];
eq               5637 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->nxt_eqe_ba = eq->buf_dma[1];
eq               5642 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->bt_l1[i] = dma_alloc_coherent(dev, bt_chk_sz,
eq               5643 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 							  &(eq->l1_dma[i]),
eq               5645 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			if (!eq->bt_l1[i])
eq               5647 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*(eq->bt_l0 + i) = eq->l1_dma[i];
eq               5654 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					eqe_alloc = (buf_chk_sz / eq->eqe_size)
eq               5656 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					size = (eq->entries - eqe_alloc)
eq               5657 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 						* eq->eqe_size;
eq               5659 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				eq->buf[idx] = dma_alloc_coherent(dev, size,
eq               5660 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 								  &(eq->buf_dma[idx]),
eq               5662 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				if (!eq->buf[idx])
eq               5665 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				*(eq->bt_l1[i] + j) = eq->buf_dma[idx];
eq               5677 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->cur_eqe_ba = eq->buf_dma[0];
eq               5679 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->nxt_eqe_ba = eq->buf_dma[1];
eq               5682 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->l0_last_num = i + 1;
eq               5684 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->l1_last_num = j + 1;
eq               5689 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	dma_free_coherent(dev, bt_chk_sz, eq->bt_l0, eq->l0_dma);
eq               5690 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->bt_l0 = NULL;
eq               5691 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->l0_dma = 0;
eq               5693 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		dma_free_coherent(dev, bt_chk_sz, eq->bt_l1[i],
eq               5694 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				  eq->l1_dma[i]);
eq               5698 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			dma_free_coherent(dev, buf_chk_sz, eq->buf[idx],
eq               5699 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					  eq->buf_dma[idx]);
eq               5705 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	dma_free_coherent(dev, bt_chk_sz, eq->bt_l0, eq->l0_dma);
eq               5706 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->bt_l0 = NULL;
eq               5707 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->l0_dma = 0;
eq               5711 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			dma_free_coherent(dev, buf_chk_sz, eq->buf[i],
eq               5712 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					  eq->buf_dma[i]);
eq               5717 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			dma_free_coherent(dev, bt_chk_sz, eq->bt_l1[i],
eq               5718 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					  eq->l1_dma[i]);
eq               5726 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 						  eq->buf[idx],
eq               5727 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 						  eq->buf_dma[idx]);
eq               5733 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq->bt_l1);
eq               5734 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->bt_l1 = NULL;
eq               5737 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq->l1_dma);
eq               5738 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->l1_dma = NULL;
eq               5741 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq->buf);
eq               5742 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->buf = NULL;
eq               5745 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq->buf_dma);
eq               5746 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq->buf_dma = NULL;
eq               5752 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				 struct hns_roce_eq *eq,
eq               5768 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->buf_list = kzalloc(sizeof(struct hns_roce_buf_list),
eq               5770 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (!eq->buf_list) {
eq               5775 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->buf_list->buf = dma_alloc_coherent(dev, buf_chk_sz,
eq               5776 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 						       &(eq->buf_list->map),
eq               5778 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (!eq->buf_list->buf) {
eq               5784 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		ret = hns_roce_mhop_alloc_eq(hr_dev, eq);
eq               5791 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	hns_roce_config_eqc(hr_dev, eq, mailbox->buf);
eq               5793 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	ret = hns_roce_cmd_mbox(hr_dev, mailbox->dma, 0, eq->eqn, 0,
eq               5806 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		dma_free_coherent(dev, buf_chk_sz, eq->buf_list->buf,
eq               5807 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				  eq->buf_list->map);
eq               5809 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		hns_roce_mhop_free_eq(hr_dev, eq);
eq               5814 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq->buf_list);
eq               5860 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			ret = request_irq(eq_table->eq[j - other_num].irq,
eq               5863 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					  &eq_table->eq[j - other_num]);
eq               5865 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			ret = request_irq(eq_table->eq[j - other_num].irq,
eq               5868 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					  &eq_table->eq[j - other_num]);
eq               5882 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			free_irq(eq_table->eq[j - other_num].irq,
eq               5883 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				 &eq_table->eq[j - other_num]);
eq               5905 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		free_irq(hr_dev->eq_table.eq[i].irq, &hr_dev->eq_table.eq[i]);
eq               5915 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_eq *eq;
eq               5932 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	eq_table->eq = kcalloc(eq_num, sizeof(*eq_table->eq), GFP_KERNEL);
eq               5933 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (!eq_table->eq)
eq               5938 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq = &eq_table->eq[i];
eq               5939 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->hr_dev = hr_dev;
eq               5940 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		eq->eqn = i;
eq               5944 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->type_flag = HNS_ROCE_CEQ;
eq               5945 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->entries = hr_dev->caps.ceqe_depth;
eq               5946 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->eqe_size = HNS_ROCE_CEQ_ENTRY_SIZE;
eq               5947 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->irq = hr_dev->irq[i + other_num + aeq_num];
eq               5948 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->eq_max_cnt = HNS_ROCE_CEQ_DEFAULT_BURST_NUM;
eq               5949 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->eq_period = HNS_ROCE_CEQ_DEFAULT_INTERVAL;
eq               5953 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->type_flag = HNS_ROCE_AEQ;
eq               5954 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->entries = hr_dev->caps.aeqe_depth;
eq               5955 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->eqe_size = HNS_ROCE_AEQ_ENTRY_SIZE;
eq               5956 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->irq = hr_dev->irq[i - comp_num + other_num];
eq               5957 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->eq_max_cnt = HNS_ROCE_AEQ_DEFAULT_BURST_NUM;
eq               5958 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			eq->eq_period = HNS_ROCE_AEQ_DEFAULT_INTERVAL;
eq               5961 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		ret = hns_roce_v2_create_eq(hr_dev, eq, eq_cmd);
eq               5996 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		hns_roce_v2_free_eq(hr_dev, &eq_table->eq[i]);
eq               5997 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq_table->eq);
eq               6018 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		hns_roce_v2_free_eq(hr_dev, &eq_table->eq[i]);
eq               6021 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	kfree(eq_table->eq);
eq               2436 drivers/infiniband/hw/mlx4/main.c 	int i, j, eq = 0, total_eqs = 0;
eq               2448 drivers/infiniband/hw/mlx4/main.c 			ibdev->eq_table[eq] = total_eqs;
eq               2450 drivers/infiniband/hw/mlx4/main.c 					    &ibdev->eq_table[eq]))
eq               2451 drivers/infiniband/hw/mlx4/main.c 				eq++;
eq               2453 drivers/infiniband/hw/mlx4/main.c 				ibdev->eq_table[eq] = -1;
eq               2457 drivers/infiniband/hw/mlx4/main.c 	for (i = eq; i < dev->caps.num_comp_vectors;
eq               2462 drivers/infiniband/hw/mlx4/main.c 	ibdev->ib_dev.num_comp_vectors = eq;
eq                 76 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_ib_pf_eq	*eq;
eq               1355 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_ib_pf_eq *eq = pfault->eq;
eq               1357 drivers/infiniband/hw/mlx5/odp.c 	mlx5_ib_pfault(eq->dev, pfault);
eq               1358 drivers/infiniband/hw/mlx5/odp.c 	mempool_free(pfault, eq->pool);
eq               1361 drivers/infiniband/hw/mlx5/odp.c static void mlx5_ib_eq_pf_process(struct mlx5_ib_pf_eq *eq)
eq               1368 drivers/infiniband/hw/mlx5/odp.c 	while ((eqe = mlx5_eq_get_eqe(eq->core, cc))) {
eq               1369 drivers/infiniband/hw/mlx5/odp.c 		pfault = mempool_alloc(eq->pool, GFP_ATOMIC);
eq               1371 drivers/infiniband/hw/mlx5/odp.c 			schedule_work(&eq->work);
eq               1379 drivers/infiniband/hw/mlx5/odp.c 		mlx5_ib_dbg(eq->dev,
eq               1399 drivers/infiniband/hw/mlx5/odp.c 			mlx5_ib_dbg(eq->dev,
eq               1403 drivers/infiniband/hw/mlx5/odp.c 			mlx5_ib_dbg(eq->dev,
eq               1422 drivers/infiniband/hw/mlx5/odp.c 			mlx5_ib_dbg(eq->dev,
eq               1430 drivers/infiniband/hw/mlx5/odp.c 			mlx5_ib_warn(eq->dev,
eq               1438 drivers/infiniband/hw/mlx5/odp.c 		pfault->eq = eq;
eq               1440 drivers/infiniband/hw/mlx5/odp.c 		queue_work(eq->wq, &pfault->work);
eq               1442 drivers/infiniband/hw/mlx5/odp.c 		cc = mlx5_eq_update_cc(eq->core, ++cc);
eq               1445 drivers/infiniband/hw/mlx5/odp.c 	mlx5_eq_update_ci(eq->core, cc, 1);
eq               1451 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_ib_pf_eq *eq =
eq               1455 drivers/infiniband/hw/mlx5/odp.c 	if (spin_trylock_irqsave(&eq->lock, flags)) {
eq               1456 drivers/infiniband/hw/mlx5/odp.c 		mlx5_ib_eq_pf_process(eq);
eq               1457 drivers/infiniband/hw/mlx5/odp.c 		spin_unlock_irqrestore(&eq->lock, flags);
eq               1459 drivers/infiniband/hw/mlx5/odp.c 		schedule_work(&eq->work);
eq               1477 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_ib_pf_eq *eq =
eq               1480 drivers/infiniband/hw/mlx5/odp.c 	mempool_refill(eq->pool);
eq               1482 drivers/infiniband/hw/mlx5/odp.c 	spin_lock_irq(&eq->lock);
eq               1483 drivers/infiniband/hw/mlx5/odp.c 	mlx5_ib_eq_pf_process(eq);
eq               1484 drivers/infiniband/hw/mlx5/odp.c 	spin_unlock_irq(&eq->lock);
eq               1493 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_create_pf_eq(struct mlx5_ib_dev *dev, struct mlx5_ib_pf_eq *eq)
eq               1498 drivers/infiniband/hw/mlx5/odp.c 	INIT_WORK(&eq->work, mlx5_ib_eq_pf_action);
eq               1499 drivers/infiniband/hw/mlx5/odp.c 	spin_lock_init(&eq->lock);
eq               1500 drivers/infiniband/hw/mlx5/odp.c 	eq->dev = dev;
eq               1502 drivers/infiniband/hw/mlx5/odp.c 	eq->pool = mempool_create_kmalloc_pool(MLX5_IB_NUM_PF_DRAIN,
eq               1504 drivers/infiniband/hw/mlx5/odp.c 	if (!eq->pool)
eq               1507 drivers/infiniband/hw/mlx5/odp.c 	eq->wq = alloc_workqueue("mlx5_ib_page_fault",
eq               1510 drivers/infiniband/hw/mlx5/odp.c 	if (!eq->wq) {
eq               1515 drivers/infiniband/hw/mlx5/odp.c 	eq->irq_nb.notifier_call = mlx5_ib_eq_pf_int;
eq               1521 drivers/infiniband/hw/mlx5/odp.c 	eq->core = mlx5_eq_create_generic(dev->mdev, &param);
eq               1522 drivers/infiniband/hw/mlx5/odp.c 	if (IS_ERR(eq->core)) {
eq               1523 drivers/infiniband/hw/mlx5/odp.c 		err = PTR_ERR(eq->core);
eq               1526 drivers/infiniband/hw/mlx5/odp.c 	err = mlx5_eq_enable(dev->mdev, eq->core, &eq->irq_nb);
eq               1534 drivers/infiniband/hw/mlx5/odp.c 	mlx5_eq_destroy_generic(dev->mdev, eq->core);
eq               1536 drivers/infiniband/hw/mlx5/odp.c 	destroy_workqueue(eq->wq);
eq               1538 drivers/infiniband/hw/mlx5/odp.c 	mempool_destroy(eq->pool);
eq               1543 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_destroy_pf_eq(struct mlx5_ib_dev *dev, struct mlx5_ib_pf_eq *eq)
eq               1547 drivers/infiniband/hw/mlx5/odp.c 	mlx5_eq_disable(dev->mdev, eq->core, &eq->irq_nb);
eq               1548 drivers/infiniband/hw/mlx5/odp.c 	err = mlx5_eq_destroy_generic(dev->mdev, eq->core);
eq               1549 drivers/infiniband/hw/mlx5/odp.c 	cancel_work_sync(&eq->work);
eq               1550 drivers/infiniband/hw/mlx5/odp.c 	destroy_workqueue(eq->wq);
eq               1551 drivers/infiniband/hw/mlx5/odp.c 	mempool_destroy(eq->pool);
eq                831 drivers/infiniband/hw/mthca/mthca_cq.c 	cq_context->error_eqn       = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn);
eq                832 drivers/infiniband/hw/mthca/mthca_cq.c 	cq_context->comp_eqn        = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_COMP].eqn);
eq                932 drivers/infiniband/hw/mthca/mthca_cq.c 		synchronize_irq(dev->eq_table.eq[MTHCA_EQ_COMP].msi_x_vector);
eq                231 drivers/infiniband/hw/mthca/mthca_dev.h 	struct mthca_eq    eq[MTHCA_NUM_EQ];
eq                173 drivers/infiniband/hw/mthca/mthca_eq.c static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci)
eq                184 drivers/infiniband/hw/mthca/mthca_eq.c 	mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1),
eq                189 drivers/infiniband/hw/mthca/mthca_eq.c static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci)
eq                194 drivers/infiniband/hw/mthca/mthca_eq.c 		     dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8);
eq                199 drivers/infiniband/hw/mthca/mthca_eq.c static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci)
eq                202 drivers/infiniband/hw/mthca/mthca_eq.c 		arbel_set_eq_ci(dev, eq, ci);
eq                204 drivers/infiniband/hw/mthca/mthca_eq.c 		tavor_set_eq_ci(dev, eq, ci);
eq                228 drivers/infiniband/hw/mthca/mthca_eq.c static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry)
eq                230 drivers/infiniband/hw/mthca/mthca_eq.c 	unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE;
eq                231 drivers/infiniband/hw/mthca/mthca_eq.c 	return eq->page_list[off / PAGE_SIZE].buf + off % PAGE_SIZE;
eq                234 drivers/infiniband/hw/mthca/mthca_eq.c static inline struct mthca_eqe *next_eqe_sw(struct mthca_eq *eq)
eq                237 drivers/infiniband/hw/mthca/mthca_eq.c 	eqe = get_eqe(eq, eq->cons_index);
eq                260 drivers/infiniband/hw/mthca/mthca_eq.c static int mthca_eq_int(struct mthca_dev *dev, struct mthca_eq *eq)
eq                267 drivers/infiniband/hw/mthca/mthca_eq.c 	while ((eqe = next_eqe_sw(eq))) {
eq                277 drivers/infiniband/hw/mthca/mthca_eq.c 			disarm_cq(dev, eq->eqn, disarm_cqn);
eq                349 drivers/infiniband/hw/mthca/mthca_eq.c 			mthca_warn(dev, "EQ overrun on EQN %d\n", eq->eqn);
eq                358 drivers/infiniband/hw/mthca/mthca_eq.c 				   eqe->type, eqe->subtype, eq->eqn);
eq                363 drivers/infiniband/hw/mthca/mthca_eq.c 		++eq->cons_index;
eq                379 drivers/infiniband/hw/mthca/mthca_eq.c 			set_eq_ci(dev, eq, eq->cons_index);
eq                408 drivers/infiniband/hw/mthca/mthca_eq.c 		if (ecr & dev->eq_table.eq[i].eqn_mask) {
eq                409 drivers/infiniband/hw/mthca/mthca_eq.c 			if (mthca_eq_int(dev, &dev->eq_table.eq[i]))
eq                410 drivers/infiniband/hw/mthca/mthca_eq.c 				tavor_set_eq_ci(dev, &dev->eq_table.eq[i],
eq                411 drivers/infiniband/hw/mthca/mthca_eq.c 						dev->eq_table.eq[i].cons_index);
eq                412 drivers/infiniband/hw/mthca/mthca_eq.c 			tavor_eq_req_not(dev, dev->eq_table.eq[i].eqn);
eq                420 drivers/infiniband/hw/mthca/mthca_eq.c 	struct mthca_eq  *eq  = eq_ptr;
eq                421 drivers/infiniband/hw/mthca/mthca_eq.c 	struct mthca_dev *dev = eq->dev;
eq                423 drivers/infiniband/hw/mthca/mthca_eq.c 	mthca_eq_int(dev, eq);
eq                424 drivers/infiniband/hw/mthca/mthca_eq.c 	tavor_set_eq_ci(dev, eq, eq->cons_index);
eq                425 drivers/infiniband/hw/mthca/mthca_eq.c 	tavor_eq_req_not(dev, eq->eqn);
eq                441 drivers/infiniband/hw/mthca/mthca_eq.c 		if (mthca_eq_int(dev, &dev->eq_table.eq[i])) {
eq                443 drivers/infiniband/hw/mthca/mthca_eq.c 			arbel_set_eq_ci(dev, &dev->eq_table.eq[i],
eq                444 drivers/infiniband/hw/mthca/mthca_eq.c 					dev->eq_table.eq[i].cons_index);
eq                454 drivers/infiniband/hw/mthca/mthca_eq.c 	struct mthca_eq  *eq  = eq_ptr;
eq                455 drivers/infiniband/hw/mthca/mthca_eq.c 	struct mthca_dev *dev = eq->dev;
eq                457 drivers/infiniband/hw/mthca/mthca_eq.c 	mthca_eq_int(dev, eq);
eq                458 drivers/infiniband/hw/mthca/mthca_eq.c 	arbel_set_eq_ci(dev, eq, eq->cons_index);
eq                459 drivers/infiniband/hw/mthca/mthca_eq.c 	arbel_eq_req_not(dev, eq->eqn_mask);
eq                468 drivers/infiniband/hw/mthca/mthca_eq.c 			   struct mthca_eq *eq)
eq                478 drivers/infiniband/hw/mthca/mthca_eq.c 	eq->dev  = dev;
eq                479 drivers/infiniband/hw/mthca/mthca_eq.c 	eq->nent = roundup_pow_of_two(max(nent, 2));
eq                480 drivers/infiniband/hw/mthca/mthca_eq.c 	npages = ALIGN(eq->nent * MTHCA_EQ_ENTRY_SIZE, PAGE_SIZE) / PAGE_SIZE;
eq                482 drivers/infiniband/hw/mthca/mthca_eq.c 	eq->page_list = kmalloc_array(npages, sizeof(*eq->page_list),
eq                484 drivers/infiniband/hw/mthca/mthca_eq.c 	if (!eq->page_list)
eq                488 drivers/infiniband/hw/mthca/mthca_eq.c 		eq->page_list[i].buf = NULL;
eq                500 drivers/infiniband/hw/mthca/mthca_eq.c 		eq->page_list[i].buf = dma_alloc_coherent(&dev->pdev->dev,
eq                502 drivers/infiniband/hw/mthca/mthca_eq.c 		if (!eq->page_list[i].buf)
eq                506 drivers/infiniband/hw/mthca/mthca_eq.c 		dma_unmap_addr_set(&eq->page_list[i], mapping, t);
eq                508 drivers/infiniband/hw/mthca/mthca_eq.c 		clear_page(eq->page_list[i].buf);
eq                511 drivers/infiniband/hw/mthca/mthca_eq.c 	for (i = 0; i < eq->nent; ++i)
eq                512 drivers/infiniband/hw/mthca/mthca_eq.c 		set_eqe_hw(get_eqe(eq, i));
eq                514 drivers/infiniband/hw/mthca/mthca_eq.c 	eq->eqn = mthca_alloc(&dev->eq_table.alloc);
eq                515 drivers/infiniband/hw/mthca/mthca_eq.c 	if (eq->eqn == -1)
eq                523 drivers/infiniband/hw/mthca/mthca_eq.c 				  &eq->mr);
eq                535 drivers/infiniband/hw/mthca/mthca_eq.c 	eq_context->logsize_usrpage = cpu_to_be32((ffs(eq->nent) - 1) << 24);
eq                543 drivers/infiniband/hw/mthca/mthca_eq.c 	eq_context->lkey            = cpu_to_be32(eq->mr.ibmr.lkey);
eq                545 drivers/infiniband/hw/mthca/mthca_eq.c 	err = mthca_SW2HW_EQ(dev, mailbox, eq->eqn);
eq                554 drivers/infiniband/hw/mthca/mthca_eq.c 	eq->eqn_mask   = swab32(1 << eq->eqn);
eq                555 drivers/infiniband/hw/mthca/mthca_eq.c 	eq->cons_index = 0;
eq                557 drivers/infiniband/hw/mthca/mthca_eq.c 	dev->eq_table.arm_mask |= eq->eqn_mask;
eq                560 drivers/infiniband/hw/mthca/mthca_eq.c 		  eq->eqn, eq->nent);
eq                565 drivers/infiniband/hw/mthca/mthca_eq.c 	mthca_free_mr(dev, &eq->mr);
eq                568 drivers/infiniband/hw/mthca/mthca_eq.c 	mthca_free(&dev->eq_table.alloc, eq->eqn);
eq                572 drivers/infiniband/hw/mthca/mthca_eq.c 		if (eq->page_list[i].buf)
eq                574 drivers/infiniband/hw/mthca/mthca_eq.c 					  eq->page_list[i].buf,
eq                575 drivers/infiniband/hw/mthca/mthca_eq.c 					  dma_unmap_addr(&eq->page_list[i],
eq                581 drivers/infiniband/hw/mthca/mthca_eq.c 	kfree(eq->page_list);
eq                589 drivers/infiniband/hw/mthca/mthca_eq.c 			  struct mthca_eq *eq)
eq                593 drivers/infiniband/hw/mthca/mthca_eq.c 	int npages = (eq->nent * MTHCA_EQ_ENTRY_SIZE + PAGE_SIZE - 1) /
eq                601 drivers/infiniband/hw/mthca/mthca_eq.c 	err = mthca_HW2SW_EQ(dev, mailbox, eq->eqn);
eq                605 drivers/infiniband/hw/mthca/mthca_eq.c 	dev->eq_table.arm_mask &= ~eq->eqn_mask;
eq                608 drivers/infiniband/hw/mthca/mthca_eq.c 		mthca_dbg(dev, "Dumping EQ context %02x:\n", eq->eqn);
eq                618 drivers/infiniband/hw/mthca/mthca_eq.c 	mthca_free_mr(dev, &eq->mr);
eq                621 drivers/infiniband/hw/mthca/mthca_eq.c 				    eq->page_list[i].buf,
eq                622 drivers/infiniband/hw/mthca/mthca_eq.c 				    dma_unmap_addr(&eq->page_list[i], mapping));
eq                624 drivers/infiniband/hw/mthca/mthca_eq.c 	kfree(eq->page_list);
eq                635 drivers/infiniband/hw/mthca/mthca_eq.c 		if (dev->eq_table.eq[i].have_irq) {
eq                636 drivers/infiniband/hw/mthca/mthca_eq.c 			free_irq(dev->eq_table.eq[i].msi_x_vector,
eq                637 drivers/infiniband/hw/mthca/mthca_eq.c 				 dev->eq_table.eq + i);
eq                638 drivers/infiniband/hw/mthca/mthca_eq.c 			dev->eq_table.eq[i].have_irq = 0;
eq                799 drivers/infiniband/hw/mthca/mthca_eq.c 			      &dev->eq_table.eq[MTHCA_EQ_COMP]);
eq                805 drivers/infiniband/hw/mthca/mthca_eq.c 			      &dev->eq_table.eq[MTHCA_EQ_ASYNC]);
eq                811 drivers/infiniband/hw/mthca/mthca_eq.c 			      &dev->eq_table.eq[MTHCA_EQ_CMD]);
eq                823 drivers/infiniband/hw/mthca/mthca_eq.c 			snprintf(dev->eq_table.eq[i].irq_name,
eq                827 drivers/infiniband/hw/mthca/mthca_eq.c 			err = request_irq(dev->eq_table.eq[i].msi_x_vector,
eq                831 drivers/infiniband/hw/mthca/mthca_eq.c 					  0, dev->eq_table.eq[i].irq_name,
eq                832 drivers/infiniband/hw/mthca/mthca_eq.c 					  dev->eq_table.eq + i);
eq                835 drivers/infiniband/hw/mthca/mthca_eq.c 			dev->eq_table.eq[i].have_irq = 1;
eq                838 drivers/infiniband/hw/mthca/mthca_eq.c 		snprintf(dev->eq_table.eq[0].irq_name, IB_DEVICE_NAME_MAX,
eq                844 drivers/infiniband/hw/mthca/mthca_eq.c 				  IRQF_SHARED, dev->eq_table.eq[0].irq_name, dev);
eq                851 drivers/infiniband/hw/mthca/mthca_eq.c 			   0, dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn);
eq                854 drivers/infiniband/hw/mthca/mthca_eq.c 			   dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn, err);
eq                857 drivers/infiniband/hw/mthca/mthca_eq.c 			   0, dev->eq_table.eq[MTHCA_EQ_CMD].eqn);
eq                860 drivers/infiniband/hw/mthca/mthca_eq.c 			   dev->eq_table.eq[MTHCA_EQ_CMD].eqn, err);
eq                864 drivers/infiniband/hw/mthca/mthca_eq.c 			arbel_eq_req_not(dev, dev->eq_table.eq[i].eqn_mask);
eq                866 drivers/infiniband/hw/mthca/mthca_eq.c 			tavor_eq_req_not(dev, dev->eq_table.eq[i].eqn);
eq                872 drivers/infiniband/hw/mthca/mthca_eq.c 	mthca_free_eq(dev, &dev->eq_table.eq[MTHCA_EQ_CMD]);
eq                875 drivers/infiniband/hw/mthca/mthca_eq.c 	mthca_free_eq(dev, &dev->eq_table.eq[MTHCA_EQ_ASYNC]);
eq                878 drivers/infiniband/hw/mthca/mthca_eq.c 	mthca_free_eq(dev, &dev->eq_table.eq[MTHCA_EQ_COMP]);
eq                895 drivers/infiniband/hw/mthca/mthca_eq.c 		     1, dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn);
eq                897 drivers/infiniband/hw/mthca/mthca_eq.c 		     1, dev->eq_table.eq[MTHCA_EQ_CMD].eqn);
eq                900 drivers/infiniband/hw/mthca/mthca_eq.c 		mthca_free_eq(dev, &dev->eq_table.eq[i]);
eq                762 drivers/infiniband/hw/mthca/mthca_main.c 				   dev->eq_table.eq[MTHCA_EQ_CMD].msi_x_vector);
eq                859 drivers/infiniband/hw/mthca/mthca_main.c 	mdev->eq_table.eq[MTHCA_EQ_COMP ].msi_x_vector =
eq                861 drivers/infiniband/hw/mthca/mthca_main.c 	mdev->eq_table.eq[MTHCA_EQ_ASYNC].msi_x_vector =
eq                863 drivers/infiniband/hw/mthca/mthca_main.c 	mdev->eq_table.eq[MTHCA_EQ_CMD  ].msi_x_vector =
eq                111 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static inline void *ocrdma_get_eqe(struct ocrdma_eq *eq)
eq                113 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	return eq->q.va + (eq->q.tail * sizeof(struct ocrdma_eqe));
eq                116 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static inline void ocrdma_eq_inc_tail(struct ocrdma_eq *eq)
eq                118 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	eq->q.tail = (eq->q.tail + 1) & (OCRDMA_EQ_LEN - 1);
eq                433 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static int ocrdma_mbx_create_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq)
eq                448 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma,
eq                453 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		eq->q.id = rsp->vector_eqid & 0xffff;
eq                454 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		eq->vector = (rsp->vector_eqid >> 16) & 0xffff;
eq                455 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		eq->q.created = true;
eq                461 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			    struct ocrdma_eq *eq, u16 q_len)
eq                465 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	status = ocrdma_alloc_q(dev, &eq->q, OCRDMA_EQ_LEN,
eq                470 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	status = ocrdma_mbx_create_eq(dev, eq);
eq                473 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	eq->dev = dev;
eq                474 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_ring_eq_db(dev, eq->q.id, true, true, 0);
eq                478 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_free_q(dev, &eq->q);
eq                482 drivers/infiniband/hw/ocrdma/ocrdma_hw.c int ocrdma_get_irq(struct ocrdma_dev *dev, struct ocrdma_eq *eq)
eq                489 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		irq = dev->nic_info.msix.vector_list[eq->vector];
eq                493 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static void _ocrdma_destroy_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq)
eq                495 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	if (eq->q.created) {
eq                496 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		ocrdma_mbx_delete_q(dev, &eq->q, QTYPE_EQ);
eq                497 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		ocrdma_free_q(dev, &eq->q);
eq                501 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static void ocrdma_destroy_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq)
eq                508 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_ring_eq_db(dev, eq->q.id, false, false, 0);
eq                510 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	irq = ocrdma_get_irq(dev, eq);
eq                511 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	free_irq(irq, eq);
eq                512 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	_ocrdma_destroy_eq(dev, eq);
eq                525 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 				   struct ocrdma_queue_info *eq)
eq                541 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	cmd->eqn = eq->id;
eq                996 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_eq *eq = handle;
eq                997 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_dev *dev = eq->dev;
eq               1002 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	int budget = eq->cq_cnt;
eq               1005 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		ptr = ocrdma_get_eqe(eq);
eq               1012 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			       eq->q.id, eqe.id_valid);
eq               1018 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		ocrdma_ring_eq_db(dev, eq->q.id, false, true, 1);
eq               1024 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		ocrdma_eq_inc_tail(eq);
eq               1034 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	eq->aic_obj.eq_intr_cnt++;
eq               1035 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_ring_eq_db(dev, eq->q.id, true, true, 0);
eq               3117 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static int ocrdma_mbx_modify_eqd(struct ocrdma_dev *dev, struct ocrdma_eq *eq,
eq               3132 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		cmd->cmd.set_eqd[i].eq_id = eq[i].q.id;
eq               3135 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 				(eq[i].aic_obj.prev_eqd * 65)/100;
eq               3143 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static int ocrdma_modify_eqd(struct ocrdma_dev *dev, struct ocrdma_eq *eq,
eq               3150 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			ocrdma_mbx_modify_eqd(dev, &eq[i], num_eqs);
eq               3155 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		ocrdma_mbx_modify_eqd(dev, eq, num);
eq               3164 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_eq *eq = NULL;
eq               3169 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		eq = &dev->eq_tbl[i];
eq               3170 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		if (eq->aic_obj.eq_intr_cnt > eq->aic_obj.prev_eq_intr_cnt) {
eq               3171 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			eq_intr = eq->aic_obj.eq_intr_cnt -
eq               3172 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 				  eq->aic_obj.prev_eq_intr_cnt;
eq               3174 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			    (eq->aic_obj.prev_eqd == EQ_AIC_MIN_EQD)) {
eq               3175 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 				eq->aic_obj.prev_eqd = EQ_AIC_MAX_EQD;
eq               3178 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 				   (eq->aic_obj.prev_eqd == EQ_AIC_MAX_EQD)) {
eq               3179 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 				eq->aic_obj.prev_eqd = EQ_AIC_MIN_EQD;
eq               3183 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		eq->aic_obj.prev_eq_intr_cnt = eq->aic_obj.eq_intr_cnt;
eq                150 drivers/infiniband/hw/ocrdma/ocrdma_hw.h int ocrdma_get_irq(struct ocrdma_dev *dev, struct ocrdma_eq *eq);
eq               1070 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_eq *eq = NULL;
eq               1078 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	eq = &dev->eq_tbl[indx];
eq               1079 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	irq = ocrdma_get_irq(dev, eq);
eq                315 drivers/input/input-mt.c static int adjust_dual(int *begin, int step, int *end, int eq, int mu)
eq                333 drivers/input/input-mt.c 	if (c == 0 || (c > mu && (!eq || mu > 0)))
eq               1007 drivers/misc/habanalabs/goya/goya.c 	struct hl_eq *eq;
eq               1018 drivers/misc/habanalabs/goya/goya.c 	eq = &hdev->event_queue;
eq               1023 drivers/misc/habanalabs/goya/goya.c 	WREG32(mmCPU_EQ_BASE_ADDR_LOW, lower_32_bits(eq->bus_address));
eq               1024 drivers/misc/habanalabs/goya/goya.c 	WREG32(mmCPU_EQ_BASE_ADDR_HIGH, upper_32_bits(eq->bus_address));
eq                153 drivers/misc/habanalabs/irq.c 	struct hl_eq *eq = arg;
eq                154 drivers/misc/habanalabs/irq.c 	struct hl_device *hdev = eq->hdev;
eq                159 drivers/misc/habanalabs/irq.c 	eq_base = (struct hl_eq_entry *) (uintptr_t) eq->kernel_address;
eq                163 drivers/misc/habanalabs/irq.c 			((le32_to_cpu(eq_base[eq->ci].hdr.ctl) &
eq                169 drivers/misc/habanalabs/irq.c 		eq_entry = &eq_base[eq->ci];
eq                200 drivers/misc/habanalabs/irq.c 		eq->ci = hl_eq_inc_ptr(eq->ci);
eq                202 drivers/misc/habanalabs/irq.c 		hdev->asic_funcs->update_eq_ci(hdev, eq->ci);
eq                581 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			struct sge_eth_txq *eq;
eq                583 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			eq = container_of(txq, struct sge_eth_txq, q);
eq                584 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			t4_sge_eth_txq_egress_update(q->adap, eq, -1);
eq               1324 drivers/net/ethernet/chelsio/cxgb4/sge.c int t4_sge_eth_txq_egress_update(struct adapter *adap, struct sge_eth_txq *eq,
eq               1328 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge_txq *q = &eq->q;
eq               1331 drivers/net/ethernet/chelsio/cxgb4/sge.c 	if (!q->in_use || !__netif_tx_trylock(eq->txq))
eq               1335 drivers/net/ethernet/chelsio/cxgb4/sge.c 	reclaimed = reclaim_completed_tx(adap, &eq->q, maxreclaim, true);
eq               1347 drivers/net/ethernet/chelsio/cxgb4/sge.c 	if (netif_tx_queue_stopped(eq->txq) && hw_in_use < (q->size / 2)) {
eq               1348 drivers/net/ethernet/chelsio/cxgb4/sge.c 		netif_tx_wake_queue(eq->txq);
eq               1349 drivers/net/ethernet/chelsio/cxgb4/sge.c 		eq->q.restarts++;
eq               1352 drivers/net/ethernet/chelsio/cxgb4/sge.c 	__netif_tx_unlock(eq->txq);
eq               4011 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge_eth_rxq *eq;
eq               4016 drivers/net/ethernet/chelsio/cxgb4/sge.c 		eq = &adap->sge.ethrxq[i];
eq               4017 drivers/net/ethernet/chelsio/cxgb4/sge.c 		if (eq->rspq.desc)
eq               4020 drivers/net/ethernet/chelsio/cxgb4/sge.c 				   eq->rspq.cntxt_id,
eq               4021 drivers/net/ethernet/chelsio/cxgb4/sge.c 				   eq->fl.size ? eq->fl.cntxt_id : 0xffff,
eq               4027 drivers/net/ethernet/chelsio/cxgb4/sge.c 		eq = &adap->sge.ethrxq[i];
eq               4028 drivers/net/ethernet/chelsio/cxgb4/sge.c 		if (eq->rspq.desc)
eq               4029 drivers/net/ethernet/chelsio/cxgb4/sge.c 			free_rspq_fl(adap, &eq->rspq,
eq               4030 drivers/net/ethernet/chelsio/cxgb4/sge.c 				     eq->fl.size ? &eq->fl : NULL);
eq               1157 drivers/net/ethernet/emulex/benet/be_cmds.c 		     struct be_queue_info *eq, bool no_delay, int coalesce_wm)
eq               1187 drivers/net/ethernet/emulex/benet/be_cmds.c 		AMAP_SET_BITS(struct amap_cq_context_be, eqid, ctxt, eq->id);
eq               1204 drivers/net/ethernet/emulex/benet/be_cmds.c 		AMAP_SET_BITS(struct amap_cq_context_v2, eqid, ctxt, eq->id);
eq               2397 drivers/net/ethernet/emulex/benet/be_cmds.h 		     struct be_queue_info *eq, bool no_delay,
eq               2952 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_queue_info *eq;
eq               2972 drivers/net/ethernet/emulex/benet/be_main.c 		eq = &eqo->q;
eq               2973 drivers/net/ethernet/emulex/benet/be_main.c 		rc = be_queue_alloc(adapter, eq, EVNT_Q_LEN,
eq               3120 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_queue_info *eq, *cq;
eq               3149 drivers/net/ethernet/emulex/benet/be_main.c 		eq = &adapter->eq_obj[i % adapter->num_evt_qs].q;
eq               3150 drivers/net/ethernet/emulex/benet/be_main.c 		rc = be_cmd_cq_create(adapter, cq, eq, false, 3);
eq                 26 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_EQ_NUM_PAGES(eq, pg_size)           \
eq                 27 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		(ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size))
eq                 29 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size)     ((pg_size) / (eq)->elem_size)
eq                 31 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_CONS_IDX_REG_ADDR(eq)        (((eq)->type == HINIC_AEQ) ? \
eq                 32 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			HINIC_CSR_AEQ_CONS_IDX_ADDR((eq)->q_id) : \
eq                 33 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			HINIC_CSR_CEQ_CONS_IDX_ADDR((eq)->q_id))
eq                 35 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_PROD_IDX_REG_ADDR(eq)        (((eq)->type == HINIC_AEQ) ? \
eq                 36 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			HINIC_CSR_AEQ_PROD_IDX_ADDR((eq)->q_id) : \
eq                 37 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			HINIC_CSR_CEQ_PROD_IDX_ADDR((eq)->q_id))
eq                 39 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_HI_PHYS_ADDR_REG(eq, pg_num) (((eq)->type == HINIC_AEQ) ? \
eq                 40 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			HINIC_CSR_AEQ_HI_PHYS_ADDR_REG((eq)->q_id, pg_num) : \
eq                 41 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			HINIC_CSR_CEQ_HI_PHYS_ADDR_REG((eq)->q_id, pg_num))
eq                 43 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_LO_PHYS_ADDR_REG(eq, pg_num) (((eq)->type == HINIC_AEQ) ? \
eq                 44 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			HINIC_CSR_AEQ_LO_PHYS_ADDR_REG((eq)->q_id, pg_num) : \
eq                 45 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			HINIC_CSR_CEQ_LO_PHYS_ADDR_REG((eq)->q_id, pg_num))
eq                 47 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_EQ_ELEMENT(eq, idx)         \
eq                 48 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		((eq)->virt_addr[(idx) / (eq)->num_elem_in_pg] + \
eq                 49 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		 (((idx) & ((eq)->num_elem_in_pg - 1)) * (eq)->elem_size))
eq                 51 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_AEQ_ELEM(eq, idx)           ((struct hinic_aeq_elem *) \
eq                 52 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 					GET_EQ_ELEMENT(eq, idx))
eq                 54 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_CEQ_ELEM(eq, idx)           ((u32 *) \
eq                 55 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 					 GET_EQ_ELEMENT(eq, idx))
eq                 57 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_CURR_AEQ_ELEM(eq)           GET_AEQ_ELEM(eq, (eq)->cons_idx)
eq                 59 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_CURR_CEQ_ELEM(eq)           GET_CEQ_ELEM(eq, (eq)->cons_idx)
eq                 62 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_SET_HW_PAGE_SIZE_VAL(eq)     (ilog2(PAGE_IN_4K((eq)->page_size)))
eq                 64 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define ELEMENT_SIZE_IN_32B(eq)         (((eq)->elem_size) >> 5)
eq                 65 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_SET_HW_ELEM_SIZE_VAL(eq)     (ilog2(ELEMENT_SIZE_IN_32B(eq)))
eq                 78 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define aeq_to_aeqs(eq)                 \
eq                 79 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		container_of((eq) - (eq)->q_id, struct hinic_aeqs, aeq[0])
eq                 81 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define ceq_to_ceqs(eq)                 \
eq                 82 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		container_of((eq) - (eq)->q_id, struct hinic_ceqs, ceq[0])
eq                191 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void eq_update_ci(struct hinic_eq *eq, u32 arm_state)
eq                193 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	u32 val, addr = EQ_CONS_IDX_REG_ADDR(eq);
eq                196 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	val = hinic_hwif_read_reg(eq->hwif, addr);
eq                203 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	val |= HINIC_EQ_CI_SET(eq->cons_idx, IDX)    |
eq                204 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	       HINIC_EQ_CI_SET(eq->wrapped, WRAPPED) |
eq                209 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	hinic_hwif_write_reg(eq->hwif, addr, val);
eq                216 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void aeq_irq_handler(struct hinic_eq *eq)
eq                218 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	struct hinic_aeqs *aeqs = aeq_to_aeqs(eq);
eq                228 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	for (i = 0; i < eq->q_len; i++) {
eq                229 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		aeqe_curr = GET_CURR_AEQ_ELEM(eq);
eq                235 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		if (HINIC_EQ_ELEM_DESC_GET(aeqe_desc, WRAPPED) == eq->wrapped)
eq                266 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		eq->cons_idx++;
eq                268 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		if (eq->cons_idx == eq->q_len) {
eq                269 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			eq->cons_idx = 0;
eq                270 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			eq->wrapped = !eq->wrapped;
eq                312 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void ceq_irq_handler(struct hinic_eq *eq)
eq                314 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	struct hinic_ceqs *ceqs = ceq_to_ceqs(eq);
eq                318 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	for (i = 0; i < eq->q_len; i++) {
eq                319 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		ceqe = *(GET_CURR_CEQ_ELEM(eq));
eq                325 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		if (HINIC_EQ_ELEM_DESC_GET(ceqe, WRAPPED) == eq->wrapped)
eq                330 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		eq->cons_idx++;
eq                332 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		if (eq->cons_idx == eq->q_len) {
eq                333 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			eq->cons_idx = 0;
eq                334 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			eq->wrapped = !eq->wrapped;
eq                345 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	struct hinic_eq *eq = data;
eq                347 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	if (eq->type == HINIC_AEQ)
eq                348 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		aeq_irq_handler(eq);
eq                349 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	else if (eq->type == HINIC_CEQ)
eq                350 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		ceq_irq_handler(eq);
eq                352 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq_update_ci(eq, EQ_ARMED);
eq                419 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void set_ctrl0(struct hinic_eq *eq)
eq                421 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	struct msix_entry *msix_entry = &eq->msix_entry;
eq                422 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	enum hinic_eq_type type = eq->type;
eq                427 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		addr = HINIC_CSR_AEQ_CTRL_0_ADDR(eq->q_id);
eq                429 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		val = hinic_hwif_read_reg(eq->hwif, addr);
eq                438 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			HINIC_AEQ_CTRL_0_SET(HINIC_HWIF_PCI_INTF(eq->hwif),
eq                444 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		hinic_hwif_write_reg(eq->hwif, addr, val);
eq                447 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		addr = HINIC_CSR_CEQ_CTRL_0_ADDR(eq->q_id);
eq                449 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		val = hinic_hwif_read_reg(eq->hwif, addr);
eq                460 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			HINIC_CEQ_CTRL_0_SET(HINIC_HWIF_PCI_INTF(eq->hwif),
eq                466 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		hinic_hwif_write_reg(eq->hwif, addr, val);
eq                470 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void set_ctrl1(struct hinic_eq *eq)
eq                472 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	enum hinic_eq_type type = eq->type;
eq                478 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		addr = HINIC_CSR_AEQ_CTRL_1_ADDR(eq->q_id);
eq                480 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		page_size_val = EQ_SET_HW_PAGE_SIZE_VAL(eq);
eq                481 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		elem_size = EQ_SET_HW_ELEM_SIZE_VAL(eq);
eq                483 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		val = hinic_hwif_read_reg(eq->hwif, addr);
eq                489 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		ctrl1 = HINIC_AEQ_CTRL_1_SET(eq->q_len, LEN)            |
eq                495 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		hinic_hwif_write_reg(eq->hwif, addr, val);
eq                498 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		addr = HINIC_CSR_CEQ_CTRL_1_ADDR(eq->q_id);
eq                500 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		page_size_val = EQ_SET_HW_PAGE_SIZE_VAL(eq);
eq                502 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		val = hinic_hwif_read_reg(eq->hwif, addr);
eq                507 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		ctrl1 = HINIC_CEQ_CTRL_1_SET(eq->q_len, LEN) |
eq                512 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		hinic_hwif_write_reg(eq->hwif, addr, val);
eq                520 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void set_eq_ctrls(struct hinic_eq *eq)
eq                522 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	set_ctrl0(eq);
eq                523 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	set_ctrl1(eq);
eq                531 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void aeq_elements_init(struct hinic_eq *eq, u32 init_val)
eq                536 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	for (i = 0; i < eq->q_len; i++) {
eq                537 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		aeqe = GET_AEQ_ELEM(eq, i);
eq                549 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void ceq_elements_init(struct hinic_eq *eq, u32 init_val)
eq                554 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	for (i = 0; i < eq->q_len; i++) {
eq                555 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		ceqe = GET_CEQ_ELEM(eq, i);
eq                568 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static int alloc_eq_pages(struct hinic_eq *eq)
eq                570 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	struct hinic_hwif *hwif = eq->hwif;
eq                576 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	addr_size = eq->num_pages * sizeof(*eq->dma_addr);
eq                577 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->dma_addr = devm_kzalloc(&pdev->dev, addr_size, GFP_KERNEL);
eq                578 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	if (!eq->dma_addr)
eq                581 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	addr_size = eq->num_pages * sizeof(*eq->virt_addr);
eq                582 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->virt_addr = devm_kzalloc(&pdev->dev, addr_size, GFP_KERNEL);
eq                583 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	if (!eq->virt_addr) {
eq                588 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	for (pg = 0; pg < eq->num_pages; pg++) {
eq                589 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		eq->virt_addr[pg] = dma_alloc_coherent(&pdev->dev,
eq                590 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 						       eq->page_size,
eq                591 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 						       &eq->dma_addr[pg],
eq                593 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		if (!eq->virt_addr[pg]) {
eq                598 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		addr = EQ_HI_PHYS_ADDR_REG(eq, pg);
eq                599 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		val = upper_32_bits(eq->dma_addr[pg]);
eq                603 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		addr = EQ_LO_PHYS_ADDR_REG(eq, pg);
eq                604 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		val = lower_32_bits(eq->dma_addr[pg]);
eq                609 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	init_val = HINIC_EQ_ELEM_DESC_SET(eq->wrapped, WRAPPED);
eq                611 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	if (eq->type == HINIC_AEQ)
eq                612 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		aeq_elements_init(eq, init_val);
eq                613 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	else if (eq->type == HINIC_CEQ)
eq                614 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		ceq_elements_init(eq, init_val);
eq                620 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		dma_free_coherent(&pdev->dev, eq->page_size,
eq                621 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 				  eq->virt_addr[pg],
eq                622 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 				  eq->dma_addr[pg]);
eq                624 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	devm_kfree(&pdev->dev, eq->virt_addr);
eq                627 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	devm_kfree(&pdev->dev, eq->dma_addr);
eq                635 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void free_eq_pages(struct hinic_eq *eq)
eq                637 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	struct hinic_hwif *hwif = eq->hwif;
eq                641 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	for (pg = 0; pg < eq->num_pages; pg++)
eq                642 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		dma_free_coherent(&pdev->dev, eq->page_size,
eq                643 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 				  eq->virt_addr[pg],
eq                644 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 				  eq->dma_addr[pg]);
eq                646 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	devm_kfree(&pdev->dev, eq->virt_addr);
eq                647 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	devm_kfree(&pdev->dev, eq->dma_addr);
eq                662 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static int init_eq(struct hinic_eq *eq, struct hinic_hwif *hwif,
eq                669 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->hwif = hwif;
eq                670 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->type = type;
eq                671 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->q_id = q_id;
eq                672 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->q_len = q_len;
eq                673 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->page_size = page_size;
eq                676 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	hinic_hwif_write_reg(eq->hwif, EQ_CONS_IDX_REG_ADDR(eq), 0);
eq                677 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	hinic_hwif_write_reg(eq->hwif, EQ_PROD_IDX_REG_ADDR(eq), 0);
eq                679 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->cons_idx = 0;
eq                680 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->wrapped = 0;
eq                683 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		eq->elem_size = HINIC_AEQE_SIZE;
eq                685 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		eq->elem_size = HINIC_CEQE_SIZE;
eq                691 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->num_pages = GET_EQ_NUM_PAGES(eq, page_size);
eq                692 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->num_elem_in_pg = GET_EQ_NUM_ELEMS_IN_PG(eq, page_size);
eq                694 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->msix_entry = entry;
eq                696 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	if (eq->num_elem_in_pg & (eq->num_elem_in_pg - 1)) {
eq                701 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	if (eq->num_pages > EQ_MAX_PAGES) {
eq                706 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	set_eq_ctrls(eq);
eq                707 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq_update_ci(eq, EQ_ARMED);
eq                709 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	err = alloc_eq_pages(eq);
eq                716 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		struct hinic_eq_work *aeq_work = &eq->aeq_work;
eq                720 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		tasklet_init(&eq->ceq_tasklet, ceq_tasklet,
eq                721 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 			     (unsigned long)eq);
eq                725 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	hinic_msix_attr_set(eq->hwif, eq->msix_entry.entry,
eq                734 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 				  "hinic_aeq", eq);
eq                737 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 				  "hinic_ceq", eq);
eq                747 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	free_eq_pages(eq);
eq                755 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void remove_eq(struct hinic_eq *eq)
eq                757 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	hinic_set_msix_state(eq->hwif, eq->msix_entry.entry,
eq                759 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	free_irq(eq->msix_entry.vector, eq);
eq                761 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	if (eq->type == HINIC_AEQ) {
eq                762 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		struct hinic_eq_work *aeq_work = &eq->aeq_work;
eq                766 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		hinic_hwif_write_reg(eq->hwif,
eq                767 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 				     HINIC_CSR_AEQ_CTRL_1_ADDR(eq->q_id), 0);
eq                768 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	} else if (eq->type == HINIC_CEQ) {
eq                769 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		tasklet_kill(&eq->ceq_tasklet);
eq                771 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 		hinic_hwif_write_reg(eq->hwif,
eq                772 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 				     HINIC_CSR_CEQ_CTRL_1_ADDR(eq->q_id), 0);
eq                776 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq->cons_idx = hinic_hwif_read_reg(eq->hwif, EQ_PROD_IDX_REG_ADDR(eq));
eq                777 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	eq_update_ci(eq, EQ_NOT_ARMED);
eq                779 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c 	free_eq_pages(eq);
eq                352 drivers/net/ethernet/ibm/ehea/ehea.h 	struct ehea_eq *eq;
eq                199 drivers/net/ethernet/ibm/ehea/ehea_main.c 				arr[i++].fwh = pr->eq->fw_handle;
eq               1289 drivers/net/ethernet/ibm/ehea/ehea_main.c 		ret = ibmebus_request_irq(pr->eq->attr.ist1,
eq               1294 drivers/net/ethernet/ibm/ehea/ehea_main.c 				   i, pr->eq->attr.ist1);
eq               1299 drivers/net/ethernet/ibm/ehea/ehea_main.c 			   pr->eq->attr.ist1, i);
eq               1307 drivers/net/ethernet/ibm/ehea/ehea_main.c 		u32 ist = port->port_res[i].eq->attr.ist1;
eq               1329 drivers/net/ethernet/ibm/ehea/ehea_main.c 		ibmebus_free_irq(pr->eq->attr.ist1, pr);
eq               1332 drivers/net/ethernet/ibm/ehea/ehea_main.c 			   i, pr->eq->attr.ist1);
eq               1459 drivers/net/ethernet/ibm/ehea/ehea_main.c 	pr->eq = ehea_create_eq(adapter, eq_type, EHEA_MAX_ENTRIES_EQ, 0);
eq               1460 drivers/net/ethernet/ibm/ehea/ehea_main.c 	if (!pr->eq) {
eq               1466 drivers/net/ethernet/ibm/ehea/ehea_main.c 				     pr->eq->fw_handle,
eq               1474 drivers/net/ethernet/ibm/ehea/ehea_main.c 				     pr->eq->fw_handle,
eq               1560 drivers/net/ethernet/ibm/ehea/ehea_main.c 	ehea_destroy_eq(pr->eq);
eq               1577 drivers/net/ethernet/ibm/ehea/ehea_main.c 		ehea_destroy_eq(pr->eq);
eq                236 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	struct ehea_eq *eq;
eq                238 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	eq = kzalloc(sizeof(*eq), GFP_KERNEL);
eq                239 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	if (!eq)
eq                242 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	eq->adapter = adapter;
eq                243 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	eq->attr.type = type;
eq                244 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	eq->attr.max_nr_of_eqes = max_nr_of_eqes;
eq                245 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	eq->attr.eqe_gen = eqe_gen;
eq                246 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	spin_lock_init(&eq->spinlock);
eq                249 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 					&eq->attr, &eq->fw_handle);
eq                255 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	ret = hw_queue_ctor(&eq->hw_queue, eq->attr.nr_pages,
eq                262 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	for (i = 0; i < eq->attr.nr_pages; i++) {
eq                263 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		vpage = hw_qpageit_get_inc(&eq->hw_queue);
eq                274 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 					     eq->fw_handle, rpage, 1);
eq                276 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		if (i == (eq->attr.nr_pages - 1)) {
eq                278 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 			vpage = hw_qpageit_get_inc(&eq->hw_queue);
eq                289 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	hw_qeit_reset(&eq->hw_queue);
eq                290 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	return eq;
eq                293 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	hw_queue_dtor(&eq->hw_queue);
eq                296 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	ehea_h_free_resource(adapter->handle, eq->fw_handle, FORCE_FREE);
eq                299 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	kfree(eq);
eq                303 drivers/net/ethernet/ibm/ehea/ehea_qmr.c struct ehea_eqe *ehea_poll_eq(struct ehea_eq *eq)
eq                308 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	spin_lock_irqsave(&eq->spinlock, flags);
eq                309 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	eqe = hw_eqit_eq_get_inc_valid(&eq->hw_queue);
eq                310 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	spin_unlock_irqrestore(&eq->spinlock, flags);
eq                315 drivers/net/ethernet/ibm/ehea/ehea_qmr.c static u64 ehea_destroy_eq_res(struct ehea_eq *eq, u64 force)
eq                320 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	spin_lock_irqsave(&eq->spinlock, flags);
eq                322 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	hret = ehea_h_free_resource(eq->adapter->handle, eq->fw_handle, force);
eq                323 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	spin_unlock_irqrestore(&eq->spinlock, flags);
eq                328 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	hw_queue_dtor(&eq->hw_queue);
eq                329 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	kfree(eq);
eq                334 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int ehea_destroy_eq(struct ehea_eq *eq)
eq                337 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	if (!eq)
eq                340 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	hcp_epas_dtor(&eq->epas);
eq                342 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	hret = ehea_destroy_eq_res(eq, NORMAL_FREE);
eq                344 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		ehea_error_data(eq->adapter, eq->fw_handle, &aer, &aerr);
eq                345 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		hret = ehea_destroy_eq_res(eq, FORCE_FREE);
eq                360 drivers/net/ethernet/ibm/ehea/ehea_qmr.h int ehea_destroy_eq(struct ehea_eq *eq);
eq                362 drivers/net/ethernet/ibm/ehea/ehea_qmr.h struct ehea_eqe *ehea_poll_eq(struct ehea_eq *eq);
eq                383 drivers/net/ethernet/mellanox/mlx4/cq.c 	cq_context->comp_eqn	    = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn;
eq                416 drivers/net/ethernet/mellanox/mlx4/cq.c 		&priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].tasklet_ctx;
eq                420 drivers/net/ethernet/mellanox/mlx4/cq.c 	cq->irq = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].irq;
eq                449 drivers/net/ethernet/mellanox/mlx4/cq.c 	synchronize_irq(priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq);
eq                450 drivers/net/ethernet/mellanox/mlx4/cq.c 	if (priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq !=
eq                451 drivers/net/ethernet/mellanox/mlx4/cq.c 	    priv->eq_table.eq[MLX4_EQ_ASYNC].irq)
eq                452 drivers/net/ethernet/mellanox/mlx4/cq.c 		synchronize_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq);
eq                 97 drivers/net/ethernet/mellanox/mlx4/eq.c static void eq_set_ci(struct mlx4_eq *eq, int req_not)
eq                 99 drivers/net/ethernet/mellanox/mlx4/eq.c 	__raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) |
eq                101 drivers/net/ethernet/mellanox/mlx4/eq.c 		     eq->doorbell);
eq                106 drivers/net/ethernet/mellanox/mlx4/eq.c static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry, u8 eqe_factor,
eq                110 drivers/net/ethernet/mellanox/mlx4/eq.c 	unsigned long offset = (entry & (eq->nent - 1)) * eqe_size;
eq                118 drivers/net/ethernet/mellanox/mlx4/eq.c 	return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % PAGE_SIZE;
eq                121 drivers/net/ethernet/mellanox/mlx4/eq.c static struct mlx4_eqe *next_eqe_sw(struct mlx4_eq *eq, u8 eqe_factor, u8 size)
eq                123 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size);
eq                124 drivers/net/ethernet/mellanox/mlx4/eq.c 	return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe;
eq                241 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eq *eq = &priv->eq_table.eq[vec];
eq                243 drivers/net/ethernet/mellanox/mlx4/eq.c 	if (!cpumask_available(eq->affinity_mask) ||
eq                244 drivers/net/ethernet/mellanox/mlx4/eq.c 	    cpumask_empty(eq->affinity_mask))
eq                247 drivers/net/ethernet/mellanox/mlx4/eq.c 	hint_err = irq_set_affinity_hint(eq->irq, eq->affinity_mask);
eq                494 drivers/net/ethernet/mellanox/mlx4/eq.c static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
eq                512 drivers/net/ethernet/mellanox/mlx4/eq.c 	while ((eqe = next_eqe_sw(eq, dev->caps.eqe_factor, eqe_size))) {
eq                543 drivers/net/ethernet/mellanox/mlx4/eq.c 						 eq->eqn, eq->cons_index, ret);
eq                560 drivers/net/ethernet/mellanox/mlx4/eq.c 				 eq->eqn);
eq                573 drivers/net/ethernet/mellanox/mlx4/eq.c 						  eq->eqn, eq->cons_index, ret);
eq                697 drivers/net/ethernet/mellanox/mlx4/eq.c 						 eq->eqn, eq->cons_index, ret);
eq                713 drivers/net/ethernet/mellanox/mlx4/eq.c 			mlx4_warn(dev, "EQ overrun on EQN %d\n", eq->eqn);
eq                781 drivers/net/ethernet/mellanox/mlx4/eq.c 					  eqe->type, eqe->subtype, eq->eqn,
eq                782 drivers/net/ethernet/mellanox/mlx4/eq.c 					  eq->cons_index, eqe->owner, eq->nent,
eq                785 drivers/net/ethernet/mellanox/mlx4/eq.c 					  !!(eq->cons_index & eq->nent) ? "HW" : "SW");
eq                806 drivers/net/ethernet/mellanox/mlx4/eq.c 					 eqe->type, eqe->subtype, eq->eqn,
eq                807 drivers/net/ethernet/mellanox/mlx4/eq.c 					 eq->cons_index, eqe->owner, eq->nent,
eq                809 drivers/net/ethernet/mellanox/mlx4/eq.c 					 !!(eq->cons_index & eq->nent) ? "HW" : "SW");
eq                818 drivers/net/ethernet/mellanox/mlx4/eq.c 				  eqe->type, eqe->subtype, eq->eqn,
eq                819 drivers/net/ethernet/mellanox/mlx4/eq.c 				  eq->cons_index, eqe->owner, eq->nent,
eq                822 drivers/net/ethernet/mellanox/mlx4/eq.c 				  !!(eq->cons_index & eq->nent) ? "HW" : "SW");
eq                826 drivers/net/ethernet/mellanox/mlx4/eq.c 		++eq->cons_index;
eq                838 drivers/net/ethernet/mellanox/mlx4/eq.c 			eq_set_ci(eq, 0);
eq                843 drivers/net/ethernet/mellanox/mlx4/eq.c 	eq_set_ci(eq, 1);
eq                858 drivers/net/ethernet/mellanox/mlx4/eq.c 		work |= mlx4_eq_int(dev, &priv->eq_table.eq[i]);
eq                865 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_eq  *eq  = eq_ptr;
eq                866 drivers/net/ethernet/mellanox/mlx4/eq.c 	struct mlx4_dev *dev = eq->dev;
eq                868 drivers/net/ethernet/mellanox/mlx4/eq.c 	mlx4_eq_int(dev, eq);
eq                934 drivers/net/ethernet/mellanox/mlx4/eq.c static void __iomem *mlx4_get_eq_uar(struct mlx4_dev *dev, struct mlx4_eq *eq)
eq                939 drivers/net/ethernet/mellanox/mlx4/eq.c 	index = eq->eqn / 4 - dev->caps.reserved_eqs / 4;
eq                945 drivers/net/ethernet/mellanox/mlx4/eq.c 				((eq->eqn / 4) << (dev->uar_page_shift)),
eq                949 drivers/net/ethernet/mellanox/mlx4/eq.c 				 eq->eqn);
eq                954 drivers/net/ethernet/mellanox/mlx4/eq.c 	return priv->eq_table.uar_map[index] + 0x800 + 8 * (eq->eqn % 4);
eq                970 drivers/net/ethernet/mellanox/mlx4/eq.c 			  u8 intr, struct mlx4_eq *eq)
eq                982 drivers/net/ethernet/mellanox/mlx4/eq.c 	eq->dev   = dev;
eq                983 drivers/net/ethernet/mellanox/mlx4/eq.c 	eq->nent  = roundup_pow_of_two(max(nent, 2));
eq                987 drivers/net/ethernet/mellanox/mlx4/eq.c 	npages = PAGE_ALIGN(eq->nent * dev->caps.eqe_size) / PAGE_SIZE;
eq                989 drivers/net/ethernet/mellanox/mlx4/eq.c 	eq->page_list = kmalloc_array(npages, sizeof(*eq->page_list),
eq                991 drivers/net/ethernet/mellanox/mlx4/eq.c 	if (!eq->page_list)
eq                995 drivers/net/ethernet/mellanox/mlx4/eq.c 		eq->page_list[i].buf = NULL;
eq               1007 drivers/net/ethernet/mellanox/mlx4/eq.c 		eq->page_list[i].buf = dma_alloc_coherent(&dev->persist->
eq               1011 drivers/net/ethernet/mellanox/mlx4/eq.c 		if (!eq->page_list[i].buf)
eq               1015 drivers/net/ethernet/mellanox/mlx4/eq.c 		eq->page_list[i].map = t;
eq               1018 drivers/net/ethernet/mellanox/mlx4/eq.c 	eq->eqn = mlx4_bitmap_alloc(&priv->eq_table.bitmap);
eq               1019 drivers/net/ethernet/mellanox/mlx4/eq.c 	if (eq->eqn == -1)
eq               1022 drivers/net/ethernet/mellanox/mlx4/eq.c 	eq->doorbell = mlx4_get_eq_uar(dev, eq);
eq               1023 drivers/net/ethernet/mellanox/mlx4/eq.c 	if (!eq->doorbell) {
eq               1028 drivers/net/ethernet/mellanox/mlx4/eq.c 	err = mlx4_mtt_init(dev, npages, PAGE_SHIFT, &eq->mtt);
eq               1032 drivers/net/ethernet/mellanox/mlx4/eq.c 	err = mlx4_write_mtt(dev, &eq->mtt, 0, npages, dma_list);
eq               1038 drivers/net/ethernet/mellanox/mlx4/eq.c 	eq_context->log_eq_size	  = ilog2(eq->nent);
eq               1042 drivers/net/ethernet/mellanox/mlx4/eq.c 	mtt_addr = mlx4_mtt_addr(dev, &eq->mtt);
eq               1046 drivers/net/ethernet/mellanox/mlx4/eq.c 	err = mlx4_SW2HW_EQ(dev, mailbox, eq->eqn);
eq               1055 drivers/net/ethernet/mellanox/mlx4/eq.c 	eq->cons_index = 0;
eq               1057 drivers/net/ethernet/mellanox/mlx4/eq.c 	INIT_LIST_HEAD(&eq->tasklet_ctx.list);
eq               1058 drivers/net/ethernet/mellanox/mlx4/eq.c 	INIT_LIST_HEAD(&eq->tasklet_ctx.process_list);
eq               1059 drivers/net/ethernet/mellanox/mlx4/eq.c 	spin_lock_init(&eq->tasklet_ctx.lock);
eq               1060 drivers/net/ethernet/mellanox/mlx4/eq.c 	tasklet_init(&eq->tasklet_ctx.task, mlx4_cq_tasklet_cb,
eq               1061 drivers/net/ethernet/mellanox/mlx4/eq.c 		     (unsigned long)&eq->tasklet_ctx);
eq               1066 drivers/net/ethernet/mellanox/mlx4/eq.c 	mlx4_mtt_cleanup(dev, &eq->mtt);
eq               1069 drivers/net/ethernet/mellanox/mlx4/eq.c 	mlx4_bitmap_free(&priv->eq_table.bitmap, eq->eqn, MLX4_USE_RR);
eq               1073 drivers/net/ethernet/mellanox/mlx4/eq.c 		if (eq->page_list[i].buf)
eq               1075 drivers/net/ethernet/mellanox/mlx4/eq.c 					  eq->page_list[i].buf,
eq               1076 drivers/net/ethernet/mellanox/mlx4/eq.c 					  eq->page_list[i].map);
eq               1081 drivers/net/ethernet/mellanox/mlx4/eq.c 	kfree(eq->page_list);
eq               1089 drivers/net/ethernet/mellanox/mlx4/eq.c 			 struct mlx4_eq *eq)
eq               1097 drivers/net/ethernet/mellanox/mlx4/eq.c 	int npages = PAGE_ALIGN(dev->caps.eqe_size  * eq->nent) / PAGE_SIZE;
eq               1099 drivers/net/ethernet/mellanox/mlx4/eq.c 	err = mlx4_HW2SW_EQ(dev, eq->eqn);
eq               1103 drivers/net/ethernet/mellanox/mlx4/eq.c 	synchronize_irq(eq->irq);
eq               1104 drivers/net/ethernet/mellanox/mlx4/eq.c 	tasklet_disable(&eq->tasklet_ctx.task);
eq               1106 drivers/net/ethernet/mellanox/mlx4/eq.c 	mlx4_mtt_cleanup(dev, &eq->mtt);
eq               1109 drivers/net/ethernet/mellanox/mlx4/eq.c 				  eq->page_list[i].buf,
eq               1110 drivers/net/ethernet/mellanox/mlx4/eq.c 				  eq->page_list[i].map);
eq               1112 drivers/net/ethernet/mellanox/mlx4/eq.c 	kfree(eq->page_list);
eq               1113 drivers/net/ethernet/mellanox/mlx4/eq.c 	mlx4_bitmap_free(&priv->eq_table.bitmap, eq->eqn, MLX4_USE_RR);
eq               1125 drivers/net/ethernet/mellanox/mlx4/eq.c 		if (eq_table->eq[i].have_irq) {
eq               1126 drivers/net/ethernet/mellanox/mlx4/eq.c 			free_cpumask_var(eq_table->eq[i].affinity_mask);
eq               1128 drivers/net/ethernet/mellanox/mlx4/eq.c 			irq_set_affinity_hint(eq_table->eq[i].irq, NULL);
eq               1130 drivers/net/ethernet/mellanox/mlx4/eq.c 			free_irq(eq_table->eq[i].irq, eq_table->eq + i);
eq               1131 drivers/net/ethernet/mellanox/mlx4/eq.c 			eq_table->eq[i].have_irq = 0;
eq               1163 drivers/net/ethernet/mellanox/mlx4/eq.c 	priv->eq_table.eq = kcalloc(dev->caps.num_eqs - dev->caps.reserved_eqs,
eq               1164 drivers/net/ethernet/mellanox/mlx4/eq.c 				    sizeof(*priv->eq_table.eq), GFP_KERNEL);
eq               1165 drivers/net/ethernet/mellanox/mlx4/eq.c 	if (!priv->eq_table.eq)
eq               1173 drivers/net/ethernet/mellanox/mlx4/eq.c 	kfree(mlx4_priv(dev)->eq_table.eq);
eq               1226 drivers/net/ethernet/mellanox/mlx4/eq.c 					     0, &priv->eq_table.eq[MLX4_EQ_ASYNC]);
eq               1228 drivers/net/ethernet/mellanox/mlx4/eq.c 			struct mlx4_eq	*eq = &priv->eq_table.eq[i];
eq               1230 drivers/net/ethernet/mellanox/mlx4/eq.c 			int port = find_first_bit(eq->actv_ports.ports,
eq               1248 drivers/net/ethernet/mellanox/mlx4/eq.c 					info->rmap, eq->irq);
eq               1257 drivers/net/ethernet/mellanox/mlx4/eq.c 					     eq);
eq               1274 drivers/net/ethernet/mellanox/mlx4/eq.c 		err = request_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq,
eq               1276 drivers/net/ethernet/mellanox/mlx4/eq.c 				  priv->eq_table.eq + MLX4_EQ_ASYNC);
eq               1280 drivers/net/ethernet/mellanox/mlx4/eq.c 		priv->eq_table.eq[MLX4_EQ_ASYNC].have_irq = 1;
eq               1295 drivers/net/ethernet/mellanox/mlx4/eq.c 			  priv->eq_table.eq[MLX4_EQ_ASYNC].eqn);
eq               1298 drivers/net/ethernet/mellanox/mlx4/eq.c 			   priv->eq_table.eq[MLX4_EQ_ASYNC].eqn, err);
eq               1301 drivers/net/ethernet/mellanox/mlx4/eq.c 	eq_set_ci(&priv->eq_table.eq[MLX4_EQ_ASYNC], 1);
eq               1307 drivers/net/ethernet/mellanox/mlx4/eq.c 		mlx4_free_eq(dev, &priv->eq_table.eq[--i]);
eq               1338 drivers/net/ethernet/mellanox/mlx4/eq.c 		    priv->eq_table.eq[MLX4_EQ_ASYNC].eqn);
eq               1351 drivers/net/ethernet/mellanox/mlx4/eq.c 		mlx4_free_eq(dev, &priv->eq_table.eq[i]);
eq               1385 drivers/net/ethernet/mellanox/mlx4/eq.c 			  priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn);
eq               1399 drivers/net/ethernet/mellanox/mlx4/eq.c 		    priv->eq_table.eq[MLX4_EQ_ASYNC].eqn);
eq               1415 drivers/net/ethernet/mellanox/mlx4/eq.c 	return test_bit(port - 1, priv->eq_table.eq[vector].actv_ports.ports);
eq               1427 drivers/net/ethernet/mellanox/mlx4/eq.c 				  priv->eq_table.eq[i].actv_ports.ports);
eq               1441 drivers/net/ethernet/mellanox/mlx4/eq.c 	return !!(bitmap_weight(priv->eq_table.eq[vector].actv_ports.ports,
eq               1466 drivers/net/ethernet/mellanox/mlx4/eq.c 			     priv->eq_table.eq[requested_vector].actv_ports.ports)) {
eq               1469 drivers/net/ethernet/mellanox/mlx4/eq.c 			struct mlx4_eq *eq;
eq               1475 drivers/net/ethernet/mellanox/mlx4/eq.c 			eq = &priv->eq_table.eq[requested_vector];
eq               1477 drivers/net/ethernet/mellanox/mlx4/eq.c 			    test_bit(port - 1, eq->actv_ports.ports)) {
eq               1487 drivers/net/ethernet/mellanox/mlx4/eq.c 			struct mlx4_eq *eq = &priv->eq_table.eq[i];
eq               1489 drivers/net/ethernet/mellanox/mlx4/eq.c 			if (min_ref_count_val > eq->ref_count &&
eq               1490 drivers/net/ethernet/mellanox/mlx4/eq.c 			    test_bit(port - 1, eq->actv_ports.ports)) {
eq               1491 drivers/net/ethernet/mellanox/mlx4/eq.c 				min_ref_count_val = eq->ref_count;
eq               1512 drivers/net/ethernet/mellanox/mlx4/eq.c 		err = request_irq(priv->eq_table.eq[*prequested_vector].irq,
eq               1515 drivers/net/ethernet/mellanox/mlx4/eq.c 				  priv->eq_table.eq + *prequested_vector);
eq               1524 drivers/net/ethernet/mellanox/mlx4/eq.c 			eq_set_ci(&priv->eq_table.eq[*prequested_vector], 1);
eq               1525 drivers/net/ethernet/mellanox/mlx4/eq.c 			priv->eq_table.eq[*prequested_vector].have_irq = 1;
eq               1530 drivers/net/ethernet/mellanox/mlx4/eq.c 		priv->eq_table.eq[*prequested_vector].ref_count++;
eq               1548 drivers/net/ethernet/mellanox/mlx4/eq.c 	return priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq_vec)].irq;
eq               1558 drivers/net/ethernet/mellanox/mlx4/eq.c 	priv->eq_table.eq[eq_vec].ref_count--;
eq               2772 drivers/net/ethernet/mellanox/mlx4/main.c 				  priv->eq_table.eq[MLX4_EQ_ASYNC].irq);
eq               2776 drivers/net/ethernet/mellanox/mlx4/main.c 				 priv->eq_table.eq[MLX4_EQ_ASYNC].irq);
eq               2905 drivers/net/ethernet/mellanox/mlx4/main.c 	struct mlx4_eq *eq;
eq               2921 drivers/net/ethernet/mellanox/mlx4/main.c 	eq = &priv->eq_table.eq[eqn];
eq               2923 drivers/net/ethernet/mellanox/mlx4/main.c 	if (!zalloc_cpumask_var(&eq->affinity_mask, GFP_KERNEL))
eq               2926 drivers/net/ethernet/mellanox/mlx4/main.c 	cpumask_set_cpu(requested_cpu, eq->affinity_mask);
eq               2964 drivers/net/ethernet/mellanox/mlx4/main.c 		priv->eq_table.eq[MLX4_EQ_ASYNC].irq = entries[0].vector;
eq               2965 drivers/net/ethernet/mellanox/mlx4/main.c 		bitmap_zero(priv->eq_table.eq[MLX4_EQ_ASYNC].actv_ports.ports,
eq               2972 drivers/net/ethernet/mellanox/mlx4/main.c 			priv->eq_table.eq[i].irq =
eq               2976 drivers/net/ethernet/mellanox/mlx4/main.c 				bitmap_fill(priv->eq_table.eq[i].actv_ports.ports,
eq               2983 drivers/net/ethernet/mellanox/mlx4/main.c 					priv->eq_table.eq[i].actv_ports.ports);
eq               3018 drivers/net/ethernet/mellanox/mlx4/main.c 		priv->eq_table.eq[i].irq = dev->persist->pdev->irq;
eq               3020 drivers/net/ethernet/mellanox/mlx4/main.c 			bitmap_fill(priv->eq_table.eq[i].actv_ports.ports,
eq                490 drivers/net/ethernet/mellanox/mlx4/mlx4.h 	struct mlx4_slave_eqe eq[MLX4_MFUNC_MAX_EQES];
eq                693 drivers/net/ethernet/mellanox/mlx4/mlx4.h 	struct mlx4_eq	       *eq;
eq               1623 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 				enum res_eq_states state, struct res_eq **eq)
eq               1665 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	if (!err && eq)
eq               1666 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		*eq = r;
eq               3104 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	struct res_eq *eq;
eq               3110 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	err = eq_res_start_move_to(dev, slave, res_id, RES_EQ_HW, &eq);
eq               3127 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	eq->mtt = mtt;
eq               3308 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	struct res_eq *eq;
eq               3311 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	err = eq_res_start_move_to(dev, slave, res_id, RES_EQ_RESERVED, &eq);
eq               3315 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	err = get_res(dev, slave, eq->mtt->com.res_id, RES_MTT, NULL);
eq               3323 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	atomic_dec(&eq->mtt->ref_count);
eq               3324 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	put_res(dev, slave, eq->mtt->com.res_id, RES_MTT);
eq               3331 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	put_res(dev, slave, eq->mtt->com.res_id, RES_MTT);
eq               3414 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	struct res_eq *eq;
eq               3417 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	err = get_res(dev, slave, res_id, RES_EQ, &eq);
eq               3421 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	if (eq->com.from_state != RES_EQ_HW) {
eq               5120 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	struct res_eq *eq;
eq               5132 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	list_for_each_entry_safe(eq, tmp, eq_list, com.list) {
eq               5134 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		if (eq->com.owner == slave) {
eq               5135 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 			eqn = eq->com.res_id;
eq               5136 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 			state = eq->com.from_state;
eq               5141 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					rb_erase(&eq->com.node,
eq               5143 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					list_del(&eq->com.list);
eq               5145 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					kfree(eq);
eq               5157 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					atomic_dec(&eq->mtt->ref_count);
eq                 96 drivers/net/ethernet/mellanox/mlx5/core/cq.c 	struct mlx5_eq_comp *eq;
eq                 99 drivers/net/ethernet/mellanox/mlx5/core/cq.c 	eq = mlx5_eqn2comp_eq(dev, eqn);
eq                100 drivers/net/ethernet/mellanox/mlx5/core/cq.c 	if (IS_ERR(eq))
eq                101 drivers/net/ethernet/mellanox/mlx5/core/cq.c 		return PTR_ERR(eq);
eq                112 drivers/net/ethernet/mellanox/mlx5/core/cq.c 	cq->eq         = eq;
eq                119 drivers/net/ethernet/mellanox/mlx5/core/cq.c 	cq->tasklet_ctx.priv = &eq->tasklet_ctx;
eq                123 drivers/net/ethernet/mellanox/mlx5/core/cq.c 	err = mlx5_eq_add_cq(&eq->core, cq);
eq                143 drivers/net/ethernet/mellanox/mlx5/core/cq.c 	mlx5_eq_del_cq(&eq->core, cq);
eq                162 drivers/net/ethernet/mellanox/mlx5/core/cq.c 	mlx5_eq_del_cq(&cq->eq->core, cq);
eq                287 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c static int mlx5_core_eq_query(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
eq                293 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c 	MLX5_SET(query_eq_in, in, eq_number, eq->eqn);
eq                297 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c static u64 eq_read_field(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
eq                310 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c 	err = mlx5_core_eq_query(dev, eq, out, outlen);
eq                474 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c int mlx5_debug_eq_add(struct mlx5_core_dev *dev, struct mlx5_eq *eq)
eq                482 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c 			   &eq->dbg, eq->eqn, eq_fields,
eq                483 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c 			   ARRAY_SIZE(eq_fields), eq);
eq                485 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c 		eq->dbg = NULL;
eq                490 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c void mlx5_debug_eq_remove(struct mlx5_core_dev *dev, struct mlx5_eq *eq)
eq                495 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c 	if (eq->dbg)
eq                496 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c 		rem_res_tree(eq->dbg);
eq                178 drivers/net/ethernet/mellanox/mlx5/core/en/health.c int mlx5e_health_channel_eq_recover(struct mlx5_eq_comp *eq, struct mlx5e_channel *channel)
eq                183 drivers/net/ethernet/mellanox/mlx5/core/en/health.c 		   eq->core.eqn, eq->core.cons_index, eq->core.irqn);
eq                185 drivers/net/ethernet/mellanox/mlx5/core/en/health.c 	eqe_count = mlx5_eq_poll_irq_disabled(eq);
eq                190 drivers/net/ethernet/mellanox/mlx5/core/en/health.c 		   eqe_count, eq->core.eqn);
eq                 42 drivers/net/ethernet/mellanox/mlx5/core/en/health.h int mlx5e_health_channel_eq_recover(struct mlx5_eq_comp *eq, struct mlx5e_channel *channel);
eq                190 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c 	struct mlx5_eq_comp *eq;
eq                196 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c 	eq = rq->cq.mcq.eq;
eq                197 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c 	err = mlx5e_health_channel_eq_recover(eq, rq->channel);
eq                100 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c 	struct mlx5_eq_comp *eq;
eq                105 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c 	eq = sq->cq.mcq.eq;
eq                106 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c 	err = mlx5e_health_channel_eq_recover(eq, sq->channel);
eq                114 drivers/net/ethernet/mellanox/mlx5/core/eq.c static struct mlx5_core_cq *mlx5_eq_cq_get(struct mlx5_eq *eq, u32 cqn)
eq                116 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_cq_table *table = &eq->cq_table;
eq                134 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eq *eq = &eq_comp->core;
eq                139 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eqe = next_eqe_sw(eq);
eq                153 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		cq = mlx5_eq_cq_get(eq, cqn);
eq                159 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			mlx5_core_warn(eq->dev, "Completion event for bogus CQ 0x%x\n", cqn);
eq                162 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		++eq->cons_index;
eq                164 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	} while ((++num_eqes < MLX5_EQ_POLLING_BUDGET) && (eqe = next_eqe_sw(eq)));
eq                167 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq_update_ci(eq, 1);
eq                180 drivers/net/ethernet/mellanox/mlx5/core/eq.c u32 mlx5_eq_poll_irq_disabled(struct mlx5_eq_comp *eq)
eq                184 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	disable_irq(eq->core.irqn);
eq                185 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	count_eqe = eq->core.cons_index;
eq                186 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	mlx5_eq_comp_int(&eq->irq_nb, 0, NULL);
eq                187 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	count_eqe = eq->core.cons_index - count_eqe;
eq                188 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	enable_irq(eq->core.irqn);
eq                198 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eq *eq = &eq_async->core;
eq                204 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	dev = eq->dev;
eq                207 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eqe = next_eqe_sw(eq);
eq                221 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		++eq->cons_index;
eq                223 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	} while ((++num_eqes < MLX5_EQ_POLLING_BUDGET) && (eqe = next_eqe_sw(eq)));
eq                226 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq_update_ci(eq, 1);
eq                231 drivers/net/ethernet/mellanox/mlx5/core/eq.c static void init_eq_buf(struct mlx5_eq *eq)
eq                236 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	for (i = 0; i < eq->nent; i++) {
eq                237 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		eqe = get_eqe(eq, i);
eq                243 drivers/net/ethernet/mellanox/mlx5/core/eq.c create_map_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
eq                246 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_cq_table *cq_table = &eq->cq_table;
eq                262 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq->nent = roundup_pow_of_two(param->nent + MLX5_NUM_SPARE_EQE);
eq                263 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq->cons_index = 0;
eq                264 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	err = mlx5_buf_alloc(dev, eq->nent * MLX5_EQE_SIZE, &eq->buf);
eq                268 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	init_eq_buf(eq);
eq                271 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		MLX5_FLD_SZ_BYTES(create_eq_in, pas[0]) * eq->buf.npages;
eq                280 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	mlx5_fill_page_array(&eq->buf, pas);
eq                291 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	MLX5_SET(eqc, eqc, log_eq_size, ilog2(eq->nent));
eq                295 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		 eq->buf.page_shift - MLX5_ADAPTER_PAGE_SHIFT);
eq                301 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq->vecidx = vecidx;
eq                302 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq->eqn = MLX5_GET(create_eq_out, out, eq_number);
eq                303 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq->irqn = pci_irq_vector(dev->pdev, vecidx);
eq                304 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq->dev = dev;
eq                305 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq->doorbell = priv->uar->map + MLX5_EQ_DOORBEL_OFFSET;
eq                307 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	err = mlx5_debug_eq_add(dev, eq);
eq                315 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	mlx5_cmd_destroy_eq(dev, eq->eqn);
eq                321 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	mlx5_buf_free(dev, &eq->buf);
eq                335 drivers/net/ethernet/mellanox/mlx5/core/eq.c int mlx5_eq_enable(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
eq                341 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	err = mlx5_irq_attach_nb(eq_table->irq_table, eq->vecidx, nb);
eq                343 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		eq_update_ci(eq, 1);
eq                357 drivers/net/ethernet/mellanox/mlx5/core/eq.c void mlx5_eq_disable(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
eq                362 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	mlx5_irq_detach_nb(eq_table->irq_table, eq->vecidx, nb);
eq                366 drivers/net/ethernet/mellanox/mlx5/core/eq.c static int destroy_unmap_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq)
eq                370 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	mlx5_debug_eq_remove(dev, eq);
eq                372 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	err = mlx5_cmd_destroy_eq(dev, eq->eqn);
eq                375 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			       eq->eqn);
eq                376 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	synchronize_irq(eq->irqn);
eq                378 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	mlx5_buf_free(dev, &eq->buf);
eq                383 drivers/net/ethernet/mellanox/mlx5/core/eq.c int mlx5_eq_add_cq(struct mlx5_eq *eq, struct mlx5_core_cq *cq)
eq                385 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_cq_table *table = &eq->cq_table;
eq                395 drivers/net/ethernet/mellanox/mlx5/core/eq.c void mlx5_eq_del_cq(struct mlx5_eq *eq, struct mlx5_core_cq *cq)
eq                397 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_cq_table *table = &eq->cq_table;
eq                405 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		mlx5_core_dbg(eq->dev, "cq 0x%x not found in eq 0x%x tree\n",
eq                406 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			      eq->eqn, cq->cqn);
eq                411 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		mlx5_core_dbg(eq->dev, "corruption on cqn 0x%x in eq 0x%x\n",
eq                412 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			      eq->eqn, cq->cqn);
eq                445 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			   struct mlx5_eq *eq, struct mlx5_eq_param *param)
eq                457 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	err = create_map_eq(dev, eq, param);
eq                463 drivers/net/ethernet/mellanox/mlx5/core/eq.c static int destroy_async_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq)
eq                469 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	err = destroy_unmap_eq(dev, eq);
eq                480 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eq *eq;
eq                486 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq  = &eqt->async_eq.core;
eq                490 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	mlx5_core_warn(eq->dev, "CQ error on CQN 0x%x, syndrome 0x%x\n",
eq                493 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	cq = mlx5_eq_cq_get(eq, cqn);
eq                495 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		mlx5_core_warn(eq->dev, "Async event for bogus CQ 0x%x\n", cqn);
eq                700 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eq *eq = kvzalloc(sizeof(*eq), GFP_KERNEL);
eq                703 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	if (!eq)
eq                706 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	err = create_async_eq(dev, eq, param);
eq                708 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		kvfree(eq);
eq                709 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		eq = ERR_PTR(err);
eq                712 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	return eq;
eq                716 drivers/net/ethernet/mellanox/mlx5/core/eq.c int mlx5_eq_destroy_generic(struct mlx5_core_dev *dev, struct mlx5_eq *eq)
eq                720 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	if (IS_ERR(eq))
eq                723 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	err = destroy_async_eq(dev, eq);
eq                727 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	kvfree(eq);
eq                733 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eqe *mlx5_eq_get_eqe(struct mlx5_eq *eq, u32 cc)
eq                735 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	u32 ci = eq->cons_index + cc;
eq                738 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eqe = get_eqe(eq, ci & (eq->nent - 1));
eq                739 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eqe = ((eqe->owner & 1) ^ !!(ci & eq->nent)) ? NULL : eqe;
eq                750 drivers/net/ethernet/mellanox/mlx5/core/eq.c void mlx5_eq_update_ci(struct mlx5_eq *eq, u32 cc, bool arm)
eq                752 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	__be32 __iomem *addr = eq->doorbell + (arm ? 0 : 2);
eq                755 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq->cons_index += cc;
eq                756 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	val = (eq->cons_index & 0xffffff) | (eq->eqn << 24);
eq                767 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eq_comp *eq, *n;
eq                769 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	list_for_each_entry_safe(eq, n, &table->comp_eqs_list, list) {
eq                770 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		list_del(&eq->list);
eq                771 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		mlx5_eq_disable(dev, &eq->core, &eq->irq_nb);
eq                772 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		if (destroy_unmap_eq(dev, &eq->core))
eq                774 drivers/net/ethernet/mellanox/mlx5/core/eq.c 				       eq->core.eqn);
eq                775 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		tasklet_disable(&eq->tasklet_ctx.task);
eq                776 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		kfree(eq);
eq                783 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eq_comp *eq;
eq                796 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		eq = kzalloc(sizeof(*eq), GFP_KERNEL);
eq                797 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		if (!eq) {
eq                802 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		INIT_LIST_HEAD(&eq->tasklet_ctx.list);
eq                803 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		INIT_LIST_HEAD(&eq->tasklet_ctx.process_list);
eq                804 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		spin_lock_init(&eq->tasklet_ctx.lock);
eq                805 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		tasklet_init(&eq->tasklet_ctx.task, mlx5_cq_tasklet_cb,
eq                806 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			     (unsigned long)&eq->tasklet_ctx);
eq                808 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		eq->irq_nb.notifier_call = mlx5_eq_comp_int;
eq                813 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		err = create_map_eq(dev, &eq->core, &param);
eq                815 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			kfree(eq);
eq                818 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		err = mlx5_eq_enable(dev, &eq->core, &eq->irq_nb);
eq                820 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			destroy_unmap_eq(dev, &eq->core);
eq                821 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			kfree(eq);
eq                825 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		mlx5_core_dbg(dev, "allocated completion EQN %d\n", eq->core.eqn);
eq                827 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		list_add_tail(&eq->list, &table->comp_eqs_list);
eq                841 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eq_comp *eq, *n;
eq                845 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	list_for_each_entry_safe(eq, n, &table->comp_eqs_list, list) {
eq                847 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			*eqn = eq->core.eqn;
eq                848 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			*irqn = eq->core.irqn;
eq                884 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	struct mlx5_eq_comp *eq;
eq                886 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	list_for_each_entry(eq, &table->comp_eqs_list, list) {
eq                887 drivers/net/ethernet/mellanox/mlx5/core/eq.c 		if (eq->core.eqn == eqn)
eq                888 drivers/net/ethernet/mellanox/mlx5/core/eq.c 			return eq;
eq                 50 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h static inline struct mlx5_eqe *get_eqe(struct mlx5_eq *eq, u32 entry)
eq                 52 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h 	return mlx5_buf_offset(&eq->buf, entry * MLX5_EQE_SIZE);
eq                 55 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h static inline struct mlx5_eqe *next_eqe_sw(struct mlx5_eq *eq)
eq                 57 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h 	struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1));
eq                 59 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h 	return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe;
eq                 62 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h static inline void eq_update_ci(struct mlx5_eq *eq, int arm)
eq                 64 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h 	__be32 __iomem *addr = eq->doorbell + (arm ? 0 : 2);
eq                 65 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h 	u32 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24);
eq                 77 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h int mlx5_eq_add_cq(struct mlx5_eq *eq, struct mlx5_core_cq *cq);
eq                 78 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h void mlx5_eq_del_cq(struct mlx5_eq *eq, struct mlx5_core_cq *cq);
eq                 84 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h u32 mlx5_eq_poll_irq_disabled(struct mlx5_eq_comp *eq);
eq                 88 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h int mlx5_debug_eq_add(struct mlx5_core_dev *dev, struct mlx5_eq *eq);
eq                 89 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h void mlx5_debug_eq_remove(struct mlx5_core_dev *dev, struct mlx5_eq *eq);
eq                 93 drivers/net/ethernet/mellanox/mlxsw/pci.c 		} eq;
eq                743 drivers/net/ethernet/mellanox/mlxsw/pci.c 			q->u.eq.ev_cmd_count++;
eq                749 drivers/net/ethernet/mellanox/mlxsw/pci.c 			q->u.eq.ev_comp_count++;
eq                752 drivers/net/ethernet/mellanox/mlxsw/pci.c 			q->u.eq.ev_other_count++;
eq               3535 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c 	u8 di0, dq0, ei, eq, fi, fq;
eq               3553 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c 		wlc_lcnphy_get_radio_loft(pi, &ei, &eq, &fi, &fq);
eq               3555 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c 		cc.im = (u16) eq;
eq               3558 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c 		wlc_lcnphy_get_radio_loft(pi, &ei, &eq, &fi, &fq);
eq               3852 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c 	u32 hq, lq, nq, eq, pubq;
eq               3858 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c 	eq = 0;
eq               3868 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c 	val32 = (nq << RQPN_NPQ_SHIFT) | (eq << RQPN_EPQ_SHIFT);
eq                 64 drivers/pci/controller/pcie-iproc-msi.c 	unsigned int eq;
eq                130 drivers/pci/controller/pcie-iproc-msi.c 				     unsigned int eq)
eq                134 drivers/pci/controller/pcie-iproc-msi.c 	return readl_relaxed(pcie->base + msi->reg_offsets[eq][reg]);
eq                139 drivers/pci/controller/pcie-iproc-msi.c 				       int eq, u32 val)
eq                143 drivers/pci/controller/pcie-iproc-msi.c 	writel_relaxed(val, pcie->base + msi->reg_offsets[eq][reg]);
eq                160 drivers/pci/controller/pcie-iproc-msi.c static inline unsigned int iproc_msi_eq_offset(struct iproc_msi *msi, u32 eq)
eq                163 drivers/pci/controller/pcie-iproc-msi.c 		return eq * EQ_MEM_REGION_SIZE;
eq                165 drivers/pci/controller/pcie-iproc-msi.c 		return eq * EQ_LEN * sizeof(u32);
eq                294 drivers/pci/controller/pcie-iproc-msi.c static inline u32 decode_msi_hwirq(struct iproc_msi *msi, u32 eq, u32 head)
eq                299 drivers/pci/controller/pcie-iproc-msi.c 	offs = iproc_msi_eq_offset(msi, eq) + head * sizeof(u32);
eq                317 drivers/pci/controller/pcie-iproc-msi.c 	u32 eq, head, tail, nr_events;
eq                325 drivers/pci/controller/pcie-iproc-msi.c 	eq = grp->eq;
eq                338 drivers/pci/controller/pcie-iproc-msi.c 				  eq) & IPROC_MSI_EQ_MASK;
eq                341 drivers/pci/controller/pcie-iproc-msi.c 					  eq) & IPROC_MSI_EQ_MASK;
eq                354 drivers/pci/controller/pcie-iproc-msi.c 			hwirq = decode_msi_hwirq(msi, eq, head);
eq                366 drivers/pci/controller/pcie-iproc-msi.c 		iproc_msi_write_reg(msi, IPROC_MSI_EQ_HEAD, eq, head);
eq                379 drivers/pci/controller/pcie-iproc-msi.c 	int i, eq;
eq                402 drivers/pci/controller/pcie-iproc-msi.c 	for (eq = 0; eq < msi->nr_irqs; eq++) {
eq                406 drivers/pci/controller/pcie-iproc-msi.c 		iproc_msi_write_reg(msi, IPROC_MSI_CTRL, eq, val);
eq                413 drivers/pci/controller/pcie-iproc-msi.c 			val = iproc_msi_read_reg(msi, IPROC_MSI_INTS_EN, eq);
eq                414 drivers/pci/controller/pcie-iproc-msi.c 			val |= BIT(eq);
eq                415 drivers/pci/controller/pcie-iproc-msi.c 			iproc_msi_write_reg(msi, IPROC_MSI_INTS_EN, eq, val);
eq                422 drivers/pci/controller/pcie-iproc-msi.c 	u32 eq, val;
eq                424 drivers/pci/controller/pcie-iproc-msi.c 	for (eq = 0; eq < msi->nr_irqs; eq++) {
eq                426 drivers/pci/controller/pcie-iproc-msi.c 			val = iproc_msi_read_reg(msi, IPROC_MSI_INTS_EN, eq);
eq                427 drivers/pci/controller/pcie-iproc-msi.c 			val &= ~BIT(eq);
eq                428 drivers/pci/controller/pcie-iproc-msi.c 			iproc_msi_write_reg(msi, IPROC_MSI_INTS_EN, eq, val);
eq                431 drivers/pci/controller/pcie-iproc-msi.c 		val = iproc_msi_read_reg(msi, IPROC_MSI_CTRL, eq);
eq                434 drivers/pci/controller/pcie-iproc-msi.c 		iproc_msi_write_reg(msi, IPROC_MSI_CTRL, eq, val);
eq                601 drivers/pci/controller/pcie-iproc-msi.c 		msi->grps[i].eq = i;
eq                 33 drivers/s390/char/sclp_sd.c 	u8 eq;
eq                188 drivers/s390/char/sclp_sd.c static int sclp_sd_sync(unsigned long page, u8 eq, u8 di, u64 sat, u64 sa,
eq                205 drivers/s390/char/sclp_sd.c 	evbuf->eq = eq;
eq                215 drivers/s390/char/sclp_sd.c 	pr_debug("request (eq=%d, di=%d, id=0x%08x)\n", eq, di, listener.id);
eq                260 drivers/s390/char/sclp_sd.c 			eq, di, sccb->hdr.response_code, evbuf->hdr.flags,
eq                736 drivers/scsi/be2iscsi/be_cmds.c 			  struct be_queue_info *eq, int eq_delay)
eq                741 drivers/scsi/be2iscsi/be_cmds.c 	struct be_dma_mem *q_mem = &eq->dma_mem;
eq                759 drivers/scsi/be2iscsi/be_cmds.c 					__ilog2_u32(eq->len / 256));
eq                768 drivers/scsi/be2iscsi/be_cmds.c 		eq->id = le16_to_cpu(resp->eq_id);
eq                769 drivers/scsi/be2iscsi/be_cmds.c 		eq->created = true;
eq                776 drivers/scsi/be2iscsi/be_cmds.c 			  struct be_queue_info *cq, struct be_queue_info *eq,
eq                805 drivers/scsi/be2iscsi/be_cmds.c 		AMAP_SET_BITS(struct amap_cq_context, eqid, ctxt, eq->id);
eq                820 drivers/scsi/be2iscsi/be_cmds.c 		AMAP_SET_BITS(struct amap_cq_context_v2, eqid, ctxt, eq->id);
eq                803 drivers/scsi/be2iscsi/be_cmds.h 			  struct be_queue_info *eq, int eq_delay);
eq                806 drivers/scsi/be2iscsi/be_cmds.h 			  struct be_queue_info *cq, struct be_queue_info *eq,
eq                669 drivers/scsi/be2iscsi/be_main.c 	struct be_queue_info *eq;
eq                675 drivers/scsi/be2iscsi/be_main.c 	eq = &pbe_eq->q;
eq                678 drivers/scsi/be2iscsi/be_main.c 	eqe = queue_tail_node(eq);
eq                689 drivers/scsi/be2iscsi/be_main.c 		queue_tail_inc(eq);
eq                690 drivers/scsi/be2iscsi/be_main.c 		eqe = queue_tail_node(eq);
eq                695 drivers/scsi/be2iscsi/be_main.c 		hwi_ring_eq_db(phba, eq->id, 1,	mcc_events, 1, 1);
eq                708 drivers/scsi/be2iscsi/be_main.c 	struct be_queue_info *eq;
eq                712 drivers/scsi/be2iscsi/be_main.c 	eq = &pbe_eq->q;
eq                716 drivers/scsi/be2iscsi/be_main.c 	hwi_ring_eq_db(phba, eq->id, 1,	0, 0, 1);
eq                733 drivers/scsi/be2iscsi/be_main.c 	struct be_queue_info *eq;
eq                751 drivers/scsi/be2iscsi/be_main.c 	eq = &phwi_context->be_eq[0].q;
eq                753 drivers/scsi/be2iscsi/be_main.c 	eqe = queue_tail_node(eq);
eq                765 drivers/scsi/be2iscsi/be_main.c 		queue_tail_inc(eq);
eq                766 drivers/scsi/be2iscsi/be_main.c 		eqe = queue_tail_node(eq);
eq                780 drivers/scsi/be2iscsi/be_main.c 	hwi_ring_eq_db(phba, eq->id, 0, (io_events + mcc_events), rearm, 1);
eq               2038 drivers/scsi/be2iscsi/be_main.c 	struct be_queue_info *eq;
eq               2048 drivers/scsi/be2iscsi/be_main.c 	eq = &pbe_eq->q;
eq               2049 drivers/scsi/be2iscsi/be_main.c 	eqe = queue_tail_node(eq);
eq               2053 drivers/scsi/be2iscsi/be_main.c 		queue_tail_inc(eq);
eq               2054 drivers/scsi/be2iscsi/be_main.c 		eqe = queue_tail_node(eq);
eq               2057 drivers/scsi/be2iscsi/be_main.c 	hwi_ring_eq_db(phba, eq->id, 1, io_events, 0, 1);
eq               2999 drivers/scsi/be2iscsi/be_main.c 	struct be_queue_info *eq;
eq               3012 drivers/scsi/be2iscsi/be_main.c 		eq = &phwi_context->be_eq[i].q;
eq               3013 drivers/scsi/be2iscsi/be_main.c 		mem = &eq->dma_mem;
eq               3024 drivers/scsi/be2iscsi/be_main.c 		ret = be_fill_queue(eq, phba->params.num_eq_entries,
eq               3033 drivers/scsi/be2iscsi/be_main.c 		ret = beiscsi_cmd_eq_create(&phba->ctrl, eq,
eq               3050 drivers/scsi/be2iscsi/be_main.c 		eq = &phwi_context->be_eq[i].q;
eq               3051 drivers/scsi/be2iscsi/be_main.c 		mem = &eq->dma_mem;
eq               3064 drivers/scsi/be2iscsi/be_main.c 	struct be_queue_info *cq, *eq;
eq               3076 drivers/scsi/be2iscsi/be_main.c 		eq = &phwi_context->be_eq[i].q;
eq               3099 drivers/scsi/be2iscsi/be_main.c 		ret = beiscsi_cmd_cq_create(&phba->ctrl, cq, eq, false,
eq               3109 drivers/scsi/be2iscsi/be_main.c 			    "iSCSI CQ CREATED\n", cq->id, eq->id);
eq               3586 drivers/scsi/be2iscsi/be_main.c 	struct be_queue_info *eq;
eq               3602 drivers/scsi/be2iscsi/be_main.c 		eq = &phwi_context->be_eq[i].q;
eq               3603 drivers/scsi/be2iscsi/be_main.c 		eqe = queue_tail_node(eq);
eq               3608 drivers/scsi/be2iscsi/be_main.c 			queue_tail_inc(eq);
eq               3609 drivers/scsi/be2iscsi/be_main.c 			eqe = queue_tail_node(eq);
eq               3614 drivers/scsi/be2iscsi/be_main.c 			hwi_ring_eq_db(phba, eq->id, 1,	num_processed, 1, 1);
eq               4099 drivers/scsi/be2iscsi/be_main.c 	struct be_queue_info *eq;
eq               4120 drivers/scsi/be2iscsi/be_main.c 		eq = &phwi_context->be_eq[0].q;
eq               4122 drivers/scsi/be2iscsi/be_main.c 			    "BM_%d : eq->id=%d\n", eq->id);
eq               4124 drivers/scsi/be2iscsi/be_main.c 		hwi_ring_eq_db(phba, eq->id, 0, 0, 1, 1);
eq               4127 drivers/scsi/be2iscsi/be_main.c 			eq = &phwi_context->be_eq[i].q;
eq               4129 drivers/scsi/be2iscsi/be_main.c 				    "BM_%d : eq->id=%d\n", eq->id);
eq               4130 drivers/scsi/be2iscsi/be_main.c 			hwi_ring_eq_db(phba, eq->id, 0, 0, 1, 1);
eq                990 drivers/scsi/csiostor/csio_wr.c 	csio_wr_reg32(hw, DBPRIO_V(prio) | QID_V(q->un.eq.physeqid) |
eq                411 drivers/scsi/csiostor/csio_wr.h 		struct csio_eq	eq;
eq                468 drivers/scsi/csiostor/csio_wr.h #define	csio_q_eqid(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->un.eq.eqid)
eq                472 drivers/scsi/csiostor/csio_wr.h 				((__hw)->wrm.q_arr[(__idx)]->un.eq.physeqid)
eq                479 drivers/scsi/csiostor/csio_wr.h #define	csio_q_eq_wrap(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->un.eq.wrap)
eq               1319 drivers/scsi/lpfc/lpfc.h lpfc_sli4_mod_hba_eq_delay(struct lpfc_hba *phba, struct lpfc_queue *eq,
eq               1325 drivers/scsi/lpfc/lpfc.h 	bf_set(lpfc_sliport_eqdelay_id, &reg_data, eq->queue_id);
eq               1328 drivers/scsi/lpfc/lpfc.h 	eq->q_mode = delay;
eq               5163 drivers/scsi/lpfc/lpfc_attr.c 	struct lpfc_queue *eq, *cq;
eq               5185 drivers/scsi/lpfc/lpfc_attr.c 		eq = phba->sli4_hba.hba_eq_hdl[i].eq;
eq               5186 drivers/scsi/lpfc/lpfc_attr.c 		if (!eq)
eq               5189 drivers/scsi/lpfc/lpfc_attr.c 		list_for_each_entry(cq, &eq->child_list, list)
eq               5317 drivers/scsi/lpfc/lpfc_attr.c 					cpup->eq, cpup->hdwq, cpup->phys_id,
eq               5339 drivers/scsi/lpfc/lpfc_attr.c 					cpup->eq, cpup->hdwq, cpup->phys_id,
eq                454 drivers/scsi/lpfc/lpfc_debugfs.h 	struct lpfc_queue *wq, *cq, *eq;
eq                459 drivers/scsi/lpfc/lpfc_debugfs.h 	eq = NULL;
eq                481 drivers/scsi/lpfc/lpfc_debugfs.h 		eq = phba->sli4_hba.hdwq[eqidx].hba_eq;
eq                482 drivers/scsi/lpfc/lpfc_debugfs.h 		if (cq->assoc_qid == eq->queue_id)
eq                488 drivers/scsi/lpfc/lpfc_debugfs.h 		eq = phba->sli4_hba.hdwq[0].hba_eq;
eq                495 drivers/scsi/lpfc/lpfc_debugfs.h 			eqidx, eq->queue_id);
eq                500 drivers/scsi/lpfc/lpfc_debugfs.h 			eqidx, eq->queue_id);
eq               1153 drivers/scsi/lpfc/lpfc_hw4.h 			struct eq_delay_info eq[LPFC_MAX_EQ_DELAY_EQID_CNT];
eq               1242 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_queue *eq, *eq_next;
eq               1264 drivers/scsi/lpfc/lpfc_init.c 			eq = phba->sli4_hba.hba_eq_hdl[i].eq;
eq               1265 drivers/scsi/lpfc/lpfc_init.c 			if (!eq)
eq               1267 drivers/scsi/lpfc/lpfc_init.c 			if (eq->q_mode) {
eq               1271 drivers/scsi/lpfc/lpfc_init.c 			if (eqcnt[eq->last_cpu] < 2)
eq               1272 drivers/scsi/lpfc/lpfc_init.c 				eqcnt[eq->last_cpu]++;
eq               1291 drivers/scsi/lpfc/lpfc_init.c 		list_for_each_entry_safe(eq, eq_next, &eqi->list, cpu_list) {
eq               1292 drivers/scsi/lpfc/lpfc_init.c 			if (eq->last_cpu != i) {
eq               1294 drivers/scsi/lpfc/lpfc_init.c 						      eq->last_cpu);
eq               1295 drivers/scsi/lpfc/lpfc_init.c 				list_move_tail(&eq->cpu_list, &eqi_new->list);
eq               1298 drivers/scsi/lpfc/lpfc_init.c 			if (usdelay != eq->q_mode)
eq               1299 drivers/scsi/lpfc/lpfc_init.c 				lpfc_modify_hba_eq_delay(phba, eq->hdwq, 1,
eq               8863 drivers/scsi/lpfc/lpfc_init.c 		eqcpu = lpfc_find_cpu_handle(phba, cpup->eq, LPFC_FIND_BY_EQ);
eq               9117 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_queue *eq;
eq               9136 drivers/scsi/lpfc/lpfc_init.c 		eq = phba->sli4_hba.hba_eq_hdl[idx].eq;
eq               9137 drivers/scsi/lpfc/lpfc_init.c 		lpfc_sli4_queue_free(eq);
eq               9138 drivers/scsi/lpfc/lpfc_init.c 		phba->sli4_hba.hba_eq_hdl[idx].eq = NULL;
eq               9238 drivers/scsi/lpfc/lpfc_init.c lpfc_create_wq_cq(struct lpfc_hba *phba, struct lpfc_queue *eq,
eq               9245 drivers/scsi/lpfc/lpfc_init.c 	if (!eq || !cq || !wq) {
eq               9248 drivers/scsi/lpfc/lpfc_init.c 			((eq) ? ((cq) ? "WQ" : "CQ") : "EQ"), qidx);
eq               9253 drivers/scsi/lpfc/lpfc_init.c 	rc = lpfc_cq_create(phba, cq, eq,
eq               9269 drivers/scsi/lpfc/lpfc_init.c 			qidx, cq->queue_id, qidx, eq->queue_id);
eq               9318 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_queue *eq, *childq;
eq               9326 drivers/scsi/lpfc/lpfc_init.c 		eq = phba->sli4_hba.hba_eq_hdl[qidx].eq;
eq               9327 drivers/scsi/lpfc/lpfc_init.c 		if (!eq)
eq               9330 drivers/scsi/lpfc/lpfc_init.c 		list_for_each_entry(childq, &eq->child_list, list) {
eq               9433 drivers/scsi/lpfc/lpfc_init.c 			if (qidx != cpup->eq)
eq               9443 drivers/scsi/lpfc/lpfc_init.c 						cpup->eq, (uint32_t)rc);
eq               9448 drivers/scsi/lpfc/lpfc_init.c 			phba->sli4_hba.hba_eq_hdl[cpup->eq].eq =
eq               9453 drivers/scsi/lpfc/lpfc_init.c 					cpup->eq,
eq               9718 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_queue *eq;
eq               9780 drivers/scsi/lpfc/lpfc_init.c 			eq = phba->sli4_hba.hba_eq_hdl[qidx].eq;
eq               9781 drivers/scsi/lpfc/lpfc_init.c 			lpfc_eq_destroy(phba, eq);
eq               10595 drivers/scsi/lpfc/lpfc_init.c 		    (cpup->eq == id))
eq               10661 drivers/scsi/lpfc/lpfc_init.c 		cpup->eq = LPFC_VECTOR_MAP_EMPTY;
eq               10731 drivers/scsi/lpfc/lpfc_init.c 				cpup->eq = idx;
eq               10743 drivers/scsi/lpfc/lpfc_init.c 			cpup->eq = idx;
eq               10756 drivers/scsi/lpfc/lpfc_init.c 					cpu, cpup->irq, cpup->eq, cpup->flag);
eq               10772 drivers/scsi/lpfc/lpfc_init.c 		if (cpup->eq == LPFC_VECTOR_MAP_EMPTY) {
eq               10797 drivers/scsi/lpfc/lpfc_init.c 			cpup->eq = new_cpup->eq;
eq               10823 drivers/scsi/lpfc/lpfc_init.c 		if (cpup->eq == LPFC_VECTOR_MAP_EMPTY) {
eq               10851 drivers/scsi/lpfc/lpfc_init.c 			cpup->eq = new_cpup->eq;
eq               10888 drivers/scsi/lpfc/lpfc_init.c 				cpup->hdwq, cpup->eq, cpup->irq, cpup->flag);
eq               10966 drivers/scsi/lpfc/lpfc_init.c 				cpup->hdwq, cpup->eq, cpup->irq, cpup->flag);
eq               10988 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_queue *eq;
eq               11023 drivers/scsi/lpfc/lpfc_init.c 			eq = phba->sli4_hba.hdwq[map->hdwq].hba_eq;
eq               11024 drivers/scsi/lpfc/lpfc_init.c 			list_add(&eq->_poll_list, eqlist);
eq               11092 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_queue *eq, *next;
eq               11107 drivers/scsi/lpfc/lpfc_init.c 	list_for_each_entry_safe(eq, next, &eqlist, _poll_list) {
eq               11108 drivers/scsi/lpfc/lpfc_init.c 		list_del_init(&eq->_poll_list);
eq               11109 drivers/scsi/lpfc/lpfc_init.c 		lpfc_sli4_start_polling(eq);
eq               11118 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_queue *eq, *next;
eq               11130 drivers/scsi/lpfc/lpfc_init.c 	list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list) {
eq               11131 drivers/scsi/lpfc/lpfc_init.c 		n = lpfc_find_cpu_handle(phba, eq->hdwq, LPFC_FIND_BY_HDWQ);
eq               11133 drivers/scsi/lpfc/lpfc_init.c 			lpfc_sli4_stop_polling(eq);
eq                 86 drivers/scsi/lpfc/lpfc_sli.c 				     struct lpfc_queue *eq,
eq                456 drivers/scsi/lpfc/lpfc_sli.c __lpfc_sli4_consume_eqe(struct lpfc_hba *phba, struct lpfc_queue *eq,
eq                462 drivers/scsi/lpfc/lpfc_sli.c 	eq->host_index = ((eq->host_index + 1) % eq->entry_count);
eq                465 drivers/scsi/lpfc/lpfc_sli.c 	if (phba->sli4_hba.pc_sli4_params.eqav && !eq->host_index)
eq                466 drivers/scsi/lpfc/lpfc_sli.c 		eq->qe_valid = (eq->qe_valid) ? 0 : 1;
eq                470 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_eq_flush(struct lpfc_hba *phba, struct lpfc_queue *eq)
eq                476 drivers/scsi/lpfc/lpfc_sli.c 	eqe = lpfc_sli4_eq_get(eq);
eq                478 drivers/scsi/lpfc/lpfc_sli.c 		__lpfc_sli4_consume_eqe(phba, eq, eqe);
eq                480 drivers/scsi/lpfc/lpfc_sli.c 		eqe = lpfc_sli4_eq_get(eq);
eq                484 drivers/scsi/lpfc/lpfc_sli.c 	phba->sli4_hba.sli4_write_eq_db(phba, eq, count, LPFC_QUEUE_REARM);
eq                488 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_process_eq(struct lpfc_hba *phba, struct lpfc_queue *eq,
eq                494 drivers/scsi/lpfc/lpfc_sli.c 	if (cmpxchg(&eq->queue_claimed, 0, 1) != 0)
eq                497 drivers/scsi/lpfc/lpfc_sli.c 	eqe = lpfc_sli4_eq_get(eq);
eq                499 drivers/scsi/lpfc/lpfc_sli.c 		lpfc_sli4_hba_handle_eqe(phba, eq, eqe);
eq                500 drivers/scsi/lpfc/lpfc_sli.c 		__lpfc_sli4_consume_eqe(phba, eq, eqe);
eq                503 drivers/scsi/lpfc/lpfc_sli.c 		if (!(++count % eq->max_proc_limit))
eq                506 drivers/scsi/lpfc/lpfc_sli.c 		if (!(count % eq->notify_interval)) {
eq                507 drivers/scsi/lpfc/lpfc_sli.c 			phba->sli4_hba.sli4_write_eq_db(phba, eq, consumed,
eq                512 drivers/scsi/lpfc/lpfc_sli.c 		eqe = lpfc_sli4_eq_get(eq);
eq                514 drivers/scsi/lpfc/lpfc_sli.c 	eq->EQ_processed += count;
eq                517 drivers/scsi/lpfc/lpfc_sli.c 	if (count > eq->EQ_max_eqe)
eq                518 drivers/scsi/lpfc/lpfc_sli.c 		eq->EQ_max_eqe = count;
eq                520 drivers/scsi/lpfc/lpfc_sli.c 	eq->queue_claimed = 0;
eq                524 drivers/scsi/lpfc/lpfc_sli.c 	phba->sli4_hba.sli4_write_eq_db(phba, eq, consumed, rearm);
eq               5554 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_queue *eq;
eq               5573 drivers/scsi/lpfc/lpfc_sli.c 			eq = sli4_hba->hba_eq_hdl[qidx].eq;
eq               5575 drivers/scsi/lpfc/lpfc_sli.c 			sli4_hba->sli4_write_eq_db(phba, eq,
eq               7869 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_queue *eq;
eq               7878 drivers/scsi/lpfc/lpfc_sli.c 			eq = phba->sli4_hba.hba_eq_hdl[eqidx].eq;
eq               7879 drivers/scsi/lpfc/lpfc_sli.c 			if (eq && eq->queue_id == sli4_hba->mbx_cq->assoc_qid) {
eq               7880 drivers/scsi/lpfc/lpfc_sli.c 				fpeq = eq;
eq               10066 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_queue *eq;
eq               10071 drivers/scsi/lpfc/lpfc_sli.c 		eq = phba->sli4_hba.hdwq[piocb->hba_wqidx].hba_eq;
eq               10081 drivers/scsi/lpfc/lpfc_sli.c 		lpfc_sli4_poll_eq(eq, LPFC_POLL_FASTPATH);
eq               14043 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_hba_handle_eqe(struct lpfc_hba *phba, struct lpfc_queue *eq,
eq               14047 drivers/scsi/lpfc/lpfc_sli.c 	uint32_t qidx = eq->hdwq;
eq               14234 drivers/scsi/lpfc/lpfc_sli.c 	fpeq = phba->sli4_hba.hba_eq_hdl[hba_eqidx].eq;
eq               14324 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_queue *eq;
eq               14329 drivers/scsi/lpfc/lpfc_sli.c 	list_for_each_entry_rcu(eq, &phba->poll_list, _poll_list)
eq               14330 drivers/scsi/lpfc/lpfc_sli.c 		i += lpfc_sli4_poll_eq(eq, LPFC_POLL_SLOWPATH);
eq               14338 drivers/scsi/lpfc/lpfc_sli.c inline int lpfc_sli4_poll_eq(struct lpfc_queue *eq, uint8_t path)
eq               14340 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_hba *phba = eq->phba;
eq               14352 drivers/scsi/lpfc/lpfc_sli.c 	if (READ_ONCE(eq->mode) == LPFC_EQ_POLL)
eq               14360 drivers/scsi/lpfc/lpfc_sli.c 		i = lpfc_sli4_process_eq(phba, eq, LPFC_QUEUE_NOARM);
eq               14365 drivers/scsi/lpfc/lpfc_sli.c static inline void lpfc_sli4_add_to_poll_list(struct lpfc_queue *eq)
eq               14367 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_hba *phba = eq->phba;
eq               14376 drivers/scsi/lpfc/lpfc_sli.c 	list_add_rcu(&eq->_poll_list, &phba->poll_list);
eq               14380 drivers/scsi/lpfc/lpfc_sli.c static inline void lpfc_sli4_remove_from_poll_list(struct lpfc_queue *eq)
eq               14382 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_hba *phba = eq->phba;
eq               14387 drivers/scsi/lpfc/lpfc_sli.c 	list_del_rcu(&eq->_poll_list);
eq               14396 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_queue *eq, *next;
eq               14398 drivers/scsi/lpfc/lpfc_sli.c 	list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list)
eq               14399 drivers/scsi/lpfc/lpfc_sli.c 		list_del(&eq->_poll_list);
eq               14406 drivers/scsi/lpfc/lpfc_sli.c __lpfc_sli4_switch_eqmode(struct lpfc_queue *eq, uint8_t mode)
eq               14408 drivers/scsi/lpfc/lpfc_sli.c 	if (mode == eq->mode)
eq               14421 drivers/scsi/lpfc/lpfc_sli.c 	WRITE_ONCE(eq->mode, mode);
eq               14437 drivers/scsi/lpfc/lpfc_sli.c 	mode ? lpfc_sli4_add_to_poll_list(eq) :
eq               14438 drivers/scsi/lpfc/lpfc_sli.c 	       lpfc_sli4_remove_from_poll_list(eq);
eq               14441 drivers/scsi/lpfc/lpfc_sli.c void lpfc_sli4_start_polling(struct lpfc_queue *eq)
eq               14443 drivers/scsi/lpfc/lpfc_sli.c 	__lpfc_sli4_switch_eqmode(eq, LPFC_EQ_POLL);
eq               14446 drivers/scsi/lpfc/lpfc_sli.c void lpfc_sli4_stop_polling(struct lpfc_queue *eq)
eq               14448 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_hba *phba = eq->phba;
eq               14450 drivers/scsi/lpfc/lpfc_sli.c 	__lpfc_sli4_switch_eqmode(eq, LPFC_EQ_INTERRUPT);
eq               14459 drivers/scsi/lpfc/lpfc_sli.c 	phba->sli4_hba.sli4_write_eq_db(phba, eq, 0, LPFC_QUEUE_REARM);
eq               14641 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_queue *eq;
eq               14661 drivers/scsi/lpfc/lpfc_sli.c 			eq = phba->sli4_hba.hba_eq_hdl[qidx].eq;
eq               14662 drivers/scsi/lpfc/lpfc_sli.c 			if (!eq)
eq               14665 drivers/scsi/lpfc/lpfc_sli.c 			lpfc_sli4_mod_hba_eq_delay(phba, eq, usdelay);
eq               14697 drivers/scsi/lpfc/lpfc_sli.c 		eq = phba->sli4_hba.hba_eq_hdl[qidx].eq;
eq               14698 drivers/scsi/lpfc/lpfc_sli.c 		if (!eq)
eq               14700 drivers/scsi/lpfc/lpfc_sli.c 		eq->q_mode = usdelay;
eq               14701 drivers/scsi/lpfc/lpfc_sli.c 		eq_delay->u.request.eq[cnt].eq_id = eq->queue_id;
eq               14702 drivers/scsi/lpfc/lpfc_sli.c 		eq_delay->u.request.eq[cnt].phase = 0;
eq               14703 drivers/scsi/lpfc/lpfc_sli.c 		eq_delay->u.request.eq[cnt].delay_multi = dmult;
eq               14749 drivers/scsi/lpfc/lpfc_sli.c lpfc_eq_create(struct lpfc_hba *phba, struct lpfc_queue *eq, uint32_t imax)
eq               14761 drivers/scsi/lpfc/lpfc_sli.c 	if (!eq)
eq               14777 drivers/scsi/lpfc/lpfc_sli.c 	       eq->page_count);
eq               14794 drivers/scsi/lpfc/lpfc_sli.c 	switch (eq->entry_count) {
eq               14798 drivers/scsi/lpfc/lpfc_sli.c 				eq->entry_count);
eq               14799 drivers/scsi/lpfc/lpfc_sli.c 		if (eq->entry_count < 256) {
eq               14825 drivers/scsi/lpfc/lpfc_sli.c 	list_for_each_entry(dmabuf, &eq->page_list, list) {
eq               14846 drivers/scsi/lpfc/lpfc_sli.c 	eq->type = LPFC_EQ;
eq               14847 drivers/scsi/lpfc/lpfc_sli.c 	eq->subtype = LPFC_NONE;
eq               14848 drivers/scsi/lpfc/lpfc_sli.c 	eq->queue_id = bf_get(lpfc_mbx_eq_create_q_id, &eq_create->u.response);
eq               14849 drivers/scsi/lpfc/lpfc_sli.c 	if (eq->queue_id == 0xFFFF)
eq               14851 drivers/scsi/lpfc/lpfc_sli.c 	eq->host_index = 0;
eq               14852 drivers/scsi/lpfc/lpfc_sli.c 	eq->notify_interval = LPFC_EQ_NOTIFY_INTRVL;
eq               14853 drivers/scsi/lpfc/lpfc_sli.c 	eq->max_proc_limit = LPFC_EQ_MAX_PROC_LIMIT;
eq               14882 drivers/scsi/lpfc/lpfc_sli.c 	       struct lpfc_queue *eq, uint32_t type, uint32_t subtype)
eq               14892 drivers/scsi/lpfc/lpfc_sli.c 	if (!cq || !eq)
eq               14915 drivers/scsi/lpfc/lpfc_sli.c 		       eq->queue_id);
eq               14920 drivers/scsi/lpfc/lpfc_sli.c 		       eq->queue_id);
eq               14985 drivers/scsi/lpfc/lpfc_sli.c 	list_add_tail(&cq->list, &eq->child_list);
eq               14990 drivers/scsi/lpfc/lpfc_sli.c 	cq->assoc_qid = eq->queue_id;
eq               14991 drivers/scsi/lpfc/lpfc_sli.c 	cq->assoc_qp = eq;
eq               15031 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_queue *eq;
eq               15070 drivers/scsi/lpfc/lpfc_sli.c 		eq = hdwq[idx].hba_eq;
eq               15071 drivers/scsi/lpfc/lpfc_sli.c 		if (!cq || !eq) {
eq               15133 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15137 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15141 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15145 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15149 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15153 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15157 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15161 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15165 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15169 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15173 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15177 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15181 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15185 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15189 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15193 drivers/scsi/lpfc/lpfc_sli.c 			       &cq_set->u.request, eq->queue_id);
eq               15198 drivers/scsi/lpfc/lpfc_sli.c 		list_add_tail(&cq->list, &eq->child_list);
eq               15202 drivers/scsi/lpfc/lpfc_sli.c 		cq->assoc_qid = eq->queue_id;
eq               15203 drivers/scsi/lpfc/lpfc_sli.c 		cq->assoc_qp = eq;
eq               16228 drivers/scsi/lpfc/lpfc_sli.c lpfc_eq_destroy(struct lpfc_hba *phba, struct lpfc_queue *eq)
eq               16236 drivers/scsi/lpfc/lpfc_sli.c 	if (!eq)
eq               16239 drivers/scsi/lpfc/lpfc_sli.c 	mbox = mempool_alloc(eq->phba->mbox_mem_pool, GFP_KERNEL);
eq               16248 drivers/scsi/lpfc/lpfc_sli.c 	       eq->queue_id);
eq               16249 drivers/scsi/lpfc/lpfc_sli.c 	mbox->vport = eq->phba->pport;
eq               16252 drivers/scsi/lpfc/lpfc_sli.c 	rc = lpfc_sli_issue_mbox(eq->phba, mbox, MBX_POLL);
eq               16267 drivers/scsi/lpfc/lpfc_sli.c 	list_del_init(&eq->list);
eq               16268 drivers/scsi/lpfc/lpfc_sli.c 	mempool_free(mbox, eq->phba->mbox_mem_pool);
eq                471 drivers/scsi/lpfc/lpfc_sli4.h 	struct lpfc_queue *eq;
eq                567 drivers/scsi/lpfc/lpfc_sli4.h 	uint16_t	eq;
eq                799 drivers/scsi/lpfc/lpfc_sli4.h 	void (*sli4_write_eq_db)(struct lpfc_hba *phba, struct lpfc_queue *eq,
eq               2313 drivers/soc/fsl/qbman/qman.c 	struct qm_eqcr_entry *eq;
eq               2325 drivers/soc/fsl/qbman/qman.c 		eq = qm_eqcr_start_stash(&p->p);
eq               2334 drivers/soc/fsl/qbman/qman.c 		eq = qm_eqcr_start_no_stash(&p->p);
eq               2337 drivers/soc/fsl/qbman/qman.c 	if (unlikely(!eq))
eq               2340 drivers/soc/fsl/qbman/qman.c 	qm_fqid_set(eq, fq->fqid);
eq               2341 drivers/soc/fsl/qbman/qman.c 	eq->tag = cpu_to_be32(fq_to_tag(fq));
eq               2342 drivers/soc/fsl/qbman/qman.c 	eq->fd = *fd;
eq                334 fs/befs/btree.c 	int eq;
eq                348 fs/befs/btree.c 	eq = befs_compare_strings(thiskey, keylen, findkey, findkey_len);
eq                349 fs/befs/btree.c 	if (eq < 0) {
eq                364 fs/befs/btree.c 		eq = befs_compare_strings(thiskey, keylen, findkey,
eq                367 fs/befs/btree.c 		if (eq == 0) {
eq                374 fs/befs/btree.c 		if (eq > 0)
eq                381 fs/befs/btree.c 	if (eq < 0)
eq                 63 include/linux/mlx5/cq.h 	struct mlx5_eq_comp	*eq;
eq                 24 include/linux/mlx5/eq.h mlx5_eq_destroy_generic(struct mlx5_core_dev *dev, struct mlx5_eq *eq);
eq                 25 include/linux/mlx5/eq.h int mlx5_eq_enable(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
eq                 27 include/linux/mlx5/eq.h void mlx5_eq_disable(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
eq                 30 include/linux/mlx5/eq.h struct mlx5_eqe *mlx5_eq_get_eqe(struct mlx5_eq *eq, u32 cc);
eq                 31 include/linux/mlx5/eq.h void mlx5_eq_update_ci(struct mlx5_eq *eq, u32 cc, bool arm);
eq                 41 include/linux/mlx5/eq.h static inline u32 mlx5_eq_update_cc(struct mlx5_eq *eq, u32 cc)
eq                 44 include/linux/mlx5/eq.h 		mlx5_eq_update_ci(eq, cc, 0);
eq                 77 net/caif/cfctrl.c 	bool eq =
eq                 83 net/caif/cfctrl.c 	if (!eq)
eq                150 net/dns_resolver/dns_key.c 			const char *eq;
eq                161 net/dns_resolver/dns_key.c 			eq = memchr(opt, '=', opt_len);
eq                162 net/dns_resolver/dns_key.c 			if (eq) {
eq                163 net/dns_resolver/dns_key.c 				opt_nlen = eq - opt;
eq                164 net/dns_resolver/dns_key.c 				eq++;
eq                165 net/dns_resolver/dns_key.c 				memcpy(optval, eq, next_opt - eq);
eq                166 net/dns_resolver/dns_key.c 				optval[next_opt - eq] = '\0';
eq               2564 security/selinux/ss/policydb.c 	int rc, eq;
eq               2566 security/selinux/ss/policydb.c 	eq = mls_level_eq(&r->level[1], &r->level[0]);
eq               2568 security/selinux/ss/policydb.c 	if (eq)
eq               2574 security/selinux/ss/policydb.c 	if (!eq)
eq               2586 security/selinux/ss/policydb.c 	if (!eq) {
eq                159 sound/pci/au88x0/au88x0.h 	eqlzr_t eq;
eq                 56 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                 78 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                101 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                118 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                164 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                211 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                221 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                231 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                241 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                252 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                267 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                282 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                297 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                314 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                338 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                478 sound/pci/au88x0/au88x0_eq.c 	eqhw_t *eqhw = &(vortex->eq.this04);
eq                495 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                497 sound/pci/au88x0/au88x0_eq.c 	if (eq->this28) {
eq                498 sound/pci/au88x0/au88x0_eq.c 		*gain = eq->this130[index];
eq                506 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                508 sound/pci/au88x0/au88x0_eq.c 	if (eq->this28 == 0)
eq                511 sound/pci/au88x0/au88x0_eq.c 	eq->this130[index] = gain;
eq                512 sound/pci/au88x0/au88x0_eq.c 	if (eq->this54)
eq                520 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                522 sound/pci/au88x0/au88x0_eq.c 	if (eq->this28) {
eq                523 sound/pci/au88x0/au88x0_eq.c 		*gain = eq->this130[index + eq->this10];
eq                531 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                533 sound/pci/au88x0/au88x0_eq.c 	if (eq->this28 == 0)
eq                536 sound/pci/au88x0/au88x0_eq.c 	eq->this130[index + eq->this10] = gain;
eq                537 sound/pci/au88x0/au88x0_eq.c 	if (eq->this54)
eq                547 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                550 sound/pci/au88x0/au88x0_eq.c 	if (eq->this10 == 0)
eq                557 sound/pci/au88x0/au88x0_eq.c 		    (vortex, si, &gains[si + eq->this10]))
eq                561 sound/pci/au88x0/au88x0_eq.c 	while (eq->this10 > si) ;
eq                568 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                570 sound/pci/au88x0/au88x0_eq.c 	vortex_EqHw_SetLeftGainsTarget(vortex, eq->this130);
eq                571 sound/pci/au88x0/au88x0_eq.c 	vortex_EqHw_SetRightGainsTarget(vortex, &(eq->this130[eq->this10]));
eq                579 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                582 sound/pci/au88x0/au88x0_eq.c 	if (((eq->this10) * 2 != count) || (eq->this28 == 0))
eq                586 sound/pci/au88x0/au88x0_eq.c 		eq->this130[i] = gains[i];
eq                589 sound/pci/au88x0/au88x0_eq.c 	if (eq->this54)
eq                597 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                600 sound/pci/au88x0/au88x0_eq.c 	eq->this58 = a;
eq                601 sound/pci/au88x0/au88x0_eq.c 	eq->this5c = b;
eq                602 sound/pci/au88x0/au88x0_eq.c 	if (eq->this54)
eq                603 sound/pci/au88x0/au88x0_eq.c 		eax = eq->this0e;
eq                605 sound/pci/au88x0/au88x0_eq.c 		eax = eq->this0a;
eq                606 sound/pci/au88x0/au88x0_eq.c 	ebx = (eax * eq->this58) >> 0x10;
eq                607 sound/pci/au88x0/au88x0_eq.c 	eax = (eax * eq->this5c) >> 0x10;
eq                613 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                616 sound/pci/au88x0/au88x0_eq.c 	if (eq->this54)
eq                617 sound/pci/au88x0/au88x0_eq.c 		eax = eq->this0e;
eq                619 sound/pci/au88x0/au88x0_eq.c 		eax = eq->this0a;
eq                620 sound/pci/au88x0/au88x0_eq.c 	ebx = (eax * eq->this58) >> 0x10;
eq                621 sound/pci/au88x0/au88x0_eq.c 	eax = (eax * eq->this5c) >> 0x10;
eq                633 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                635 sound/pci/au88x0/au88x0_eq.c 	if ((eq->this28) && (bp == 0)) {
eq                638 sound/pci/au88x0/au88x0_eq.c 		vortex_EqHw_SetBypassGain(vortex, eq->this08, eq->this08);
eq                641 sound/pci/au88x0/au88x0_eq.c 		vortex_EqHw_SetLeftGainsTarget(vortex, eq->this14_array);
eq                642 sound/pci/au88x0/au88x0_eq.c 		vortex_EqHw_SetRightGainsTarget(vortex, eq->this14_array);
eq                643 sound/pci/au88x0/au88x0_eq.c 		vortex_EqHw_SetBypassGain(vortex, eq->this0c, eq->this0c);
eq                650 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                653 sound/pci/au88x0/au88x0_eq.c 	memcpy(&(eq->coefset), &asEqCoefsNormal, sizeof(auxxEqCoeffSet_t));
eq                655 sound/pci/au88x0/au88x0_eq.c 	vortex_Eqlzr_SetAllBands(vortex, eq_gains_normal, eq->this10 * 2);
eq                660 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                662 sound/pci/au88x0/au88x0_eq.c 	if (eq->this10 == 0)
eq                664 sound/pci/au88x0/au88x0_eq.c 	*count = eq->this10 * 2;
eq                672 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                674 sound/pci/au88x0/au88x0_eq.c 	return (&(eq->coefset));
eq                679 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                683 sound/pci/au88x0/au88x0_eq.c 	eq->this08 = 0;		/* Bypass gain with EQ in use. */
eq                684 sound/pci/au88x0/au88x0_eq.c 	eq->this0a = 0x5999;
eq                685 sound/pci/au88x0/au88x0_eq.c 	eq->this0c = 0x5999;	/* Bypass gain with EQ disabled. */
eq                686 sound/pci/au88x0/au88x0_eq.c 	eq->this0e = 0x5999;
eq                688 sound/pci/au88x0/au88x0_eq.c 	eq->this10 = 0xa;	/* 10 eq frequency bands. */
eq                689 sound/pci/au88x0/au88x0_eq.c 	eq->this04.this04 = eq->this10;
eq                690 sound/pci/au88x0/au88x0_eq.c 	eq->this28 = 0x1;	/* if 1 => Allow read access to this130 (gains) */
eq                691 sound/pci/au88x0/au88x0_eq.c 	eq->this54 = 0x0;	/* if 1 => Dont Allow access to hardware (gains) */
eq                692 sound/pci/au88x0/au88x0_eq.c 	eq->this58 = 0xffff;
eq                693 sound/pci/au88x0/au88x0_eq.c 	eq->this5c = 0xffff;
eq                696 sound/pci/au88x0/au88x0_eq.c 	memset(eq->this14_array, 0, sizeof(eq->this14_array));
eq                703 sound/pci/au88x0/au88x0_eq.c 	vortex_EqHw_Program10Band(vortex, &(eq->coefset));
eq                704 sound/pci/au88x0/au88x0_eq.c 	vortex_Eqlzr_SetBypass(vortex, eq->this54);
eq                726 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                729 sound/pci/au88x0/au88x0_eq.c 	ucontrol->value.integer.value[0] = eq->this54 ? 0 : 1;
eq                739 sound/pci/au88x0/au88x0_eq.c 	eqlzr_t *eq = &(vortex->eq);
eq                742 sound/pci/au88x0/au88x0_eq.c 	eq->this54 = ucontrol->value.integer.value[0] ? 0 : 1;
eq                743 sound/pci/au88x0/au88x0_eq.c 	vortex_Eqlzr_SetBypass(vortex, eq->this54);
eq                789 sound/soc/codecs/wm8958-dsp2.c 	int eq = kcontrol->private_value;
eq                793 sound/soc/codecs/wm8958-dsp2.c 	ucontrol->value.integer.value[0] = wm8994->enh_eq_ena[eq];
eq                801 sound/soc/codecs/wm8958-dsp2.c 	int eq = kcontrol->private_value;
eq                805 sound/soc/codecs/wm8958-dsp2.c 	if (wm8994->enh_eq_ena[eq] == ucontrol->value.integer.value[0])
eq                814 sound/soc/codecs/wm8958-dsp2.c 	if (wm8958_dsp2_busy(wm8994, eq)) {
eq                815 sound/soc/codecs/wm8958-dsp2.c 		dev_dbg(component->dev, "DSP2 active on %d already\n", eq);
eq                819 sound/soc/codecs/wm8958-dsp2.c 	if (wm8994->mbc_ena[eq] || wm8994->vss_ena[eq] ||
eq                820 sound/soc/codecs/wm8958-dsp2.c 	    wm8994->hpf1_ena[eq] || wm8994->hpf2_ena[eq])
eq                823 sound/soc/codecs/wm8958-dsp2.c 	wm8994->enh_eq_ena[eq] = ucontrol->value.integer.value[0];
eq                825 sound/soc/codecs/wm8958-dsp2.c 	wm8958_dsp_apply(component, eq, ucontrol->value.integer.value[0]);
eq               2233 tools/lib/bpf/btf.c 	int i, eq;
eq               2306 tools/lib/bpf/btf.c 		eq = btf_dedup_is_equiv(d,
eq               2308 tools/lib/bpf/btf.c 		if (eq <= 0)
eq               2309 tools/lib/bpf/btf.c 			return eq;
eq               2324 tools/lib/bpf/btf.c 			eq = btf_dedup_is_equiv(d, cand_m->type, canon_m->type);
eq               2325 tools/lib/bpf/btf.c 			if (eq <= 0)
eq               2326 tools/lib/bpf/btf.c 				return eq;
eq               2340 tools/lib/bpf/btf.c 		eq = btf_dedup_is_equiv(d, cand_type->type, canon_type->type);
eq               2341 tools/lib/bpf/btf.c 		if (eq <= 0)
eq               2342 tools/lib/bpf/btf.c 			return eq;
eq               2347 tools/lib/bpf/btf.c 			eq = btf_dedup_is_equiv(d, cand_p->type, canon_p->type);
eq               2348 tools/lib/bpf/btf.c 			if (eq <= 0)
eq               2349 tools/lib/bpf/btf.c 				return eq;
eq               2478 tools/lib/bpf/btf.c 		int eq;
eq               2495 tools/lib/bpf/btf.c 		eq = btf_dedup_is_equiv(d, type_id, cand_id);
eq               2496 tools/lib/bpf/btf.c 		if (eq < 0)
eq               2497 tools/lib/bpf/btf.c 			return eq;
eq               2498 tools/lib/bpf/btf.c 		if (!eq)