eq 93 arch/arc/include/asm/tlb-mmu1.h xor.eq r1,r1,1 /* flip bottom bit of data index */ eq 183 arch/arm/include/asm/assembler.h asm_trace_hardirqs_on cond=eq eq 192 arch/arm/include/asm/assembler.h .irp c,,eq,ne,cs,cc,mi,pl,vs,vc,hi,ls,ge,lt,gt,le,hs,lo eq 449 arch/arm/include/asm/assembler.h .irp c,,eq,ne,cs,cc,mi,pl,vs,vc,hi,ls,ge,lt,gt,le,hs,lo eq 117 arch/arm/include/asm/futex.h "2: " TUSERCOND(str, eq) " %3, [%4]\n" eq 39 arch/arm64/include/asm/fpsimdmacros.h b.eq 9999f eq 204 arch/arm64/include/asm/fpsimdmacros.h b.eq 921f eq 68 arch/ia64/kernel/minstate.h cmp.eq pKStk,pUStk=r0,r17; /* are we in kernel mode already? */ \ eq 128 arch/ia64/kernel/minstate.h cmp.eq pNonSys,pSys=r0,r0 /* initialize pSys=0, pNonSys=1 */ \ eq 62 arch/unicore32/include/asm/assembler.h .ifc \cond, eq eq 408 arch/x86/include/asm/vmx.h #define DEBUG_REG_ACCESS_REG(eq) (((eq) >> 8) & 0xf) /* 11:8, general purpose reg. */ eq 6426 block/bfq-iosched.c struct elevator_queue *eq; eq 6428 block/bfq-iosched.c eq = elevator_alloc(q, e); eq 6429 block/bfq-iosched.c if (!eq) eq 6434 block/bfq-iosched.c kobject_put(&eq->kobj); eq 6437 block/bfq-iosched.c eq->elevator_data = bfqd; eq 6440 block/bfq-iosched.c q->elevator = eq; eq 6551 block/bfq-iosched.c kobject_put(&eq->kobj); eq 526 block/blk-mq-sched.c struct elevator_queue *eq; eq 560 block/blk-mq-sched.c eq = q->elevator; eq 562 block/blk-mq-sched.c blk_mq_exit_sched(q, eq); eq 563 block/blk-mq-sched.c kobject_put(&eq->kobj); eq 167 block/elevator.c struct elevator_queue *eq; eq 169 block/elevator.c eq = kzalloc_node(sizeof(*eq), GFP_KERNEL, q->node); eq 170 block/elevator.c if (unlikely(!eq)) eq 173 block/elevator.c eq->type = e; eq 174 block/elevator.c kobject_init(&eq->kobj, &elv_ktype); eq 175 block/elevator.c mutex_init(&eq->sysfs_lock); eq 176 block/elevator.c hash_init(eq->hash); eq 178 block/elevator.c return eq; eq 419 block/kyber-iosched.c struct elevator_queue *eq; eq 421 block/kyber-iosched.c eq = elevator_alloc(q, e); eq 422 block/kyber-iosched.c if (!eq) eq 427 block/kyber-iosched.c kobject_put(&eq->kobj); eq 433 block/kyber-iosched.c eq->elevator_data = kqd; eq 434 block/kyber-iosched.c q->elevator = eq; eq 409 block/mq-deadline.c struct elevator_queue *eq; eq 411 block/mq-deadline.c eq = elevator_alloc(q, e); eq 412 block/mq-deadline.c if (!eq) eq 417 block/mq-deadline.c kobject_put(&eq->kobj); eq 420 block/mq-deadline.c eq->elevator_data = dd; eq 435 block/mq-deadline.c q->elevator = eq; eq 1175 drivers/block/drbd/drbd_worker.c int err, eq = 0; eq 1201 drivers/block/drbd/drbd_worker.c eq = !memcmp(digest, di->digest, digest_size); eq 1205 drivers/block/drbd/drbd_worker.c if (eq) { eq 1298 drivers/block/drbd/drbd_worker.c int err, eq = 0; eq 1323 drivers/block/drbd/drbd_worker.c eq = !memcmp(digest, di->digest, digest_size); eq 1334 drivers/block/drbd/drbd_worker.c if (!eq) eq 1340 drivers/block/drbd/drbd_worker.c eq ? ID_IN_SYNC : ID_OUT_OF_SYNC); eq 49 drivers/clk/spear/clk-aux-synth.c u8 eq = rtbl[index].eq ? 1 : 2; eq 52 drivers/clk/spear/clk-aux-synth.c (rtbl[index].yscale * eq)) * 10000; eq 115 drivers/clk/spear/clk-aux-synth.c val |= (rtbl[i].eq & aux->masks->eq_sel_mask) << eq 46 drivers/clk/spear/clk.h u8 eq; eq 255 drivers/clk/spear/spear1310_clock.c {.xscale = 10, .yscale = 204, .eq = 0}, /* 12.29 MHz */ eq 256 drivers/clk/spear/spear1310_clock.c {.xscale = 4, .yscale = 21, .eq = 0}, /* 48 MHz */ eq 257 drivers/clk/spear/spear1310_clock.c {.xscale = 2, .yscale = 6, .eq = 0}, /* 83 MHz */ eq 258 drivers/clk/spear/spear1310_clock.c {.xscale = 2, .yscale = 4, .eq = 0}, /* 125 MHz */ eq 259 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 3, .eq = 1}, /* 166 MHz */ eq 260 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 2, .eq = 1}, /* 250 MHz */ eq 266 drivers/clk/spear/spear1310_clock.c {.xscale = 2, .yscale = 6, .eq = 0}, /* divided by 6 */ eq 267 drivers/clk/spear/spear1310_clock.c {.xscale = 2, .yscale = 4, .eq = 0}, /* divided by 4 */ eq 268 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 3, .eq = 1}, /* divided by 3 */ eq 269 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 2, .eq = 1}, /* divided by 2 */ eq 314 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 12, .eq = 0}, /* 2.048 MHz, smp freq = 8Khz */ eq 315 drivers/clk/spear/spear1310_clock.c {.xscale = 11, .yscale = 96, .eq = 0}, /* 2.816 MHz, smp freq = 11Khz */ eq 316 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 6, .eq = 0}, /* 4.096 MHz, smp freq = 16Khz */ eq 317 drivers/clk/spear/spear1310_clock.c {.xscale = 11, .yscale = 48, .eq = 0}, /* 5.632 MHz, smp freq = 22Khz */ eq 323 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 3, .eq = 0}, eq 326 drivers/clk/spear/spear1310_clock.c {.xscale = 17, .yscale = 37, .eq = 0}, /* 11.289 MHz, smp freq = 44Khz*/ eq 328 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 2, .eq = 0}, /* 12.288 MHz */ eq 334 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 4, .eq = 0}, /* 1.53 MHz */ eq 335 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 2, .eq = 0}, /* 3.07 Mhz */ eq 342 drivers/clk/spear/spear1310_clock.c {.xscale = 1, .yscale = 31, .eq = 0}, /* 2.68 MHz */ eq 343 drivers/clk/spear/spear1310_clock.c {.xscale = 2, .yscale = 21, .eq = 0}, /* 7.94 MHz */ eq 344 drivers/clk/spear/spear1310_clock.c {.xscale = 4, .yscale = 21, .eq = 0}, /* 15.87 MHz */ eq 345 drivers/clk/spear/spear1310_clock.c {.xscale = 10, .yscale = 42, .eq = 0}, /* 19.84 MHz */ eq 267 drivers/clk/spear/spear1340_clock.c {.xscale = 5, .yscale = 122, .eq = 0}, eq 269 drivers/clk/spear/spear1340_clock.c {.xscale = 10, .yscale = 204, .eq = 0}, eq 271 drivers/clk/spear/spear1340_clock.c {.xscale = 4, .yscale = 25, .eq = 0}, eq 273 drivers/clk/spear/spear1340_clock.c {.xscale = 4, .yscale = 21, .eq = 0}, eq 275 drivers/clk/spear/spear1340_clock.c {.xscale = 5, .yscale = 18, .eq = 0}, eq 277 drivers/clk/spear/spear1340_clock.c {.xscale = 2, .yscale = 6, .eq = 0}, eq 279 drivers/clk/spear/spear1340_clock.c {.xscale = 5, .yscale = 12, .eq = 0}, eq 281 drivers/clk/spear/spear1340_clock.c {.xscale = 2, .yscale = 4, .eq = 0}, eq 283 drivers/clk/spear/spear1340_clock.c {.xscale = 5, .yscale = 18, .eq = 1}, eq 285 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 3, .eq = 1}, eq 287 drivers/clk/spear/spear1340_clock.c {.xscale = 5, .yscale = 12, .eq = 1}, eq 289 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 2, .eq = 1}, eq 295 drivers/clk/spear/spear1340_clock.c {.xscale = 2, .yscale = 6, .eq = 0}, /* divided by 6 */ eq 296 drivers/clk/spear/spear1340_clock.c {.xscale = 2, .yscale = 4, .eq = 0}, /* divided by 4 */ eq 297 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 3, .eq = 1}, /* divided by 3 */ eq 298 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 2, .eq = 1}, /* divided by 2 */ eq 353 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 12, .eq = 0}, /* 2.048 MHz, smp freq = 8Khz */ eq 354 drivers/clk/spear/spear1340_clock.c {.xscale = 11, .yscale = 96, .eq = 0}, /* 2.816 MHz, smp freq = 11Khz */ eq 355 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 6, .eq = 0}, /* 4.096 MHz, smp freq = 16Khz */ eq 356 drivers/clk/spear/spear1340_clock.c {.xscale = 11, .yscale = 48, .eq = 0}, /* 5.632 MHz, smp freq = 22Khz */ eq 362 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 3, .eq = 0}, eq 365 drivers/clk/spear/spear1340_clock.c {.xscale = 17, .yscale = 37, .eq = 0}, /* 11.289 MHz, smp freq = 44Khz*/ eq 366 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 2, .eq = 0}, /* 12.288 MHz, smp freq = 48Khz*/ eq 372 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 4, .eq = 0}, eq 373 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 2, .eq = 0}, eq 380 drivers/clk/spear/spear1340_clock.c {.xscale = 1, .yscale = 31, .eq = 0}, /* 2.68 MHz */ eq 381 drivers/clk/spear/spear1340_clock.c {.xscale = 2, .yscale = 21, .eq = 0}, /* 7.94 MHz */ eq 382 drivers/clk/spear/spear1340_clock.c {.xscale = 4, .yscale = 21, .eq = 0}, /* 15.87 MHz */ eq 383 drivers/clk/spear/spear1340_clock.c {.xscale = 10, .yscale = 42, .eq = 0}, /* 19.84 MHz */ eq 109 drivers/clk/spear/spear3xx_clock.c {.xscale = 1, .yscale = 81, .eq = 0}, /* 2.049 MHz */ eq 110 drivers/clk/spear/spear3xx_clock.c {.xscale = 1, .yscale = 59, .eq = 0}, /* 2.822 MHz */ eq 111 drivers/clk/spear/spear3xx_clock.c {.xscale = 2, .yscale = 81, .eq = 0}, /* 4.098 MHz */ eq 112 drivers/clk/spear/spear3xx_clock.c {.xscale = 3, .yscale = 89, .eq = 0}, /* 5.644 MHz */ eq 113 drivers/clk/spear/spear3xx_clock.c {.xscale = 4, .yscale = 81, .eq = 0}, /* 8.197 MHz */ eq 114 drivers/clk/spear/spear3xx_clock.c {.xscale = 4, .yscale = 59, .eq = 0}, /* 11.254 MHz */ eq 115 drivers/clk/spear/spear3xx_clock.c {.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */ eq 116 drivers/clk/spear/spear3xx_clock.c {.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */ eq 117 drivers/clk/spear/spear3xx_clock.c {.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */ eq 118 drivers/clk/spear/spear3xx_clock.c {.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */ eq 93 drivers/clk/spear/spear6xx_clock.c {.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */ eq 94 drivers/clk/spear/spear6xx_clock.c {.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */ eq 95 drivers/clk/spear/spear6xx_clock.c {.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */ eq 96 drivers/clk/spear/spear6xx_clock.c {.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */ eq 167 drivers/firmware/broadcom/bcm47xx_nvram.c char *var, *value, *end, *eq; eq 183 drivers/firmware/broadcom/bcm47xx_nvram.c eq = strchr(var, '='); eq 184 drivers/firmware/broadcom/bcm47xx_nvram.c if (!eq) eq 186 drivers/firmware/broadcom/bcm47xx_nvram.c value = eq + 1; eq 187 drivers/firmware/broadcom/bcm47xx_nvram.c if (eq - var == strlen(name) && eq 188 drivers/firmware/broadcom/bcm47xx_nvram.c strncmp(var, name, eq - var) == 0) eq 766 drivers/infiniband/hw/hfi1/pcie.c static int load_eq_table(struct hfi1_devdata *dd, const u8 eq[11][3], u8 fs, eq 780 drivers/infiniband/hw/hfi1/pcie.c c_minus1 = eq[i][PREC] / div; eq 781 drivers/infiniband/hw/hfi1/pcie.c c0 = fs - (eq[i][PREC] / div) - (eq[i][POST] / div); eq 782 drivers/infiniband/hw/hfi1/pcie.c c_plus1 = eq[i][POST] / div; eq 802 drivers/infiniband/hw/hfi1/pcie.c i, (u32)eq[i][0], (u32)eq[i][1], eq 803 drivers/infiniband/hw/hfi1/pcie.c (u32)eq[i][2]); eq 998 drivers/infiniband/hw/hfi1/pcie.c const u8 (*eq)[3]; eq 1136 drivers/infiniband/hw/hfi1/pcie.c eq = discrete_preliminary_eq; eq 1146 drivers/infiniband/hw/hfi1/pcie.c eq = integrated_preliminary_eq; eq 1157 drivers/infiniband/hw/hfi1/pcie.c ret = load_eq_table(dd, eq, fs, div); eq 196 drivers/infiniband/hw/hns/hns_roce_cq.c synchronize_irq(hr_dev->eq_table.eq[hr_cq->vector].irq); eq 791 drivers/infiniband/hw/hns/hns_roce_device.h struct hns_roce_eq *eq; eq 3694 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static void set_eq_cons_index_v1(struct hns_roce_eq *eq, int req_not) eq 3696 drivers/infiniband/hw/hns/hns_roce_hw_v1.c roce_raw_write((eq->cons_index & HNS_ROCE_V1_CONS_IDX_M) | eq 3697 drivers/infiniband/hw/hns/hns_roce_hw_v1.c (req_not << eq->log_entries), eq->doorbell); eq 3862 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_aeqe *get_aeqe_v1(struct hns_roce_eq *eq, u32 entry) eq 3864 drivers/infiniband/hw/hns/hns_roce_hw_v1.c unsigned long off = (entry & (eq->entries - 1)) * eq 3868 drivers/infiniband/hw/hns/hns_roce_hw_v1.c (eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) + eq 3872 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_aeqe *next_aeqe_sw_v1(struct hns_roce_eq *eq) eq 3874 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_aeqe *aeqe = get_aeqe_v1(eq, eq->cons_index); eq 3877 drivers/infiniband/hw/hns/hns_roce_hw_v1.c !!(eq->cons_index & eq->entries)) ? aeqe : NULL; eq 3881 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_eq *eq) eq 3888 drivers/infiniband/hw/hns/hns_roce_hw_v1.c while ((aeqe = next_aeqe_sw_v1(eq))) { eq 3952 drivers/infiniband/hw/hns/hns_roce_hw_v1.c event_type, eq->eqn, eq->cons_index); eq 3956 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->cons_index++; eq 3959 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (eq->cons_index > 2 * hr_dev->caps.aeqe_depth - 1) { eq 3961 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->cons_index = 0; eq 3965 drivers/infiniband/hw/hns/hns_roce_hw_v1.c set_eq_cons_index_v1(eq, 0); eq 3970 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_ceqe *get_ceqe_v1(struct hns_roce_eq *eq, u32 entry) eq 3972 drivers/infiniband/hw/hns/hns_roce_hw_v1.c unsigned long off = (entry & (eq->entries - 1)) * eq 3976 drivers/infiniband/hw/hns/hns_roce_hw_v1.c (eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) + eq 3980 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_ceqe *next_ceqe_sw_v1(struct hns_roce_eq *eq) eq 3982 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_ceqe *ceqe = get_ceqe_v1(eq, eq->cons_index); eq 3986 drivers/infiniband/hw/hns/hns_roce_hw_v1.c (!!(eq->cons_index & eq->entries)) ? ceqe : NULL; eq 3990 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_eq *eq) eq 3996 drivers/infiniband/hw/hns/hns_roce_hw_v1.c while ((ceqe = next_ceqe_sw_v1(eq))) { eq 4008 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ++eq->cons_index; eq 4011 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (eq->cons_index > eq 4013 drivers/infiniband/hw/hns/hns_roce_hw_v1.c dev_warn(&eq->hr_dev->pdev->dev, eq 4015 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->cons_index = 0; eq 4019 drivers/infiniband/hw/hns/hns_roce_hw_v1.c set_eq_cons_index_v1(eq, 0); eq 4026 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_eq *eq = eq_ptr; eq 4027 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_dev *hr_dev = eq->hr_dev; eq 4030 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (eq->type_flag == HNS_ROCE_CEQ) eq 4032 drivers/infiniband/hw/hns/hns_roce_hw_v1.c int_work = hns_roce_v1_ceq_int(hr_dev, eq); eq 4035 drivers/infiniband/hw/hns/hns_roce_hw_v1.c int_work = hns_roce_v1_aeq_int(hr_dev, eq); eq 4179 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_eq *eq) eq 4181 drivers/infiniband/hw/hns/hns_roce_hw_v1.c int npages = (PAGE_ALIGN(eq->eqe_size * eq->entries) + eq 4185 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (!eq->buf_list) eq 4190 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->buf_list[i].buf, eq->buf_list[i].map); eq 4192 drivers/infiniband/hw/hns/hns_roce_hw_v1.c kfree(eq->buf_list); eq 4221 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_eq *eq) eq 4223 drivers/infiniband/hw/hns/hns_roce_hw_v1.c void __iomem *eqc = hr_dev->eq_table.eqc_base[eq->eqn]; eq 4236 drivers/infiniband/hw/hns/hns_roce_hw_v1.c num_bas = (PAGE_ALIGN(eq->entries * eq->eqe_size) + eq 4239 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if ((eq->entries * eq->eqe_size) > HNS_ROCE_BA_SIZE) { eq 4241 drivers/infiniband/hw/hns/hns_roce_hw_v1.c (eq->entries * eq->eqe_size), HNS_ROCE_BA_SIZE, eq 4246 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->buf_list = kcalloc(num_bas, sizeof(*eq->buf_list), GFP_KERNEL); eq 4247 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (!eq->buf_list) eq 4251 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->buf_list[i].buf = dma_alloc_coherent(dev, HNS_ROCE_BA_SIZE, eq 4254 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (!eq->buf_list[i].buf) { eq 4259 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->buf_list[i].map = tmp_dma_addr; eq 4261 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->cons_index = 0; eq 4267 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->log_entries); eq 4272 drivers/infiniband/hw/hns/hns_roce_hw_v1.c writel((u32)(eq->buf_list[0].map >> 12), eqc + 4); eq 4282 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->buf_list[0].map >> 44); eq 4298 drivers/infiniband/hw/hns/hns_roce_hw_v1.c dma_free_coherent(dev, HNS_ROCE_BA_SIZE, eq->buf_list[i].buf, eq 4299 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->buf_list[i].map); eq 4301 drivers/infiniband/hw/hns/hns_roce_hw_v1.c kfree(eq->buf_list); eq 4309 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_eq *eq; eq 4318 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq_table->eq = kcalloc(eq_num, sizeof(*eq_table->eq), GFP_KERNEL); eq 4319 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (!eq_table->eq) eq 4330 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq = &eq_table->eq[i]; eq 4331 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->hr_dev = hr_dev; eq 4332 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->eqn = i; eq 4333 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->irq = hr_dev->irq[i]; eq 4334 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->log_page_size = PAGE_SHIFT; eq 4341 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->type_flag = HNS_ROCE_CEQ; eq 4342 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->doorbell = hr_dev->reg_base + eq 4345 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->entries = hr_dev->caps.ceqe_depth; eq 4346 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->log_entries = ilog2(eq->entries); eq 4347 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->eqe_size = HNS_ROCE_CEQ_ENTRY_SIZE; eq 4352 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->type_flag = HNS_ROCE_AEQ; eq 4353 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->doorbell = hr_dev->reg_base + eq 4355 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->entries = hr_dev->caps.aeqe_depth; eq 4356 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->log_entries = ilog2(eq->entries); eq 4357 drivers/infiniband/hw/hns/hns_roce_hw_v1.c eq->eqe_size = HNS_ROCE_AEQ_ENTRY_SIZE; eq 4373 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ret = hns_roce_v1_create_eq(hr_dev, &eq_table->eq[i]); eq 4385 drivers/infiniband/hw/hns/hns_roce_hw_v1.c &eq_table->eq[j]); eq 4404 drivers/infiniband/hw/hns/hns_roce_hw_v1.c free_irq(hr_dev->irq[j], &eq_table->eq[j]); eq 4408 drivers/infiniband/hw/hns/hns_roce_hw_v1.c hns_roce_v1_free_eq(hr_dev, &eq_table->eq[i]); eq 4413 drivers/infiniband/hw/hns/hns_roce_hw_v1.c kfree(eq_table->eq); eq 4431 drivers/infiniband/hw/hns/hns_roce_hw_v1.c free_irq(hr_dev->irq[i], &eq_table->eq[i]); eq 4433 drivers/infiniband/hw/hns/hns_roce_hw_v1.c hns_roce_v1_free_eq(hr_dev, &eq_table->eq[i]); eq 4439 drivers/infiniband/hw/hns/hns_roce_hw_v1.c kfree(eq_table->eq); eq 4931 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_eq *eq, eq 4944 drivers/infiniband/hw/hns/hns_roce_hw_v2.c irq_work->event_type = eq->event_type; eq 4945 drivers/infiniband/hw/hns/hns_roce_hw_v2.c irq_work->sub_type = eq->sub_type; eq 4949 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static void set_eq_cons_index_v2(struct hns_roce_eq *eq) eq 4951 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_dev *hr_dev = eq->hr_dev; eq 4957 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (eq->type_flag == HNS_ROCE_AEQ) { eq 4960 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->arm_st == HNS_ROCE_V2_EQ_ALWAYS_ARMED ? eq 4965 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_V2_EQ_DB_TAG_S, eq->eqn); eq 4969 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->arm_st == HNS_ROCE_V2_EQ_ALWAYS_ARMED ? eq 4976 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (eq->cons_index & HNS_ROCE_V2_CONS_IDX_M)); eq 4978 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hns_roce_write64(hr_dev, doorbell, eq->doorbell); eq 4981 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_aeqe *get_aeqe_v2(struct hns_roce_eq *eq, u32 entry) eq 4986 drivers/infiniband/hw/hns/hns_roce_hw_v2.c buf_chk_sz = 1 << (eq->eqe_buf_pg_sz + PAGE_SHIFT); eq 4987 drivers/infiniband/hw/hns/hns_roce_hw_v2.c off = (entry & (eq->entries - 1)) * HNS_ROCE_AEQ_ENTRY_SIZE; eq 4989 drivers/infiniband/hw/hns/hns_roce_hw_v2.c return (struct hns_roce_aeqe *)((char *)(eq->buf_list->buf) + eq 4993 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_aeqe *mhop_get_aeqe(struct hns_roce_eq *eq, u32 entry) eq 4998 drivers/infiniband/hw/hns/hns_roce_hw_v2.c buf_chk_sz = 1 << (eq->eqe_buf_pg_sz + PAGE_SHIFT); eq 5000 drivers/infiniband/hw/hns/hns_roce_hw_v2.c off = (entry & (eq->entries - 1)) * HNS_ROCE_AEQ_ENTRY_SIZE; eq 5002 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (eq->hop_num == HNS_ROCE_HOP_NUM_0) eq 5003 drivers/infiniband/hw/hns/hns_roce_hw_v2.c return (struct hns_roce_aeqe *)((u8 *)(eq->bt_l0) + eq 5007 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (eq->buf[off / buf_chk_sz]) + off % buf_chk_sz); eq 5010 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_aeqe *next_aeqe_sw_v2(struct hns_roce_eq *eq) eq 5014 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->hop_num) eq 5015 drivers/infiniband/hw/hns/hns_roce_hw_v2.c aeqe = get_aeqe_v2(eq, eq->cons_index); eq 5017 drivers/infiniband/hw/hns/hns_roce_hw_v2.c aeqe = mhop_get_aeqe(eq, eq->cons_index); eq 5020 drivers/infiniband/hw/hns/hns_roce_hw_v2.c !!(eq->cons_index & eq->entries)) ? aeqe : NULL; eq 5024 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_eq *eq) eq 5027 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_aeqe *aeqe = next_aeqe_sw_v2(eq); eq 5090 drivers/infiniband/hw/hns/hns_roce_hw_v2.c event_type, eq->eqn, eq->cons_index); eq 5094 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->event_type = event_type; eq 5095 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->sub_type = sub_type; eq 5096 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ++eq->cons_index; eq 5099 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (eq->cons_index > (2 * eq->entries - 1)) eq 5100 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->cons_index = 0; eq 5102 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hns_roce_v2_init_irq_work(hr_dev, eq, qpn, cqn); eq 5104 drivers/infiniband/hw/hns/hns_roce_hw_v2.c aeqe = next_aeqe_sw_v2(eq); eq 5107 drivers/infiniband/hw/hns/hns_roce_hw_v2.c set_eq_cons_index_v2(eq); eq 5111 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_ceqe *get_ceqe_v2(struct hns_roce_eq *eq, u32 entry) eq 5116 drivers/infiniband/hw/hns/hns_roce_hw_v2.c buf_chk_sz = 1 << (eq->eqe_buf_pg_sz + PAGE_SHIFT); eq 5117 drivers/infiniband/hw/hns/hns_roce_hw_v2.c off = (entry & (eq->entries - 1)) * HNS_ROCE_CEQ_ENTRY_SIZE; eq 5119 drivers/infiniband/hw/hns/hns_roce_hw_v2.c return (struct hns_roce_ceqe *)((char *)(eq->buf_list->buf) + eq 5123 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_ceqe *mhop_get_ceqe(struct hns_roce_eq *eq, u32 entry) eq 5128 drivers/infiniband/hw/hns/hns_roce_hw_v2.c buf_chk_sz = 1 << (eq->eqe_buf_pg_sz + PAGE_SHIFT); eq 5130 drivers/infiniband/hw/hns/hns_roce_hw_v2.c off = (entry & (eq->entries - 1)) * HNS_ROCE_CEQ_ENTRY_SIZE; eq 5132 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (eq->hop_num == HNS_ROCE_HOP_NUM_0) eq 5133 drivers/infiniband/hw/hns/hns_roce_hw_v2.c return (struct hns_roce_ceqe *)((u8 *)(eq->bt_l0) + eq 5136 drivers/infiniband/hw/hns/hns_roce_hw_v2.c return (struct hns_roce_ceqe *)((u8 *)(eq->buf[off / eq 5140 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_ceqe *next_ceqe_sw_v2(struct hns_roce_eq *eq) eq 5144 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->hop_num) eq 5145 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ceqe = get_ceqe_v2(eq, eq->cons_index); eq 5147 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ceqe = mhop_get_ceqe(eq, eq->cons_index); eq 5150 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (!!(eq->cons_index & eq->entries)) ? ceqe : NULL; eq 5154 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_eq *eq) eq 5157 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_ceqe *ceqe = next_ceqe_sw_v2(eq); eq 5173 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ++eq->cons_index; eq 5176 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (eq->cons_index > (EQ_DEPTH_COEFF * eq->entries - 1)) { eq 5178 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->cons_index = 0; eq 5181 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ceqe = next_ceqe_sw_v2(eq); eq 5184 drivers/infiniband/hw/hns/hns_roce_hw_v2.c set_eq_cons_index_v2(eq); eq 5191 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_eq *eq = eq_ptr; eq 5192 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_dev *hr_dev = eq->hr_dev; eq 5195 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (eq->type_flag == HNS_ROCE_CEQ) eq 5197 drivers/infiniband/hw/hns/hns_roce_hw_v2.c int_work = hns_roce_v2_ceq_int(hr_dev, eq); eq 5200 drivers/infiniband/hw/hns/hns_roce_hw_v2.c int_work = hns_roce_v2_aeq_int(hr_dev, eq); eq 5310 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_eq *eq) eq 5327 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, (unsigned int)(eq->entries * eq 5328 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eqe_size), eq->bt_l0, eq->l0_dma); eq 5332 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, bt_chk_sz, eq->bt_l0, eq->l0_dma); eq 5334 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (i = 0; i < eq->l0_last_num; i++) { eq 5335 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (i == eq->l0_last_num - 1) { eq 5336 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eqe_alloc = i * (buf_chk_sz / eq->eqe_size); eq 5337 drivers/infiniband/hw/hns/hns_roce_hw_v2.c size = (eq->entries - eqe_alloc) * eq->eqe_size; eq 5338 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, size, eq->buf[i], eq 5339 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[i]); eq 5342 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, buf_chk_sz, eq->buf[i], eq 5343 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[i]); eq 5346 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (i = 0; i < eq->l0_last_num; i++) { eq 5347 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, bt_chk_sz, eq->bt_l1[i], eq 5348 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->l1_dma[i]); eq 5352 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if ((i == eq->l0_last_num - 1) eq 5353 drivers/infiniband/hw/hns/hns_roce_hw_v2.c && j == eq->l1_last_num - 1) { eq 5354 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eqe_alloc = (buf_chk_sz / eq->eqe_size) eq 5356 drivers/infiniband/hw/hns/hns_roce_hw_v2.c size = (eq->entries - eqe_alloc) eq 5357 drivers/infiniband/hw/hns/hns_roce_hw_v2.c * eq->eqe_size; eq 5359 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf[idx], eq 5360 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[idx]); eq 5363 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, buf_chk_sz, eq->buf[idx], eq 5364 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[idx]); eq 5368 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq->buf_dma); eq 5369 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq->buf); eq 5370 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq->l1_dma); eq 5371 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq->bt_l1); eq 5372 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma = NULL; eq 5373 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf = NULL; eq 5374 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->l1_dma = NULL; eq 5375 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->bt_l1 = NULL; eq 5379 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_eq *eq) eq 5383 drivers/infiniband/hw/hns/hns_roce_hw_v2.c buf_chk_sz = 1 << (eq->eqe_buf_pg_sz + PAGE_SHIFT); eq 5386 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hns_roce_mhop_free_eq(hr_dev, eq); eq 5390 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(hr_dev->dev, buf_chk_sz, eq->buf_list->buf, eq 5391 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_list->map); eq 5392 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq->buf_list); eq 5396 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_eq *eq, eq 5405 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->doorbell = hr_dev->reg_base + ROCEE_VF_EQ_DB_CFG0_REG; eq 5406 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->hop_num = hr_dev->caps.eqe_hop_num; eq 5407 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->cons_index = 0; eq 5408 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->over_ignore = HNS_ROCE_V2_EQ_OVER_IGNORE_0; eq 5409 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->coalesce = HNS_ROCE_V2_EQ_COALESCE_0; eq 5410 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->arm_st = HNS_ROCE_V2_EQ_ALWAYS_ARMED; eq 5411 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eqe_ba_pg_sz = hr_dev->caps.eqe_ba_pg_sz; eq 5412 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eqe_buf_pg_sz = hr_dev->caps.eqe_buf_pg_sz; eq 5413 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->shift = ilog2((unsigned int)eq->entries); eq 5415 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->hop_num) eq 5416 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eqe_ba = eq->buf_list->map; eq 5418 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eqe_ba = eq->l0_dma; eq 5429 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_HOP_NUM_S, eq->hop_num); eq 5434 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_OVER_IGNORE_S, eq->over_ignore); eq 5439 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_COALESCE_S, eq->coalesce); eq 5444 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_ARM_ST_S, eq->arm_st); eq 5449 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_EQN_S, eq->eqn); eq 5461 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eqe_ba_pg_sz + PG_SHIFT_OFFSET); eq 5467 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eqe_buf_pg_sz + PG_SHIFT_OFFSET); eq 5478 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_MAX_CNT_S, eq->eq_max_cnt); eq 5483 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_PERIOD_S, eq->eq_period); eq 5494 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_EQE_BA_L_S, eq->eqe_ba >> 3); eq 5499 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_EQE_BA_H_S, eq->eqe_ba >> 35); eq 5504 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_SHIFT_S, eq->shift); eq 5515 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_CUR_EQE_BA_L_S, eq->cur_eqe_ba >> 12); eq 5520 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_CUR_EQE_BA_M_S, eq->cur_eqe_ba >> 28); eq 5525 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_CUR_EQE_BA_H_S, eq->cur_eqe_ba >> 60); eq 5536 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_NXT_EQE_BA_L_S, eq->nxt_eqe_ba >> 12); eq 5541 drivers/infiniband/hw/hns/hns_roce_hw_v2.c HNS_ROCE_EQC_NXT_EQE_BA_H_S, eq->nxt_eqe_ba >> 44); eq 5545 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_eq *eq) eq 5567 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ba_num = DIV_ROUND_UP(PAGE_ALIGN(eq->entries * eq->eqe_size), eq 5572 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (eq->entries > buf_chk_sz / eq->eqe_size) { eq 5574 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->entries); eq 5577 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->bt_l0 = dma_alloc_coherent(dev, eq->entries * eq->eqe_size, eq 5578 drivers/infiniband/hw/hns/hns_roce_hw_v2.c &(eq->l0_dma), GFP_KERNEL); eq 5579 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->bt_l0) eq 5582 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->cur_eqe_ba = eq->l0_dma; eq 5583 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->nxt_eqe_ba = 0; eq 5588 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma = kcalloc(ba_num, sizeof(*eq->buf_dma), GFP_KERNEL); eq 5589 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->buf_dma) eq 5591 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf = kcalloc(ba_num, sizeof(*eq->buf), GFP_KERNEL); eq 5592 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->buf) eq 5596 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->l1_dma = kcalloc(bt_num, sizeof(*eq->l1_dma), GFP_KERNEL); eq 5597 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->l1_dma) eq 5600 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->bt_l1 = kcalloc(bt_num, sizeof(*eq->bt_l1), GFP_KERNEL); eq 5601 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->bt_l1) eq 5606 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->bt_l0 = dma_alloc_coherent(dev, bt_chk_sz, &eq->l0_dma, GFP_KERNEL); eq 5607 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->bt_l0) eq 5620 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eqe_alloc = i * (buf_chk_sz / eq->eqe_size); eq 5621 drivers/infiniband/hw/hns/hns_roce_hw_v2.c size = (eq->entries - eqe_alloc) * eq->eqe_size; eq 5623 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf[i] = dma_alloc_coherent(dev, size, eq 5624 drivers/infiniband/hw/hns/hns_roce_hw_v2.c &(eq->buf_dma[i]), eq 5626 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->buf[i]) eq 5629 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *(eq->bt_l0 + i) = eq->buf_dma[i]; eq 5635 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->cur_eqe_ba = eq->buf_dma[0]; eq 5637 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->nxt_eqe_ba = eq->buf_dma[1]; eq 5642 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->bt_l1[i] = dma_alloc_coherent(dev, bt_chk_sz, eq 5643 drivers/infiniband/hw/hns/hns_roce_hw_v2.c &(eq->l1_dma[i]), eq 5645 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->bt_l1[i]) eq 5647 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *(eq->bt_l0 + i) = eq->l1_dma[i]; eq 5654 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eqe_alloc = (buf_chk_sz / eq->eqe_size) eq 5656 drivers/infiniband/hw/hns/hns_roce_hw_v2.c size = (eq->entries - eqe_alloc) eq 5657 drivers/infiniband/hw/hns/hns_roce_hw_v2.c * eq->eqe_size; eq 5659 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf[idx] = dma_alloc_coherent(dev, size, eq 5660 drivers/infiniband/hw/hns/hns_roce_hw_v2.c &(eq->buf_dma[idx]), eq 5662 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->buf[idx]) eq 5665 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *(eq->bt_l1[i] + j) = eq->buf_dma[idx]; eq 5677 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->cur_eqe_ba = eq->buf_dma[0]; eq 5679 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->nxt_eqe_ba = eq->buf_dma[1]; eq 5682 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->l0_last_num = i + 1; eq 5684 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->l1_last_num = j + 1; eq 5689 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, bt_chk_sz, eq->bt_l0, eq->l0_dma); eq 5690 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->bt_l0 = NULL; eq 5691 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->l0_dma = 0; eq 5693 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, bt_chk_sz, eq->bt_l1[i], eq 5694 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->l1_dma[i]); eq 5698 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, buf_chk_sz, eq->buf[idx], eq 5699 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[idx]); eq 5705 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, bt_chk_sz, eq->bt_l0, eq->l0_dma); eq 5706 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->bt_l0 = NULL; eq 5707 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->l0_dma = 0; eq 5711 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, buf_chk_sz, eq->buf[i], eq 5712 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[i]); eq 5717 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, bt_chk_sz, eq->bt_l1[i], eq 5718 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->l1_dma[i]); eq 5726 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf[idx], eq 5727 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[idx]); eq 5733 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq->bt_l1); eq 5734 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->bt_l1 = NULL; eq 5737 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq->l1_dma); eq 5738 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->l1_dma = NULL; eq 5741 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq->buf); eq 5742 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf = NULL; eq 5745 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq->buf_dma); eq 5746 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma = NULL; eq 5752 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_eq *eq, eq 5768 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_list = kzalloc(sizeof(struct hns_roce_buf_list), eq 5770 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->buf_list) { eq 5775 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_list->buf = dma_alloc_coherent(dev, buf_chk_sz, eq 5776 drivers/infiniband/hw/hns/hns_roce_hw_v2.c &(eq->buf_list->map), eq 5778 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->buf_list->buf) { eq 5784 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ret = hns_roce_mhop_alloc_eq(hr_dev, eq); eq 5791 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hns_roce_config_eqc(hr_dev, eq, mailbox->buf); eq 5793 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ret = hns_roce_cmd_mbox(hr_dev, mailbox->dma, 0, eq->eqn, 0, eq 5806 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, buf_chk_sz, eq->buf_list->buf, eq 5807 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_list->map); eq 5809 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hns_roce_mhop_free_eq(hr_dev, eq); eq 5814 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq->buf_list); eq 5860 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ret = request_irq(eq_table->eq[j - other_num].irq, eq 5863 drivers/infiniband/hw/hns/hns_roce_hw_v2.c &eq_table->eq[j - other_num]); eq 5865 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ret = request_irq(eq_table->eq[j - other_num].irq, eq 5868 drivers/infiniband/hw/hns/hns_roce_hw_v2.c &eq_table->eq[j - other_num]); eq 5882 drivers/infiniband/hw/hns/hns_roce_hw_v2.c free_irq(eq_table->eq[j - other_num].irq, eq 5883 drivers/infiniband/hw/hns/hns_roce_hw_v2.c &eq_table->eq[j - other_num]); eq 5905 drivers/infiniband/hw/hns/hns_roce_hw_v2.c free_irq(hr_dev->eq_table.eq[i].irq, &hr_dev->eq_table.eq[i]); eq 5915 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_eq *eq; eq 5932 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq_table->eq = kcalloc(eq_num, sizeof(*eq_table->eq), GFP_KERNEL); eq 5933 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq_table->eq) eq 5938 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq = &eq_table->eq[i]; eq 5939 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->hr_dev = hr_dev; eq 5940 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eqn = i; eq 5944 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->type_flag = HNS_ROCE_CEQ; eq 5945 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->entries = hr_dev->caps.ceqe_depth; eq 5946 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eqe_size = HNS_ROCE_CEQ_ENTRY_SIZE; eq 5947 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->irq = hr_dev->irq[i + other_num + aeq_num]; eq 5948 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eq_max_cnt = HNS_ROCE_CEQ_DEFAULT_BURST_NUM; eq 5949 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eq_period = HNS_ROCE_CEQ_DEFAULT_INTERVAL; eq 5953 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->type_flag = HNS_ROCE_AEQ; eq 5954 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->entries = hr_dev->caps.aeqe_depth; eq 5955 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eqe_size = HNS_ROCE_AEQ_ENTRY_SIZE; eq 5956 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->irq = hr_dev->irq[i - comp_num + other_num]; eq 5957 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eq_max_cnt = HNS_ROCE_AEQ_DEFAULT_BURST_NUM; eq 5958 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->eq_period = HNS_ROCE_AEQ_DEFAULT_INTERVAL; eq 5961 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ret = hns_roce_v2_create_eq(hr_dev, eq, eq_cmd); eq 5996 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hns_roce_v2_free_eq(hr_dev, &eq_table->eq[i]); eq 5997 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq_table->eq); eq 6018 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hns_roce_v2_free_eq(hr_dev, &eq_table->eq[i]); eq 6021 drivers/infiniband/hw/hns/hns_roce_hw_v2.c kfree(eq_table->eq); eq 2436 drivers/infiniband/hw/mlx4/main.c int i, j, eq = 0, total_eqs = 0; eq 2448 drivers/infiniband/hw/mlx4/main.c ibdev->eq_table[eq] = total_eqs; eq 2450 drivers/infiniband/hw/mlx4/main.c &ibdev->eq_table[eq])) eq 2451 drivers/infiniband/hw/mlx4/main.c eq++; eq 2453 drivers/infiniband/hw/mlx4/main.c ibdev->eq_table[eq] = -1; eq 2457 drivers/infiniband/hw/mlx4/main.c for (i = eq; i < dev->caps.num_comp_vectors; eq 2462 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.num_comp_vectors = eq; eq 76 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_pf_eq *eq; eq 1355 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_pf_eq *eq = pfault->eq; eq 1357 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_pfault(eq->dev, pfault); eq 1358 drivers/infiniband/hw/mlx5/odp.c mempool_free(pfault, eq->pool); eq 1361 drivers/infiniband/hw/mlx5/odp.c static void mlx5_ib_eq_pf_process(struct mlx5_ib_pf_eq *eq) eq 1368 drivers/infiniband/hw/mlx5/odp.c while ((eqe = mlx5_eq_get_eqe(eq->core, cc))) { eq 1369 drivers/infiniband/hw/mlx5/odp.c pfault = mempool_alloc(eq->pool, GFP_ATOMIC); eq 1371 drivers/infiniband/hw/mlx5/odp.c schedule_work(&eq->work); eq 1379 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_dbg(eq->dev, eq 1399 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_dbg(eq->dev, eq 1403 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_dbg(eq->dev, eq 1422 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_dbg(eq->dev, eq 1430 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_warn(eq->dev, eq 1438 drivers/infiniband/hw/mlx5/odp.c pfault->eq = eq; eq 1440 drivers/infiniband/hw/mlx5/odp.c queue_work(eq->wq, &pfault->work); eq 1442 drivers/infiniband/hw/mlx5/odp.c cc = mlx5_eq_update_cc(eq->core, ++cc); eq 1445 drivers/infiniband/hw/mlx5/odp.c mlx5_eq_update_ci(eq->core, cc, 1); eq 1451 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_pf_eq *eq = eq 1455 drivers/infiniband/hw/mlx5/odp.c if (spin_trylock_irqsave(&eq->lock, flags)) { eq 1456 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_eq_pf_process(eq); eq 1457 drivers/infiniband/hw/mlx5/odp.c spin_unlock_irqrestore(&eq->lock, flags); eq 1459 drivers/infiniband/hw/mlx5/odp.c schedule_work(&eq->work); eq 1477 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_pf_eq *eq = eq 1480 drivers/infiniband/hw/mlx5/odp.c mempool_refill(eq->pool); eq 1482 drivers/infiniband/hw/mlx5/odp.c spin_lock_irq(&eq->lock); eq 1483 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_eq_pf_process(eq); eq 1484 drivers/infiniband/hw/mlx5/odp.c spin_unlock_irq(&eq->lock); eq 1493 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_create_pf_eq(struct mlx5_ib_dev *dev, struct mlx5_ib_pf_eq *eq) eq 1498 drivers/infiniband/hw/mlx5/odp.c INIT_WORK(&eq->work, mlx5_ib_eq_pf_action); eq 1499 drivers/infiniband/hw/mlx5/odp.c spin_lock_init(&eq->lock); eq 1500 drivers/infiniband/hw/mlx5/odp.c eq->dev = dev; eq 1502 drivers/infiniband/hw/mlx5/odp.c eq->pool = mempool_create_kmalloc_pool(MLX5_IB_NUM_PF_DRAIN, eq 1504 drivers/infiniband/hw/mlx5/odp.c if (!eq->pool) eq 1507 drivers/infiniband/hw/mlx5/odp.c eq->wq = alloc_workqueue("mlx5_ib_page_fault", eq 1510 drivers/infiniband/hw/mlx5/odp.c if (!eq->wq) { eq 1515 drivers/infiniband/hw/mlx5/odp.c eq->irq_nb.notifier_call = mlx5_ib_eq_pf_int; eq 1521 drivers/infiniband/hw/mlx5/odp.c eq->core = mlx5_eq_create_generic(dev->mdev, ¶m); eq 1522 drivers/infiniband/hw/mlx5/odp.c if (IS_ERR(eq->core)) { eq 1523 drivers/infiniband/hw/mlx5/odp.c err = PTR_ERR(eq->core); eq 1526 drivers/infiniband/hw/mlx5/odp.c err = mlx5_eq_enable(dev->mdev, eq->core, &eq->irq_nb); eq 1534 drivers/infiniband/hw/mlx5/odp.c mlx5_eq_destroy_generic(dev->mdev, eq->core); eq 1536 drivers/infiniband/hw/mlx5/odp.c destroy_workqueue(eq->wq); eq 1538 drivers/infiniband/hw/mlx5/odp.c mempool_destroy(eq->pool); eq 1543 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_destroy_pf_eq(struct mlx5_ib_dev *dev, struct mlx5_ib_pf_eq *eq) eq 1547 drivers/infiniband/hw/mlx5/odp.c mlx5_eq_disable(dev->mdev, eq->core, &eq->irq_nb); eq 1548 drivers/infiniband/hw/mlx5/odp.c err = mlx5_eq_destroy_generic(dev->mdev, eq->core); eq 1549 drivers/infiniband/hw/mlx5/odp.c cancel_work_sync(&eq->work); eq 1550 drivers/infiniband/hw/mlx5/odp.c destroy_workqueue(eq->wq); eq 1551 drivers/infiniband/hw/mlx5/odp.c mempool_destroy(eq->pool); eq 831 drivers/infiniband/hw/mthca/mthca_cq.c cq_context->error_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn); eq 832 drivers/infiniband/hw/mthca/mthca_cq.c cq_context->comp_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_COMP].eqn); eq 932 drivers/infiniband/hw/mthca/mthca_cq.c synchronize_irq(dev->eq_table.eq[MTHCA_EQ_COMP].msi_x_vector); eq 231 drivers/infiniband/hw/mthca/mthca_dev.h struct mthca_eq eq[MTHCA_NUM_EQ]; eq 173 drivers/infiniband/hw/mthca/mthca_eq.c static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) eq 184 drivers/infiniband/hw/mthca/mthca_eq.c mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), eq 189 drivers/infiniband/hw/mthca/mthca_eq.c static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) eq 194 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8); eq 199 drivers/infiniband/hw/mthca/mthca_eq.c static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) eq 202 drivers/infiniband/hw/mthca/mthca_eq.c arbel_set_eq_ci(dev, eq, ci); eq 204 drivers/infiniband/hw/mthca/mthca_eq.c tavor_set_eq_ci(dev, eq, ci); eq 228 drivers/infiniband/hw/mthca/mthca_eq.c static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry) eq 230 drivers/infiniband/hw/mthca/mthca_eq.c unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; eq 231 drivers/infiniband/hw/mthca/mthca_eq.c return eq->page_list[off / PAGE_SIZE].buf + off % PAGE_SIZE; eq 234 drivers/infiniband/hw/mthca/mthca_eq.c static inline struct mthca_eqe *next_eqe_sw(struct mthca_eq *eq) eq 237 drivers/infiniband/hw/mthca/mthca_eq.c eqe = get_eqe(eq, eq->cons_index); eq 260 drivers/infiniband/hw/mthca/mthca_eq.c static int mthca_eq_int(struct mthca_dev *dev, struct mthca_eq *eq) eq 267 drivers/infiniband/hw/mthca/mthca_eq.c while ((eqe = next_eqe_sw(eq))) { eq 277 drivers/infiniband/hw/mthca/mthca_eq.c disarm_cq(dev, eq->eqn, disarm_cqn); eq 349 drivers/infiniband/hw/mthca/mthca_eq.c mthca_warn(dev, "EQ overrun on EQN %d\n", eq->eqn); eq 358 drivers/infiniband/hw/mthca/mthca_eq.c eqe->type, eqe->subtype, eq->eqn); eq 363 drivers/infiniband/hw/mthca/mthca_eq.c ++eq->cons_index; eq 379 drivers/infiniband/hw/mthca/mthca_eq.c set_eq_ci(dev, eq, eq->cons_index); eq 408 drivers/infiniband/hw/mthca/mthca_eq.c if (ecr & dev->eq_table.eq[i].eqn_mask) { eq 409 drivers/infiniband/hw/mthca/mthca_eq.c if (mthca_eq_int(dev, &dev->eq_table.eq[i])) eq 410 drivers/infiniband/hw/mthca/mthca_eq.c tavor_set_eq_ci(dev, &dev->eq_table.eq[i], eq 411 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_table.eq[i].cons_index); eq 412 drivers/infiniband/hw/mthca/mthca_eq.c tavor_eq_req_not(dev, dev->eq_table.eq[i].eqn); eq 420 drivers/infiniband/hw/mthca/mthca_eq.c struct mthca_eq *eq = eq_ptr; eq 421 drivers/infiniband/hw/mthca/mthca_eq.c struct mthca_dev *dev = eq->dev; eq 423 drivers/infiniband/hw/mthca/mthca_eq.c mthca_eq_int(dev, eq); eq 424 drivers/infiniband/hw/mthca/mthca_eq.c tavor_set_eq_ci(dev, eq, eq->cons_index); eq 425 drivers/infiniband/hw/mthca/mthca_eq.c tavor_eq_req_not(dev, eq->eqn); eq 441 drivers/infiniband/hw/mthca/mthca_eq.c if (mthca_eq_int(dev, &dev->eq_table.eq[i])) { eq 443 drivers/infiniband/hw/mthca/mthca_eq.c arbel_set_eq_ci(dev, &dev->eq_table.eq[i], eq 444 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_table.eq[i].cons_index); eq 454 drivers/infiniband/hw/mthca/mthca_eq.c struct mthca_eq *eq = eq_ptr; eq 455 drivers/infiniband/hw/mthca/mthca_eq.c struct mthca_dev *dev = eq->dev; eq 457 drivers/infiniband/hw/mthca/mthca_eq.c mthca_eq_int(dev, eq); eq 458 drivers/infiniband/hw/mthca/mthca_eq.c arbel_set_eq_ci(dev, eq, eq->cons_index); eq 459 drivers/infiniband/hw/mthca/mthca_eq.c arbel_eq_req_not(dev, eq->eqn_mask); eq 468 drivers/infiniband/hw/mthca/mthca_eq.c struct mthca_eq *eq) eq 478 drivers/infiniband/hw/mthca/mthca_eq.c eq->dev = dev; eq 479 drivers/infiniband/hw/mthca/mthca_eq.c eq->nent = roundup_pow_of_two(max(nent, 2)); eq 480 drivers/infiniband/hw/mthca/mthca_eq.c npages = ALIGN(eq->nent * MTHCA_EQ_ENTRY_SIZE, PAGE_SIZE) / PAGE_SIZE; eq 482 drivers/infiniband/hw/mthca/mthca_eq.c eq->page_list = kmalloc_array(npages, sizeof(*eq->page_list), eq 484 drivers/infiniband/hw/mthca/mthca_eq.c if (!eq->page_list) eq 488 drivers/infiniband/hw/mthca/mthca_eq.c eq->page_list[i].buf = NULL; eq 500 drivers/infiniband/hw/mthca/mthca_eq.c eq->page_list[i].buf = dma_alloc_coherent(&dev->pdev->dev, eq 502 drivers/infiniband/hw/mthca/mthca_eq.c if (!eq->page_list[i].buf) eq 506 drivers/infiniband/hw/mthca/mthca_eq.c dma_unmap_addr_set(&eq->page_list[i], mapping, t); eq 508 drivers/infiniband/hw/mthca/mthca_eq.c clear_page(eq->page_list[i].buf); eq 511 drivers/infiniband/hw/mthca/mthca_eq.c for (i = 0; i < eq->nent; ++i) eq 512 drivers/infiniband/hw/mthca/mthca_eq.c set_eqe_hw(get_eqe(eq, i)); eq 514 drivers/infiniband/hw/mthca/mthca_eq.c eq->eqn = mthca_alloc(&dev->eq_table.alloc); eq 515 drivers/infiniband/hw/mthca/mthca_eq.c if (eq->eqn == -1) eq 523 drivers/infiniband/hw/mthca/mthca_eq.c &eq->mr); eq 535 drivers/infiniband/hw/mthca/mthca_eq.c eq_context->logsize_usrpage = cpu_to_be32((ffs(eq->nent) - 1) << 24); eq 543 drivers/infiniband/hw/mthca/mthca_eq.c eq_context->lkey = cpu_to_be32(eq->mr.ibmr.lkey); eq 545 drivers/infiniband/hw/mthca/mthca_eq.c err = mthca_SW2HW_EQ(dev, mailbox, eq->eqn); eq 554 drivers/infiniband/hw/mthca/mthca_eq.c eq->eqn_mask = swab32(1 << eq->eqn); eq 555 drivers/infiniband/hw/mthca/mthca_eq.c eq->cons_index = 0; eq 557 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_table.arm_mask |= eq->eqn_mask; eq 560 drivers/infiniband/hw/mthca/mthca_eq.c eq->eqn, eq->nent); eq 565 drivers/infiniband/hw/mthca/mthca_eq.c mthca_free_mr(dev, &eq->mr); eq 568 drivers/infiniband/hw/mthca/mthca_eq.c mthca_free(&dev->eq_table.alloc, eq->eqn); eq 572 drivers/infiniband/hw/mthca/mthca_eq.c if (eq->page_list[i].buf) eq 574 drivers/infiniband/hw/mthca/mthca_eq.c eq->page_list[i].buf, eq 575 drivers/infiniband/hw/mthca/mthca_eq.c dma_unmap_addr(&eq->page_list[i], eq 581 drivers/infiniband/hw/mthca/mthca_eq.c kfree(eq->page_list); eq 589 drivers/infiniband/hw/mthca/mthca_eq.c struct mthca_eq *eq) eq 593 drivers/infiniband/hw/mthca/mthca_eq.c int npages = (eq->nent * MTHCA_EQ_ENTRY_SIZE + PAGE_SIZE - 1) / eq 601 drivers/infiniband/hw/mthca/mthca_eq.c err = mthca_HW2SW_EQ(dev, mailbox, eq->eqn); eq 605 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_table.arm_mask &= ~eq->eqn_mask; eq 608 drivers/infiniband/hw/mthca/mthca_eq.c mthca_dbg(dev, "Dumping EQ context %02x:\n", eq->eqn); eq 618 drivers/infiniband/hw/mthca/mthca_eq.c mthca_free_mr(dev, &eq->mr); eq 621 drivers/infiniband/hw/mthca/mthca_eq.c eq->page_list[i].buf, eq 622 drivers/infiniband/hw/mthca/mthca_eq.c dma_unmap_addr(&eq->page_list[i], mapping)); eq 624 drivers/infiniband/hw/mthca/mthca_eq.c kfree(eq->page_list); eq 635 drivers/infiniband/hw/mthca/mthca_eq.c if (dev->eq_table.eq[i].have_irq) { eq 636 drivers/infiniband/hw/mthca/mthca_eq.c free_irq(dev->eq_table.eq[i].msi_x_vector, eq 637 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_table.eq + i); eq 638 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_table.eq[i].have_irq = 0; eq 799 drivers/infiniband/hw/mthca/mthca_eq.c &dev->eq_table.eq[MTHCA_EQ_COMP]); eq 805 drivers/infiniband/hw/mthca/mthca_eq.c &dev->eq_table.eq[MTHCA_EQ_ASYNC]); eq 811 drivers/infiniband/hw/mthca/mthca_eq.c &dev->eq_table.eq[MTHCA_EQ_CMD]); eq 823 drivers/infiniband/hw/mthca/mthca_eq.c snprintf(dev->eq_table.eq[i].irq_name, eq 827 drivers/infiniband/hw/mthca/mthca_eq.c err = request_irq(dev->eq_table.eq[i].msi_x_vector, eq 831 drivers/infiniband/hw/mthca/mthca_eq.c 0, dev->eq_table.eq[i].irq_name, eq 832 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_table.eq + i); eq 835 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_table.eq[i].have_irq = 1; eq 838 drivers/infiniband/hw/mthca/mthca_eq.c snprintf(dev->eq_table.eq[0].irq_name, IB_DEVICE_NAME_MAX, eq 844 drivers/infiniband/hw/mthca/mthca_eq.c IRQF_SHARED, dev->eq_table.eq[0].irq_name, dev); eq 851 drivers/infiniband/hw/mthca/mthca_eq.c 0, dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn); eq 854 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn, err); eq 857 drivers/infiniband/hw/mthca/mthca_eq.c 0, dev->eq_table.eq[MTHCA_EQ_CMD].eqn); eq 860 drivers/infiniband/hw/mthca/mthca_eq.c dev->eq_table.eq[MTHCA_EQ_CMD].eqn, err); eq 864 drivers/infiniband/hw/mthca/mthca_eq.c arbel_eq_req_not(dev, dev->eq_table.eq[i].eqn_mask); eq 866 drivers/infiniband/hw/mthca/mthca_eq.c tavor_eq_req_not(dev, dev->eq_table.eq[i].eqn); eq 872 drivers/infiniband/hw/mthca/mthca_eq.c mthca_free_eq(dev, &dev->eq_table.eq[MTHCA_EQ_CMD]); eq 875 drivers/infiniband/hw/mthca/mthca_eq.c mthca_free_eq(dev, &dev->eq_table.eq[MTHCA_EQ_ASYNC]); eq 878 drivers/infiniband/hw/mthca/mthca_eq.c mthca_free_eq(dev, &dev->eq_table.eq[MTHCA_EQ_COMP]); eq 895 drivers/infiniband/hw/mthca/mthca_eq.c 1, dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn); eq 897 drivers/infiniband/hw/mthca/mthca_eq.c 1, dev->eq_table.eq[MTHCA_EQ_CMD].eqn); eq 900 drivers/infiniband/hw/mthca/mthca_eq.c mthca_free_eq(dev, &dev->eq_table.eq[i]); eq 762 drivers/infiniband/hw/mthca/mthca_main.c dev->eq_table.eq[MTHCA_EQ_CMD].msi_x_vector); eq 859 drivers/infiniband/hw/mthca/mthca_main.c mdev->eq_table.eq[MTHCA_EQ_COMP ].msi_x_vector = eq 861 drivers/infiniband/hw/mthca/mthca_main.c mdev->eq_table.eq[MTHCA_EQ_ASYNC].msi_x_vector = eq 863 drivers/infiniband/hw/mthca/mthca_main.c mdev->eq_table.eq[MTHCA_EQ_CMD ].msi_x_vector = eq 111 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static inline void *ocrdma_get_eqe(struct ocrdma_eq *eq) eq 113 drivers/infiniband/hw/ocrdma/ocrdma_hw.c return eq->q.va + (eq->q.tail * sizeof(struct ocrdma_eqe)); eq 116 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static inline void ocrdma_eq_inc_tail(struct ocrdma_eq *eq) eq 118 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->q.tail = (eq->q.tail + 1) & (OCRDMA_EQ_LEN - 1); eq 433 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static int ocrdma_mbx_create_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) eq 448 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma, eq 453 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->q.id = rsp->vector_eqid & 0xffff; eq 454 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->vector = (rsp->vector_eqid >> 16) & 0xffff; eq 455 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->q.created = true; eq 461 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_eq *eq, u16 q_len) eq 465 drivers/infiniband/hw/ocrdma/ocrdma_hw.c status = ocrdma_alloc_q(dev, &eq->q, OCRDMA_EQ_LEN, eq 470 drivers/infiniband/hw/ocrdma/ocrdma_hw.c status = ocrdma_mbx_create_eq(dev, eq); eq 473 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->dev = dev; eq 474 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_ring_eq_db(dev, eq->q.id, true, true, 0); eq 478 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_free_q(dev, &eq->q); eq 482 drivers/infiniband/hw/ocrdma/ocrdma_hw.c int ocrdma_get_irq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) eq 489 drivers/infiniband/hw/ocrdma/ocrdma_hw.c irq = dev->nic_info.msix.vector_list[eq->vector]; eq 493 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static void _ocrdma_destroy_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) eq 495 drivers/infiniband/hw/ocrdma/ocrdma_hw.c if (eq->q.created) { eq 496 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_mbx_delete_q(dev, &eq->q, QTYPE_EQ); eq 497 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_free_q(dev, &eq->q); eq 501 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static void ocrdma_destroy_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) eq 508 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_ring_eq_db(dev, eq->q.id, false, false, 0); eq 510 drivers/infiniband/hw/ocrdma/ocrdma_hw.c irq = ocrdma_get_irq(dev, eq); eq 511 drivers/infiniband/hw/ocrdma/ocrdma_hw.c free_irq(irq, eq); eq 512 drivers/infiniband/hw/ocrdma/ocrdma_hw.c _ocrdma_destroy_eq(dev, eq); eq 525 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_queue_info *eq) eq 541 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cmd->eqn = eq->id; eq 996 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_eq *eq = handle; eq 997 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_dev *dev = eq->dev; eq 1002 drivers/infiniband/hw/ocrdma/ocrdma_hw.c int budget = eq->cq_cnt; eq 1005 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ptr = ocrdma_get_eqe(eq); eq 1012 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->q.id, eqe.id_valid); eq 1018 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_ring_eq_db(dev, eq->q.id, false, true, 1); eq 1024 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_eq_inc_tail(eq); eq 1034 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->aic_obj.eq_intr_cnt++; eq 1035 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_ring_eq_db(dev, eq->q.id, true, true, 0); eq 3117 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static int ocrdma_mbx_modify_eqd(struct ocrdma_dev *dev, struct ocrdma_eq *eq, eq 3132 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cmd->cmd.set_eqd[i].eq_id = eq[i].q.id; eq 3135 drivers/infiniband/hw/ocrdma/ocrdma_hw.c (eq[i].aic_obj.prev_eqd * 65)/100; eq 3143 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static int ocrdma_modify_eqd(struct ocrdma_dev *dev, struct ocrdma_eq *eq, eq 3150 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_mbx_modify_eqd(dev, &eq[i], num_eqs); eq 3155 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_mbx_modify_eqd(dev, eq, num); eq 3164 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_eq *eq = NULL; eq 3169 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq = &dev->eq_tbl[i]; eq 3170 drivers/infiniband/hw/ocrdma/ocrdma_hw.c if (eq->aic_obj.eq_intr_cnt > eq->aic_obj.prev_eq_intr_cnt) { eq 3171 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq_intr = eq->aic_obj.eq_intr_cnt - eq 3172 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->aic_obj.prev_eq_intr_cnt; eq 3174 drivers/infiniband/hw/ocrdma/ocrdma_hw.c (eq->aic_obj.prev_eqd == EQ_AIC_MIN_EQD)) { eq 3175 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->aic_obj.prev_eqd = EQ_AIC_MAX_EQD; eq 3178 drivers/infiniband/hw/ocrdma/ocrdma_hw.c (eq->aic_obj.prev_eqd == EQ_AIC_MAX_EQD)) { eq 3179 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->aic_obj.prev_eqd = EQ_AIC_MIN_EQD; eq 3183 drivers/infiniband/hw/ocrdma/ocrdma_hw.c eq->aic_obj.prev_eq_intr_cnt = eq->aic_obj.eq_intr_cnt; eq 150 drivers/infiniband/hw/ocrdma/ocrdma_hw.h int ocrdma_get_irq(struct ocrdma_dev *dev, struct ocrdma_eq *eq); eq 1070 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_eq *eq = NULL; eq 1078 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c eq = &dev->eq_tbl[indx]; eq 1079 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c irq = ocrdma_get_irq(dev, eq); eq 315 drivers/input/input-mt.c static int adjust_dual(int *begin, int step, int *end, int eq, int mu) eq 333 drivers/input/input-mt.c if (c == 0 || (c > mu && (!eq || mu > 0))) eq 1007 drivers/misc/habanalabs/goya/goya.c struct hl_eq *eq; eq 1018 drivers/misc/habanalabs/goya/goya.c eq = &hdev->event_queue; eq 1023 drivers/misc/habanalabs/goya/goya.c WREG32(mmCPU_EQ_BASE_ADDR_LOW, lower_32_bits(eq->bus_address)); eq 1024 drivers/misc/habanalabs/goya/goya.c WREG32(mmCPU_EQ_BASE_ADDR_HIGH, upper_32_bits(eq->bus_address)); eq 153 drivers/misc/habanalabs/irq.c struct hl_eq *eq = arg; eq 154 drivers/misc/habanalabs/irq.c struct hl_device *hdev = eq->hdev; eq 159 drivers/misc/habanalabs/irq.c eq_base = (struct hl_eq_entry *) (uintptr_t) eq->kernel_address; eq 163 drivers/misc/habanalabs/irq.c ((le32_to_cpu(eq_base[eq->ci].hdr.ctl) & eq 169 drivers/misc/habanalabs/irq.c eq_entry = &eq_base[eq->ci]; eq 200 drivers/misc/habanalabs/irq.c eq->ci = hl_eq_inc_ptr(eq->ci); eq 202 drivers/misc/habanalabs/irq.c hdev->asic_funcs->update_eq_ci(hdev, eq->ci); eq 581 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge_eth_txq *eq; eq 583 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c eq = container_of(txq, struct sge_eth_txq, q); eq 584 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c t4_sge_eth_txq_egress_update(q->adap, eq, -1); eq 1324 drivers/net/ethernet/chelsio/cxgb4/sge.c int t4_sge_eth_txq_egress_update(struct adapter *adap, struct sge_eth_txq *eq, eq 1328 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_txq *q = &eq->q; eq 1331 drivers/net/ethernet/chelsio/cxgb4/sge.c if (!q->in_use || !__netif_tx_trylock(eq->txq)) eq 1335 drivers/net/ethernet/chelsio/cxgb4/sge.c reclaimed = reclaim_completed_tx(adap, &eq->q, maxreclaim, true); eq 1347 drivers/net/ethernet/chelsio/cxgb4/sge.c if (netif_tx_queue_stopped(eq->txq) && hw_in_use < (q->size / 2)) { eq 1348 drivers/net/ethernet/chelsio/cxgb4/sge.c netif_tx_wake_queue(eq->txq); eq 1349 drivers/net/ethernet/chelsio/cxgb4/sge.c eq->q.restarts++; eq 1352 drivers/net/ethernet/chelsio/cxgb4/sge.c __netif_tx_unlock(eq->txq); eq 4011 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_eth_rxq *eq; eq 4016 drivers/net/ethernet/chelsio/cxgb4/sge.c eq = &adap->sge.ethrxq[i]; eq 4017 drivers/net/ethernet/chelsio/cxgb4/sge.c if (eq->rspq.desc) eq 4020 drivers/net/ethernet/chelsio/cxgb4/sge.c eq->rspq.cntxt_id, eq 4021 drivers/net/ethernet/chelsio/cxgb4/sge.c eq->fl.size ? eq->fl.cntxt_id : 0xffff, eq 4027 drivers/net/ethernet/chelsio/cxgb4/sge.c eq = &adap->sge.ethrxq[i]; eq 4028 drivers/net/ethernet/chelsio/cxgb4/sge.c if (eq->rspq.desc) eq 4029 drivers/net/ethernet/chelsio/cxgb4/sge.c free_rspq_fl(adap, &eq->rspq, eq 4030 drivers/net/ethernet/chelsio/cxgb4/sge.c eq->fl.size ? &eq->fl : NULL); eq 1157 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_queue_info *eq, bool no_delay, int coalesce_wm) eq 1187 drivers/net/ethernet/emulex/benet/be_cmds.c AMAP_SET_BITS(struct amap_cq_context_be, eqid, ctxt, eq->id); eq 1204 drivers/net/ethernet/emulex/benet/be_cmds.c AMAP_SET_BITS(struct amap_cq_context_v2, eqid, ctxt, eq->id); eq 2397 drivers/net/ethernet/emulex/benet/be_cmds.h struct be_queue_info *eq, bool no_delay, eq 2952 drivers/net/ethernet/emulex/benet/be_main.c struct be_queue_info *eq; eq 2972 drivers/net/ethernet/emulex/benet/be_main.c eq = &eqo->q; eq 2973 drivers/net/ethernet/emulex/benet/be_main.c rc = be_queue_alloc(adapter, eq, EVNT_Q_LEN, eq 3120 drivers/net/ethernet/emulex/benet/be_main.c struct be_queue_info *eq, *cq; eq 3149 drivers/net/ethernet/emulex/benet/be_main.c eq = &adapter->eq_obj[i % adapter->num_evt_qs].q; eq 3150 drivers/net/ethernet/emulex/benet/be_main.c rc = be_cmd_cq_create(adapter, cq, eq, false, 3); eq 26 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_EQ_NUM_PAGES(eq, pg_size) \ eq 27 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c (ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size)) eq 29 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size) ((pg_size) / (eq)->elem_size) eq 31 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_CONS_IDX_REG_ADDR(eq) (((eq)->type == HINIC_AEQ) ? \ eq 32 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CSR_AEQ_CONS_IDX_ADDR((eq)->q_id) : \ eq 33 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CSR_CEQ_CONS_IDX_ADDR((eq)->q_id)) eq 35 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_PROD_IDX_REG_ADDR(eq) (((eq)->type == HINIC_AEQ) ? \ eq 36 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CSR_AEQ_PROD_IDX_ADDR((eq)->q_id) : \ eq 37 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CSR_CEQ_PROD_IDX_ADDR((eq)->q_id)) eq 39 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_HI_PHYS_ADDR_REG(eq, pg_num) (((eq)->type == HINIC_AEQ) ? \ eq 40 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CSR_AEQ_HI_PHYS_ADDR_REG((eq)->q_id, pg_num) : \ eq 41 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CSR_CEQ_HI_PHYS_ADDR_REG((eq)->q_id, pg_num)) eq 43 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_LO_PHYS_ADDR_REG(eq, pg_num) (((eq)->type == HINIC_AEQ) ? \ eq 44 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CSR_AEQ_LO_PHYS_ADDR_REG((eq)->q_id, pg_num) : \ eq 45 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CSR_CEQ_LO_PHYS_ADDR_REG((eq)->q_id, pg_num)) eq 47 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_EQ_ELEMENT(eq, idx) \ eq 48 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c ((eq)->virt_addr[(idx) / (eq)->num_elem_in_pg] + \ eq 49 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c (((idx) & ((eq)->num_elem_in_pg - 1)) * (eq)->elem_size)) eq 51 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_AEQ_ELEM(eq, idx) ((struct hinic_aeq_elem *) \ eq 52 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c GET_EQ_ELEMENT(eq, idx)) eq 54 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_CEQ_ELEM(eq, idx) ((u32 *) \ eq 55 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c GET_EQ_ELEMENT(eq, idx)) eq 57 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_CURR_AEQ_ELEM(eq) GET_AEQ_ELEM(eq, (eq)->cons_idx) eq 59 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_CURR_CEQ_ELEM(eq) GET_CEQ_ELEM(eq, (eq)->cons_idx) eq 62 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_SET_HW_PAGE_SIZE_VAL(eq) (ilog2(PAGE_IN_4K((eq)->page_size))) eq 64 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define ELEMENT_SIZE_IN_32B(eq) (((eq)->elem_size) >> 5) eq 65 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define EQ_SET_HW_ELEM_SIZE_VAL(eq) (ilog2(ELEMENT_SIZE_IN_32B(eq))) eq 78 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define aeq_to_aeqs(eq) \ eq 79 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c container_of((eq) - (eq)->q_id, struct hinic_aeqs, aeq[0]) eq 81 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define ceq_to_ceqs(eq) \ eq 82 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c container_of((eq) - (eq)->q_id, struct hinic_ceqs, ceq[0]) eq 191 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void eq_update_ci(struct hinic_eq *eq, u32 arm_state) eq 193 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c u32 val, addr = EQ_CONS_IDX_REG_ADDR(eq); eq 196 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c val = hinic_hwif_read_reg(eq->hwif, addr); eq 203 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c val |= HINIC_EQ_CI_SET(eq->cons_idx, IDX) | eq 204 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_EQ_CI_SET(eq->wrapped, WRAPPED) | eq 209 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_hwif_write_reg(eq->hwif, addr, val); eq 216 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void aeq_irq_handler(struct hinic_eq *eq) eq 218 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c struct hinic_aeqs *aeqs = aeq_to_aeqs(eq); eq 228 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c for (i = 0; i < eq->q_len; i++) { eq 229 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c aeqe_curr = GET_CURR_AEQ_ELEM(eq); eq 235 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (HINIC_EQ_ELEM_DESC_GET(aeqe_desc, WRAPPED) == eq->wrapped) eq 266 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->cons_idx++; eq 268 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (eq->cons_idx == eq->q_len) { eq 269 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->cons_idx = 0; eq 270 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->wrapped = !eq->wrapped; eq 312 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void ceq_irq_handler(struct hinic_eq *eq) eq 314 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c struct hinic_ceqs *ceqs = ceq_to_ceqs(eq); eq 318 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c for (i = 0; i < eq->q_len; i++) { eq 319 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c ceqe = *(GET_CURR_CEQ_ELEM(eq)); eq 325 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (HINIC_EQ_ELEM_DESC_GET(ceqe, WRAPPED) == eq->wrapped) eq 330 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->cons_idx++; eq 332 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (eq->cons_idx == eq->q_len) { eq 333 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->cons_idx = 0; eq 334 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->wrapped = !eq->wrapped; eq 345 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c struct hinic_eq *eq = data; eq 347 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (eq->type == HINIC_AEQ) eq 348 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c aeq_irq_handler(eq); eq 349 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c else if (eq->type == HINIC_CEQ) eq 350 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c ceq_irq_handler(eq); eq 352 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq_update_ci(eq, EQ_ARMED); eq 419 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void set_ctrl0(struct hinic_eq *eq) eq 421 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c struct msix_entry *msix_entry = &eq->msix_entry; eq 422 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c enum hinic_eq_type type = eq->type; eq 427 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c addr = HINIC_CSR_AEQ_CTRL_0_ADDR(eq->q_id); eq 429 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c val = hinic_hwif_read_reg(eq->hwif, addr); eq 438 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_AEQ_CTRL_0_SET(HINIC_HWIF_PCI_INTF(eq->hwif), eq 444 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_hwif_write_reg(eq->hwif, addr, val); eq 447 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c addr = HINIC_CSR_CEQ_CTRL_0_ADDR(eq->q_id); eq 449 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c val = hinic_hwif_read_reg(eq->hwif, addr); eq 460 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CEQ_CTRL_0_SET(HINIC_HWIF_PCI_INTF(eq->hwif), eq 466 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_hwif_write_reg(eq->hwif, addr, val); eq 470 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void set_ctrl1(struct hinic_eq *eq) eq 472 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c enum hinic_eq_type type = eq->type; eq 478 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c addr = HINIC_CSR_AEQ_CTRL_1_ADDR(eq->q_id); eq 480 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c page_size_val = EQ_SET_HW_PAGE_SIZE_VAL(eq); eq 481 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c elem_size = EQ_SET_HW_ELEM_SIZE_VAL(eq); eq 483 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c val = hinic_hwif_read_reg(eq->hwif, addr); eq 489 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c ctrl1 = HINIC_AEQ_CTRL_1_SET(eq->q_len, LEN) | eq 495 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_hwif_write_reg(eq->hwif, addr, val); eq 498 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c addr = HINIC_CSR_CEQ_CTRL_1_ADDR(eq->q_id); eq 500 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c page_size_val = EQ_SET_HW_PAGE_SIZE_VAL(eq); eq 502 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c val = hinic_hwif_read_reg(eq->hwif, addr); eq 507 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c ctrl1 = HINIC_CEQ_CTRL_1_SET(eq->q_len, LEN) | eq 512 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_hwif_write_reg(eq->hwif, addr, val); eq 520 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void set_eq_ctrls(struct hinic_eq *eq) eq 522 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c set_ctrl0(eq); eq 523 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c set_ctrl1(eq); eq 531 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void aeq_elements_init(struct hinic_eq *eq, u32 init_val) eq 536 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c for (i = 0; i < eq->q_len; i++) { eq 537 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c aeqe = GET_AEQ_ELEM(eq, i); eq 549 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void ceq_elements_init(struct hinic_eq *eq, u32 init_val) eq 554 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c for (i = 0; i < eq->q_len; i++) { eq 555 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c ceqe = GET_CEQ_ELEM(eq, i); eq 568 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static int alloc_eq_pages(struct hinic_eq *eq) eq 570 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c struct hinic_hwif *hwif = eq->hwif; eq 576 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c addr_size = eq->num_pages * sizeof(*eq->dma_addr); eq 577 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->dma_addr = devm_kzalloc(&pdev->dev, addr_size, GFP_KERNEL); eq 578 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (!eq->dma_addr) eq 581 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c addr_size = eq->num_pages * sizeof(*eq->virt_addr); eq 582 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->virt_addr = devm_kzalloc(&pdev->dev, addr_size, GFP_KERNEL); eq 583 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (!eq->virt_addr) { eq 588 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c for (pg = 0; pg < eq->num_pages; pg++) { eq 589 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->virt_addr[pg] = dma_alloc_coherent(&pdev->dev, eq 590 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->page_size, eq 591 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c &eq->dma_addr[pg], eq 593 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (!eq->virt_addr[pg]) { eq 598 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c addr = EQ_HI_PHYS_ADDR_REG(eq, pg); eq 599 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c val = upper_32_bits(eq->dma_addr[pg]); eq 603 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c addr = EQ_LO_PHYS_ADDR_REG(eq, pg); eq 604 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c val = lower_32_bits(eq->dma_addr[pg]); eq 609 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c init_val = HINIC_EQ_ELEM_DESC_SET(eq->wrapped, WRAPPED); eq 611 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (eq->type == HINIC_AEQ) eq 612 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c aeq_elements_init(eq, init_val); eq 613 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c else if (eq->type == HINIC_CEQ) eq 614 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c ceq_elements_init(eq, init_val); eq 620 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c dma_free_coherent(&pdev->dev, eq->page_size, eq 621 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->virt_addr[pg], eq 622 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->dma_addr[pg]); eq 624 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c devm_kfree(&pdev->dev, eq->virt_addr); eq 627 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c devm_kfree(&pdev->dev, eq->dma_addr); eq 635 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void free_eq_pages(struct hinic_eq *eq) eq 637 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c struct hinic_hwif *hwif = eq->hwif; eq 641 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c for (pg = 0; pg < eq->num_pages; pg++) eq 642 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c dma_free_coherent(&pdev->dev, eq->page_size, eq 643 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->virt_addr[pg], eq 644 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->dma_addr[pg]); eq 646 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c devm_kfree(&pdev->dev, eq->virt_addr); eq 647 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c devm_kfree(&pdev->dev, eq->dma_addr); eq 662 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static int init_eq(struct hinic_eq *eq, struct hinic_hwif *hwif, eq 669 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->hwif = hwif; eq 670 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->type = type; eq 671 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->q_id = q_id; eq 672 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->q_len = q_len; eq 673 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->page_size = page_size; eq 676 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_hwif_write_reg(eq->hwif, EQ_CONS_IDX_REG_ADDR(eq), 0); eq 677 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_hwif_write_reg(eq->hwif, EQ_PROD_IDX_REG_ADDR(eq), 0); eq 679 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->cons_idx = 0; eq 680 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->wrapped = 0; eq 683 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->elem_size = HINIC_AEQE_SIZE; eq 685 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->elem_size = HINIC_CEQE_SIZE; eq 691 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->num_pages = GET_EQ_NUM_PAGES(eq, page_size); eq 692 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->num_elem_in_pg = GET_EQ_NUM_ELEMS_IN_PG(eq, page_size); eq 694 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->msix_entry = entry; eq 696 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (eq->num_elem_in_pg & (eq->num_elem_in_pg - 1)) { eq 701 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (eq->num_pages > EQ_MAX_PAGES) { eq 706 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c set_eq_ctrls(eq); eq 707 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq_update_ci(eq, EQ_ARMED); eq 709 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c err = alloc_eq_pages(eq); eq 716 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c struct hinic_eq_work *aeq_work = &eq->aeq_work; eq 720 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c tasklet_init(&eq->ceq_tasklet, ceq_tasklet, eq 721 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c (unsigned long)eq); eq 725 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_msix_attr_set(eq->hwif, eq->msix_entry.entry, eq 734 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c "hinic_aeq", eq); eq 737 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c "hinic_ceq", eq); eq 747 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c free_eq_pages(eq); eq 755 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c static void remove_eq(struct hinic_eq *eq) eq 757 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_set_msix_state(eq->hwif, eq->msix_entry.entry, eq 759 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c free_irq(eq->msix_entry.vector, eq); eq 761 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c if (eq->type == HINIC_AEQ) { eq 762 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c struct hinic_eq_work *aeq_work = &eq->aeq_work; eq 766 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_hwif_write_reg(eq->hwif, eq 767 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CSR_AEQ_CTRL_1_ADDR(eq->q_id), 0); eq 768 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c } else if (eq->type == HINIC_CEQ) { eq 769 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c tasklet_kill(&eq->ceq_tasklet); eq 771 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_hwif_write_reg(eq->hwif, eq 772 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c HINIC_CSR_CEQ_CTRL_1_ADDR(eq->q_id), 0); eq 776 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->cons_idx = hinic_hwif_read_reg(eq->hwif, EQ_PROD_IDX_REG_ADDR(eq)); eq 777 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq_update_ci(eq, EQ_NOT_ARMED); eq 779 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c free_eq_pages(eq); eq 352 drivers/net/ethernet/ibm/ehea/ehea.h struct ehea_eq *eq; eq 199 drivers/net/ethernet/ibm/ehea/ehea_main.c arr[i++].fwh = pr->eq->fw_handle; eq 1289 drivers/net/ethernet/ibm/ehea/ehea_main.c ret = ibmebus_request_irq(pr->eq->attr.ist1, eq 1294 drivers/net/ethernet/ibm/ehea/ehea_main.c i, pr->eq->attr.ist1); eq 1299 drivers/net/ethernet/ibm/ehea/ehea_main.c pr->eq->attr.ist1, i); eq 1307 drivers/net/ethernet/ibm/ehea/ehea_main.c u32 ist = port->port_res[i].eq->attr.ist1; eq 1329 drivers/net/ethernet/ibm/ehea/ehea_main.c ibmebus_free_irq(pr->eq->attr.ist1, pr); eq 1332 drivers/net/ethernet/ibm/ehea/ehea_main.c i, pr->eq->attr.ist1); eq 1459 drivers/net/ethernet/ibm/ehea/ehea_main.c pr->eq = ehea_create_eq(adapter, eq_type, EHEA_MAX_ENTRIES_EQ, 0); eq 1460 drivers/net/ethernet/ibm/ehea/ehea_main.c if (!pr->eq) { eq 1466 drivers/net/ethernet/ibm/ehea/ehea_main.c pr->eq->fw_handle, eq 1474 drivers/net/ethernet/ibm/ehea/ehea_main.c pr->eq->fw_handle, eq 1560 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_destroy_eq(pr->eq); eq 1577 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_destroy_eq(pr->eq); eq 236 drivers/net/ethernet/ibm/ehea/ehea_qmr.c struct ehea_eq *eq; eq 238 drivers/net/ethernet/ibm/ehea/ehea_qmr.c eq = kzalloc(sizeof(*eq), GFP_KERNEL); eq 239 drivers/net/ethernet/ibm/ehea/ehea_qmr.c if (!eq) eq 242 drivers/net/ethernet/ibm/ehea/ehea_qmr.c eq->adapter = adapter; eq 243 drivers/net/ethernet/ibm/ehea/ehea_qmr.c eq->attr.type = type; eq 244 drivers/net/ethernet/ibm/ehea/ehea_qmr.c eq->attr.max_nr_of_eqes = max_nr_of_eqes; eq 245 drivers/net/ethernet/ibm/ehea/ehea_qmr.c eq->attr.eqe_gen = eqe_gen; eq 246 drivers/net/ethernet/ibm/ehea/ehea_qmr.c spin_lock_init(&eq->spinlock); eq 249 drivers/net/ethernet/ibm/ehea/ehea_qmr.c &eq->attr, &eq->fw_handle); eq 255 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ret = hw_queue_ctor(&eq->hw_queue, eq->attr.nr_pages, eq 262 drivers/net/ethernet/ibm/ehea/ehea_qmr.c for (i = 0; i < eq->attr.nr_pages; i++) { eq 263 drivers/net/ethernet/ibm/ehea/ehea_qmr.c vpage = hw_qpageit_get_inc(&eq->hw_queue); eq 274 drivers/net/ethernet/ibm/ehea/ehea_qmr.c eq->fw_handle, rpage, 1); eq 276 drivers/net/ethernet/ibm/ehea/ehea_qmr.c if (i == (eq->attr.nr_pages - 1)) { eq 278 drivers/net/ethernet/ibm/ehea/ehea_qmr.c vpage = hw_qpageit_get_inc(&eq->hw_queue); eq 289 drivers/net/ethernet/ibm/ehea/ehea_qmr.c hw_qeit_reset(&eq->hw_queue); eq 290 drivers/net/ethernet/ibm/ehea/ehea_qmr.c return eq; eq 293 drivers/net/ethernet/ibm/ehea/ehea_qmr.c hw_queue_dtor(&eq->hw_queue); eq 296 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ehea_h_free_resource(adapter->handle, eq->fw_handle, FORCE_FREE); eq 299 drivers/net/ethernet/ibm/ehea/ehea_qmr.c kfree(eq); eq 303 drivers/net/ethernet/ibm/ehea/ehea_qmr.c struct ehea_eqe *ehea_poll_eq(struct ehea_eq *eq) eq 308 drivers/net/ethernet/ibm/ehea/ehea_qmr.c spin_lock_irqsave(&eq->spinlock, flags); eq 309 drivers/net/ethernet/ibm/ehea/ehea_qmr.c eqe = hw_eqit_eq_get_inc_valid(&eq->hw_queue); eq 310 drivers/net/ethernet/ibm/ehea/ehea_qmr.c spin_unlock_irqrestore(&eq->spinlock, flags); eq 315 drivers/net/ethernet/ibm/ehea/ehea_qmr.c static u64 ehea_destroy_eq_res(struct ehea_eq *eq, u64 force) eq 320 drivers/net/ethernet/ibm/ehea/ehea_qmr.c spin_lock_irqsave(&eq->spinlock, flags); eq 322 drivers/net/ethernet/ibm/ehea/ehea_qmr.c hret = ehea_h_free_resource(eq->adapter->handle, eq->fw_handle, force); eq 323 drivers/net/ethernet/ibm/ehea/ehea_qmr.c spin_unlock_irqrestore(&eq->spinlock, flags); eq 328 drivers/net/ethernet/ibm/ehea/ehea_qmr.c hw_queue_dtor(&eq->hw_queue); eq 329 drivers/net/ethernet/ibm/ehea/ehea_qmr.c kfree(eq); eq 334 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int ehea_destroy_eq(struct ehea_eq *eq) eq 337 drivers/net/ethernet/ibm/ehea/ehea_qmr.c if (!eq) eq 340 drivers/net/ethernet/ibm/ehea/ehea_qmr.c hcp_epas_dtor(&eq->epas); eq 342 drivers/net/ethernet/ibm/ehea/ehea_qmr.c hret = ehea_destroy_eq_res(eq, NORMAL_FREE); eq 344 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ehea_error_data(eq->adapter, eq->fw_handle, &aer, &aerr); eq 345 drivers/net/ethernet/ibm/ehea/ehea_qmr.c hret = ehea_destroy_eq_res(eq, FORCE_FREE); eq 360 drivers/net/ethernet/ibm/ehea/ehea_qmr.h int ehea_destroy_eq(struct ehea_eq *eq); eq 362 drivers/net/ethernet/ibm/ehea/ehea_qmr.h struct ehea_eqe *ehea_poll_eq(struct ehea_eq *eq); eq 383 drivers/net/ethernet/mellanox/mlx4/cq.c cq_context->comp_eqn = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn; eq 416 drivers/net/ethernet/mellanox/mlx4/cq.c &priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].tasklet_ctx; eq 420 drivers/net/ethernet/mellanox/mlx4/cq.c cq->irq = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].irq; eq 449 drivers/net/ethernet/mellanox/mlx4/cq.c synchronize_irq(priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq); eq 450 drivers/net/ethernet/mellanox/mlx4/cq.c if (priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq != eq 451 drivers/net/ethernet/mellanox/mlx4/cq.c priv->eq_table.eq[MLX4_EQ_ASYNC].irq) eq 452 drivers/net/ethernet/mellanox/mlx4/cq.c synchronize_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq); eq 97 drivers/net/ethernet/mellanox/mlx4/eq.c static void eq_set_ci(struct mlx4_eq *eq, int req_not) eq 99 drivers/net/ethernet/mellanox/mlx4/eq.c __raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) | eq 101 drivers/net/ethernet/mellanox/mlx4/eq.c eq->doorbell); eq 106 drivers/net/ethernet/mellanox/mlx4/eq.c static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry, u8 eqe_factor, eq 110 drivers/net/ethernet/mellanox/mlx4/eq.c unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; eq 118 drivers/net/ethernet/mellanox/mlx4/eq.c return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % PAGE_SIZE; eq 121 drivers/net/ethernet/mellanox/mlx4/eq.c static struct mlx4_eqe *next_eqe_sw(struct mlx4_eq *eq, u8 eqe_factor, u8 size) eq 123 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size); eq 124 drivers/net/ethernet/mellanox/mlx4/eq.c return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; eq 241 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eq *eq = &priv->eq_table.eq[vec]; eq 243 drivers/net/ethernet/mellanox/mlx4/eq.c if (!cpumask_available(eq->affinity_mask) || eq 244 drivers/net/ethernet/mellanox/mlx4/eq.c cpumask_empty(eq->affinity_mask)) eq 247 drivers/net/ethernet/mellanox/mlx4/eq.c hint_err = irq_set_affinity_hint(eq->irq, eq->affinity_mask); eq 494 drivers/net/ethernet/mellanox/mlx4/eq.c static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq) eq 512 drivers/net/ethernet/mellanox/mlx4/eq.c while ((eqe = next_eqe_sw(eq, dev->caps.eqe_factor, eqe_size))) { eq 543 drivers/net/ethernet/mellanox/mlx4/eq.c eq->eqn, eq->cons_index, ret); eq 560 drivers/net/ethernet/mellanox/mlx4/eq.c eq->eqn); eq 573 drivers/net/ethernet/mellanox/mlx4/eq.c eq->eqn, eq->cons_index, ret); eq 697 drivers/net/ethernet/mellanox/mlx4/eq.c eq->eqn, eq->cons_index, ret); eq 713 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_warn(dev, "EQ overrun on EQN %d\n", eq->eqn); eq 781 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type, eqe->subtype, eq->eqn, eq 782 drivers/net/ethernet/mellanox/mlx4/eq.c eq->cons_index, eqe->owner, eq->nent, eq 785 drivers/net/ethernet/mellanox/mlx4/eq.c !!(eq->cons_index & eq->nent) ? "HW" : "SW"); eq 806 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type, eqe->subtype, eq->eqn, eq 807 drivers/net/ethernet/mellanox/mlx4/eq.c eq->cons_index, eqe->owner, eq->nent, eq 809 drivers/net/ethernet/mellanox/mlx4/eq.c !!(eq->cons_index & eq->nent) ? "HW" : "SW"); eq 818 drivers/net/ethernet/mellanox/mlx4/eq.c eqe->type, eqe->subtype, eq->eqn, eq 819 drivers/net/ethernet/mellanox/mlx4/eq.c eq->cons_index, eqe->owner, eq->nent, eq 822 drivers/net/ethernet/mellanox/mlx4/eq.c !!(eq->cons_index & eq->nent) ? "HW" : "SW"); eq 826 drivers/net/ethernet/mellanox/mlx4/eq.c ++eq->cons_index; eq 838 drivers/net/ethernet/mellanox/mlx4/eq.c eq_set_ci(eq, 0); eq 843 drivers/net/ethernet/mellanox/mlx4/eq.c eq_set_ci(eq, 1); eq 858 drivers/net/ethernet/mellanox/mlx4/eq.c work |= mlx4_eq_int(dev, &priv->eq_table.eq[i]); eq 865 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eq *eq = eq_ptr; eq 866 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_dev *dev = eq->dev; eq 868 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_eq_int(dev, eq); eq 934 drivers/net/ethernet/mellanox/mlx4/eq.c static void __iomem *mlx4_get_eq_uar(struct mlx4_dev *dev, struct mlx4_eq *eq) eq 939 drivers/net/ethernet/mellanox/mlx4/eq.c index = eq->eqn / 4 - dev->caps.reserved_eqs / 4; eq 945 drivers/net/ethernet/mellanox/mlx4/eq.c ((eq->eqn / 4) << (dev->uar_page_shift)), eq 949 drivers/net/ethernet/mellanox/mlx4/eq.c eq->eqn); eq 954 drivers/net/ethernet/mellanox/mlx4/eq.c return priv->eq_table.uar_map[index] + 0x800 + 8 * (eq->eqn % 4); eq 970 drivers/net/ethernet/mellanox/mlx4/eq.c u8 intr, struct mlx4_eq *eq) eq 982 drivers/net/ethernet/mellanox/mlx4/eq.c eq->dev = dev; eq 983 drivers/net/ethernet/mellanox/mlx4/eq.c eq->nent = roundup_pow_of_two(max(nent, 2)); eq 987 drivers/net/ethernet/mellanox/mlx4/eq.c npages = PAGE_ALIGN(eq->nent * dev->caps.eqe_size) / PAGE_SIZE; eq 989 drivers/net/ethernet/mellanox/mlx4/eq.c eq->page_list = kmalloc_array(npages, sizeof(*eq->page_list), eq 991 drivers/net/ethernet/mellanox/mlx4/eq.c if (!eq->page_list) eq 995 drivers/net/ethernet/mellanox/mlx4/eq.c eq->page_list[i].buf = NULL; eq 1007 drivers/net/ethernet/mellanox/mlx4/eq.c eq->page_list[i].buf = dma_alloc_coherent(&dev->persist-> eq 1011 drivers/net/ethernet/mellanox/mlx4/eq.c if (!eq->page_list[i].buf) eq 1015 drivers/net/ethernet/mellanox/mlx4/eq.c eq->page_list[i].map = t; eq 1018 drivers/net/ethernet/mellanox/mlx4/eq.c eq->eqn = mlx4_bitmap_alloc(&priv->eq_table.bitmap); eq 1019 drivers/net/ethernet/mellanox/mlx4/eq.c if (eq->eqn == -1) eq 1022 drivers/net/ethernet/mellanox/mlx4/eq.c eq->doorbell = mlx4_get_eq_uar(dev, eq); eq 1023 drivers/net/ethernet/mellanox/mlx4/eq.c if (!eq->doorbell) { eq 1028 drivers/net/ethernet/mellanox/mlx4/eq.c err = mlx4_mtt_init(dev, npages, PAGE_SHIFT, &eq->mtt); eq 1032 drivers/net/ethernet/mellanox/mlx4/eq.c err = mlx4_write_mtt(dev, &eq->mtt, 0, npages, dma_list); eq 1038 drivers/net/ethernet/mellanox/mlx4/eq.c eq_context->log_eq_size = ilog2(eq->nent); eq 1042 drivers/net/ethernet/mellanox/mlx4/eq.c mtt_addr = mlx4_mtt_addr(dev, &eq->mtt); eq 1046 drivers/net/ethernet/mellanox/mlx4/eq.c err = mlx4_SW2HW_EQ(dev, mailbox, eq->eqn); eq 1055 drivers/net/ethernet/mellanox/mlx4/eq.c eq->cons_index = 0; eq 1057 drivers/net/ethernet/mellanox/mlx4/eq.c INIT_LIST_HEAD(&eq->tasklet_ctx.list); eq 1058 drivers/net/ethernet/mellanox/mlx4/eq.c INIT_LIST_HEAD(&eq->tasklet_ctx.process_list); eq 1059 drivers/net/ethernet/mellanox/mlx4/eq.c spin_lock_init(&eq->tasklet_ctx.lock); eq 1060 drivers/net/ethernet/mellanox/mlx4/eq.c tasklet_init(&eq->tasklet_ctx.task, mlx4_cq_tasklet_cb, eq 1061 drivers/net/ethernet/mellanox/mlx4/eq.c (unsigned long)&eq->tasklet_ctx); eq 1066 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_mtt_cleanup(dev, &eq->mtt); eq 1069 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_bitmap_free(&priv->eq_table.bitmap, eq->eqn, MLX4_USE_RR); eq 1073 drivers/net/ethernet/mellanox/mlx4/eq.c if (eq->page_list[i].buf) eq 1075 drivers/net/ethernet/mellanox/mlx4/eq.c eq->page_list[i].buf, eq 1076 drivers/net/ethernet/mellanox/mlx4/eq.c eq->page_list[i].map); eq 1081 drivers/net/ethernet/mellanox/mlx4/eq.c kfree(eq->page_list); eq 1089 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eq *eq) eq 1097 drivers/net/ethernet/mellanox/mlx4/eq.c int npages = PAGE_ALIGN(dev->caps.eqe_size * eq->nent) / PAGE_SIZE; eq 1099 drivers/net/ethernet/mellanox/mlx4/eq.c err = mlx4_HW2SW_EQ(dev, eq->eqn); eq 1103 drivers/net/ethernet/mellanox/mlx4/eq.c synchronize_irq(eq->irq); eq 1104 drivers/net/ethernet/mellanox/mlx4/eq.c tasklet_disable(&eq->tasklet_ctx.task); eq 1106 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_mtt_cleanup(dev, &eq->mtt); eq 1109 drivers/net/ethernet/mellanox/mlx4/eq.c eq->page_list[i].buf, eq 1110 drivers/net/ethernet/mellanox/mlx4/eq.c eq->page_list[i].map); eq 1112 drivers/net/ethernet/mellanox/mlx4/eq.c kfree(eq->page_list); eq 1113 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_bitmap_free(&priv->eq_table.bitmap, eq->eqn, MLX4_USE_RR); eq 1125 drivers/net/ethernet/mellanox/mlx4/eq.c if (eq_table->eq[i].have_irq) { eq 1126 drivers/net/ethernet/mellanox/mlx4/eq.c free_cpumask_var(eq_table->eq[i].affinity_mask); eq 1128 drivers/net/ethernet/mellanox/mlx4/eq.c irq_set_affinity_hint(eq_table->eq[i].irq, NULL); eq 1130 drivers/net/ethernet/mellanox/mlx4/eq.c free_irq(eq_table->eq[i].irq, eq_table->eq + i); eq 1131 drivers/net/ethernet/mellanox/mlx4/eq.c eq_table->eq[i].have_irq = 0; eq 1163 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq = kcalloc(dev->caps.num_eqs - dev->caps.reserved_eqs, eq 1164 drivers/net/ethernet/mellanox/mlx4/eq.c sizeof(*priv->eq_table.eq), GFP_KERNEL); eq 1165 drivers/net/ethernet/mellanox/mlx4/eq.c if (!priv->eq_table.eq) eq 1173 drivers/net/ethernet/mellanox/mlx4/eq.c kfree(mlx4_priv(dev)->eq_table.eq); eq 1226 drivers/net/ethernet/mellanox/mlx4/eq.c 0, &priv->eq_table.eq[MLX4_EQ_ASYNC]); eq 1228 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eq *eq = &priv->eq_table.eq[i]; eq 1230 drivers/net/ethernet/mellanox/mlx4/eq.c int port = find_first_bit(eq->actv_ports.ports, eq 1248 drivers/net/ethernet/mellanox/mlx4/eq.c info->rmap, eq->irq); eq 1257 drivers/net/ethernet/mellanox/mlx4/eq.c eq); eq 1274 drivers/net/ethernet/mellanox/mlx4/eq.c err = request_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq, eq 1276 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq + MLX4_EQ_ASYNC); eq 1280 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[MLX4_EQ_ASYNC].have_irq = 1; eq 1295 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[MLX4_EQ_ASYNC].eqn); eq 1298 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[MLX4_EQ_ASYNC].eqn, err); eq 1301 drivers/net/ethernet/mellanox/mlx4/eq.c eq_set_ci(&priv->eq_table.eq[MLX4_EQ_ASYNC], 1); eq 1307 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_free_eq(dev, &priv->eq_table.eq[--i]); eq 1338 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[MLX4_EQ_ASYNC].eqn); eq 1351 drivers/net/ethernet/mellanox/mlx4/eq.c mlx4_free_eq(dev, &priv->eq_table.eq[i]); eq 1385 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn); eq 1399 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[MLX4_EQ_ASYNC].eqn); eq 1415 drivers/net/ethernet/mellanox/mlx4/eq.c return test_bit(port - 1, priv->eq_table.eq[vector].actv_ports.ports); eq 1427 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[i].actv_ports.ports); eq 1441 drivers/net/ethernet/mellanox/mlx4/eq.c return !!(bitmap_weight(priv->eq_table.eq[vector].actv_ports.ports, eq 1466 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[requested_vector].actv_ports.ports)) { eq 1469 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eq *eq; eq 1475 drivers/net/ethernet/mellanox/mlx4/eq.c eq = &priv->eq_table.eq[requested_vector]; eq 1477 drivers/net/ethernet/mellanox/mlx4/eq.c test_bit(port - 1, eq->actv_ports.ports)) { eq 1487 drivers/net/ethernet/mellanox/mlx4/eq.c struct mlx4_eq *eq = &priv->eq_table.eq[i]; eq 1489 drivers/net/ethernet/mellanox/mlx4/eq.c if (min_ref_count_val > eq->ref_count && eq 1490 drivers/net/ethernet/mellanox/mlx4/eq.c test_bit(port - 1, eq->actv_ports.ports)) { eq 1491 drivers/net/ethernet/mellanox/mlx4/eq.c min_ref_count_val = eq->ref_count; eq 1512 drivers/net/ethernet/mellanox/mlx4/eq.c err = request_irq(priv->eq_table.eq[*prequested_vector].irq, eq 1515 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq + *prequested_vector); eq 1524 drivers/net/ethernet/mellanox/mlx4/eq.c eq_set_ci(&priv->eq_table.eq[*prequested_vector], 1); eq 1525 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[*prequested_vector].have_irq = 1; eq 1530 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[*prequested_vector].ref_count++; eq 1548 drivers/net/ethernet/mellanox/mlx4/eq.c return priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq_vec)].irq; eq 1558 drivers/net/ethernet/mellanox/mlx4/eq.c priv->eq_table.eq[eq_vec].ref_count--; eq 2772 drivers/net/ethernet/mellanox/mlx4/main.c priv->eq_table.eq[MLX4_EQ_ASYNC].irq); eq 2776 drivers/net/ethernet/mellanox/mlx4/main.c priv->eq_table.eq[MLX4_EQ_ASYNC].irq); eq 2905 drivers/net/ethernet/mellanox/mlx4/main.c struct mlx4_eq *eq; eq 2921 drivers/net/ethernet/mellanox/mlx4/main.c eq = &priv->eq_table.eq[eqn]; eq 2923 drivers/net/ethernet/mellanox/mlx4/main.c if (!zalloc_cpumask_var(&eq->affinity_mask, GFP_KERNEL)) eq 2926 drivers/net/ethernet/mellanox/mlx4/main.c cpumask_set_cpu(requested_cpu, eq->affinity_mask); eq 2964 drivers/net/ethernet/mellanox/mlx4/main.c priv->eq_table.eq[MLX4_EQ_ASYNC].irq = entries[0].vector; eq 2965 drivers/net/ethernet/mellanox/mlx4/main.c bitmap_zero(priv->eq_table.eq[MLX4_EQ_ASYNC].actv_ports.ports, eq 2972 drivers/net/ethernet/mellanox/mlx4/main.c priv->eq_table.eq[i].irq = eq 2976 drivers/net/ethernet/mellanox/mlx4/main.c bitmap_fill(priv->eq_table.eq[i].actv_ports.ports, eq 2983 drivers/net/ethernet/mellanox/mlx4/main.c priv->eq_table.eq[i].actv_ports.ports); eq 3018 drivers/net/ethernet/mellanox/mlx4/main.c priv->eq_table.eq[i].irq = dev->persist->pdev->irq; eq 3020 drivers/net/ethernet/mellanox/mlx4/main.c bitmap_fill(priv->eq_table.eq[i].actv_ports.ports, eq 490 drivers/net/ethernet/mellanox/mlx4/mlx4.h struct mlx4_slave_eqe eq[MLX4_MFUNC_MAX_EQES]; eq 693 drivers/net/ethernet/mellanox/mlx4/mlx4.h struct mlx4_eq *eq; eq 1623 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c enum res_eq_states state, struct res_eq **eq) eq 1665 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c if (!err && eq) eq 1666 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c *eq = r; eq 3104 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct res_eq *eq; eq 3110 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c err = eq_res_start_move_to(dev, slave, res_id, RES_EQ_HW, &eq); eq 3127 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c eq->mtt = mtt; eq 3308 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct res_eq *eq; eq 3311 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c err = eq_res_start_move_to(dev, slave, res_id, RES_EQ_RESERVED, &eq); eq 3315 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c err = get_res(dev, slave, eq->mtt->com.res_id, RES_MTT, NULL); eq 3323 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c atomic_dec(&eq->mtt->ref_count); eq 3324 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c put_res(dev, slave, eq->mtt->com.res_id, RES_MTT); eq 3331 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c put_res(dev, slave, eq->mtt->com.res_id, RES_MTT); eq 3414 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct res_eq *eq; eq 3417 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c err = get_res(dev, slave, res_id, RES_EQ, &eq); eq 3421 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c if (eq->com.from_state != RES_EQ_HW) { eq 5120 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c struct res_eq *eq; eq 5132 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c list_for_each_entry_safe(eq, tmp, eq_list, com.list) { eq 5134 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c if (eq->com.owner == slave) { eq 5135 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c eqn = eq->com.res_id; eq 5136 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c state = eq->com.from_state; eq 5141 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rb_erase(&eq->com.node, eq 5143 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c list_del(&eq->com.list); eq 5145 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c kfree(eq); eq 5157 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c atomic_dec(&eq->mtt->ref_count); eq 96 drivers/net/ethernet/mellanox/mlx5/core/cq.c struct mlx5_eq_comp *eq; eq 99 drivers/net/ethernet/mellanox/mlx5/core/cq.c eq = mlx5_eqn2comp_eq(dev, eqn); eq 100 drivers/net/ethernet/mellanox/mlx5/core/cq.c if (IS_ERR(eq)) eq 101 drivers/net/ethernet/mellanox/mlx5/core/cq.c return PTR_ERR(eq); eq 112 drivers/net/ethernet/mellanox/mlx5/core/cq.c cq->eq = eq; eq 119 drivers/net/ethernet/mellanox/mlx5/core/cq.c cq->tasklet_ctx.priv = &eq->tasklet_ctx; eq 123 drivers/net/ethernet/mellanox/mlx5/core/cq.c err = mlx5_eq_add_cq(&eq->core, cq); eq 143 drivers/net/ethernet/mellanox/mlx5/core/cq.c mlx5_eq_del_cq(&eq->core, cq); eq 162 drivers/net/ethernet/mellanox/mlx5/core/cq.c mlx5_eq_del_cq(&cq->eq->core, cq); eq 287 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c static int mlx5_core_eq_query(struct mlx5_core_dev *dev, struct mlx5_eq *eq, eq 293 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c MLX5_SET(query_eq_in, in, eq_number, eq->eqn); eq 297 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c static u64 eq_read_field(struct mlx5_core_dev *dev, struct mlx5_eq *eq, eq 310 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c err = mlx5_core_eq_query(dev, eq, out, outlen); eq 474 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c int mlx5_debug_eq_add(struct mlx5_core_dev *dev, struct mlx5_eq *eq) eq 482 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c &eq->dbg, eq->eqn, eq_fields, eq 483 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c ARRAY_SIZE(eq_fields), eq); eq 485 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c eq->dbg = NULL; eq 490 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c void mlx5_debug_eq_remove(struct mlx5_core_dev *dev, struct mlx5_eq *eq) eq 495 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c if (eq->dbg) eq 496 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c rem_res_tree(eq->dbg); eq 178 drivers/net/ethernet/mellanox/mlx5/core/en/health.c int mlx5e_health_channel_eq_recover(struct mlx5_eq_comp *eq, struct mlx5e_channel *channel) eq 183 drivers/net/ethernet/mellanox/mlx5/core/en/health.c eq->core.eqn, eq->core.cons_index, eq->core.irqn); eq 185 drivers/net/ethernet/mellanox/mlx5/core/en/health.c eqe_count = mlx5_eq_poll_irq_disabled(eq); eq 190 drivers/net/ethernet/mellanox/mlx5/core/en/health.c eqe_count, eq->core.eqn); eq 42 drivers/net/ethernet/mellanox/mlx5/core/en/health.h int mlx5e_health_channel_eq_recover(struct mlx5_eq_comp *eq, struct mlx5e_channel *channel); eq 190 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c struct mlx5_eq_comp *eq; eq 196 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c eq = rq->cq.mcq.eq; eq 197 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c err = mlx5e_health_channel_eq_recover(eq, rq->channel); eq 100 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c struct mlx5_eq_comp *eq; eq 105 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c eq = sq->cq.mcq.eq; eq 106 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c err = mlx5e_health_channel_eq_recover(eq, sq->channel); eq 114 drivers/net/ethernet/mellanox/mlx5/core/eq.c static struct mlx5_core_cq *mlx5_eq_cq_get(struct mlx5_eq *eq, u32 cqn) eq 116 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_cq_table *table = &eq->cq_table; eq 134 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eq *eq = &eq_comp->core; eq 139 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = next_eqe_sw(eq); eq 153 drivers/net/ethernet/mellanox/mlx5/core/eq.c cq = mlx5_eq_cq_get(eq, cqn); eq 159 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_core_warn(eq->dev, "Completion event for bogus CQ 0x%x\n", cqn); eq 162 drivers/net/ethernet/mellanox/mlx5/core/eq.c ++eq->cons_index; eq 164 drivers/net/ethernet/mellanox/mlx5/core/eq.c } while ((++num_eqes < MLX5_EQ_POLLING_BUDGET) && (eqe = next_eqe_sw(eq))); eq 167 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq_update_ci(eq, 1); eq 180 drivers/net/ethernet/mellanox/mlx5/core/eq.c u32 mlx5_eq_poll_irq_disabled(struct mlx5_eq_comp *eq) eq 184 drivers/net/ethernet/mellanox/mlx5/core/eq.c disable_irq(eq->core.irqn); eq 185 drivers/net/ethernet/mellanox/mlx5/core/eq.c count_eqe = eq->core.cons_index; eq 186 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_eq_comp_int(&eq->irq_nb, 0, NULL); eq 187 drivers/net/ethernet/mellanox/mlx5/core/eq.c count_eqe = eq->core.cons_index - count_eqe; eq 188 drivers/net/ethernet/mellanox/mlx5/core/eq.c enable_irq(eq->core.irqn); eq 198 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eq *eq = &eq_async->core; eq 204 drivers/net/ethernet/mellanox/mlx5/core/eq.c dev = eq->dev; eq 207 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = next_eqe_sw(eq); eq 221 drivers/net/ethernet/mellanox/mlx5/core/eq.c ++eq->cons_index; eq 223 drivers/net/ethernet/mellanox/mlx5/core/eq.c } while ((++num_eqes < MLX5_EQ_POLLING_BUDGET) && (eqe = next_eqe_sw(eq))); eq 226 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq_update_ci(eq, 1); eq 231 drivers/net/ethernet/mellanox/mlx5/core/eq.c static void init_eq_buf(struct mlx5_eq *eq) eq 236 drivers/net/ethernet/mellanox/mlx5/core/eq.c for (i = 0; i < eq->nent; i++) { eq 237 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = get_eqe(eq, i); eq 243 drivers/net/ethernet/mellanox/mlx5/core/eq.c create_map_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq, eq 246 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_cq_table *cq_table = &eq->cq_table; eq 262 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->nent = roundup_pow_of_two(param->nent + MLX5_NUM_SPARE_EQE); eq 263 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->cons_index = 0; eq 264 drivers/net/ethernet/mellanox/mlx5/core/eq.c err = mlx5_buf_alloc(dev, eq->nent * MLX5_EQE_SIZE, &eq->buf); eq 268 drivers/net/ethernet/mellanox/mlx5/core/eq.c init_eq_buf(eq); eq 271 drivers/net/ethernet/mellanox/mlx5/core/eq.c MLX5_FLD_SZ_BYTES(create_eq_in, pas[0]) * eq->buf.npages; eq 280 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_fill_page_array(&eq->buf, pas); eq 291 drivers/net/ethernet/mellanox/mlx5/core/eq.c MLX5_SET(eqc, eqc, log_eq_size, ilog2(eq->nent)); eq 295 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->buf.page_shift - MLX5_ADAPTER_PAGE_SHIFT); eq 301 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->vecidx = vecidx; eq 302 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->eqn = MLX5_GET(create_eq_out, out, eq_number); eq 303 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->irqn = pci_irq_vector(dev->pdev, vecidx); eq 304 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->dev = dev; eq 305 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->doorbell = priv->uar->map + MLX5_EQ_DOORBEL_OFFSET; eq 307 drivers/net/ethernet/mellanox/mlx5/core/eq.c err = mlx5_debug_eq_add(dev, eq); eq 315 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_cmd_destroy_eq(dev, eq->eqn); eq 321 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_buf_free(dev, &eq->buf); eq 335 drivers/net/ethernet/mellanox/mlx5/core/eq.c int mlx5_eq_enable(struct mlx5_core_dev *dev, struct mlx5_eq *eq, eq 341 drivers/net/ethernet/mellanox/mlx5/core/eq.c err = mlx5_irq_attach_nb(eq_table->irq_table, eq->vecidx, nb); eq 343 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq_update_ci(eq, 1); eq 357 drivers/net/ethernet/mellanox/mlx5/core/eq.c void mlx5_eq_disable(struct mlx5_core_dev *dev, struct mlx5_eq *eq, eq 362 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_irq_detach_nb(eq_table->irq_table, eq->vecidx, nb); eq 366 drivers/net/ethernet/mellanox/mlx5/core/eq.c static int destroy_unmap_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq) eq 370 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_debug_eq_remove(dev, eq); eq 372 drivers/net/ethernet/mellanox/mlx5/core/eq.c err = mlx5_cmd_destroy_eq(dev, eq->eqn); eq 375 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->eqn); eq 376 drivers/net/ethernet/mellanox/mlx5/core/eq.c synchronize_irq(eq->irqn); eq 378 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_buf_free(dev, &eq->buf); eq 383 drivers/net/ethernet/mellanox/mlx5/core/eq.c int mlx5_eq_add_cq(struct mlx5_eq *eq, struct mlx5_core_cq *cq) eq 385 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_cq_table *table = &eq->cq_table; eq 395 drivers/net/ethernet/mellanox/mlx5/core/eq.c void mlx5_eq_del_cq(struct mlx5_eq *eq, struct mlx5_core_cq *cq) eq 397 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_cq_table *table = &eq->cq_table; eq 405 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_core_dbg(eq->dev, "cq 0x%x not found in eq 0x%x tree\n", eq 406 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->eqn, cq->cqn); eq 411 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_core_dbg(eq->dev, "corruption on cqn 0x%x in eq 0x%x\n", eq 412 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->eqn, cq->cqn); eq 445 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eq *eq, struct mlx5_eq_param *param) eq 457 drivers/net/ethernet/mellanox/mlx5/core/eq.c err = create_map_eq(dev, eq, param); eq 463 drivers/net/ethernet/mellanox/mlx5/core/eq.c static int destroy_async_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq) eq 469 drivers/net/ethernet/mellanox/mlx5/core/eq.c err = destroy_unmap_eq(dev, eq); eq 480 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eq *eq; eq 486 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq = &eqt->async_eq.core; eq 490 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_core_warn(eq->dev, "CQ error on CQN 0x%x, syndrome 0x%x\n", eq 493 drivers/net/ethernet/mellanox/mlx5/core/eq.c cq = mlx5_eq_cq_get(eq, cqn); eq 495 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_core_warn(eq->dev, "Async event for bogus CQ 0x%x\n", cqn); eq 700 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eq *eq = kvzalloc(sizeof(*eq), GFP_KERNEL); eq 703 drivers/net/ethernet/mellanox/mlx5/core/eq.c if (!eq) eq 706 drivers/net/ethernet/mellanox/mlx5/core/eq.c err = create_async_eq(dev, eq, param); eq 708 drivers/net/ethernet/mellanox/mlx5/core/eq.c kvfree(eq); eq 709 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq = ERR_PTR(err); eq 712 drivers/net/ethernet/mellanox/mlx5/core/eq.c return eq; eq 716 drivers/net/ethernet/mellanox/mlx5/core/eq.c int mlx5_eq_destroy_generic(struct mlx5_core_dev *dev, struct mlx5_eq *eq) eq 720 drivers/net/ethernet/mellanox/mlx5/core/eq.c if (IS_ERR(eq)) eq 723 drivers/net/ethernet/mellanox/mlx5/core/eq.c err = destroy_async_eq(dev, eq); eq 727 drivers/net/ethernet/mellanox/mlx5/core/eq.c kvfree(eq); eq 733 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eqe *mlx5_eq_get_eqe(struct mlx5_eq *eq, u32 cc) eq 735 drivers/net/ethernet/mellanox/mlx5/core/eq.c u32 ci = eq->cons_index + cc; eq 738 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = get_eqe(eq, ci & (eq->nent - 1)); eq 739 drivers/net/ethernet/mellanox/mlx5/core/eq.c eqe = ((eqe->owner & 1) ^ !!(ci & eq->nent)) ? NULL : eqe; eq 750 drivers/net/ethernet/mellanox/mlx5/core/eq.c void mlx5_eq_update_ci(struct mlx5_eq *eq, u32 cc, bool arm) eq 752 drivers/net/ethernet/mellanox/mlx5/core/eq.c __be32 __iomem *addr = eq->doorbell + (arm ? 0 : 2); eq 755 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->cons_index += cc; eq 756 drivers/net/ethernet/mellanox/mlx5/core/eq.c val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); eq 767 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eq_comp *eq, *n; eq 769 drivers/net/ethernet/mellanox/mlx5/core/eq.c list_for_each_entry_safe(eq, n, &table->comp_eqs_list, list) { eq 770 drivers/net/ethernet/mellanox/mlx5/core/eq.c list_del(&eq->list); eq 771 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_eq_disable(dev, &eq->core, &eq->irq_nb); eq 772 drivers/net/ethernet/mellanox/mlx5/core/eq.c if (destroy_unmap_eq(dev, &eq->core)) eq 774 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->core.eqn); eq 775 drivers/net/ethernet/mellanox/mlx5/core/eq.c tasklet_disable(&eq->tasklet_ctx.task); eq 776 drivers/net/ethernet/mellanox/mlx5/core/eq.c kfree(eq); eq 783 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eq_comp *eq; eq 796 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq = kzalloc(sizeof(*eq), GFP_KERNEL); eq 797 drivers/net/ethernet/mellanox/mlx5/core/eq.c if (!eq) { eq 802 drivers/net/ethernet/mellanox/mlx5/core/eq.c INIT_LIST_HEAD(&eq->tasklet_ctx.list); eq 803 drivers/net/ethernet/mellanox/mlx5/core/eq.c INIT_LIST_HEAD(&eq->tasklet_ctx.process_list); eq 804 drivers/net/ethernet/mellanox/mlx5/core/eq.c spin_lock_init(&eq->tasklet_ctx.lock); eq 805 drivers/net/ethernet/mellanox/mlx5/core/eq.c tasklet_init(&eq->tasklet_ctx.task, mlx5_cq_tasklet_cb, eq 806 drivers/net/ethernet/mellanox/mlx5/core/eq.c (unsigned long)&eq->tasklet_ctx); eq 808 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->irq_nb.notifier_call = mlx5_eq_comp_int; eq 813 drivers/net/ethernet/mellanox/mlx5/core/eq.c err = create_map_eq(dev, &eq->core, ¶m); eq 815 drivers/net/ethernet/mellanox/mlx5/core/eq.c kfree(eq); eq 818 drivers/net/ethernet/mellanox/mlx5/core/eq.c err = mlx5_eq_enable(dev, &eq->core, &eq->irq_nb); eq 820 drivers/net/ethernet/mellanox/mlx5/core/eq.c destroy_unmap_eq(dev, &eq->core); eq 821 drivers/net/ethernet/mellanox/mlx5/core/eq.c kfree(eq); eq 825 drivers/net/ethernet/mellanox/mlx5/core/eq.c mlx5_core_dbg(dev, "allocated completion EQN %d\n", eq->core.eqn); eq 827 drivers/net/ethernet/mellanox/mlx5/core/eq.c list_add_tail(&eq->list, &table->comp_eqs_list); eq 841 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eq_comp *eq, *n; eq 845 drivers/net/ethernet/mellanox/mlx5/core/eq.c list_for_each_entry_safe(eq, n, &table->comp_eqs_list, list) { eq 847 drivers/net/ethernet/mellanox/mlx5/core/eq.c *eqn = eq->core.eqn; eq 848 drivers/net/ethernet/mellanox/mlx5/core/eq.c *irqn = eq->core.irqn; eq 884 drivers/net/ethernet/mellanox/mlx5/core/eq.c struct mlx5_eq_comp *eq; eq 886 drivers/net/ethernet/mellanox/mlx5/core/eq.c list_for_each_entry(eq, &table->comp_eqs_list, list) { eq 887 drivers/net/ethernet/mellanox/mlx5/core/eq.c if (eq->core.eqn == eqn) eq 888 drivers/net/ethernet/mellanox/mlx5/core/eq.c return eq; eq 50 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h static inline struct mlx5_eqe *get_eqe(struct mlx5_eq *eq, u32 entry) eq 52 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h return mlx5_buf_offset(&eq->buf, entry * MLX5_EQE_SIZE); eq 55 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h static inline struct mlx5_eqe *next_eqe_sw(struct mlx5_eq *eq) eq 57 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); eq 59 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; eq 62 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h static inline void eq_update_ci(struct mlx5_eq *eq, int arm) eq 64 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h __be32 __iomem *addr = eq->doorbell + (arm ? 0 : 2); eq 65 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h u32 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); eq 77 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h int mlx5_eq_add_cq(struct mlx5_eq *eq, struct mlx5_core_cq *cq); eq 78 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h void mlx5_eq_del_cq(struct mlx5_eq *eq, struct mlx5_core_cq *cq); eq 84 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h u32 mlx5_eq_poll_irq_disabled(struct mlx5_eq_comp *eq); eq 88 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h int mlx5_debug_eq_add(struct mlx5_core_dev *dev, struct mlx5_eq *eq); eq 89 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h void mlx5_debug_eq_remove(struct mlx5_core_dev *dev, struct mlx5_eq *eq); eq 93 drivers/net/ethernet/mellanox/mlxsw/pci.c } eq; eq 743 drivers/net/ethernet/mellanox/mlxsw/pci.c q->u.eq.ev_cmd_count++; eq 749 drivers/net/ethernet/mellanox/mlxsw/pci.c q->u.eq.ev_comp_count++; eq 752 drivers/net/ethernet/mellanox/mlxsw/pci.c q->u.eq.ev_other_count++; eq 3535 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c u8 di0, dq0, ei, eq, fi, fq; eq 3553 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c wlc_lcnphy_get_radio_loft(pi, &ei, &eq, &fi, &fq); eq 3555 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c cc.im = (u16) eq; eq 3558 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c wlc_lcnphy_get_radio_loft(pi, &ei, &eq, &fi, &fq); eq 3852 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c u32 hq, lq, nq, eq, pubq; eq 3858 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c eq = 0; eq 3868 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c val32 = (nq << RQPN_NPQ_SHIFT) | (eq << RQPN_EPQ_SHIFT); eq 64 drivers/pci/controller/pcie-iproc-msi.c unsigned int eq; eq 130 drivers/pci/controller/pcie-iproc-msi.c unsigned int eq) eq 134 drivers/pci/controller/pcie-iproc-msi.c return readl_relaxed(pcie->base + msi->reg_offsets[eq][reg]); eq 139 drivers/pci/controller/pcie-iproc-msi.c int eq, u32 val) eq 143 drivers/pci/controller/pcie-iproc-msi.c writel_relaxed(val, pcie->base + msi->reg_offsets[eq][reg]); eq 160 drivers/pci/controller/pcie-iproc-msi.c static inline unsigned int iproc_msi_eq_offset(struct iproc_msi *msi, u32 eq) eq 163 drivers/pci/controller/pcie-iproc-msi.c return eq * EQ_MEM_REGION_SIZE; eq 165 drivers/pci/controller/pcie-iproc-msi.c return eq * EQ_LEN * sizeof(u32); eq 294 drivers/pci/controller/pcie-iproc-msi.c static inline u32 decode_msi_hwirq(struct iproc_msi *msi, u32 eq, u32 head) eq 299 drivers/pci/controller/pcie-iproc-msi.c offs = iproc_msi_eq_offset(msi, eq) + head * sizeof(u32); eq 317 drivers/pci/controller/pcie-iproc-msi.c u32 eq, head, tail, nr_events; eq 325 drivers/pci/controller/pcie-iproc-msi.c eq = grp->eq; eq 338 drivers/pci/controller/pcie-iproc-msi.c eq) & IPROC_MSI_EQ_MASK; eq 341 drivers/pci/controller/pcie-iproc-msi.c eq) & IPROC_MSI_EQ_MASK; eq 354 drivers/pci/controller/pcie-iproc-msi.c hwirq = decode_msi_hwirq(msi, eq, head); eq 366 drivers/pci/controller/pcie-iproc-msi.c iproc_msi_write_reg(msi, IPROC_MSI_EQ_HEAD, eq, head); eq 379 drivers/pci/controller/pcie-iproc-msi.c int i, eq; eq 402 drivers/pci/controller/pcie-iproc-msi.c for (eq = 0; eq < msi->nr_irqs; eq++) { eq 406 drivers/pci/controller/pcie-iproc-msi.c iproc_msi_write_reg(msi, IPROC_MSI_CTRL, eq, val); eq 413 drivers/pci/controller/pcie-iproc-msi.c val = iproc_msi_read_reg(msi, IPROC_MSI_INTS_EN, eq); eq 414 drivers/pci/controller/pcie-iproc-msi.c val |= BIT(eq); eq 415 drivers/pci/controller/pcie-iproc-msi.c iproc_msi_write_reg(msi, IPROC_MSI_INTS_EN, eq, val); eq 422 drivers/pci/controller/pcie-iproc-msi.c u32 eq, val; eq 424 drivers/pci/controller/pcie-iproc-msi.c for (eq = 0; eq < msi->nr_irqs; eq++) { eq 426 drivers/pci/controller/pcie-iproc-msi.c val = iproc_msi_read_reg(msi, IPROC_MSI_INTS_EN, eq); eq 427 drivers/pci/controller/pcie-iproc-msi.c val &= ~BIT(eq); eq 428 drivers/pci/controller/pcie-iproc-msi.c iproc_msi_write_reg(msi, IPROC_MSI_INTS_EN, eq, val); eq 431 drivers/pci/controller/pcie-iproc-msi.c val = iproc_msi_read_reg(msi, IPROC_MSI_CTRL, eq); eq 434 drivers/pci/controller/pcie-iproc-msi.c iproc_msi_write_reg(msi, IPROC_MSI_CTRL, eq, val); eq 601 drivers/pci/controller/pcie-iproc-msi.c msi->grps[i].eq = i; eq 33 drivers/s390/char/sclp_sd.c u8 eq; eq 188 drivers/s390/char/sclp_sd.c static int sclp_sd_sync(unsigned long page, u8 eq, u8 di, u64 sat, u64 sa, eq 205 drivers/s390/char/sclp_sd.c evbuf->eq = eq; eq 215 drivers/s390/char/sclp_sd.c pr_debug("request (eq=%d, di=%d, id=0x%08x)\n", eq, di, listener.id); eq 260 drivers/s390/char/sclp_sd.c eq, di, sccb->hdr.response_code, evbuf->hdr.flags, eq 736 drivers/scsi/be2iscsi/be_cmds.c struct be_queue_info *eq, int eq_delay) eq 741 drivers/scsi/be2iscsi/be_cmds.c struct be_dma_mem *q_mem = &eq->dma_mem; eq 759 drivers/scsi/be2iscsi/be_cmds.c __ilog2_u32(eq->len / 256)); eq 768 drivers/scsi/be2iscsi/be_cmds.c eq->id = le16_to_cpu(resp->eq_id); eq 769 drivers/scsi/be2iscsi/be_cmds.c eq->created = true; eq 776 drivers/scsi/be2iscsi/be_cmds.c struct be_queue_info *cq, struct be_queue_info *eq, eq 805 drivers/scsi/be2iscsi/be_cmds.c AMAP_SET_BITS(struct amap_cq_context, eqid, ctxt, eq->id); eq 820 drivers/scsi/be2iscsi/be_cmds.c AMAP_SET_BITS(struct amap_cq_context_v2, eqid, ctxt, eq->id); eq 803 drivers/scsi/be2iscsi/be_cmds.h struct be_queue_info *eq, int eq_delay); eq 806 drivers/scsi/be2iscsi/be_cmds.h struct be_queue_info *cq, struct be_queue_info *eq, eq 669 drivers/scsi/be2iscsi/be_main.c struct be_queue_info *eq; eq 675 drivers/scsi/be2iscsi/be_main.c eq = &pbe_eq->q; eq 678 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eq 689 drivers/scsi/be2iscsi/be_main.c queue_tail_inc(eq); eq 690 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eq 695 drivers/scsi/be2iscsi/be_main.c hwi_ring_eq_db(phba, eq->id, 1, mcc_events, 1, 1); eq 708 drivers/scsi/be2iscsi/be_main.c struct be_queue_info *eq; eq 712 drivers/scsi/be2iscsi/be_main.c eq = &pbe_eq->q; eq 716 drivers/scsi/be2iscsi/be_main.c hwi_ring_eq_db(phba, eq->id, 1, 0, 0, 1); eq 733 drivers/scsi/be2iscsi/be_main.c struct be_queue_info *eq; eq 751 drivers/scsi/be2iscsi/be_main.c eq = &phwi_context->be_eq[0].q; eq 753 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eq 765 drivers/scsi/be2iscsi/be_main.c queue_tail_inc(eq); eq 766 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eq 780 drivers/scsi/be2iscsi/be_main.c hwi_ring_eq_db(phba, eq->id, 0, (io_events + mcc_events), rearm, 1); eq 2038 drivers/scsi/be2iscsi/be_main.c struct be_queue_info *eq; eq 2048 drivers/scsi/be2iscsi/be_main.c eq = &pbe_eq->q; eq 2049 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eq 2053 drivers/scsi/be2iscsi/be_main.c queue_tail_inc(eq); eq 2054 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eq 2057 drivers/scsi/be2iscsi/be_main.c hwi_ring_eq_db(phba, eq->id, 1, io_events, 0, 1); eq 2999 drivers/scsi/be2iscsi/be_main.c struct be_queue_info *eq; eq 3012 drivers/scsi/be2iscsi/be_main.c eq = &phwi_context->be_eq[i].q; eq 3013 drivers/scsi/be2iscsi/be_main.c mem = &eq->dma_mem; eq 3024 drivers/scsi/be2iscsi/be_main.c ret = be_fill_queue(eq, phba->params.num_eq_entries, eq 3033 drivers/scsi/be2iscsi/be_main.c ret = beiscsi_cmd_eq_create(&phba->ctrl, eq, eq 3050 drivers/scsi/be2iscsi/be_main.c eq = &phwi_context->be_eq[i].q; eq 3051 drivers/scsi/be2iscsi/be_main.c mem = &eq->dma_mem; eq 3064 drivers/scsi/be2iscsi/be_main.c struct be_queue_info *cq, *eq; eq 3076 drivers/scsi/be2iscsi/be_main.c eq = &phwi_context->be_eq[i].q; eq 3099 drivers/scsi/be2iscsi/be_main.c ret = beiscsi_cmd_cq_create(&phba->ctrl, cq, eq, false, eq 3109 drivers/scsi/be2iscsi/be_main.c "iSCSI CQ CREATED\n", cq->id, eq->id); eq 3586 drivers/scsi/be2iscsi/be_main.c struct be_queue_info *eq; eq 3602 drivers/scsi/be2iscsi/be_main.c eq = &phwi_context->be_eq[i].q; eq 3603 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eq 3608 drivers/scsi/be2iscsi/be_main.c queue_tail_inc(eq); eq 3609 drivers/scsi/be2iscsi/be_main.c eqe = queue_tail_node(eq); eq 3614 drivers/scsi/be2iscsi/be_main.c hwi_ring_eq_db(phba, eq->id, 1, num_processed, 1, 1); eq 4099 drivers/scsi/be2iscsi/be_main.c struct be_queue_info *eq; eq 4120 drivers/scsi/be2iscsi/be_main.c eq = &phwi_context->be_eq[0].q; eq 4122 drivers/scsi/be2iscsi/be_main.c "BM_%d : eq->id=%d\n", eq->id); eq 4124 drivers/scsi/be2iscsi/be_main.c hwi_ring_eq_db(phba, eq->id, 0, 0, 1, 1); eq 4127 drivers/scsi/be2iscsi/be_main.c eq = &phwi_context->be_eq[i].q; eq 4129 drivers/scsi/be2iscsi/be_main.c "BM_%d : eq->id=%d\n", eq->id); eq 4130 drivers/scsi/be2iscsi/be_main.c hwi_ring_eq_db(phba, eq->id, 0, 0, 1, 1); eq 990 drivers/scsi/csiostor/csio_wr.c csio_wr_reg32(hw, DBPRIO_V(prio) | QID_V(q->un.eq.physeqid) | eq 411 drivers/scsi/csiostor/csio_wr.h struct csio_eq eq; eq 468 drivers/scsi/csiostor/csio_wr.h #define csio_q_eqid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.eqid) eq 472 drivers/scsi/csiostor/csio_wr.h ((__hw)->wrm.q_arr[(__idx)]->un.eq.physeqid) eq 479 drivers/scsi/csiostor/csio_wr.h #define csio_q_eq_wrap(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.wrap) eq 1319 drivers/scsi/lpfc/lpfc.h lpfc_sli4_mod_hba_eq_delay(struct lpfc_hba *phba, struct lpfc_queue *eq, eq 1325 drivers/scsi/lpfc/lpfc.h bf_set(lpfc_sliport_eqdelay_id, ®_data, eq->queue_id); eq 1328 drivers/scsi/lpfc/lpfc.h eq->q_mode = delay; eq 5163 drivers/scsi/lpfc/lpfc_attr.c struct lpfc_queue *eq, *cq; eq 5185 drivers/scsi/lpfc/lpfc_attr.c eq = phba->sli4_hba.hba_eq_hdl[i].eq; eq 5186 drivers/scsi/lpfc/lpfc_attr.c if (!eq) eq 5189 drivers/scsi/lpfc/lpfc_attr.c list_for_each_entry(cq, &eq->child_list, list) eq 5317 drivers/scsi/lpfc/lpfc_attr.c cpup->eq, cpup->hdwq, cpup->phys_id, eq 5339 drivers/scsi/lpfc/lpfc_attr.c cpup->eq, cpup->hdwq, cpup->phys_id, eq 454 drivers/scsi/lpfc/lpfc_debugfs.h struct lpfc_queue *wq, *cq, *eq; eq 459 drivers/scsi/lpfc/lpfc_debugfs.h eq = NULL; eq 481 drivers/scsi/lpfc/lpfc_debugfs.h eq = phba->sli4_hba.hdwq[eqidx].hba_eq; eq 482 drivers/scsi/lpfc/lpfc_debugfs.h if (cq->assoc_qid == eq->queue_id) eq 488 drivers/scsi/lpfc/lpfc_debugfs.h eq = phba->sli4_hba.hdwq[0].hba_eq; eq 495 drivers/scsi/lpfc/lpfc_debugfs.h eqidx, eq->queue_id); eq 500 drivers/scsi/lpfc/lpfc_debugfs.h eqidx, eq->queue_id); eq 1153 drivers/scsi/lpfc/lpfc_hw4.h struct eq_delay_info eq[LPFC_MAX_EQ_DELAY_EQID_CNT]; eq 1242 drivers/scsi/lpfc/lpfc_init.c struct lpfc_queue *eq, *eq_next; eq 1264 drivers/scsi/lpfc/lpfc_init.c eq = phba->sli4_hba.hba_eq_hdl[i].eq; eq 1265 drivers/scsi/lpfc/lpfc_init.c if (!eq) eq 1267 drivers/scsi/lpfc/lpfc_init.c if (eq->q_mode) { eq 1271 drivers/scsi/lpfc/lpfc_init.c if (eqcnt[eq->last_cpu] < 2) eq 1272 drivers/scsi/lpfc/lpfc_init.c eqcnt[eq->last_cpu]++; eq 1291 drivers/scsi/lpfc/lpfc_init.c list_for_each_entry_safe(eq, eq_next, &eqi->list, cpu_list) { eq 1292 drivers/scsi/lpfc/lpfc_init.c if (eq->last_cpu != i) { eq 1294 drivers/scsi/lpfc/lpfc_init.c eq->last_cpu); eq 1295 drivers/scsi/lpfc/lpfc_init.c list_move_tail(&eq->cpu_list, &eqi_new->list); eq 1298 drivers/scsi/lpfc/lpfc_init.c if (usdelay != eq->q_mode) eq 1299 drivers/scsi/lpfc/lpfc_init.c lpfc_modify_hba_eq_delay(phba, eq->hdwq, 1, eq 8863 drivers/scsi/lpfc/lpfc_init.c eqcpu = lpfc_find_cpu_handle(phba, cpup->eq, LPFC_FIND_BY_EQ); eq 9117 drivers/scsi/lpfc/lpfc_init.c struct lpfc_queue *eq; eq 9136 drivers/scsi/lpfc/lpfc_init.c eq = phba->sli4_hba.hba_eq_hdl[idx].eq; eq 9137 drivers/scsi/lpfc/lpfc_init.c lpfc_sli4_queue_free(eq); eq 9138 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.hba_eq_hdl[idx].eq = NULL; eq 9238 drivers/scsi/lpfc/lpfc_init.c lpfc_create_wq_cq(struct lpfc_hba *phba, struct lpfc_queue *eq, eq 9245 drivers/scsi/lpfc/lpfc_init.c if (!eq || !cq || !wq) { eq 9248 drivers/scsi/lpfc/lpfc_init.c ((eq) ? ((cq) ? "WQ" : "CQ") : "EQ"), qidx); eq 9253 drivers/scsi/lpfc/lpfc_init.c rc = lpfc_cq_create(phba, cq, eq, eq 9269 drivers/scsi/lpfc/lpfc_init.c qidx, cq->queue_id, qidx, eq->queue_id); eq 9318 drivers/scsi/lpfc/lpfc_init.c struct lpfc_queue *eq, *childq; eq 9326 drivers/scsi/lpfc/lpfc_init.c eq = phba->sli4_hba.hba_eq_hdl[qidx].eq; eq 9327 drivers/scsi/lpfc/lpfc_init.c if (!eq) eq 9330 drivers/scsi/lpfc/lpfc_init.c list_for_each_entry(childq, &eq->child_list, list) { eq 9433 drivers/scsi/lpfc/lpfc_init.c if (qidx != cpup->eq) eq 9443 drivers/scsi/lpfc/lpfc_init.c cpup->eq, (uint32_t)rc); eq 9448 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.hba_eq_hdl[cpup->eq].eq = eq 9453 drivers/scsi/lpfc/lpfc_init.c cpup->eq, eq 9718 drivers/scsi/lpfc/lpfc_init.c struct lpfc_queue *eq; eq 9780 drivers/scsi/lpfc/lpfc_init.c eq = phba->sli4_hba.hba_eq_hdl[qidx].eq; eq 9781 drivers/scsi/lpfc/lpfc_init.c lpfc_eq_destroy(phba, eq); eq 10595 drivers/scsi/lpfc/lpfc_init.c (cpup->eq == id)) eq 10661 drivers/scsi/lpfc/lpfc_init.c cpup->eq = LPFC_VECTOR_MAP_EMPTY; eq 10731 drivers/scsi/lpfc/lpfc_init.c cpup->eq = idx; eq 10743 drivers/scsi/lpfc/lpfc_init.c cpup->eq = idx; eq 10756 drivers/scsi/lpfc/lpfc_init.c cpu, cpup->irq, cpup->eq, cpup->flag); eq 10772 drivers/scsi/lpfc/lpfc_init.c if (cpup->eq == LPFC_VECTOR_MAP_EMPTY) { eq 10797 drivers/scsi/lpfc/lpfc_init.c cpup->eq = new_cpup->eq; eq 10823 drivers/scsi/lpfc/lpfc_init.c if (cpup->eq == LPFC_VECTOR_MAP_EMPTY) { eq 10851 drivers/scsi/lpfc/lpfc_init.c cpup->eq = new_cpup->eq; eq 10888 drivers/scsi/lpfc/lpfc_init.c cpup->hdwq, cpup->eq, cpup->irq, cpup->flag); eq 10966 drivers/scsi/lpfc/lpfc_init.c cpup->hdwq, cpup->eq, cpup->irq, cpup->flag); eq 10988 drivers/scsi/lpfc/lpfc_init.c struct lpfc_queue *eq; eq 11023 drivers/scsi/lpfc/lpfc_init.c eq = phba->sli4_hba.hdwq[map->hdwq].hba_eq; eq 11024 drivers/scsi/lpfc/lpfc_init.c list_add(&eq->_poll_list, eqlist); eq 11092 drivers/scsi/lpfc/lpfc_init.c struct lpfc_queue *eq, *next; eq 11107 drivers/scsi/lpfc/lpfc_init.c list_for_each_entry_safe(eq, next, &eqlist, _poll_list) { eq 11108 drivers/scsi/lpfc/lpfc_init.c list_del_init(&eq->_poll_list); eq 11109 drivers/scsi/lpfc/lpfc_init.c lpfc_sli4_start_polling(eq); eq 11118 drivers/scsi/lpfc/lpfc_init.c struct lpfc_queue *eq, *next; eq 11130 drivers/scsi/lpfc/lpfc_init.c list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list) { eq 11131 drivers/scsi/lpfc/lpfc_init.c n = lpfc_find_cpu_handle(phba, eq->hdwq, LPFC_FIND_BY_HDWQ); eq 11133 drivers/scsi/lpfc/lpfc_init.c lpfc_sli4_stop_polling(eq); eq 86 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *eq, eq 456 drivers/scsi/lpfc/lpfc_sli.c __lpfc_sli4_consume_eqe(struct lpfc_hba *phba, struct lpfc_queue *eq, eq 462 drivers/scsi/lpfc/lpfc_sli.c eq->host_index = ((eq->host_index + 1) % eq->entry_count); eq 465 drivers/scsi/lpfc/lpfc_sli.c if (phba->sli4_hba.pc_sli4_params.eqav && !eq->host_index) eq 466 drivers/scsi/lpfc/lpfc_sli.c eq->qe_valid = (eq->qe_valid) ? 0 : 1; eq 470 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_eq_flush(struct lpfc_hba *phba, struct lpfc_queue *eq) eq 476 drivers/scsi/lpfc/lpfc_sli.c eqe = lpfc_sli4_eq_get(eq); eq 478 drivers/scsi/lpfc/lpfc_sli.c __lpfc_sli4_consume_eqe(phba, eq, eqe); eq 480 drivers/scsi/lpfc/lpfc_sli.c eqe = lpfc_sli4_eq_get(eq); eq 484 drivers/scsi/lpfc/lpfc_sli.c phba->sli4_hba.sli4_write_eq_db(phba, eq, count, LPFC_QUEUE_REARM); eq 488 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_process_eq(struct lpfc_hba *phba, struct lpfc_queue *eq, eq 494 drivers/scsi/lpfc/lpfc_sli.c if (cmpxchg(&eq->queue_claimed, 0, 1) != 0) eq 497 drivers/scsi/lpfc/lpfc_sli.c eqe = lpfc_sli4_eq_get(eq); eq 499 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_hba_handle_eqe(phba, eq, eqe); eq 500 drivers/scsi/lpfc/lpfc_sli.c __lpfc_sli4_consume_eqe(phba, eq, eqe); eq 503 drivers/scsi/lpfc/lpfc_sli.c if (!(++count % eq->max_proc_limit)) eq 506 drivers/scsi/lpfc/lpfc_sli.c if (!(count % eq->notify_interval)) { eq 507 drivers/scsi/lpfc/lpfc_sli.c phba->sli4_hba.sli4_write_eq_db(phba, eq, consumed, eq 512 drivers/scsi/lpfc/lpfc_sli.c eqe = lpfc_sli4_eq_get(eq); eq 514 drivers/scsi/lpfc/lpfc_sli.c eq->EQ_processed += count; eq 517 drivers/scsi/lpfc/lpfc_sli.c if (count > eq->EQ_max_eqe) eq 518 drivers/scsi/lpfc/lpfc_sli.c eq->EQ_max_eqe = count; eq 520 drivers/scsi/lpfc/lpfc_sli.c eq->queue_claimed = 0; eq 524 drivers/scsi/lpfc/lpfc_sli.c phba->sli4_hba.sli4_write_eq_db(phba, eq, consumed, rearm); eq 5554 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *eq; eq 5573 drivers/scsi/lpfc/lpfc_sli.c eq = sli4_hba->hba_eq_hdl[qidx].eq; eq 5575 drivers/scsi/lpfc/lpfc_sli.c sli4_hba->sli4_write_eq_db(phba, eq, eq 7869 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *eq; eq 7878 drivers/scsi/lpfc/lpfc_sli.c eq = phba->sli4_hba.hba_eq_hdl[eqidx].eq; eq 7879 drivers/scsi/lpfc/lpfc_sli.c if (eq && eq->queue_id == sli4_hba->mbx_cq->assoc_qid) { eq 7880 drivers/scsi/lpfc/lpfc_sli.c fpeq = eq; eq 10066 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *eq; eq 10071 drivers/scsi/lpfc/lpfc_sli.c eq = phba->sli4_hba.hdwq[piocb->hba_wqidx].hba_eq; eq 10081 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_poll_eq(eq, LPFC_POLL_FASTPATH); eq 14043 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_hba_handle_eqe(struct lpfc_hba *phba, struct lpfc_queue *eq, eq 14047 drivers/scsi/lpfc/lpfc_sli.c uint32_t qidx = eq->hdwq; eq 14234 drivers/scsi/lpfc/lpfc_sli.c fpeq = phba->sli4_hba.hba_eq_hdl[hba_eqidx].eq; eq 14324 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *eq; eq 14329 drivers/scsi/lpfc/lpfc_sli.c list_for_each_entry_rcu(eq, &phba->poll_list, _poll_list) eq 14330 drivers/scsi/lpfc/lpfc_sli.c i += lpfc_sli4_poll_eq(eq, LPFC_POLL_SLOWPATH); eq 14338 drivers/scsi/lpfc/lpfc_sli.c inline int lpfc_sli4_poll_eq(struct lpfc_queue *eq, uint8_t path) eq 14340 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_hba *phba = eq->phba; eq 14352 drivers/scsi/lpfc/lpfc_sli.c if (READ_ONCE(eq->mode) == LPFC_EQ_POLL) eq 14360 drivers/scsi/lpfc/lpfc_sli.c i = lpfc_sli4_process_eq(phba, eq, LPFC_QUEUE_NOARM); eq 14365 drivers/scsi/lpfc/lpfc_sli.c static inline void lpfc_sli4_add_to_poll_list(struct lpfc_queue *eq) eq 14367 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_hba *phba = eq->phba; eq 14376 drivers/scsi/lpfc/lpfc_sli.c list_add_rcu(&eq->_poll_list, &phba->poll_list); eq 14380 drivers/scsi/lpfc/lpfc_sli.c static inline void lpfc_sli4_remove_from_poll_list(struct lpfc_queue *eq) eq 14382 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_hba *phba = eq->phba; eq 14387 drivers/scsi/lpfc/lpfc_sli.c list_del_rcu(&eq->_poll_list); eq 14396 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *eq, *next; eq 14398 drivers/scsi/lpfc/lpfc_sli.c list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list) eq 14399 drivers/scsi/lpfc/lpfc_sli.c list_del(&eq->_poll_list); eq 14406 drivers/scsi/lpfc/lpfc_sli.c __lpfc_sli4_switch_eqmode(struct lpfc_queue *eq, uint8_t mode) eq 14408 drivers/scsi/lpfc/lpfc_sli.c if (mode == eq->mode) eq 14421 drivers/scsi/lpfc/lpfc_sli.c WRITE_ONCE(eq->mode, mode); eq 14437 drivers/scsi/lpfc/lpfc_sli.c mode ? lpfc_sli4_add_to_poll_list(eq) : eq 14438 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_remove_from_poll_list(eq); eq 14441 drivers/scsi/lpfc/lpfc_sli.c void lpfc_sli4_start_polling(struct lpfc_queue *eq) eq 14443 drivers/scsi/lpfc/lpfc_sli.c __lpfc_sli4_switch_eqmode(eq, LPFC_EQ_POLL); eq 14446 drivers/scsi/lpfc/lpfc_sli.c void lpfc_sli4_stop_polling(struct lpfc_queue *eq) eq 14448 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_hba *phba = eq->phba; eq 14450 drivers/scsi/lpfc/lpfc_sli.c __lpfc_sli4_switch_eqmode(eq, LPFC_EQ_INTERRUPT); eq 14459 drivers/scsi/lpfc/lpfc_sli.c phba->sli4_hba.sli4_write_eq_db(phba, eq, 0, LPFC_QUEUE_REARM); eq 14641 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *eq; eq 14661 drivers/scsi/lpfc/lpfc_sli.c eq = phba->sli4_hba.hba_eq_hdl[qidx].eq; eq 14662 drivers/scsi/lpfc/lpfc_sli.c if (!eq) eq 14665 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli4_mod_hba_eq_delay(phba, eq, usdelay); eq 14697 drivers/scsi/lpfc/lpfc_sli.c eq = phba->sli4_hba.hba_eq_hdl[qidx].eq; eq 14698 drivers/scsi/lpfc/lpfc_sli.c if (!eq) eq 14700 drivers/scsi/lpfc/lpfc_sli.c eq->q_mode = usdelay; eq 14701 drivers/scsi/lpfc/lpfc_sli.c eq_delay->u.request.eq[cnt].eq_id = eq->queue_id; eq 14702 drivers/scsi/lpfc/lpfc_sli.c eq_delay->u.request.eq[cnt].phase = 0; eq 14703 drivers/scsi/lpfc/lpfc_sli.c eq_delay->u.request.eq[cnt].delay_multi = dmult; eq 14749 drivers/scsi/lpfc/lpfc_sli.c lpfc_eq_create(struct lpfc_hba *phba, struct lpfc_queue *eq, uint32_t imax) eq 14761 drivers/scsi/lpfc/lpfc_sli.c if (!eq) eq 14777 drivers/scsi/lpfc/lpfc_sli.c eq->page_count); eq 14794 drivers/scsi/lpfc/lpfc_sli.c switch (eq->entry_count) { eq 14798 drivers/scsi/lpfc/lpfc_sli.c eq->entry_count); eq 14799 drivers/scsi/lpfc/lpfc_sli.c if (eq->entry_count < 256) { eq 14825 drivers/scsi/lpfc/lpfc_sli.c list_for_each_entry(dmabuf, &eq->page_list, list) { eq 14846 drivers/scsi/lpfc/lpfc_sli.c eq->type = LPFC_EQ; eq 14847 drivers/scsi/lpfc/lpfc_sli.c eq->subtype = LPFC_NONE; eq 14848 drivers/scsi/lpfc/lpfc_sli.c eq->queue_id = bf_get(lpfc_mbx_eq_create_q_id, &eq_create->u.response); eq 14849 drivers/scsi/lpfc/lpfc_sli.c if (eq->queue_id == 0xFFFF) eq 14851 drivers/scsi/lpfc/lpfc_sli.c eq->host_index = 0; eq 14852 drivers/scsi/lpfc/lpfc_sli.c eq->notify_interval = LPFC_EQ_NOTIFY_INTRVL; eq 14853 drivers/scsi/lpfc/lpfc_sli.c eq->max_proc_limit = LPFC_EQ_MAX_PROC_LIMIT; eq 14882 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *eq, uint32_t type, uint32_t subtype) eq 14892 drivers/scsi/lpfc/lpfc_sli.c if (!cq || !eq) eq 14915 drivers/scsi/lpfc/lpfc_sli.c eq->queue_id); eq 14920 drivers/scsi/lpfc/lpfc_sli.c eq->queue_id); eq 14985 drivers/scsi/lpfc/lpfc_sli.c list_add_tail(&cq->list, &eq->child_list); eq 14990 drivers/scsi/lpfc/lpfc_sli.c cq->assoc_qid = eq->queue_id; eq 14991 drivers/scsi/lpfc/lpfc_sli.c cq->assoc_qp = eq; eq 15031 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *eq; eq 15070 drivers/scsi/lpfc/lpfc_sli.c eq = hdwq[idx].hba_eq; eq 15071 drivers/scsi/lpfc/lpfc_sli.c if (!cq || !eq) { eq 15133 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15137 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15141 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15145 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15149 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15153 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15157 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15161 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15165 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15169 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15173 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15177 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15181 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15185 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15189 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15193 drivers/scsi/lpfc/lpfc_sli.c &cq_set->u.request, eq->queue_id); eq 15198 drivers/scsi/lpfc/lpfc_sli.c list_add_tail(&cq->list, &eq->child_list); eq 15202 drivers/scsi/lpfc/lpfc_sli.c cq->assoc_qid = eq->queue_id; eq 15203 drivers/scsi/lpfc/lpfc_sli.c cq->assoc_qp = eq; eq 16228 drivers/scsi/lpfc/lpfc_sli.c lpfc_eq_destroy(struct lpfc_hba *phba, struct lpfc_queue *eq) eq 16236 drivers/scsi/lpfc/lpfc_sli.c if (!eq) eq 16239 drivers/scsi/lpfc/lpfc_sli.c mbox = mempool_alloc(eq->phba->mbox_mem_pool, GFP_KERNEL); eq 16248 drivers/scsi/lpfc/lpfc_sli.c eq->queue_id); eq 16249 drivers/scsi/lpfc/lpfc_sli.c mbox->vport = eq->phba->pport; eq 16252 drivers/scsi/lpfc/lpfc_sli.c rc = lpfc_sli_issue_mbox(eq->phba, mbox, MBX_POLL); eq 16267 drivers/scsi/lpfc/lpfc_sli.c list_del_init(&eq->list); eq 16268 drivers/scsi/lpfc/lpfc_sli.c mempool_free(mbox, eq->phba->mbox_mem_pool); eq 471 drivers/scsi/lpfc/lpfc_sli4.h struct lpfc_queue *eq; eq 567 drivers/scsi/lpfc/lpfc_sli4.h uint16_t eq; eq 799 drivers/scsi/lpfc/lpfc_sli4.h void (*sli4_write_eq_db)(struct lpfc_hba *phba, struct lpfc_queue *eq, eq 2313 drivers/soc/fsl/qbman/qman.c struct qm_eqcr_entry *eq; eq 2325 drivers/soc/fsl/qbman/qman.c eq = qm_eqcr_start_stash(&p->p); eq 2334 drivers/soc/fsl/qbman/qman.c eq = qm_eqcr_start_no_stash(&p->p); eq 2337 drivers/soc/fsl/qbman/qman.c if (unlikely(!eq)) eq 2340 drivers/soc/fsl/qbman/qman.c qm_fqid_set(eq, fq->fqid); eq 2341 drivers/soc/fsl/qbman/qman.c eq->tag = cpu_to_be32(fq_to_tag(fq)); eq 2342 drivers/soc/fsl/qbman/qman.c eq->fd = *fd; eq 334 fs/befs/btree.c int eq; eq 348 fs/befs/btree.c eq = befs_compare_strings(thiskey, keylen, findkey, findkey_len); eq 349 fs/befs/btree.c if (eq < 0) { eq 364 fs/befs/btree.c eq = befs_compare_strings(thiskey, keylen, findkey, eq 367 fs/befs/btree.c if (eq == 0) { eq 374 fs/befs/btree.c if (eq > 0) eq 381 fs/befs/btree.c if (eq < 0) eq 63 include/linux/mlx5/cq.h struct mlx5_eq_comp *eq; eq 24 include/linux/mlx5/eq.h mlx5_eq_destroy_generic(struct mlx5_core_dev *dev, struct mlx5_eq *eq); eq 25 include/linux/mlx5/eq.h int mlx5_eq_enable(struct mlx5_core_dev *dev, struct mlx5_eq *eq, eq 27 include/linux/mlx5/eq.h void mlx5_eq_disable(struct mlx5_core_dev *dev, struct mlx5_eq *eq, eq 30 include/linux/mlx5/eq.h struct mlx5_eqe *mlx5_eq_get_eqe(struct mlx5_eq *eq, u32 cc); eq 31 include/linux/mlx5/eq.h void mlx5_eq_update_ci(struct mlx5_eq *eq, u32 cc, bool arm); eq 41 include/linux/mlx5/eq.h static inline u32 mlx5_eq_update_cc(struct mlx5_eq *eq, u32 cc) eq 44 include/linux/mlx5/eq.h mlx5_eq_update_ci(eq, cc, 0); eq 77 net/caif/cfctrl.c bool eq = eq 83 net/caif/cfctrl.c if (!eq) eq 150 net/dns_resolver/dns_key.c const char *eq; eq 161 net/dns_resolver/dns_key.c eq = memchr(opt, '=', opt_len); eq 162 net/dns_resolver/dns_key.c if (eq) { eq 163 net/dns_resolver/dns_key.c opt_nlen = eq - opt; eq 164 net/dns_resolver/dns_key.c eq++; eq 165 net/dns_resolver/dns_key.c memcpy(optval, eq, next_opt - eq); eq 166 net/dns_resolver/dns_key.c optval[next_opt - eq] = '\0'; eq 2564 security/selinux/ss/policydb.c int rc, eq; eq 2566 security/selinux/ss/policydb.c eq = mls_level_eq(&r->level[1], &r->level[0]); eq 2568 security/selinux/ss/policydb.c if (eq) eq 2574 security/selinux/ss/policydb.c if (!eq) eq 2586 security/selinux/ss/policydb.c if (!eq) { eq 159 sound/pci/au88x0/au88x0.h eqlzr_t eq; eq 56 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 78 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 101 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 118 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 164 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 211 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 221 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 231 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 241 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 252 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 267 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 282 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 297 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 314 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 338 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 478 sound/pci/au88x0/au88x0_eq.c eqhw_t *eqhw = &(vortex->eq.this04); eq 495 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 497 sound/pci/au88x0/au88x0_eq.c if (eq->this28) { eq 498 sound/pci/au88x0/au88x0_eq.c *gain = eq->this130[index]; eq 506 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 508 sound/pci/au88x0/au88x0_eq.c if (eq->this28 == 0) eq 511 sound/pci/au88x0/au88x0_eq.c eq->this130[index] = gain; eq 512 sound/pci/au88x0/au88x0_eq.c if (eq->this54) eq 520 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 522 sound/pci/au88x0/au88x0_eq.c if (eq->this28) { eq 523 sound/pci/au88x0/au88x0_eq.c *gain = eq->this130[index + eq->this10]; eq 531 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 533 sound/pci/au88x0/au88x0_eq.c if (eq->this28 == 0) eq 536 sound/pci/au88x0/au88x0_eq.c eq->this130[index + eq->this10] = gain; eq 537 sound/pci/au88x0/au88x0_eq.c if (eq->this54) eq 547 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 550 sound/pci/au88x0/au88x0_eq.c if (eq->this10 == 0) eq 557 sound/pci/au88x0/au88x0_eq.c (vortex, si, &gains[si + eq->this10])) eq 561 sound/pci/au88x0/au88x0_eq.c while (eq->this10 > si) ; eq 568 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 570 sound/pci/au88x0/au88x0_eq.c vortex_EqHw_SetLeftGainsTarget(vortex, eq->this130); eq 571 sound/pci/au88x0/au88x0_eq.c vortex_EqHw_SetRightGainsTarget(vortex, &(eq->this130[eq->this10])); eq 579 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 582 sound/pci/au88x0/au88x0_eq.c if (((eq->this10) * 2 != count) || (eq->this28 == 0)) eq 586 sound/pci/au88x0/au88x0_eq.c eq->this130[i] = gains[i]; eq 589 sound/pci/au88x0/au88x0_eq.c if (eq->this54) eq 597 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 600 sound/pci/au88x0/au88x0_eq.c eq->this58 = a; eq 601 sound/pci/au88x0/au88x0_eq.c eq->this5c = b; eq 602 sound/pci/au88x0/au88x0_eq.c if (eq->this54) eq 603 sound/pci/au88x0/au88x0_eq.c eax = eq->this0e; eq 605 sound/pci/au88x0/au88x0_eq.c eax = eq->this0a; eq 606 sound/pci/au88x0/au88x0_eq.c ebx = (eax * eq->this58) >> 0x10; eq 607 sound/pci/au88x0/au88x0_eq.c eax = (eax * eq->this5c) >> 0x10; eq 613 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 616 sound/pci/au88x0/au88x0_eq.c if (eq->this54) eq 617 sound/pci/au88x0/au88x0_eq.c eax = eq->this0e; eq 619 sound/pci/au88x0/au88x0_eq.c eax = eq->this0a; eq 620 sound/pci/au88x0/au88x0_eq.c ebx = (eax * eq->this58) >> 0x10; eq 621 sound/pci/au88x0/au88x0_eq.c eax = (eax * eq->this5c) >> 0x10; eq 633 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 635 sound/pci/au88x0/au88x0_eq.c if ((eq->this28) && (bp == 0)) { eq 638 sound/pci/au88x0/au88x0_eq.c vortex_EqHw_SetBypassGain(vortex, eq->this08, eq->this08); eq 641 sound/pci/au88x0/au88x0_eq.c vortex_EqHw_SetLeftGainsTarget(vortex, eq->this14_array); eq 642 sound/pci/au88x0/au88x0_eq.c vortex_EqHw_SetRightGainsTarget(vortex, eq->this14_array); eq 643 sound/pci/au88x0/au88x0_eq.c vortex_EqHw_SetBypassGain(vortex, eq->this0c, eq->this0c); eq 650 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 653 sound/pci/au88x0/au88x0_eq.c memcpy(&(eq->coefset), &asEqCoefsNormal, sizeof(auxxEqCoeffSet_t)); eq 655 sound/pci/au88x0/au88x0_eq.c vortex_Eqlzr_SetAllBands(vortex, eq_gains_normal, eq->this10 * 2); eq 660 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 662 sound/pci/au88x0/au88x0_eq.c if (eq->this10 == 0) eq 664 sound/pci/au88x0/au88x0_eq.c *count = eq->this10 * 2; eq 672 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 674 sound/pci/au88x0/au88x0_eq.c return (&(eq->coefset)); eq 679 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 683 sound/pci/au88x0/au88x0_eq.c eq->this08 = 0; /* Bypass gain with EQ in use. */ eq 684 sound/pci/au88x0/au88x0_eq.c eq->this0a = 0x5999; eq 685 sound/pci/au88x0/au88x0_eq.c eq->this0c = 0x5999; /* Bypass gain with EQ disabled. */ eq 686 sound/pci/au88x0/au88x0_eq.c eq->this0e = 0x5999; eq 688 sound/pci/au88x0/au88x0_eq.c eq->this10 = 0xa; /* 10 eq frequency bands. */ eq 689 sound/pci/au88x0/au88x0_eq.c eq->this04.this04 = eq->this10; eq 690 sound/pci/au88x0/au88x0_eq.c eq->this28 = 0x1; /* if 1 => Allow read access to this130 (gains) */ eq 691 sound/pci/au88x0/au88x0_eq.c eq->this54 = 0x0; /* if 1 => Dont Allow access to hardware (gains) */ eq 692 sound/pci/au88x0/au88x0_eq.c eq->this58 = 0xffff; eq 693 sound/pci/au88x0/au88x0_eq.c eq->this5c = 0xffff; eq 696 sound/pci/au88x0/au88x0_eq.c memset(eq->this14_array, 0, sizeof(eq->this14_array)); eq 703 sound/pci/au88x0/au88x0_eq.c vortex_EqHw_Program10Band(vortex, &(eq->coefset)); eq 704 sound/pci/au88x0/au88x0_eq.c vortex_Eqlzr_SetBypass(vortex, eq->this54); eq 726 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 729 sound/pci/au88x0/au88x0_eq.c ucontrol->value.integer.value[0] = eq->this54 ? 0 : 1; eq 739 sound/pci/au88x0/au88x0_eq.c eqlzr_t *eq = &(vortex->eq); eq 742 sound/pci/au88x0/au88x0_eq.c eq->this54 = ucontrol->value.integer.value[0] ? 0 : 1; eq 743 sound/pci/au88x0/au88x0_eq.c vortex_Eqlzr_SetBypass(vortex, eq->this54); eq 789 sound/soc/codecs/wm8958-dsp2.c int eq = kcontrol->private_value; eq 793 sound/soc/codecs/wm8958-dsp2.c ucontrol->value.integer.value[0] = wm8994->enh_eq_ena[eq]; eq 801 sound/soc/codecs/wm8958-dsp2.c int eq = kcontrol->private_value; eq 805 sound/soc/codecs/wm8958-dsp2.c if (wm8994->enh_eq_ena[eq] == ucontrol->value.integer.value[0]) eq 814 sound/soc/codecs/wm8958-dsp2.c if (wm8958_dsp2_busy(wm8994, eq)) { eq 815 sound/soc/codecs/wm8958-dsp2.c dev_dbg(component->dev, "DSP2 active on %d already\n", eq); eq 819 sound/soc/codecs/wm8958-dsp2.c if (wm8994->mbc_ena[eq] || wm8994->vss_ena[eq] || eq 820 sound/soc/codecs/wm8958-dsp2.c wm8994->hpf1_ena[eq] || wm8994->hpf2_ena[eq]) eq 823 sound/soc/codecs/wm8958-dsp2.c wm8994->enh_eq_ena[eq] = ucontrol->value.integer.value[0]; eq 825 sound/soc/codecs/wm8958-dsp2.c wm8958_dsp_apply(component, eq, ucontrol->value.integer.value[0]); eq 2233 tools/lib/bpf/btf.c int i, eq; eq 2306 tools/lib/bpf/btf.c eq = btf_dedup_is_equiv(d, eq 2308 tools/lib/bpf/btf.c if (eq <= 0) eq 2309 tools/lib/bpf/btf.c return eq; eq 2324 tools/lib/bpf/btf.c eq = btf_dedup_is_equiv(d, cand_m->type, canon_m->type); eq 2325 tools/lib/bpf/btf.c if (eq <= 0) eq 2326 tools/lib/bpf/btf.c return eq; eq 2340 tools/lib/bpf/btf.c eq = btf_dedup_is_equiv(d, cand_type->type, canon_type->type); eq 2341 tools/lib/bpf/btf.c if (eq <= 0) eq 2342 tools/lib/bpf/btf.c return eq; eq 2347 tools/lib/bpf/btf.c eq = btf_dedup_is_equiv(d, cand_p->type, canon_p->type); eq 2348 tools/lib/bpf/btf.c if (eq <= 0) eq 2349 tools/lib/bpf/btf.c return eq; eq 2478 tools/lib/bpf/btf.c int eq; eq 2495 tools/lib/bpf/btf.c eq = btf_dedup_is_equiv(d, type_id, cand_id); eq 2496 tools/lib/bpf/btf.c if (eq < 0) eq 2497 tools/lib/bpf/btf.c return eq; eq 2498 tools/lib/bpf/btf.c if (!eq)