neq 70 crypto/memneq.c unsigned long neq = 0; neq 74 crypto/memneq.c neq |= *(unsigned long *)a ^ *(unsigned long *)b; neq 75 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 82 crypto/memneq.c neq |= *(unsigned char *)a ^ *(unsigned char *)b; neq 83 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 88 crypto/memneq.c return neq; neq 94 crypto/memneq.c unsigned long neq = 0; neq 98 crypto/memneq.c neq |= *(unsigned long *)(a) ^ *(unsigned long *)(b); neq 99 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 100 crypto/memneq.c neq |= *(unsigned long *)(a+8) ^ *(unsigned long *)(b+8); neq 101 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 103 crypto/memneq.c neq |= *(unsigned int *)(a) ^ *(unsigned int *)(b); neq 104 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 105 crypto/memneq.c neq |= *(unsigned int *)(a+4) ^ *(unsigned int *)(b+4); neq 106 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 107 crypto/memneq.c neq |= *(unsigned int *)(a+8) ^ *(unsigned int *)(b+8); neq 108 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 109 crypto/memneq.c neq |= *(unsigned int *)(a+12) ^ *(unsigned int *)(b+12); neq 110 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 114 crypto/memneq.c neq |= *(unsigned char *)(a) ^ *(unsigned char *)(b); neq 115 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 116 crypto/memneq.c neq |= *(unsigned char *)(a+1) ^ *(unsigned char *)(b+1); neq 117 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 118 crypto/memneq.c neq |= *(unsigned char *)(a+2) ^ *(unsigned char *)(b+2); neq 119 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 120 crypto/memneq.c neq |= *(unsigned char *)(a+3) ^ *(unsigned char *)(b+3); neq 121 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 122 crypto/memneq.c neq |= *(unsigned char *)(a+4) ^ *(unsigned char *)(b+4); neq 123 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 124 crypto/memneq.c neq |= *(unsigned char *)(a+5) ^ *(unsigned char *)(b+5); neq 125 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 126 crypto/memneq.c neq |= *(unsigned char *)(a+6) ^ *(unsigned char *)(b+6); neq 127 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 128 crypto/memneq.c neq |= *(unsigned char *)(a+7) ^ *(unsigned char *)(b+7); neq 129 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 130 crypto/memneq.c neq |= *(unsigned char *)(a+8) ^ *(unsigned char *)(b+8); neq 131 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 132 crypto/memneq.c neq |= *(unsigned char *)(a+9) ^ *(unsigned char *)(b+9); neq 133 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 134 crypto/memneq.c neq |= *(unsigned char *)(a+10) ^ *(unsigned char *)(b+10); neq 135 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 136 crypto/memneq.c neq |= *(unsigned char *)(a+11) ^ *(unsigned char *)(b+11); neq 137 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 138 crypto/memneq.c neq |= *(unsigned char *)(a+12) ^ *(unsigned char *)(b+12); neq 139 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 140 crypto/memneq.c neq |= *(unsigned char *)(a+13) ^ *(unsigned char *)(b+13); neq 141 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 142 crypto/memneq.c neq |= *(unsigned char *)(a+14) ^ *(unsigned char *)(b+14); neq 143 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 144 crypto/memneq.c neq |= *(unsigned char *)(a+15) ^ *(unsigned char *)(b+15); neq 145 crypto/memneq.c OPTIMIZER_HIDE_VAR(neq); neq 148 crypto/memneq.c return neq; neq 147 drivers/media/dvb-core/dvb_demux.c u8 neq = 0; neq 156 drivers/media/dvb-core/dvb_demux.c neq |= f->maskandnotmode[i] & xor; neq 159 drivers/media/dvb-core/dvb_demux.c if (f->doneq && !neq) neq 298 drivers/media/pci/ttpci/av7110.c u8 xor, neq = 0; neq 302 drivers/media/pci/ttpci/av7110.c neq |= dvbdmxfilter->maskandnotmode[i] & xor; neq 304 drivers/media/pci/ttpci/av7110.c if (!neq) neq 329 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h unsigned int neq; /* N egress Qs */ neq 2588 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c S("Egress Queues", "%d", neq); neq 5081 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c int niqflint, neq, avail_eth_qsets; neq 5109 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c neq = adap->params.pfres.neq / 2; neq 5110 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c avail_eth_qsets = min(niqflint, neq); neq 2920 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c pfres->neq = FW_PFVF_CMD_NEQ_G(word); neq 2360 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c S("Egress Queues", "%d", neq); neq 2559 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c if (vfres->neq < ethqsets*2) neq 2560 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c ethqsets = vfres->neq/2; neq 246 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h unsigned int neq; /* N egress Qs */ neq 1119 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c vfres->neq = FW_PFVF_CMD_NEQ_G(word); neq 382 drivers/net/ethernet/ibm/ehea/ehea.h struct ehea_eq *neq; /* notification event queue */ neq 211 drivers/net/ethernet/ibm/ehea/ehea_main.c arr[i++].fwh = adapter->neq->fw_handle; neq 1221 drivers/net/ethernet/ibm/ehea/ehea_main.c eqe = ehea_poll_eq(adapter->neq); neq 1227 drivers/net/ethernet/ibm/ehea/ehea_main.c eqe = ehea_poll_eq(adapter->neq); neq 1236 drivers/net/ethernet/ibm/ehea/ehea_main.c adapter->neq->fw_handle, event_mask); neq 3412 drivers/net/ethernet/ibm/ehea/ehea_main.c adapter->neq = ehea_create_eq(adapter, neq 3414 drivers/net/ethernet/ibm/ehea/ehea_main.c if (!adapter->neq) { neq 3433 drivers/net/ethernet/ibm/ehea/ehea_main.c ret = ibmebus_request_irq(adapter->neq->attr.ist1, neq 3458 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_destroy_eq(adapter->neq); neq 3482 drivers/net/ethernet/ibm/ehea/ehea_main.c ibmebus_free_irq(adapter->neq->attr.ist1, adapter); neq 3485 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_destroy_eq(adapter->neq); neq 95 drivers/soc/fsl/qbman/qman_test_api.c bool neq = qm_fd_addr_get64(a) != qm_fd_addr_get64(b); neq 97 drivers/soc/fsl/qbman/qman_test_api.c neq |= qm_fd_get_format(a) != qm_fd_get_format(b); neq 98 drivers/soc/fsl/qbman/qman_test_api.c neq |= a->cfg != b->cfg; neq 99 drivers/soc/fsl/qbman/qman_test_api.c neq |= a->cmd != b->cmd; neq 101 drivers/soc/fsl/qbman/qman_test_api.c return neq;