cnum 392 arch/ia64/kernel/perfmon.c typedef int (*pfm_reg_check_t)(struct task_struct *task, pfm_context_t *ctx, unsigned int cnum, unsigned long *val, struct pt_regs *regs); cnum 405 arch/ia64/kernel/perfmon.c #define PMC_PM(cnum, val) (((val) >> (pmu_conf->pmc_desc[cnum].pm_pos)) & 0x1) cnum 2772 arch/ia64/kernel/perfmon.c unsigned int cnum, reg_flags, flags, pmc_type; cnum 2803 arch/ia64/kernel/perfmon.c cnum = req->reg_num; cnum 2811 arch/ia64/kernel/perfmon.c if (cnum >= PMU_MAX_PMCS) { cnum 2812 arch/ia64/kernel/perfmon.c DPRINT(("pmc%u is invalid\n", cnum)); cnum 2816 arch/ia64/kernel/perfmon.c pmc_type = pmu_conf->pmc_desc[cnum].type; cnum 2817 arch/ia64/kernel/perfmon.c pmc_pm = (value >> pmu_conf->pmc_desc[cnum].pm_pos) & 0x1; cnum 2827 arch/ia64/kernel/perfmon.c DPRINT(("pmc%u is unimplemented or no-access pmc_type=%x\n", cnum, pmc_type)); cnum 2830 arch/ia64/kernel/perfmon.c wr_func = pmu_conf->pmc_desc[cnum].write_check; cnum 2836 arch/ia64/kernel/perfmon.c if (is_monitor && value != PMC_DFL_VAL(cnum) && is_system ^ pmc_pm) { cnum 2838 arch/ia64/kernel/perfmon.c cnum, cnum 2859 arch/ia64/kernel/perfmon.c DPRINT(("invalid smpl_pmds 0x%lx for pmc%u\n", smpl_pmds, cnum)); cnum 2865 arch/ia64/kernel/perfmon.c DPRINT(("invalid reset_pmds 0x%lx for pmc%u\n", reset_pmds, cnum)); cnum 2870 arch/ia64/kernel/perfmon.c DPRINT(("cannot set ovfl_notify or random on pmc%u\n", cnum)); cnum 2880 arch/ia64/kernel/perfmon.c ret = (*wr_func)(task, ctx, cnum, &value, regs); cnum 2901 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].flags = flags; cnum 2903 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].reset_pmds[0] = reset_pmds; cnum 2904 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].smpl_pmds[0] = smpl_pmds; cnum 2905 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].eventid = req->reg_smpl_eventid; cnum 2924 arch/ia64/kernel/perfmon.c if (state == PFM_CTX_MASKED) ctx->ctx_ovfl_regs[0] &= ~1UL << cnum; cnum 2931 arch/ia64/kernel/perfmon.c CTX_USED_PMD(ctx, pmu_conf->pmc_desc[cnum].dep_pmd[0]); cnum 2945 arch/ia64/kernel/perfmon.c if (is_monitor) CTX_USED_MONITOR(ctx, 1UL << cnum); cnum 2950 arch/ia64/kernel/perfmon.c ctx->ctx_pmcs[cnum] = value; cnum 2956 arch/ia64/kernel/perfmon.c if (is_system == 0) ctx->th_pmcs[cnum] = value; cnum 2962 arch/ia64/kernel/perfmon.c ia64_set_pmc(cnum, value); cnum 2973 arch/ia64/kernel/perfmon.c ctx->ctx_reload_pmcs[0] |= 1UL << cnum; cnum 2979 arch/ia64/kernel/perfmon.c cnum, cnum 2986 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].eventid, cnum 3011 arch/ia64/kernel/perfmon.c unsigned int cnum; cnum 3046 arch/ia64/kernel/perfmon.c cnum = req->reg_num; cnum 3049 arch/ia64/kernel/perfmon.c if (!PMD_IS_IMPL(cnum)) { cnum 3050 arch/ia64/kernel/perfmon.c DPRINT(("pmd[%u] is unimplemented or invalid\n", cnum)); cnum 3053 arch/ia64/kernel/perfmon.c is_counting = PMD_IS_COUNTING(cnum); cnum 3054 arch/ia64/kernel/perfmon.c wr_func = pmu_conf->pmd_desc[cnum].write_check; cnum 3062 arch/ia64/kernel/perfmon.c ret = (*wr_func)(task, ctx, cnum, &v, regs); cnum 3086 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].lval = value; cnum 3099 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].long_reset = req->reg_long_reset; cnum 3100 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].short_reset = req->reg_short_reset; cnum 3105 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].seed = req->reg_random_seed; cnum 3106 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].mask = req->reg_random_mask; cnum 3111 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].val = value; cnum 3119 arch/ia64/kernel/perfmon.c CTX_USED_PMD(ctx, PMD_PMD_DEP(cnum)); cnum 3124 arch/ia64/kernel/perfmon.c CTX_USED_PMD(ctx, RDEP(cnum)); cnum 3131 arch/ia64/kernel/perfmon.c ctx->ctx_ovfl_regs[0] &= ~1UL << cnum; cnum 3138 arch/ia64/kernel/perfmon.c if (is_system == 0) ctx->th_pmds[cnum] = hw_value; cnum 3144 arch/ia64/kernel/perfmon.c ia64_set_pmd(cnum, hw_value); cnum 3152 arch/ia64/kernel/perfmon.c ctx->ctx_reload_pmds[0] |= 1UL << cnum; cnum 3159 arch/ia64/kernel/perfmon.c cnum, cnum 3164 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].val, cnum 3165 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].short_reset, cnum 3166 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].long_reset, cnum 3167 arch/ia64/kernel/perfmon.c PMC_OVFL_NOTIFY(ctx, cnum) ? 'Y':'N', cnum 3168 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].seed, cnum 3169 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].mask, cnum 3171 arch/ia64/kernel/perfmon.c ctx->ctx_pmds[cnum].reset_pmds[0], cnum 3207 arch/ia64/kernel/perfmon.c unsigned int cnum, reg_flags = 0; cnum 3257 arch/ia64/kernel/perfmon.c cnum = req->reg_num; cnum 3260 arch/ia64/kernel/perfmon.c if (unlikely(!PMD_IS_IMPL(cnum))) goto error; cnum 3269 arch/ia64/kernel/perfmon.c if (unlikely(!CTX_IS_USED_PMD(ctx, cnum))) goto error; cnum 3271 arch/ia64/kernel/perfmon.c sval = ctx->ctx_pmds[cnum].val; cnum 3272 arch/ia64/kernel/perfmon.c lval = ctx->ctx_pmds[cnum].lval; cnum 3273 arch/ia64/kernel/perfmon.c is_counting = PMD_IS_COUNTING(cnum); cnum 3281 arch/ia64/kernel/perfmon.c val = ia64_get_pmd(cnum); cnum 3288 arch/ia64/kernel/perfmon.c val = is_loaded ? ctx->th_pmds[cnum] : 0UL; cnum 3290 arch/ia64/kernel/perfmon.c rd_func = pmu_conf->pmd_desc[cnum].read_check; cnum 3305 arch/ia64/kernel/perfmon.c ret = (*rd_func)(ctx->ctx_task, ctx, cnum, &v, regs); cnum 3313 arch/ia64/kernel/perfmon.c DPRINT(("pmd[%u]=0x%lx\n", cnum, val)); cnum 4059 arch/ia64/kernel/perfmon.c unsigned int cnum; cnum 4065 arch/ia64/kernel/perfmon.c cnum = req->reg_num; cnum 4067 arch/ia64/kernel/perfmon.c if (!PMC_IS_IMPL(cnum)) goto abort_mission; cnum 4069 arch/ia64/kernel/perfmon.c req->reg_value = PMC_DFL_VAL(cnum); cnum 4073 arch/ia64/kernel/perfmon.c DPRINT(("pmc_reset_val pmc[%u]=0x%lx\n", cnum, req->reg_value)); cnum 9 arch/ia64/kernel/perfmon_itanium.h static int pfm_ita_pmc_check(struct task_struct *task, pfm_context_t *ctx, unsigned int cnum, unsigned long *val, struct pt_regs *regs); cnum 52 arch/ia64/kernel/perfmon_itanium.h pfm_ita_pmc_check(struct task_struct *task, pfm_context_t *ctx, unsigned int cnum, unsigned long *val, struct pt_regs *regs) cnum 66 arch/ia64/kernel/perfmon_itanium.h if (cnum == 13 && is_loaded && ((*val & 0x1) == 0UL) && ctx->ctx_fl_using_dbreg == 0) { cnum 68 arch/ia64/kernel/perfmon_itanium.h DPRINT(("pmc[%d]=0x%lx has active pmc13.ta cleared, clearing ibr\n", cnum, *val)); cnum 85 arch/ia64/kernel/perfmon_itanium.h if (cnum == 11 && is_loaded && ((*val >> 28)& 0x1) == 0 && ctx->ctx_fl_using_dbreg == 0) { cnum 87 arch/ia64/kernel/perfmon_itanium.h DPRINT(("pmc[%d]=0x%lx has active pmc11.pt cleared, clearing dbr\n", cnum, *val)); cnum 9 arch/ia64/kernel/perfmon_mckinley.h static int pfm_mck_pmc_check(struct task_struct *task, pfm_context_t *ctx, unsigned int cnum, unsigned long *val, struct pt_regs *regs); cnum 57 arch/ia64/kernel/perfmon_mckinley.h pfm_mck_reserved(unsigned int cnum, unsigned long *val, struct pt_regs *regs) cnum 62 arch/ia64/kernel/perfmon_mckinley.h tmp1 = ival & PMC_RSVD_MASK(cnum); cnum 65 arch/ia64/kernel/perfmon_mckinley.h tmp2 = PMC_DFL_VAL(cnum) & ~PMC_RSVD_MASK(cnum); cnum 70 arch/ia64/kernel/perfmon_mckinley.h cnum, ival, PMC_RSVD_MASK(cnum), PMC_DFL_VAL(cnum), *val)); cnum 78 arch/ia64/kernel/perfmon_mckinley.h pfm_mck_pmc_check(struct task_struct *task, pfm_context_t *ctx, unsigned int cnum, unsigned long *val, struct pt_regs *regs) cnum 85 arch/ia64/kernel/perfmon_mckinley.h pfm_mck_reserved(cnum, val, regs); cnum 102 arch/ia64/kernel/perfmon_mckinley.h DPRINT(("cnum=%u val=0x%lx, using_dbreg=%d loaded=%d\n", cnum, *val, ctx->ctx_fl_using_dbreg, is_loaded)); cnum 104 arch/ia64/kernel/perfmon_mckinley.h if (cnum == 13 && is_loaded cnum 107 arch/ia64/kernel/perfmon_mckinley.h DPRINT(("pmc[%d]=0x%lx has active pmc13 settings, clearing dbr\n", cnum, *val)); cnum 123 arch/ia64/kernel/perfmon_mckinley.h if (cnum == 14 && is_loaded && ((*val & 0x2222UL) != 0x2222UL) && ctx->ctx_fl_using_dbreg == 0) { cnum 125 arch/ia64/kernel/perfmon_mckinley.h DPRINT(("pmc[%d]=0x%lx has active pmc14 settings, clearing ibr\n", cnum, *val)); cnum 139 arch/ia64/kernel/perfmon_mckinley.h switch(cnum) { cnum 9 arch/ia64/kernel/perfmon_montecito.h static int pfm_mont_pmc_check(struct task_struct *task, pfm_context_t *ctx, unsigned int cnum, unsigned long *val, struct pt_regs *regs); cnum 135 arch/ia64/kernel/perfmon_montecito.h pfm_mont_reserved(unsigned int cnum, unsigned long *val, struct pt_regs *regs) cnum 140 arch/ia64/kernel/perfmon_montecito.h tmp1 = ival & PMC_RSVD_MASK(cnum); cnum 143 arch/ia64/kernel/perfmon_montecito.h tmp2 = PMC_DFL_VAL(cnum) & ~PMC_RSVD_MASK(cnum); cnum 148 arch/ia64/kernel/perfmon_montecito.h cnum, ival, PMC_RSVD_MASK(cnum), PMC_DFL_VAL(cnum), *val)); cnum 156 arch/ia64/kernel/perfmon_montecito.h pfm_mont_pmc_check(struct task_struct *task, pfm_context_t *ctx, unsigned int cnum, unsigned long *val, struct pt_regs *regs) cnum 165 arch/ia64/kernel/perfmon_montecito.h pfm_mont_reserved(cnum, val, regs); cnum 186 arch/ia64/kernel/perfmon_montecito.h DPRINT(("cnum=%u val=0x%lx, using_dbreg=%d loaded=%d\n", cnum, tmpval, ctx->ctx_fl_using_dbreg, is_loaded)); cnum 188 arch/ia64/kernel/perfmon_montecito.h if (cnum == 41 && is_loaded cnum 191 arch/ia64/kernel/perfmon_montecito.h DPRINT(("pmc[%d]=0x%lx has active pmc41 settings, clearing dbr\n", cnum, tmpval)); cnum 209 arch/ia64/kernel/perfmon_montecito.h if (cnum == 38 && is_loaded && ((tmpval & 0x492UL) != 0x492UL) && ctx->ctx_fl_using_dbreg == 0) { cnum 224 arch/ia64/kernel/perfmon_montecito.h switch(cnum) { cnum 120 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c int cnum; cnum 123 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c for (cnum = 0; cnum < tchannels; cnum++) { cnum 124 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c tx_ctl_reg = readl(ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cnum)); cnum 127 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cnum)); cnum 151 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c int cnum; cnum 154 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c for (cnum = 0; cnum < tchannels; cnum++) { cnum 155 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c tx_ctl_reg = readl(ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cnum)); cnum 157 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c writel(tx_ctl_reg, ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cnum)); cnum 163 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c int cnum; cnum 166 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c for (cnum = 0; cnum < rchannels; cnum++) { cnum 167 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c rx_ctl_reg = readl(ioaddr + SXGBE_DMA_CHA_RXCTL_REG(cnum)); cnum 170 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c ioaddr + SXGBE_DMA_CHA_RXCTL_REG(cnum)); cnum 176 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c int cnum; cnum 179 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c for (cnum = 0; cnum < rchannels; cnum++) { cnum 180 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c rx_ctl_reg = readl(ioaddr + SXGBE_DMA_CHA_RXCTL_REG(cnum)); cnum 182 drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c writel(rx_ctl_reg, ioaddr + SXGBE_DMA_CHA_RXCTL_REG(cnum)); cnum 139 drivers/nvme/host/fc.c u32 cnum; cnum 543 drivers/nvme/host/fc.c "Attempting reconnect\n", ctrl->cnum); cnum 763 drivers/nvme/host/fc.c "Reconnect", ctrl->cnum); cnum 778 drivers/nvme/host/fc.c ctrl->cnum); cnum 845 drivers/nvme/host/fc.c ctrl->cnum); cnum 1616 drivers/nvme/host/fc.c ctrl->cnum, freq->status); cnum 1649 drivers/nvme/host/fc.c ctrl->cnum, freq->transferred_length, cnum 1672 drivers/nvme/host/fc.c ctrl->cnum, be16_to_cpu(op->rsp_iu.iu_len), cnum 1689 drivers/nvme/host/fc.c ctrl->cnum, freq->rcv_rsplen); cnum 2040 drivers/nvme/host/fc.c ida_simple_remove(&nvme_fc_ctrl_cnt, ctrl->cnum); cnum 2099 drivers/nvme/host/fc.c ctrl->cnum, errmsg); cnum 2101 drivers/nvme/host/fc.c "NVME-FC{%d}: resetting controller\n", ctrl->cnum); cnum 2640 drivers/nvme/host/fc.c ctrl->cnum, ctrl->lport->localport.port_name, cnum 2878 drivers/nvme/host/fc.c ctrl->cnum, status); cnum 2887 drivers/nvme/host/fc.c ctrl->cnum, recon_delay / HZ); cnum 2897 drivers/nvme/host/fc.c ctrl->cnum, ctrl->ctrl.nr_reconnects); cnum 2902 drivers/nvme/host/fc.c ctrl->cnum, portptr->dev_loss_tmo); cnum 2931 drivers/nvme/host/fc.c "to CONNECTING\n", ctrl->cnum); cnum 2955 drivers/nvme/host/fc.c ctrl->cnum); cnum 3004 drivers/nvme/host/fc.c ctrl->cnum); cnum 3087 drivers/nvme/host/fc.c ctrl->cnum = idx; cnum 3169 drivers/nvme/host/fc.c "NVME-FC{%d}: failed to init ctrl state\n", ctrl->cnum); cnum 3179 drivers/nvme/host/fc.c ctrl->cnum); cnum 3187 drivers/nvme/host/fc.c ctrl->cnum, ctrl->ctrl.opts->subsysnqn); cnum 3226 drivers/nvme/host/fc.c ida_simple_remove(&nvme_fc_ctrl_cnt, ctrl->cnum); cnum 3507 drivers/nvme/host/fc.c ctrl->cnum); cnum 3359 drivers/scsi/aacraid/aachba.c if (qd.cnum == -1) { cnum 3362 drivers/scsi/aacraid/aachba.c qd.cnum = qd.id; cnum 3364 drivers/scsi/aacraid/aachba.c if (qd.cnum < 0 || qd.cnum >= dev->maximum_num_containers) cnum 3368 drivers/scsi/aacraid/aachba.c qd.id = CONTAINER_TO_ID(qd.cnum); cnum 3369 drivers/scsi/aacraid/aachba.c qd.lun = CONTAINER_TO_LUN(qd.cnum); cnum 3373 drivers/scsi/aacraid/aachba.c qd.valid = fsa_dev_ptr[qd.cnum].valid != 0; cnum 3374 drivers/scsi/aacraid/aachba.c qd.locked = fsa_dev_ptr[qd.cnum].locked; cnum 3375 drivers/scsi/aacraid/aachba.c qd.deleted = fsa_dev_ptr[qd.cnum].deleted; cnum 3377 drivers/scsi/aacraid/aachba.c if (fsa_dev_ptr[qd.cnum].devname[0] == '\0') cnum 3382 drivers/scsi/aacraid/aachba.c strlcpy(qd.name, fsa_dev_ptr[qd.cnum].devname, cnum 3383 drivers/scsi/aacraid/aachba.c min(sizeof(qd.name), sizeof(fsa_dev_ptr[qd.cnum].devname) + 1)); cnum 3402 drivers/scsi/aacraid/aachba.c if (dd.cnum >= dev->maximum_num_containers) cnum 3407 drivers/scsi/aacraid/aachba.c fsa_dev_ptr[dd.cnum].deleted = 1; cnum 3411 drivers/scsi/aacraid/aachba.c fsa_dev_ptr[dd.cnum].valid = 0; cnum 3427 drivers/scsi/aacraid/aachba.c if (dd.cnum >= dev->maximum_num_containers) cnum 3432 drivers/scsi/aacraid/aachba.c if (fsa_dev_ptr[dd.cnum].locked) cnum 3438 drivers/scsi/aacraid/aachba.c fsa_dev_ptr[dd.cnum].valid = 0; cnum 3439 drivers/scsi/aacraid/aachba.c fsa_dev_ptr[dd.cnum].devname[0] = '\0'; cnum 2313 drivers/scsi/aacraid/aacraid.h s32 cnum; cnum 2327 drivers/scsi/aacraid/aacraid.h u32 cnum; cnum 130 net/bluetooth/bnep/bnep.h __u32 cnum; cnum 703 net/bluetooth/bnep/core.c if (++n >= req->cnum) cnum 708 net/bluetooth/bnep/core.c req->cnum = n; cnum 104 net/bluetooth/bnep/sock.c if (cl.cnum <= 0) cnum 151 net/bluetooth/bnep/sock.c if (get_user(cl.cnum, p) || get_user(uci, p + 1)) cnum 156 net/bluetooth/bnep/sock.c if (cl.cnum <= 0) cnum 161 net/bluetooth/bnep/sock.c if (!err && put_user(cl.cnum, p)) cnum 57 net/bluetooth/cmtp/cmtp.h __u32 cnum; cnum 466 net/bluetooth/cmtp/core.c if (++n >= req->cnum) cnum 471 net/bluetooth/cmtp/core.c req->cnum = n; cnum 116 net/bluetooth/cmtp/sock.c if (cl.cnum <= 0) cnum 154 net/bluetooth/cmtp/sock.c if (get_user(cl.cnum, p) || get_user(uci, p + 1)) cnum 159 net/bluetooth/cmtp/sock.c if (cl.cnum <= 0) cnum 164 net/bluetooth/cmtp/sock.c if (!err && put_user(cl.cnum, p)) cnum 1436 net/bluetooth/hidp/core.c if (++n >= req->cnum) cnum 1441 net/bluetooth/hidp/core.c req->cnum = n; cnum 121 net/bluetooth/hidp/hidp.h __u32 cnum; cnum 102 net/bluetooth/hidp/sock.c if (cl.cnum <= 0) cnum 157 net/bluetooth/hidp/sock.c if (get_user(cl.cnum, p) || get_user(uci, p + 1)) cnum 162 net/bluetooth/hidp/sock.c if (cl.cnum <= 0) cnum 167 net/bluetooth/hidp/sock.c if (!err && put_user(cl.cnum, p))