chp 85 arch/s390/include/uapi/asm/chsc.h struct chp_id chp; chp 35 drivers/infiniband/hw/cxgb3/iwch_cq.c static int __iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, chp 45 drivers/infiniband/hw/cxgb3/iwch_cq.c ret = cxio_poll_cq(wq, &(chp->cq), &cqe, &cqe_flushed, &cookie, chp 47 drivers/infiniband/hw/cxgb3/iwch_cq.c if (t3a_device(chp->rhp) && credit) { chp 49 drivers/infiniband/hw/cxgb3/iwch_cq.c credit, chp->cq.cqid); chp 50 drivers/infiniband/hw/cxgb3/iwch_cq.c cxio_hal_cq_op(&rhp->rdev, &chp->cq, CQ_CREDIT_UPDATE, credit); chp 175 drivers/infiniband/hw/cxgb3/iwch_cq.c static int iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, chp 182 drivers/infiniband/hw/cxgb3/iwch_cq.c rd_cqe = cxio_next_cqe(&chp->cq); chp 190 drivers/infiniband/hw/cxgb3/iwch_cq.c ret = __iwch_poll_cq_one(rhp, chp, qhp, wc); chp 193 drivers/infiniband/hw/cxgb3/iwch_cq.c ret = __iwch_poll_cq_one(rhp, chp, NULL, wc); chp 201 drivers/infiniband/hw/cxgb3/iwch_cq.c struct iwch_cq *chp; chp 206 drivers/infiniband/hw/cxgb3/iwch_cq.c chp = to_iwch_cq(ibcq); chp 207 drivers/infiniband/hw/cxgb3/iwch_cq.c rhp = chp->rhp; chp 209 drivers/infiniband/hw/cxgb3/iwch_cq.c spin_lock_irqsave(&chp->lock, flags); chp 218 drivers/infiniband/hw/cxgb3/iwch_cq.c err = iwch_poll_cq_one(rhp, chp, wc + npolled); chp 223 drivers/infiniband/hw/cxgb3/iwch_cq.c spin_unlock_irqrestore(&chp->lock, flags); chp 41 drivers/infiniband/hw/cxgb3/iwch_ev.c static void post_qp_event(struct iwch_dev *rnicp, struct iwch_cq *chp, chp 90 drivers/infiniband/hw/cxgb3/iwch_ev.c event.device = chp->ibcq.device; chp 92 drivers/infiniband/hw/cxgb3/iwch_ev.c event.element.cq = &chp->ibcq; chp 99 drivers/infiniband/hw/cxgb3/iwch_ev.c spin_lock_irqsave(&chp->comp_handler_lock, flag); chp 100 drivers/infiniband/hw/cxgb3/iwch_ev.c (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); chp 101 drivers/infiniband/hw/cxgb3/iwch_ev.c spin_unlock_irqrestore(&chp->comp_handler_lock, flag); chp 111 drivers/infiniband/hw/cxgb3/iwch_ev.c struct iwch_cq *chp; chp 118 drivers/infiniband/hw/cxgb3/iwch_ev.c chp = get_chp(rnicp, cqid); chp 120 drivers/infiniband/hw/cxgb3/iwch_ev.c if (!chp || !qhp) { chp 130 drivers/infiniband/hw/cxgb3/iwch_ev.c atomic_inc(&chp->refcnt); chp 146 drivers/infiniband/hw/cxgb3/iwch_ev.c post_qp_event(rnicp, chp, rsp_msg, chp 156 drivers/infiniband/hw/cxgb3/iwch_ev.c post_qp_event(rnicp, chp, rsp_msg, IB_EVENT_QP_REQ_ERR, 1); chp 163 drivers/infiniband/hw/cxgb3/iwch_ev.c post_qp_event(rnicp, chp, rsp_msg, IB_EVENT_QP_REQ_ERR, 1); chp 177 drivers/infiniband/hw/cxgb3/iwch_ev.c spin_lock_irqsave(&chp->comp_handler_lock, flag); chp 178 drivers/infiniband/hw/cxgb3/iwch_ev.c (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); chp 179 drivers/infiniband/hw/cxgb3/iwch_ev.c spin_unlock_irqrestore(&chp->comp_handler_lock, flag); chp 190 drivers/infiniband/hw/cxgb3/iwch_ev.c post_qp_event(rnicp, chp, rsp_msg, IB_EVENT_QP_ACCESS_ERR, 1); chp 197 drivers/infiniband/hw/cxgb3/iwch_ev.c post_qp_event(rnicp, chp, rsp_msg, IB_EVENT_DEVICE_FATAL, 1); chp 217 drivers/infiniband/hw/cxgb3/iwch_ev.c post_qp_event(rnicp, chp, rsp_msg, IB_EVENT_QP_FATAL, 1); chp 223 drivers/infiniband/hw/cxgb3/iwch_ev.c post_qp_event(rnicp, chp, rsp_msg, IB_EVENT_QP_FATAL, 1); chp 227 drivers/infiniband/hw/cxgb3/iwch_ev.c if (atomic_dec_and_test(&chp->refcnt)) chp 228 drivers/infiniband/hw/cxgb3/iwch_ev.c wake_up(&chp->wait); chp 93 drivers/infiniband/hw/cxgb3/iwch_provider.c struct iwch_cq *chp; chp 96 drivers/infiniband/hw/cxgb3/iwch_provider.c chp = to_iwch_cq(ib_cq); chp 98 drivers/infiniband/hw/cxgb3/iwch_provider.c xa_erase_irq(&chp->rhp->cqs, chp->cq.cqid); chp 99 drivers/infiniband/hw/cxgb3/iwch_provider.c atomic_dec(&chp->refcnt); chp 100 drivers/infiniband/hw/cxgb3/iwch_provider.c wait_event(chp->wait, !atomic_read(&chp->refcnt)); chp 102 drivers/infiniband/hw/cxgb3/iwch_provider.c cxio_destroy_cq(&chp->rhp->rdev, &chp->cq); chp 112 drivers/infiniband/hw/cxgb3/iwch_provider.c struct iwch_cq *chp = to_iwch_cq(ibcq); chp 127 drivers/infiniband/hw/cxgb3/iwch_provider.c chp->user_rptr_addr = (u32 __user *)(unsigned long)ureq.user_rptr_addr; chp 145 drivers/infiniband/hw/cxgb3/iwch_provider.c chp->cq.size_log2 = ilog2(entries); chp 147 drivers/infiniband/hw/cxgb3/iwch_provider.c if (cxio_create_cq(&rhp->rdev, &chp->cq, !udata)) chp 150 drivers/infiniband/hw/cxgb3/iwch_provider.c chp->rhp = rhp; chp 151 drivers/infiniband/hw/cxgb3/iwch_provider.c chp->ibcq.cqe = 1 << chp->cq.size_log2; chp 152 drivers/infiniband/hw/cxgb3/iwch_provider.c spin_lock_init(&chp->lock); chp 153 drivers/infiniband/hw/cxgb3/iwch_provider.c spin_lock_init(&chp->comp_handler_lock); chp 154 drivers/infiniband/hw/cxgb3/iwch_provider.c atomic_set(&chp->refcnt, 1); chp 155 drivers/infiniband/hw/cxgb3/iwch_provider.c init_waitqueue_head(&chp->wait); chp 156 drivers/infiniband/hw/cxgb3/iwch_provider.c if (xa_store_irq(&rhp->cqs, chp->cq.cqid, chp, GFP_KERNEL)) { chp 157 drivers/infiniband/hw/cxgb3/iwch_provider.c cxio_destroy_cq(&chp->rhp->rdev, &chp->cq); chp 168 drivers/infiniband/hw/cxgb3/iwch_provider.c iwch_destroy_cq(&chp->ibcq, udata); chp 171 drivers/infiniband/hw/cxgb3/iwch_provider.c uresp.cqid = chp->cq.cqid; chp 172 drivers/infiniband/hw/cxgb3/iwch_provider.c uresp.size_log2 = chp->cq.size_log2; chp 178 drivers/infiniband/hw/cxgb3/iwch_provider.c mm->addr = virt_to_phys(chp->cq.queue); chp 194 drivers/infiniband/hw/cxgb3/iwch_provider.c iwch_destroy_cq(&chp->ibcq, udata); chp 200 drivers/infiniband/hw/cxgb3/iwch_provider.c chp->cq.cqid, chp, (1 << chp->cq.size_log2), chp 201 drivers/infiniband/hw/cxgb3/iwch_provider.c &chp->cq.dma_addr); chp 208 drivers/infiniband/hw/cxgb3/iwch_provider.c struct iwch_cq *chp; chp 214 drivers/infiniband/hw/cxgb3/iwch_provider.c chp = to_iwch_cq(ibcq); chp 215 drivers/infiniband/hw/cxgb3/iwch_provider.c rhp = chp->rhp; chp 220 drivers/infiniband/hw/cxgb3/iwch_provider.c if (chp->user_rptr_addr) { chp 221 drivers/infiniband/hw/cxgb3/iwch_provider.c if (get_user(rptr, chp->user_rptr_addr)) chp 223 drivers/infiniband/hw/cxgb3/iwch_provider.c spin_lock_irqsave(&chp->lock, flag); chp 224 drivers/infiniband/hw/cxgb3/iwch_provider.c chp->cq.rptr = rptr; chp 226 drivers/infiniband/hw/cxgb3/iwch_provider.c spin_lock_irqsave(&chp->lock, flag); chp 227 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s rptr 0x%x\n", __func__, chp->cq.rptr); chp 228 drivers/infiniband/hw/cxgb3/iwch_provider.c err = cxio_hal_cq_op(&rhp->rdev, &chp->cq, cq_op, 0); chp 229 drivers/infiniband/hw/cxgb3/iwch_provider.c spin_unlock_irqrestore(&chp->lock, flag); chp 231 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_err("Error %d rearming CQID 0x%x\n", err, chp->cq.cqid); chp 242 drivers/infiniband/hw/cxgb4/cq.c struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq); chp 243 drivers/infiniband/hw/cxgb4/cq.c struct t4_cq *cq = &chp->cq; chp 335 drivers/infiniband/hw/cxgb4/cq.c void c4iw_flush_hw_cq(struct c4iw_cq *chp, struct c4iw_qp *flush_qhp) chp 342 drivers/infiniband/hw/cxgb4/cq.c pr_debug("cqid 0x%x\n", chp->cq.cqid); chp 343 drivers/infiniband/hw/cxgb4/cq.c ret = t4_next_hw_cqe(&chp->cq, &hw_cqe); chp 351 drivers/infiniband/hw/cxgb4/cq.c qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); chp 407 drivers/infiniband/hw/cxgb4/cq.c flush_completed_wrs(&qhp->wq, &chp->cq); chp 409 drivers/infiniband/hw/cxgb4/cq.c swcqe = &chp->cq.sw_queue[chp->cq.sw_pidx]; chp 412 drivers/infiniband/hw/cxgb4/cq.c t4_swcq_produce(&chp->cq); chp 415 drivers/infiniband/hw/cxgb4/cq.c t4_hwcq_consume(&chp->cq); chp 416 drivers/infiniband/hw/cxgb4/cq.c ret = t4_next_hw_cqe(&chp->cq, &hw_cqe); chp 754 drivers/infiniband/hw/cxgb4/cq.c static int __c4iw_poll_cq_one(struct c4iw_cq *chp, struct c4iw_qp *qhp, chp 764 drivers/infiniband/hw/cxgb4/cq.c ret = poll_cq(wq, &(chp->cq), &cqe, &cqe_flushed, &cookie, &credit, chp 921 drivers/infiniband/hw/cxgb4/cq.c static int c4iw_poll_cq_one(struct c4iw_cq *chp, struct ib_wc *wc) chp 928 drivers/infiniband/hw/cxgb4/cq.c ret = t4_next_cqe(&chp->cq, &rd_cqe); chp 933 drivers/infiniband/hw/cxgb4/cq.c qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); chp 939 drivers/infiniband/hw/cxgb4/cq.c ret = __c4iw_poll_cq_one(chp, qhp, wc, srq); chp 944 drivers/infiniband/hw/cxgb4/cq.c ret = __c4iw_poll_cq_one(chp, NULL, wc, NULL); chp 951 drivers/infiniband/hw/cxgb4/cq.c struct c4iw_cq *chp; chp 956 drivers/infiniband/hw/cxgb4/cq.c chp = to_c4iw_cq(ibcq); chp 958 drivers/infiniband/hw/cxgb4/cq.c spin_lock_irqsave(&chp->lock, flags); chp 961 drivers/infiniband/hw/cxgb4/cq.c err = c4iw_poll_cq_one(chp, wc + npolled); chp 966 drivers/infiniband/hw/cxgb4/cq.c spin_unlock_irqrestore(&chp->lock, flags); chp 972 drivers/infiniband/hw/cxgb4/cq.c struct c4iw_cq *chp; chp 976 drivers/infiniband/hw/cxgb4/cq.c chp = to_c4iw_cq(ib_cq); chp 978 drivers/infiniband/hw/cxgb4/cq.c xa_erase_irq(&chp->rhp->cqs, chp->cq.cqid); chp 979 drivers/infiniband/hw/cxgb4/cq.c atomic_dec(&chp->refcnt); chp 980 drivers/infiniband/hw/cxgb4/cq.c wait_event(chp->wait, !atomic_read(&chp->refcnt)); chp 984 drivers/infiniband/hw/cxgb4/cq.c destroy_cq(&chp->rhp->rdev, &chp->cq, chp 985 drivers/infiniband/hw/cxgb4/cq.c ucontext ? &ucontext->uctx : &chp->cq.rdev->uctx, chp 986 drivers/infiniband/hw/cxgb4/cq.c chp->destroy_skb, chp->wr_waitp); chp 987 drivers/infiniband/hw/cxgb4/cq.c c4iw_put_wr_wait(chp->wr_waitp); chp 997 drivers/infiniband/hw/cxgb4/cq.c struct c4iw_cq *chp = to_c4iw_cq(ibcq); chp 1018 drivers/infiniband/hw/cxgb4/cq.c chp->wr_waitp = c4iw_alloc_wr_wait(GFP_KERNEL); chp 1019 drivers/infiniband/hw/cxgb4/cq.c if (!chp->wr_waitp) { chp 1023 drivers/infiniband/hw/cxgb4/cq.c c4iw_init_wr_wait(chp->wr_waitp); chp 1026 drivers/infiniband/hw/cxgb4/cq.c chp->destroy_skb = alloc_skb(wr_len, GFP_KERNEL); chp 1027 drivers/infiniband/hw/cxgb4/cq.c if (!chp->destroy_skb) { chp 1056 drivers/infiniband/hw/cxgb4/cq.c (sizeof(*chp->cq.queue) / 2) : sizeof(*chp->cq.queue)); chp 1064 drivers/infiniband/hw/cxgb4/cq.c chp->cq.size = hwentries; chp 1065 drivers/infiniband/hw/cxgb4/cq.c chp->cq.memsize = memsize; chp 1066 drivers/infiniband/hw/cxgb4/cq.c chp->cq.vector = vector; chp 1068 drivers/infiniband/hw/cxgb4/cq.c ret = create_cq(&rhp->rdev, &chp->cq, chp 1070 drivers/infiniband/hw/cxgb4/cq.c chp->wr_waitp); chp 1074 drivers/infiniband/hw/cxgb4/cq.c chp->rhp = rhp; chp 1075 drivers/infiniband/hw/cxgb4/cq.c chp->cq.size--; /* status page */ chp 1076 drivers/infiniband/hw/cxgb4/cq.c chp->ibcq.cqe = entries - 2; chp 1077 drivers/infiniband/hw/cxgb4/cq.c spin_lock_init(&chp->lock); chp 1078 drivers/infiniband/hw/cxgb4/cq.c spin_lock_init(&chp->comp_handler_lock); chp 1079 drivers/infiniband/hw/cxgb4/cq.c atomic_set(&chp->refcnt, 1); chp 1080 drivers/infiniband/hw/cxgb4/cq.c init_waitqueue_head(&chp->wait); chp 1081 drivers/infiniband/hw/cxgb4/cq.c ret = xa_insert_irq(&rhp->cqs, chp->cq.cqid, chp, GFP_KERNEL); chp 1096 drivers/infiniband/hw/cxgb4/cq.c uresp.cqid = chp->cq.cqid; chp 1097 drivers/infiniband/hw/cxgb4/cq.c uresp.size = chp->cq.size; chp 1098 drivers/infiniband/hw/cxgb4/cq.c uresp.memsize = chp->cq.memsize; chp 1118 drivers/infiniband/hw/cxgb4/cq.c mm->addr = virt_to_phys(chp->cq.queue); chp 1119 drivers/infiniband/hw/cxgb4/cq.c mm->len = chp->cq.memsize; chp 1123 drivers/infiniband/hw/cxgb4/cq.c mm2->addr = chp->cq.bar2_pa; chp 1129 drivers/infiniband/hw/cxgb4/cq.c chp->cq.cqid, chp, chp->cq.size, chp->cq.memsize, chp 1130 drivers/infiniband/hw/cxgb4/cq.c &chp->cq.dma_addr); chp 1137 drivers/infiniband/hw/cxgb4/cq.c xa_erase_irq(&rhp->cqs, chp->cq.cqid); chp 1139 drivers/infiniband/hw/cxgb4/cq.c destroy_cq(&chp->rhp->rdev, &chp->cq, chp 1141 drivers/infiniband/hw/cxgb4/cq.c chp->destroy_skb, chp->wr_waitp); chp 1143 drivers/infiniband/hw/cxgb4/cq.c kfree_skb(chp->destroy_skb); chp 1145 drivers/infiniband/hw/cxgb4/cq.c c4iw_put_wr_wait(chp->wr_waitp); chp 1152 drivers/infiniband/hw/cxgb4/cq.c struct c4iw_cq *chp; chp 1156 drivers/infiniband/hw/cxgb4/cq.c chp = to_c4iw_cq(ibcq); chp 1157 drivers/infiniband/hw/cxgb4/cq.c spin_lock_irqsave(&chp->lock, flag); chp 1158 drivers/infiniband/hw/cxgb4/cq.c t4_arm_cq(&chp->cq, chp 1161 drivers/infiniband/hw/cxgb4/cq.c ret = t4_cq_notempty(&chp->cq); chp 1162 drivers/infiniband/hw/cxgb4/cq.c spin_unlock_irqrestore(&chp->lock, flag); chp 87 drivers/infiniband/hw/cxgb4/ev.c static void post_qp_event(struct c4iw_dev *dev, struct c4iw_cq *chp, chp 105 drivers/infiniband/hw/cxgb4/ev.c event.device = chp->ibcq.device; chp 107 drivers/infiniband/hw/cxgb4/ev.c event.element.cq = &chp->ibcq; chp 113 drivers/infiniband/hw/cxgb4/ev.c if (t4_clear_cq_armed(&chp->cq)) { chp 114 drivers/infiniband/hw/cxgb4/ev.c spin_lock_irqsave(&chp->comp_handler_lock, flag); chp 115 drivers/infiniband/hw/cxgb4/ev.c (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); chp 116 drivers/infiniband/hw/cxgb4/ev.c spin_unlock_irqrestore(&chp->comp_handler_lock, flag); chp 122 drivers/infiniband/hw/cxgb4/ev.c struct c4iw_cq *chp; chp 142 drivers/infiniband/hw/cxgb4/ev.c chp = get_chp(dev, cqid); chp 143 drivers/infiniband/hw/cxgb4/ev.c if (!chp) { chp 154 drivers/infiniband/hw/cxgb4/ev.c atomic_inc(&chp->refcnt); chp 160 drivers/infiniband/hw/cxgb4/ev.c post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_REQ_ERR); chp 179 drivers/infiniband/hw/cxgb4/ev.c post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_ACCESS_ERR); chp 186 drivers/infiniband/hw/cxgb4/ev.c post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_DEVICE_FATAL); chp 206 drivers/infiniband/hw/cxgb4/ev.c post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_FATAL); chp 212 drivers/infiniband/hw/cxgb4/ev.c post_qp_event(dev, chp, qhp, err_cqe, IB_EVENT_QP_FATAL); chp 216 drivers/infiniband/hw/cxgb4/ev.c if (atomic_dec_and_test(&chp->refcnt)) chp 217 drivers/infiniband/hw/cxgb4/ev.c wake_up(&chp->wait); chp 225 drivers/infiniband/hw/cxgb4/ev.c struct c4iw_cq *chp; chp 229 drivers/infiniband/hw/cxgb4/ev.c chp = xa_load(&dev->cqs, qid); chp 230 drivers/infiniband/hw/cxgb4/ev.c if (chp) { chp 231 drivers/infiniband/hw/cxgb4/ev.c atomic_inc(&chp->refcnt); chp 233 drivers/infiniband/hw/cxgb4/ev.c t4_clear_cq_armed(&chp->cq); chp 234 drivers/infiniband/hw/cxgb4/ev.c spin_lock_irqsave(&chp->comp_handler_lock, flag); chp 235 drivers/infiniband/hw/cxgb4/ev.c (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); chp 236 drivers/infiniband/hw/cxgb4/ev.c spin_unlock_irqrestore(&chp->comp_handler_lock, flag); chp 237 drivers/infiniband/hw/cxgb4/ev.c if (atomic_dec_and_test(&chp->refcnt)) chp 238 drivers/infiniband/hw/cxgb4/ev.c wake_up(&chp->wait); chp 1020 drivers/infiniband/hw/cxgb4/iw_cxgb4.h void c4iw_flush_hw_cq(struct c4iw_cq *chp, struct c4iw_qp *flush_qhp); chp 379 drivers/infiniband/hw/cxgb4/restrack.c struct c4iw_cq *chp = to_c4iw_cq(ibcq); chp 395 drivers/infiniband/hw/cxgb4/restrack.c spin_lock_irq(&chp->lock); chp 398 drivers/infiniband/hw/cxgb4/restrack.c cq = chp->cq; chp 402 drivers/infiniband/hw/cxgb4/restrack.c hwcqes[0] = chp->cq.queue[idx]; chp 405 drivers/infiniband/hw/cxgb4/restrack.c hwcqes[1] = chp->cq.queue[idx]; chp 409 drivers/infiniband/hw/cxgb4/restrack.c swcqes[0] = chp->cq.sw_queue[cq.sw_cidx]; chp 412 drivers/infiniband/hw/cxgb4/restrack.c swcqes[1] = chp->cq.sw_queue[idx]; chp 416 drivers/infiniband/hw/cxgb4/restrack.c spin_unlock_irq(&chp->lock); chp 437 drivers/media/usb/pvrusb2/pvrusb2-dvb.c static void pvr2_dvb_internal_check(struct pvr2_channel *chp) chp 440 drivers/media/usb/pvrusb2/pvrusb2-dvb.c adap = container_of(chp, struct pvr2_dvb_adapter, channel); chp 744 drivers/media/usb/pvrusb2/pvrusb2-sysfs.c static void pvr2_sysfs_internal_check(struct pvr2_channel *chp) chp 747 drivers/media/usb/pvrusb2/pvrusb2-sysfs.c sfp = container_of(chp,struct pvr2_sysfs,channel); chp 893 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c static void pvr2_v4l2_internal_check(struct pvr2_channel *chp) chp 896 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c vp = container_of(chp,struct pvr2_v4l2,channel); chp 3841 drivers/s390/block/dasd.c int chp, oldopm, hpfpm, ifccpm; chp 3848 drivers/s390/block/dasd.c for (chp = 0; chp < 8; chp++) { chp 3849 drivers/s390/block/dasd.c if (path_event[chp] & PE_PATH_GONE) { chp 3850 drivers/s390/block/dasd.c dasd_path_notoper(device, chp); chp 3852 drivers/s390/block/dasd.c if (path_event[chp] & PE_PATH_AVAILABLE) { chp 3853 drivers/s390/block/dasd.c dasd_path_available(device, chp); chp 3856 drivers/s390/block/dasd.c if (path_event[chp] & PE_PATHGROUP_ESTABLISHED) { chp 3857 drivers/s390/block/dasd.c if (!dasd_path_is_operational(device, chp) && chp 3858 drivers/s390/block/dasd.c !dasd_path_need_verify(device, chp)) { chp 3864 drivers/s390/block/dasd.c dasd_path_available(device, chp); chp 896 drivers/s390/block/dasd_int.h static inline int dasd_path_is_operational(struct dasd_device *device, int chp) chp 898 drivers/s390/block/dasd_int.h return test_bit(DASD_PATH_OPERATIONAL, &device->path[chp].flags); chp 901 drivers/s390/block/dasd_int.h static inline int dasd_path_need_verify(struct dasd_device *device, int chp) chp 903 drivers/s390/block/dasd_int.h return test_bit(DASD_PATH_TBV, &device->path[chp].flags); chp 906 drivers/s390/block/dasd_int.h static inline void dasd_path_verify(struct dasd_device *device, int chp) chp 908 drivers/s390/block/dasd_int.h __set_bit(DASD_PATH_TBV, &device->path[chp].flags); chp 911 drivers/s390/block/dasd_int.h static inline void dasd_path_clear_verify(struct dasd_device *device, int chp) chp 913 drivers/s390/block/dasd_int.h __clear_bit(DASD_PATH_TBV, &device->path[chp].flags); chp 918 drivers/s390/block/dasd_int.h int chp; chp 920 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 921 drivers/s390/block/dasd_int.h dasd_path_clear_verify(device, chp); chp 924 drivers/s390/block/dasd_int.h static inline void dasd_path_operational(struct dasd_device *device, int chp) chp 926 drivers/s390/block/dasd_int.h __set_bit(DASD_PATH_OPERATIONAL, &device->path[chp].flags); chp 927 drivers/s390/block/dasd_int.h device->opm |= (0x80 >> chp); chp 930 drivers/s390/block/dasd_int.h static inline void dasd_path_nonpreferred(struct dasd_device *device, int chp) chp 932 drivers/s390/block/dasd_int.h __set_bit(DASD_PATH_NPP, &device->path[chp].flags); chp 935 drivers/s390/block/dasd_int.h static inline int dasd_path_is_nonpreferred(struct dasd_device *device, int chp) chp 937 drivers/s390/block/dasd_int.h return test_bit(DASD_PATH_NPP, &device->path[chp].flags); chp 941 drivers/s390/block/dasd_int.h int chp) chp 943 drivers/s390/block/dasd_int.h __clear_bit(DASD_PATH_NPP, &device->path[chp].flags); chp 946 drivers/s390/block/dasd_int.h static inline void dasd_path_preferred(struct dasd_device *device, int chp) chp 948 drivers/s390/block/dasd_int.h __set_bit(DASD_PATH_PP, &device->path[chp].flags); chp 951 drivers/s390/block/dasd_int.h static inline int dasd_path_is_preferred(struct dasd_device *device, int chp) chp 953 drivers/s390/block/dasd_int.h return test_bit(DASD_PATH_PP, &device->path[chp].flags); chp 957 drivers/s390/block/dasd_int.h int chp) chp 959 drivers/s390/block/dasd_int.h __clear_bit(DASD_PATH_PP, &device->path[chp].flags); chp 962 drivers/s390/block/dasd_int.h static inline void dasd_path_clear_oper(struct dasd_device *device, int chp) chp 964 drivers/s390/block/dasd_int.h __clear_bit(DASD_PATH_OPERATIONAL, &device->path[chp].flags); chp 965 drivers/s390/block/dasd_int.h device->opm &= ~(0x80 >> chp); chp 968 drivers/s390/block/dasd_int.h static inline void dasd_path_clear_cable(struct dasd_device *device, int chp) chp 970 drivers/s390/block/dasd_int.h __clear_bit(DASD_PATH_MISCABLED, &device->path[chp].flags); chp 973 drivers/s390/block/dasd_int.h static inline void dasd_path_cuir(struct dasd_device *device, int chp) chp 975 drivers/s390/block/dasd_int.h __set_bit(DASD_PATH_CUIR, &device->path[chp].flags); chp 978 drivers/s390/block/dasd_int.h static inline int dasd_path_is_cuir(struct dasd_device *device, int chp) chp 980 drivers/s390/block/dasd_int.h return test_bit(DASD_PATH_CUIR, &device->path[chp].flags); chp 983 drivers/s390/block/dasd_int.h static inline void dasd_path_clear_cuir(struct dasd_device *device, int chp) chp 985 drivers/s390/block/dasd_int.h __clear_bit(DASD_PATH_CUIR, &device->path[chp].flags); chp 988 drivers/s390/block/dasd_int.h static inline void dasd_path_ifcc(struct dasd_device *device, int chp) chp 990 drivers/s390/block/dasd_int.h set_bit(DASD_PATH_IFCC, &device->path[chp].flags); chp 993 drivers/s390/block/dasd_int.h static inline int dasd_path_is_ifcc(struct dasd_device *device, int chp) chp 995 drivers/s390/block/dasd_int.h return test_bit(DASD_PATH_IFCC, &device->path[chp].flags); chp 998 drivers/s390/block/dasd_int.h static inline void dasd_path_clear_ifcc(struct dasd_device *device, int chp) chp 1000 drivers/s390/block/dasd_int.h clear_bit(DASD_PATH_IFCC, &device->path[chp].flags); chp 1003 drivers/s390/block/dasd_int.h static inline void dasd_path_clear_nohpf(struct dasd_device *device, int chp) chp 1005 drivers/s390/block/dasd_int.h __clear_bit(DASD_PATH_NOHPF, &device->path[chp].flags); chp 1008 drivers/s390/block/dasd_int.h static inline void dasd_path_miscabled(struct dasd_device *device, int chp) chp 1010 drivers/s390/block/dasd_int.h __set_bit(DASD_PATH_MISCABLED, &device->path[chp].flags); chp 1013 drivers/s390/block/dasd_int.h static inline int dasd_path_is_miscabled(struct dasd_device *device, int chp) chp 1015 drivers/s390/block/dasd_int.h return test_bit(DASD_PATH_MISCABLED, &device->path[chp].flags); chp 1018 drivers/s390/block/dasd_int.h static inline void dasd_path_nohpf(struct dasd_device *device, int chp) chp 1020 drivers/s390/block/dasd_int.h __set_bit(DASD_PATH_NOHPF, &device->path[chp].flags); chp 1023 drivers/s390/block/dasd_int.h static inline int dasd_path_is_nohpf(struct dasd_device *device, int chp) chp 1025 drivers/s390/block/dasd_int.h return test_bit(DASD_PATH_NOHPF, &device->path[chp].flags); chp 1040 drivers/s390/block/dasd_int.h int chp; chp 1043 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1044 drivers/s390/block/dasd_int.h if (dasd_path_need_verify(device, chp)) chp 1045 drivers/s390/block/dasd_int.h tbvpm |= 0x80 >> chp; chp 1051 drivers/s390/block/dasd_int.h int chp; chp 1054 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) { chp 1055 drivers/s390/block/dasd_int.h if (dasd_path_is_nonpreferred(device, chp)) chp 1056 drivers/s390/block/dasd_int.h npm |= 0x80 >> chp; chp 1063 drivers/s390/block/dasd_int.h int chp; chp 1066 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1067 drivers/s390/block/dasd_int.h if (dasd_path_is_preferred(device, chp)) chp 1068 drivers/s390/block/dasd_int.h ppm |= 0x80 >> chp; chp 1074 drivers/s390/block/dasd_int.h int chp; chp 1077 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1078 drivers/s390/block/dasd_int.h if (dasd_path_is_miscabled(device, chp)) chp 1079 drivers/s390/block/dasd_int.h cablepm |= 0x80 >> chp; chp 1085 drivers/s390/block/dasd_int.h int chp; chp 1088 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1089 drivers/s390/block/dasd_int.h if (dasd_path_is_cuir(device, chp)) chp 1090 drivers/s390/block/dasd_int.h cuirpm |= 0x80 >> chp; chp 1096 drivers/s390/block/dasd_int.h int chp; chp 1099 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1100 drivers/s390/block/dasd_int.h if (dasd_path_is_ifcc(device, chp)) chp 1101 drivers/s390/block/dasd_int.h ifccpm |= 0x80 >> chp; chp 1107 drivers/s390/block/dasd_int.h int chp; chp 1110 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1111 drivers/s390/block/dasd_int.h if (dasd_path_is_nohpf(device, chp)) chp 1112 drivers/s390/block/dasd_int.h hpfpm |= 0x80 >> chp; chp 1122 drivers/s390/block/dasd_int.h int chp; chp 1124 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1125 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) chp 1126 drivers/s390/block/dasd_int.h dasd_path_verify(device, chp); chp 1131 drivers/s390/block/dasd_int.h int chp; chp 1134 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1135 drivers/s390/block/dasd_int.h if (dasd_path_is_nohpf(device, chp) || chp 1136 drivers/s390/block/dasd_int.h dasd_path_is_ifcc(device, chp) || chp 1137 drivers/s390/block/dasd_int.h dasd_path_is_cuir(device, chp) || chp 1138 drivers/s390/block/dasd_int.h dasd_path_is_miscabled(device, chp)) chp 1139 drivers/s390/block/dasd_int.h nopm |= 0x80 >> chp; chp 1145 drivers/s390/block/dasd_int.h int chp; chp 1147 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1148 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) { chp 1149 drivers/s390/block/dasd_int.h dasd_path_operational(device, chp); chp 1154 drivers/s390/block/dasd_int.h dasd_path_clear_nohpf(device, chp); chp 1155 drivers/s390/block/dasd_int.h dasd_path_clear_cuir(device, chp); chp 1156 drivers/s390/block/dasd_int.h dasd_path_clear_cable(device, chp); chp 1157 drivers/s390/block/dasd_int.h dasd_path_clear_ifcc(device, chp); chp 1163 drivers/s390/block/dasd_int.h int chp; chp 1165 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1166 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) chp 1167 drivers/s390/block/dasd_int.h dasd_path_miscabled(device, chp); chp 1172 drivers/s390/block/dasd_int.h int chp; chp 1174 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1175 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) chp 1176 drivers/s390/block/dasd_int.h dasd_path_cuir(device, chp); chp 1181 drivers/s390/block/dasd_int.h int chp; chp 1183 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1184 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) chp 1185 drivers/s390/block/dasd_int.h dasd_path_ifcc(device, chp); chp 1190 drivers/s390/block/dasd_int.h int chp; chp 1192 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1193 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) chp 1194 drivers/s390/block/dasd_int.h dasd_path_nonpreferred(device, chp); chp 1199 drivers/s390/block/dasd_int.h int chp; chp 1201 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1202 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) chp 1203 drivers/s390/block/dasd_int.h dasd_path_nohpf(device, chp); chp 1208 drivers/s390/block/dasd_int.h int chp; chp 1210 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1211 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) chp 1212 drivers/s390/block/dasd_int.h dasd_path_preferred(device, chp); chp 1221 drivers/s390/block/dasd_int.h int chp; chp 1223 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1224 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) chp 1225 drivers/s390/block/dasd_int.h dasd_path_verify(device, chp); chp 1227 drivers/s390/block/dasd_int.h dasd_path_clear_verify(device, chp); chp 1232 drivers/s390/block/dasd_int.h int chp; chp 1234 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) { chp 1235 drivers/s390/block/dasd_int.h dasd_path_clear_oper(device, chp); chp 1236 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) { chp 1237 drivers/s390/block/dasd_int.h dasd_path_operational(device, chp); chp 1242 drivers/s390/block/dasd_int.h dasd_path_clear_nohpf(device, chp); chp 1243 drivers/s390/block/dasd_int.h dasd_path_clear_cuir(device, chp); chp 1244 drivers/s390/block/dasd_int.h dasd_path_clear_cable(device, chp); chp 1245 drivers/s390/block/dasd_int.h dasd_path_clear_ifcc(device, chp); chp 1256 drivers/s390/block/dasd_int.h int chp; chp 1258 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) { chp 1259 drivers/s390/block/dasd_int.h if (pm & (0x80 >> chp)) chp 1260 drivers/s390/block/dasd_int.h dasd_path_clear_oper(device, chp); chp 1268 drivers/s390/block/dasd_int.h static inline void dasd_path_available(struct dasd_device *device, int chp) chp 1270 drivers/s390/block/dasd_int.h dasd_path_clear_oper(device, chp); chp 1271 drivers/s390/block/dasd_int.h dasd_path_verify(device, chp); chp 1274 drivers/s390/block/dasd_int.h static inline void dasd_path_notoper(struct dasd_device *device, int chp) chp 1276 drivers/s390/block/dasd_int.h dasd_path_clear_oper(device, chp); chp 1277 drivers/s390/block/dasd_int.h dasd_path_clear_preferred(device, chp); chp 1278 drivers/s390/block/dasd_int.h dasd_path_clear_nonpreferred(device, chp); chp 1286 drivers/s390/block/dasd_int.h int chp; chp 1288 drivers/s390/block/dasd_int.h for (chp = 0; chp < 8; chp++) chp 1289 drivers/s390/block/dasd_int.h dasd_path_notoper(device, chp); chp 331 drivers/s390/cio/ccwreq.c int rc = -ENODEV, chp; chp 336 drivers/s390/cio/ccwreq.c for (chp = 0; chp < 8; chp++) { chp 337 drivers/s390/cio/ccwreq.c if ((0x80 >> chp) & sch->schib.pmcw.lpum) chp 343 drivers/s390/cio/ccwreq.c sch->schib.pmcw.chpid[chp]); chp 135 drivers/s390/cio/chp.c struct channel_path *chp; chp 139 drivers/s390/cio/chp.c chp = to_channelpath(device); chp 140 drivers/s390/cio/chp.c if (chp->cmg == -1) chp 143 drivers/s390/cio/chp.c return memory_read_from_buffer(buf, count, &off, &chp->cmg_chars, chp 144 drivers/s390/cio/chp.c sizeof(chp->cmg_chars)); chp 182 drivers/s390/cio/chp.c struct channel_path *chp; chp 188 drivers/s390/cio/chp.c chp = to_channelpath(device); chp 189 drivers/s390/cio/chp.c css = to_css(chp->dev.parent); chp 196 drivers/s390/cio/chp.c chp_measurement_copy_block((struct cmg_entry *)buf, css, chp->chpid); chp 210 drivers/s390/cio/chp.c void chp_remove_cmg_attr(struct channel_path *chp) chp 212 drivers/s390/cio/chp.c device_remove_bin_file(&chp->dev, &chp_measurement_chars_attr); chp 213 drivers/s390/cio/chp.c device_remove_bin_file(&chp->dev, &chp_measurement_attr); chp 216 drivers/s390/cio/chp.c int chp_add_cmg_attr(struct channel_path *chp) chp 220 drivers/s390/cio/chp.c ret = device_create_bin_file(&chp->dev, &chp_measurement_chars_attr); chp 223 drivers/s390/cio/chp.c ret = device_create_bin_file(&chp->dev, &chp_measurement_attr); chp 225 drivers/s390/cio/chp.c device_remove_bin_file(&chp->dev, &chp_measurement_chars_attr); chp 235 drivers/s390/cio/chp.c struct channel_path *chp = to_channelpath(dev); chp 238 drivers/s390/cio/chp.c mutex_lock(&chp->lock); chp 239 drivers/s390/cio/chp.c status = chp->state; chp 240 drivers/s390/cio/chp.c mutex_unlock(&chp->lock); chp 314 drivers/s390/cio/chp.c struct channel_path *chp = to_channelpath(dev); chp 317 drivers/s390/cio/chp.c mutex_lock(&chp->lock); chp 318 drivers/s390/cio/chp.c type = chp->desc.desc; chp 319 drivers/s390/cio/chp.c mutex_unlock(&chp->lock); chp 328 drivers/s390/cio/chp.c struct channel_path *chp = to_channelpath(dev); chp 330 drivers/s390/cio/chp.c if (!chp) chp 332 drivers/s390/cio/chp.c if (chp->cmg == -1) /* channel measurements not available */ chp 334 drivers/s390/cio/chp.c return sprintf(buf, "%x\n", chp->cmg); chp 342 drivers/s390/cio/chp.c struct channel_path *chp = to_channelpath(dev); chp 344 drivers/s390/cio/chp.c if (!chp) chp 346 drivers/s390/cio/chp.c if (chp->shared == -1) /* channel measurements not available */ chp 348 drivers/s390/cio/chp.c return sprintf(buf, "%x\n", chp->shared); chp 356 drivers/s390/cio/chp.c struct channel_path *chp = to_channelpath(dev); chp 359 drivers/s390/cio/chp.c mutex_lock(&chp->lock); chp 360 drivers/s390/cio/chp.c if (chp->desc_fmt1.flags & 0x10) chp 361 drivers/s390/cio/chp.c rc = sprintf(buf, "%04x\n", chp->desc_fmt1.chid); chp 364 drivers/s390/cio/chp.c mutex_unlock(&chp->lock); chp 373 drivers/s390/cio/chp.c struct channel_path *chp = to_channelpath(dev); chp 376 drivers/s390/cio/chp.c mutex_lock(&chp->lock); chp 377 drivers/s390/cio/chp.c if (chp->desc_fmt1.flags & 0x10) chp 378 drivers/s390/cio/chp.c rc = sprintf(buf, "%x\n", chp->desc_fmt1.flags & 0x8 ? 1 : 0); chp 381 drivers/s390/cio/chp.c mutex_unlock(&chp->lock); chp 391 drivers/s390/cio/chp.c struct channel_path *chp = to_channelpath(kobj_to_dev(kobj)); chp 394 drivers/s390/cio/chp.c mutex_lock(&chp->lock); chp 395 drivers/s390/cio/chp.c rc = memory_read_from_buffer(buf, count, &off, chp->desc_fmt3.util_str, chp 396 drivers/s390/cio/chp.c sizeof(chp->desc_fmt3.util_str)); chp 397 drivers/s390/cio/chp.c mutex_unlock(&chp->lock); chp 444 drivers/s390/cio/chp.c int chp_update_desc(struct channel_path *chp) chp 448 drivers/s390/cio/chp.c rc = chsc_determine_fmt0_channel_path_desc(chp->chpid, &chp->desc); chp 456 drivers/s390/cio/chp.c chsc_determine_fmt1_channel_path_desc(chp->chpid, &chp->desc_fmt1); chp 457 drivers/s390/cio/chp.c chsc_determine_fmt3_channel_path_desc(chp->chpid, &chp->desc_fmt3); chp 458 drivers/s390/cio/chp.c chsc_get_channel_measurement_chars(chp); chp 473 drivers/s390/cio/chp.c struct channel_path *chp; chp 480 drivers/s390/cio/chp.c chp = kzalloc(sizeof(struct channel_path), GFP_KERNEL); chp 481 drivers/s390/cio/chp.c if (!chp) { chp 486 drivers/s390/cio/chp.c chp->chpid = chpid; chp 487 drivers/s390/cio/chp.c chp->state = 1; chp 488 drivers/s390/cio/chp.c chp->dev.parent = &css->device; chp 489 drivers/s390/cio/chp.c chp->dev.groups = chp_attr_groups; chp 490 drivers/s390/cio/chp.c chp->dev.release = chp_release; chp 491 drivers/s390/cio/chp.c mutex_init(&chp->lock); chp 494 drivers/s390/cio/chp.c ret = chp_update_desc(chp); chp 497 drivers/s390/cio/chp.c if ((chp->desc.flags & 0x80) == 0) { chp 501 drivers/s390/cio/chp.c dev_set_name(&chp->dev, "chp%x.%02x", chpid.cssid, chpid.id); chp 504 drivers/s390/cio/chp.c ret = device_register(&chp->dev); chp 508 drivers/s390/cio/chp.c put_device(&chp->dev); chp 513 drivers/s390/cio/chp.c ret = chp_add_cmg_attr(chp); chp 515 drivers/s390/cio/chp.c device_unregister(&chp->dev); chp 519 drivers/s390/cio/chp.c css->chps[chpid.id] = chp; chp 522 drivers/s390/cio/chp.c kfree(chp); chp 537 drivers/s390/cio/chp.c struct channel_path *chp; chp 540 drivers/s390/cio/chp.c chp = chpid_to_chp(chpid); chp 541 drivers/s390/cio/chp.c if (!chp) chp 547 drivers/s390/cio/chp.c mutex_lock(&chp->lock); chp 548 drivers/s390/cio/chp.c memcpy(desc, &chp->desc, sizeof(*desc)); chp 549 drivers/s390/cio/chp.c mutex_unlock(&chp->lock); chp 66 drivers/s390/cio/chp.h void chp_remove_cmg_attr(struct channel_path *chp); chp 67 drivers/s390/cio/chp.h int chp_add_cmg_attr(struct channel_path *chp); chp 68 drivers/s390/cio/chp.h int chp_update_desc(struct channel_path *chp); chp 231 drivers/s390/cio/chsc.c struct channel_path *chp = chpid_to_chp(chpid); chp 245 drivers/s390/cio/chsc.c mutex_lock(&chp->lock); chp 246 drivers/s390/cio/chsc.c chp_update_desc(chp); chp 247 drivers/s390/cio/chsc.c mutex_unlock(&chp->lock); chp 426 drivers/s390/cio/chsc.c struct channel_path *chp; chp 445 drivers/s390/cio/chsc.c chp = chpid_to_chp(chpid); chp 446 drivers/s390/cio/chsc.c mutex_lock(&chp->lock); chp 447 drivers/s390/cio/chsc.c chp_update_desc(chp); chp 448 drivers/s390/cio/chsc.c mutex_unlock(&chp->lock); chp 466 drivers/s390/cio/chsc.c struct channel_path *chp; chp 483 drivers/s390/cio/chsc.c chp = chpid_to_chp(chpid); chp 484 drivers/s390/cio/chsc.c if (!chp) { chp 488 drivers/s390/cio/chsc.c mutex_lock(&chp->lock); chp 489 drivers/s390/cio/chsc.c chp_update_desc(chp); chp 490 drivers/s390/cio/chsc.c mutex_unlock(&chp->lock); chp 693 drivers/s390/cio/chsc.c struct channel_path *chp = chpid_to_chp(chpid); chp 706 drivers/s390/cio/chsc.c mutex_lock(&chp->lock); chp 707 drivers/s390/cio/chsc.c chp_update_desc(chp); chp 708 drivers/s390/cio/chsc.c mutex_unlock(&chp->lock); chp 754 drivers/s390/cio/chsc.c struct channel_path *chp = chpid_to_chp(chpid); chp 763 drivers/s390/cio/chsc.c chp_update_desc(chp); chp 962 drivers/s390/cio/chsc.c chsc_initialize_cmg_chars(struct channel_path *chp, u8 cmcv, chp 970 drivers/s390/cio/chsc.c chp->cmg_chars.values[i] = chars->values[i]; chp 972 drivers/s390/cio/chsc.c chp->cmg_chars.values[i] = 0; chp 976 drivers/s390/cio/chsc.c int chsc_get_channel_measurement_chars(struct channel_path *chp) chp 1002 drivers/s390/cio/chsc.c chp->shared = -1; chp 1003 drivers/s390/cio/chsc.c chp->cmg = -1; chp 1013 drivers/s390/cio/chsc.c scmc_area->first_chpid = chp->chpid.id; chp 1014 drivers/s390/cio/chsc.c scmc_area->last_chpid = chp->chpid.id; chp 1031 drivers/s390/cio/chsc.c chp->cmg = scmc_area->cmg; chp 1032 drivers/s390/cio/chsc.c chp->shared = scmc_area->shared; chp 1033 drivers/s390/cio/chsc.c if (chp->cmg != 2 && chp->cmg != 3) { chp 1037 drivers/s390/cio/chsc.c chsc_initialize_cmg_chars(chp, scmc_area->cmcv, chp 163 drivers/s390/cio/chsc.h int chsc_get_channel_measurement_chars(struct channel_path *chp); chp 755 drivers/s390/cio/chsc_sch.c chpid_parm->cssid = ccl->req.chpid.chp.cssid; chp 756 drivers/s390/cio/chsc_sch.c chpid_parm->chpid = ccl->req.chpid.chp.id; chp 403 drivers/s390/cio/css.c int chp; chp 405 drivers/s390/cio/css.c for (chp = 0; chp < 8; chp++) { chp 406 drivers/s390/cio/css.c mask = 0x80 >> chp; chp 408 drivers/s390/cio/css.c ret += sprintf(buf + ret, "%02x ", ssd->chpid[chp].id); chp 1350 drivers/s390/cio/css.c struct channel_path *chp; chp 1355 drivers/s390/cio/css.c chp = chpid_to_chp(chpid); chp 1356 drivers/s390/cio/css.c if (chp) chp 1357 drivers/s390/cio/css.c chp_update_desc(chp); chp 443 drivers/s390/cio/device_fsm.c int chp, mask; chp 445 drivers/s390/cio/device_fsm.c for (chp = 0, mask = 0x80; chp < 8; chp++, mask >>= 1) { chp 446 drivers/s390/cio/device_fsm.c path_event[chp] = PE_NONE; chp 448 drivers/s390/cio/device_fsm.c path_event[chp] |= PE_PATH_GONE; chp 450 drivers/s390/cio/device_fsm.c path_event[chp] |= PE_PATH_AVAILABLE; chp 452 drivers/s390/cio/device_fsm.c path_event[chp] |= PE_PATHGROUP_ESTABLISHED; chp 497 drivers/s390/cio/device_ops.c struct channel_path *chp; chp 503 drivers/s390/cio/device_ops.c chp = chpid_to_chp(chpid); chp 505 drivers/s390/cio/device_ops.c util_str = kmalloc(sizeof(chp->desc_fmt3.util_str), GFP_KERNEL); chp 509 drivers/s390/cio/device_ops.c mutex_lock(&chp->lock); chp 510 drivers/s390/cio/device_ops.c memcpy(util_str, chp->desc_fmt3.util_str, sizeof(chp->desc_fmt3.util_str)); chp 511 drivers/s390/cio/device_ops.c mutex_unlock(&chp->lock); chp 643 drivers/s390/cio/device_ops.c struct channel_path *chp; chp 658 drivers/s390/cio/device_ops.c chp = chpid_to_chp(chpid); chp 659 drivers/s390/cio/device_ops.c if (!chp) chp 662 drivers/s390/cio/device_ops.c mutex_lock(&chp->lock); chp 663 drivers/s390/cio/device_ops.c if (!chp->desc_fmt1.f) { chp 664 drivers/s390/cio/device_ops.c mutex_unlock(&chp->lock); chp 667 drivers/s390/cio/device_ops.c if (!chp->desc_fmt1.r) chp 669 drivers/s390/cio/device_ops.c mdc = mdc ? min_t(int, mdc, chp->desc_fmt1.mdc) : chp 670 drivers/s390/cio/device_ops.c chp->desc_fmt1.mdc; chp 671 drivers/s390/cio/device_ops.c mutex_unlock(&chp->lock); chp 88 include/sound/vx_core.h void (*set_clock_source)(struct vx_core *chp, int src); chp 65 lib/earlycpio.c unsigned int ch[C_NFIELDS], *chp, v; chp 81 lib/earlycpio.c chp = ch; chp 102 lib/earlycpio.c *chp++ = v;