/linux-4.4.14/fs/xfs/ |
H A D | xfs_icreate_item.c | 63 struct xfs_icreate_item *icp = ICR_ITEM(lip); xfs_icreate_item_format() local 67 &icp->ic_format, xfs_icreate_item_format() 92 struct xfs_icreate_item *icp = ICR_ITEM(lip); xfs_icreate_item_unlock() local 94 if (icp->ic_item.li_flags & XFS_LI_ABORTED) xfs_icreate_item_unlock() 95 kmem_zone_free(xfs_icreate_zone, icp); xfs_icreate_item_unlock() 109 struct xfs_icreate_item *icp = ICR_ITEM(lip); xfs_icreate_item_committed() local 111 kmem_zone_free(xfs_icreate_zone, icp); xfs_icreate_item_committed() 169 struct xfs_icreate_item *icp; xfs_icreate_log() local 171 icp = kmem_zone_zalloc(xfs_icreate_zone, KM_SLEEP); xfs_icreate_log() 173 xfs_log_item_init(tp->t_mountp, &icp->ic_item, XFS_LI_ICREATE, xfs_icreate_log() 176 icp->ic_format.icl_type = XFS_LI_ICREATE; xfs_icreate_log() 177 icp->ic_format.icl_size = 1; /* single vector */ xfs_icreate_log() 178 icp->ic_format.icl_ag = cpu_to_be32(agno); xfs_icreate_log() 179 icp->ic_format.icl_agbno = cpu_to_be32(agbno); xfs_icreate_log() 180 icp->ic_format.icl_count = cpu_to_be32(count); xfs_icreate_log() 181 icp->ic_format.icl_isize = cpu_to_be32(inode_size); xfs_icreate_log() 182 icp->ic_format.icl_length = cpu_to_be32(length); xfs_icreate_log() 183 icp->ic_format.icl_gen = cpu_to_be32(generation); xfs_icreate_log() 185 xfs_trans_add_item(tp, &icp->ic_item); xfs_icreate_log() 187 icp->ic_item.li_desc->lid_flags |= XFS_LID_DIRTY; xfs_icreate_log()
|
/linux-4.4.14/arch/powerpc/kvm/ |
H A D | book3s_hv_rm_xics.c | 26 static void icp_rm_deliver_irq(struct kvmppc_xics *xics, struct kvmppc_icp *icp, 31 struct kvmppc_ics *ics, struct kvmppc_icp *icp) ics_rm_check_resend() 44 icp_rm_deliver_irq(xics, icp, state->number); ics_rm_check_resend() 56 struct kvmppc_icp *this_icp = this_vcpu->arch.icp; icp_rm_set_vcpu_irq() 89 static inline bool icp_rm_try_update(struct kvmppc_icp *icp, icp_rm_try_update() argument 100 success = cmpxchg64(&icp->state.raw, old.raw, new.raw) == old.raw; icp_rm_try_update() 120 icp_rm_set_vcpu_irq(icp->vcpu, this_vcpu); icp_rm_try_update() 123 this_vcpu->arch.icp->rm_dbgstate = new; icp_rm_try_update() 124 this_vcpu->arch.icp->rm_dbgtgt = icp->vcpu; icp_rm_try_update() 131 struct kvmppc_icp *icp) check_too_hard() 133 return (xics->real_mode_dbg || icp->rm_action) ? H_TOO_HARD : H_SUCCESS; check_too_hard() 137 struct kvmppc_icp *icp) icp_rm_check_resend() 143 for_each_set_bit(icsid, icp->resend_map, xics->max_icsid + 1) { icp_rm_check_resend() 146 if (!test_and_clear_bit(icsid, icp->resend_map)) icp_rm_check_resend() 150 ics_rm_check_resend(xics, ics, icp); icp_rm_check_resend() 154 static bool icp_rm_try_to_deliver(struct kvmppc_icp *icp, u32 irq, u8 priority, icp_rm_try_to_deliver() argument 161 old_state = new_state = READ_ONCE(icp->state); icp_rm_try_to_deliver() 187 } while (!icp_rm_try_update(icp, old_state, new_state)); icp_rm_try_to_deliver() 192 static void icp_rm_deliver_irq(struct kvmppc_xics *xics, struct kvmppc_icp *icp, icp_rm_deliver_irq() argument 229 if (!icp || state->server != icp->server_num) { icp_rm_deliver_irq() 230 icp = kvmppc_xics_find_server(xics->kvm, state->server); icp_rm_deliver_irq() 231 if (!icp) { icp_rm_deliver_irq() 277 if (icp_rm_try_to_deliver(icp, new_irq, state->priority, &reject)) { icp_rm_deliver_irq() 291 set_bit(ics->icsid, icp->resend_map); icp_rm_deliver_irq() 301 if (!icp->state.need_resend) { icp_rm_deliver_irq() 310 static void icp_rm_down_cppr(struct kvmppc_xics *xics, struct kvmppc_icp *icp, icp_rm_down_cppr() argument 346 old_state = new_state = READ_ONCE(icp->state); icp_rm_down_cppr() 370 } while (!icp_rm_try_update(icp, old_state, new_state)); icp_rm_down_cppr() 378 icp->n_check_resend++; icp_rm_down_cppr() 379 icp_rm_check_resend(xics, icp); icp_rm_down_cppr() 388 struct kvmppc_icp *icp = vcpu->arch.icp; kvmppc_rm_h_xirr() local 395 icp_rm_clr_vcpu_irq(icp->vcpu); kvmppc_rm_h_xirr() 405 old_state = new_state = READ_ONCE(icp->state); kvmppc_rm_h_xirr() 414 } while (!icp_rm_try_update(icp, old_state, new_state)); kvmppc_rm_h_xirr() 419 return check_too_hard(xics, icp); kvmppc_rm_h_xirr() 427 struct kvmppc_icp *icp, *this_icp = vcpu->arch.icp; kvmppc_rm_h_ipi() local 437 icp = this_icp; kvmppc_rm_h_ipi() 439 icp = kvmppc_xics_find_server(vcpu->kvm, server); kvmppc_rm_h_ipi() 440 if (!icp) kvmppc_rm_h_ipi() 471 old_state = new_state = READ_ONCE(icp->state); kvmppc_rm_h_ipi() 492 } while (!icp_rm_try_update(icp, old_state, new_state)); kvmppc_rm_h_ipi() 497 icp_rm_deliver_irq(xics, icp, reject); kvmppc_rm_h_ipi() 503 icp_rm_check_resend(xics, icp); kvmppc_rm_h_ipi() 513 struct kvmppc_icp *icp = vcpu->arch.icp; kvmppc_rm_h_cppr() local 526 if (cppr > icp->state.cppr) { kvmppc_rm_h_cppr() 527 icp_rm_down_cppr(xics, icp, cppr); kvmppc_rm_h_cppr() 529 } else if (cppr == icp->state.cppr) kvmppc_rm_h_cppr() 543 icp_rm_clr_vcpu_irq(icp->vcpu); kvmppc_rm_h_cppr() 546 old_state = new_state = READ_ONCE(icp->state); kvmppc_rm_h_cppr() 557 } while (!icp_rm_try_update(icp, old_state, new_state)); kvmppc_rm_h_cppr() 564 icp->n_reject++; kvmppc_rm_h_cppr() 565 icp_rm_deliver_irq(xics, icp, reject); kvmppc_rm_h_cppr() 568 return check_too_hard(xics, icp); kvmppc_rm_h_cppr() 574 struct kvmppc_icp *icp = vcpu->arch.icp; kvmppc_rm_h_eoi() local 597 icp_rm_down_cppr(xics, icp, xirr >> 24); kvmppc_rm_h_eoi() 615 icp->n_reject++; kvmppc_rm_h_eoi() 616 icp_rm_deliver_irq(xics, icp, irq); kvmppc_rm_h_eoi() 620 icp->rm_action |= XICS_RM_NOTIFY_EOI; kvmppc_rm_h_eoi() 621 icp->rm_eoied_irq = irq; kvmppc_rm_h_eoi() 624 return check_too_hard(xics, icp); kvmppc_rm_h_eoi() 30 ics_rm_check_resend(struct kvmppc_xics *xics, struct kvmppc_ics *ics, struct kvmppc_icp *icp) ics_rm_check_resend() argument 130 check_too_hard(struct kvmppc_xics *xics, struct kvmppc_icp *icp) check_too_hard() argument 136 icp_rm_check_resend(struct kvmppc_xics *xics, struct kvmppc_icp *icp) icp_rm_check_resend() argument
|
H A D | book3s_xics.c | 65 static void icp_deliver_irq(struct kvmppc_xics *xics, struct kvmppc_icp *icp, 109 struct kvmppc_icp *icp) ics_check_resend() 129 icp_deliver_irq(xics, icp, state->number); ics_check_resend() 166 struct kvmppc_icp *icp; kvmppc_xics_set_xive() local 179 icp = kvmppc_xics_find_server(kvm, server); kvmppc_xics_set_xive() 180 if (!icp) kvmppc_xics_set_xive() 188 icp_deliver_irq(xics, icp, irq); kvmppc_xics_set_xive() 222 struct kvmppc_icp *icp; kvmppc_xics_int_on() local 235 icp = kvmppc_xics_find_server(kvm, state->server); kvmppc_xics_int_on() 236 if (!icp) kvmppc_xics_int_on() 241 icp_deliver_irq(xics, icp, irq); kvmppc_xics_int_on() 268 static inline bool icp_try_update(struct kvmppc_icp *icp, icp_try_update() argument 279 success = cmpxchg64(&icp->state.raw, old.raw, new.raw) == old.raw; icp_try_update() 284 icp->server_num, icp_try_update() 306 kvmppc_book3s_queue_irqprio(icp->vcpu, icp_try_update() 309 kvmppc_fast_vcpu_kick(icp->vcpu); icp_try_update() 316 struct kvmppc_icp *icp) icp_check_resend() 322 for_each_set_bit(icsid, icp->resend_map, xics->max_icsid + 1) { icp_check_resend() 325 if (!test_and_clear_bit(icsid, icp->resend_map)) icp_check_resend() 329 ics_check_resend(xics, ics, icp); icp_check_resend() 333 static bool icp_try_to_deliver(struct kvmppc_icp *icp, u32 irq, u8 priority, icp_try_to_deliver() argument 340 icp->server_num); icp_try_to_deliver() 343 old_state = new_state = READ_ONCE(icp->state); icp_try_to_deliver() 369 } while (!icp_try_update(icp, old_state, new_state, false)); icp_try_to_deliver() 374 static void icp_deliver_irq(struct kvmppc_xics *xics, struct kvmppc_icp *icp, icp_deliver_irq() argument 412 if (!icp || state->server != icp->server_num) { icp_deliver_irq() 413 icp = kvmppc_xics_find_server(xics->kvm, state->server); icp_deliver_irq() 414 if (!icp) { icp_deliver_irq() 461 if (icp_try_to_deliver(icp, new_irq, state->priority, &reject)) { icp_deliver_irq() 476 set_bit(ics->icsid, icp->resend_map); icp_deliver_irq() 486 if (!icp->state.need_resend) { icp_deliver_irq() 497 static void icp_down_cppr(struct kvmppc_xics *xics, struct kvmppc_icp *icp, icp_down_cppr() argument 533 old_state = new_state = READ_ONCE(icp->state); icp_down_cppr() 559 } while (!icp_try_update(icp, old_state, new_state, true)); icp_down_cppr() 567 icp_check_resend(xics, icp); icp_down_cppr() 573 struct kvmppc_icp *icp = vcpu->arch.icp; kvmppc_h_xirr() local 577 kvmppc_book3s_dequeue_irqprio(icp->vcpu, kvmppc_h_xirr() 588 old_state = new_state = READ_ONCE(icp->state); kvmppc_h_xirr() 597 } while (!icp_try_update(icp, old_state, new_state, true)); kvmppc_h_xirr() 609 struct kvmppc_icp *icp; kvmppc_h_ipi() local 617 icp = vcpu->arch.icp; kvmppc_h_ipi() 618 local = icp->server_num == server; kvmppc_h_ipi() 620 icp = kvmppc_xics_find_server(vcpu->kvm, server); kvmppc_h_ipi() 621 if (!icp) kvmppc_h_ipi() 655 old_state = new_state = READ_ONCE(icp->state); kvmppc_h_ipi() 676 } while (!icp_try_update(icp, old_state, new_state, local)); kvmppc_h_ipi() 680 icp_deliver_irq(xics, icp, reject); kvmppc_h_ipi() 684 icp_check_resend(xics, icp); kvmppc_h_ipi() 692 struct kvmppc_icp *icp; kvmppc_h_ipoll() local 694 icp = vcpu->arch.icp; kvmppc_h_ipoll() 695 if (icp->server_num != server) { kvmppc_h_ipoll() 696 icp = kvmppc_xics_find_server(vcpu->kvm, server); kvmppc_h_ipoll() 697 if (!icp) kvmppc_h_ipoll() 700 state = READ_ONCE(icp->state); kvmppc_h_ipoll() 710 struct kvmppc_icp *icp = vcpu->arch.icp; kvmppc_h_cppr() local 722 if (cppr > icp->state.cppr) kvmppc_h_cppr() 723 icp_down_cppr(xics, icp, cppr); kvmppc_h_cppr() 724 else if (cppr == icp->state.cppr) kvmppc_h_cppr() 738 kvmppc_book3s_dequeue_irqprio(icp->vcpu, kvmppc_h_cppr() 742 old_state = new_state = READ_ONCE(icp->state); kvmppc_h_cppr() 753 } while (!icp_try_update(icp, old_state, new_state, true)); kvmppc_h_cppr() 760 icp_deliver_irq(xics, icp, reject); kvmppc_h_cppr() 766 struct kvmppc_icp *icp = vcpu->arch.icp; kvmppc_h_eoi() local 788 icp_down_cppr(xics, icp, xirr >> 24); kvmppc_h_eoi() 808 icp_deliver_irq(xics, icp, irq); kvmppc_h_eoi() 818 struct kvmppc_icp *icp = vcpu->arch.icp; kvmppc_xics_rm_complete() local 821 hcall, icp->rm_action, icp->rm_dbgstate.raw, icp->rm_dbgtgt); kvmppc_xics_rm_complete() 823 if (icp->rm_action & XICS_RM_KICK_VCPU) { kvmppc_xics_rm_complete() 824 icp->n_rm_kick_vcpu++; kvmppc_xics_rm_complete() 825 kvmppc_fast_vcpu_kick(icp->rm_kick_target); kvmppc_xics_rm_complete() 827 if (icp->rm_action & XICS_RM_CHECK_RESEND) { kvmppc_xics_rm_complete() 828 icp->n_rm_check_resend++; kvmppc_xics_rm_complete() 829 icp_check_resend(xics, icp->rm_resend_icp); kvmppc_xics_rm_complete() 831 if (icp->rm_action & XICS_RM_REJECT) { kvmppc_xics_rm_complete() 832 icp->n_rm_reject++; kvmppc_xics_rm_complete() 833 icp_deliver_irq(xics, icp, icp->rm_reject); kvmppc_xics_rm_complete() 835 if (icp->rm_action & XICS_RM_NOTIFY_EOI) { kvmppc_xics_rm_complete() 836 icp->n_rm_notify_eoi++; kvmppc_xics_rm_complete() 837 kvm_notify_acked_irq(vcpu->kvm, 0, icp->rm_eoied_irq); kvmppc_xics_rm_complete() 840 icp->rm_action = 0; kvmppc_xics_rm_complete() 852 if (!xics || !vcpu->arch.icp) kvmppc_xics_hcall() 919 struct kvmppc_icp *icp = vcpu->arch.icp; kvm_for_each_vcpu() local 922 if (!icp) kvm_for_each_vcpu() 925 state.raw = READ_ONCE(icp->state.raw); kvm_for_each_vcpu() 927 icp->server_num, state.xisr, kvm_for_each_vcpu() 930 t_rm_kick_vcpu += icp->n_rm_kick_vcpu; kvm_for_each_vcpu() 931 t_rm_notify_eoi += icp->n_rm_notify_eoi; kvm_for_each_vcpu() 932 t_rm_check_resend += icp->n_rm_check_resend; kvm_for_each_vcpu() 933 t_rm_reject += icp->n_rm_reject; kvm_for_each_vcpu() 934 t_check_resend += icp->n_check_resend; kvm_for_each_vcpu() 935 t_reject += icp->n_reject; kvm_for_each_vcpu() 1038 struct kvmppc_icp *icp; kvmppc_xics_create_icp() local 1046 icp = kzalloc(sizeof(struct kvmppc_icp), GFP_KERNEL); kvmppc_xics_create_icp() 1047 if (!icp) kvmppc_xics_create_icp() 1050 icp->vcpu = vcpu; kvmppc_xics_create_icp() 1051 icp->server_num = server_num; kvmppc_xics_create_icp() 1052 icp->state.mfrr = MASKED; kvmppc_xics_create_icp() 1053 icp->state.pending_pri = MASKED; kvmppc_xics_create_icp() 1054 vcpu->arch.icp = icp; kvmppc_xics_create_icp() 1063 struct kvmppc_icp *icp = vcpu->arch.icp; kvmppc_xics_get_icp() local 1066 if (!icp) kvmppc_xics_get_icp() 1068 state = icp->state; kvmppc_xics_get_icp() 1077 struct kvmppc_icp *icp = vcpu->arch.icp; kvmppc_xics_set_icp() local 1086 if (!icp || !xics) kvmppc_xics_set_icp() 1120 kvmppc_book3s_dequeue_irqprio(icp->vcpu, kvmppc_xics_set_icp() 1134 old_state = READ_ONCE(icp->state); kvmppc_xics_set_icp() 1143 } while (!icp_try_update(icp, old_state, new_state, false)); kvmppc_xics_set_icp() 1146 icp_check_resend(xics, icp); kvmppc_xics_set_icp() 1383 if (!vcpu->arch.icp) kvmppc_xics_free_icp() 1385 kfree(vcpu->arch.icp); kvmppc_xics_free_icp() 1386 vcpu->arch.icp = NULL; kvmppc_xics_free_icp() 108 ics_check_resend(struct kvmppc_xics *xics, struct kvmppc_ics *ics, struct kvmppc_icp *icp) ics_check_resend() argument 315 icp_check_resend(struct kvmppc_xics *xics, struct kvmppc_icp *icp) icp_check_resend() argument
|
H A D | book3s_xics.h | 120 if (vcpu->arch.icp && nr == vcpu->arch.icp->server_num) kvm_for_each_vcpu() 121 return vcpu->arch.icp; kvm_for_each_vcpu()
|
H A D | book3s.c | 572 if (!vcpu->arch.icp) { kvmppc_get_one_reg() 645 if (!vcpu->arch.icp) { kvmppc_set_one_reg()
|
/linux-4.4.14/drivers/net/slip/ |
H A D | slhc.c | 220 * icp and isize are the original packet. 222 * cpp is initially a pointer to icp. If the copy is used, 227 slhc_compress(struct slcompress *comp, unsigned char *icp, int isize, slhc_compress() argument 250 ip = (struct iphdr *) icp; slhc_compress() 460 memcpy(cp+deltaS,icp+hlen,isize-hlen); slhc_compress() 478 memcpy(ocp, icp, isize); slhc_compress() 487 slhc_uncompress(struct slcompress *comp, unsigned char *icp, int isize) slhc_uncompress() argument 495 unsigned char *cp = icp; slhc_uncompress() 597 len = isize - (cp - icp); slhc_uncompress() 604 memmove(icp + hdrlen, cp, len - hdrlen); slhc_uncompress() 606 cp = icp; slhc_uncompress() 615 put_unaligned(ip_fast_csum(icp, ip->ihl), slhc_uncompress() 616 &((struct iphdr *)icp)->check); slhc_uncompress() 634 slhc_remember(struct slcompress *comp, unsigned char *icp, int isize) slhc_remember() argument 647 ihl = icp[0] & 0xf; slhc_remember() 653 index = icp[9]; slhc_remember() 654 icp[9] = IPPROTO_TCP; slhc_remember() 656 if (ip_fast_csum(icp, ihl)) { slhc_remember() 669 memcpy(&cs->cs_ip,icp,20); slhc_remember() 670 memcpy(&cs->cs_tcp,icp + ihl*4,20); slhc_remember() 672 memcpy(cs->cs_ipopt, icp + sizeof(struct iphdr), (ihl - 5) * 4); slhc_remember() 674 memcpy(cs->cs_tcpopt, icp + ihl*4 + sizeof(struct tcphdr), (cs->cs_tcp.doff - 5) * 4); slhc_remember() 702 slhc_uncompress(struct slcompress *comp, unsigned char *icp, int isize) slhc_uncompress() argument 708 slhc_compress(struct slcompress *comp, unsigned char *icp, int isize, slhc_compress() argument 716 slhc_remember(struct slcompress *comp, unsigned char *icp, int isize) slhc_remember() argument
|
H A D | slip.c | 375 static void sl_encaps(struct slip *sl, unsigned char *icp, int len) sl_encaps() argument 387 p = icp; sl_encaps()
|
/linux-4.4.14/include/net/ |
H A D | slhc_vj.h | 177 int slhc_compress(struct slcompress *comp, unsigned char *icp, int isize, 179 int slhc_uncompress(struct slcompress *comp, unsigned char *icp, int isize); 180 int slhc_remember(struct slcompress *comp, unsigned char *icp, int isize);
|
/linux-4.4.14/drivers/iommu/ |
H A D | msm_iommu.c | 687 static int __init get_tex_class(int icp, int ocp, int mt, int nos) get_tex_class() argument 703 if (icp == c_icp && ocp == c_ocp && c_mt == mt && c_nos == nos) get_tex_class()
|
/linux-4.4.14/drivers/net/wan/ |
H A D | x25_asy.c | 221 static void x25_asy_encaps(struct x25_asy *sl, unsigned char *icp, int len) x25_asy_encaps() argument 236 p = icp; x25_asy_encaps()
|
/linux-4.4.14/drivers/tty/serial/ |
H A D | sunsu.c | 709 unsigned int icp; sunsu_startup() local 713 icp = (up->port.iobase & 0xfe0) | 0x01f; sunsu_startup() 714 outb_p(0x80, icp); sunsu_startup() 715 (void) inb_p(icp); sunsu_startup()
|
/linux-4.4.14/drivers/net/hamradio/ |
H A D | 6pack.c | 164 static void sp_encaps(struct sixpack *sp, unsigned char *icp, int len) sp_encaps() argument 166 unsigned char *msg, *p = icp; sp_encaps()
|
H A D | mkiss.c | 433 static void ax_encaps(struct net_device *dev, unsigned char *icp, int len) ax_encaps() argument 450 p = icp; ax_encaps()
|
/linux-4.4.14/drivers/tty/serial/8250/ |
H A D | 8250_port.c | 2032 unsigned int icp; serial8250_do_startup() local 2036 icp = (port->iobase & 0xfe0) | 0x01f; serial8250_do_startup() 2037 outb_p(0x80, icp); serial8250_do_startup() 2038 inb_p(icp); serial8250_do_startup()
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | kvm_host.h | 651 struct kvmppc_icp *icp; /* XICS presentation controller */ member in struct:kvmppc_slb::kvm_vcpu_arch
|