sop 655 arch/mips/include/asm/octeon/cvmx-ipd-defs.h uint64_t sop:1; sop 681 arch/mips/include/asm/octeon/cvmx-ipd-defs.h uint64_t sop:1; sop 787 arch/mips/include/asm/octeon/cvmx-ipd-defs.h uint64_t sop:1; sop 813 arch/mips/include/asm/octeon/cvmx-ipd-defs.h uint64_t sop:1; sop 1148 arch/mips/include/asm/octeon/cvmx-ipd-defs.h uint64_t sop:64; sop 1150 arch/mips/include/asm/octeon/cvmx-ipd-defs.h uint64_t sop:64; sop 225 arch/mips/include/asm/octeon/cvmx-pko-defs.h uint64_t sop:1; sop 235 arch/mips/include/asm/octeon/cvmx-pko-defs.h uint64_t sop:1; sop 114 arch/mips/include/asm/octeon/cvmx-srxx-defs.h uint64_t sop:1; sop 122 arch/mips/include/asm/octeon/cvmx-srxx-defs.h uint64_t sop:1; sop 257 drivers/input/mouse/cyapa_gen5.c u8 sop; /* Start of packet, must be 01h */ sop 1076 drivers/input/mouse/cyapa_gen5.c bl_packet_start->sop = PIP_SOP_KEY; sop 1349 drivers/input/mouse/cyapa_gen5.c bl_packet_start->sop = PIP_SOP_KEY; sop 187 drivers/net/ethernet/chelsio/cxgb/sge.c u8 sop; /* is next entry start of packet? */ sop 630 drivers/net/ethernet/chelsio/cxgb/sge.c if (q->sop) sop 631 drivers/net/ethernet/chelsio/cxgb/sge.c q->sop = 0; sop 635 drivers/net/ethernet/chelsio/cxgb/sge.c q->sop = 1; sop 684 drivers/net/ethernet/chelsio/cxgb/sge.c q->sop = 1; sop 106 drivers/net/ethernet/cisco/enic/cq_enet_desc.h u8 *ingress_port, u8 *fcoe, u8 *eop, u8 *sop, u8 *rss_type, sop 131 drivers/net/ethernet/cisco/enic/cq_enet_desc.h *sop = (completed_index_flags & CQ_ENET_RQ_DESC_FLAGS_SOP) ? sop 363 drivers/net/ethernet/cisco/enic/enic_main.c if (buf->sop) sop 1337 drivers/net/ethernet/cisco/enic/enic_main.c u8 type, color, eop, sop, ingress_port, vlan_stripped; sop 1353 drivers/net/ethernet/cisco/enic/enic_main.c &ingress_port, &fcoe, &eop, &sop, &rss_type, sop 47 drivers/net/ethernet/cisco/enic/enic_res.h int offload_mode, int cq_entry, int sop, int eop, int loopback) sop 65 drivers/net/ethernet/cisco/enic/enic_res.h vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop, desc_skip_cnt, sop 59 drivers/net/ethernet/cisco/enic/vnic_wq.h int sop; sop 133 drivers/net/ethernet/cisco/enic/vnic_wq.h unsigned int len, int sop, int eop, sop 139 drivers/net/ethernet/cisco/enic/vnic_wq.h buf->sop = sop; sop 4520 drivers/net/ethernet/marvell/sky2.c int sop; sop 4557 drivers/net/ethernet/marvell/sky2.c sop = 1; sop 4563 drivers/net/ethernet/marvell/sky2.c if (sop) sop 4565 drivers/net/ethernet/marvell/sky2.c sop = 0; sop 4596 drivers/net/ethernet/marvell/sky2.c sop = 1; sop 233 drivers/net/vmxnet3/vmxnet3_defs.h u32 sop:1; /* Start of Packet */ sop 241 drivers/net/vmxnet3/vmxnet3_defs.h u32 sop:1; /* Start of Packet */ sop 1333 drivers/net/vmxnet3/vmxnet3_drv.c if (rcd->sop) { /* first buf of the pkt */ sop 1346 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(!(rcd->sop && rcd->eop)); sop 203 drivers/pci/controller/pcie-altera.c bool sop = false; sop 214 drivers/pci/controller/pcie-altera.c if ((ctrl & RP_RXCPL_SOP) || (ctrl & RP_RXCPL_EOP) || sop) { sop 219 drivers/pci/controller/pcie-altera.c sop = true; sop 95 drivers/scsi/fnic/cq_enet_desc.h u8 *ingress_port, u8 *fcoe, u8 *eop, u8 *sop, u8 *rss_type, sop 116 drivers/scsi/fnic/cq_enet_desc.h *sop = (completed_index_flags & CQ_ENET_RQ_DESC_FLAGS_SOP) ? sop 92 drivers/scsi/fnic/cq_exch_desc.h u8 *sop, sop 110 drivers/scsi/fnic/cq_exch_desc.h *sop = (desc_ptr->completed_index_eop_sop_prt & sop 841 drivers/scsi/fnic/fnic_fcs.c u8 type, color, eop, sop, ingress_port, vlan_stripped; sop 865 drivers/scsi/fnic/fnic_fcs.c &eop, &sop, &fcoe_fc_crc_ok, &exchange_id, sop 878 drivers/scsi/fnic/fnic_fcs.c &ingress_port, &fcoe, &eop, &sop, sop 35 drivers/scsi/fnic/fnic_res.h int cq_entry, int sop, int eop) sop 51 drivers/scsi/fnic/fnic_res.h vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop); sop 75 drivers/scsi/fnic/vnic_wq.h int sop; sop 117 drivers/scsi/fnic/vnic_wq.h unsigned int len, int sop, int eop) sop 121 drivers/scsi/fnic/vnic_wq.h buf->sop = sop; sop 56 drivers/scsi/snic/vnic_wq.h int sop; sop 104 drivers/scsi/snic/vnic_wq.h unsigned int len, int sop, int eop) sop 108 drivers/scsi/snic/vnic_wq.h buf->sop = sop; sop 358 fs/nfsd/nfs4state.c nfs4_get_stateowner(struct nfs4_stateowner *sop) sop 360 fs/nfsd/nfs4state.c atomic_inc(&sop->so_count); sop 361 fs/nfsd/nfs4state.c return sop; sop 365 fs/nfsd/nfs4state.c same_owner_str(struct nfs4_stateowner *sop, struct xdr_netobj *owner) sop 367 fs/nfsd/nfs4state.c return (sop->so_owner.len == owner->len) && sop 368 fs/nfsd/nfs4state.c 0 == memcmp(sop->so_owner.data, owner->data, owner->len); sop 1246 fs/nfsd/nfs4state.c static inline void nfs4_free_stateowner(struct nfs4_stateowner *sop) sop 1248 fs/nfsd/nfs4state.c kfree(sop->so_owner.data); sop 1249 fs/nfsd/nfs4state.c sop->so_ops->so_free(sop); sop 1252 fs/nfsd/nfs4state.c static void nfs4_put_stateowner(struct nfs4_stateowner *sop) sop 1254 fs/nfsd/nfs4state.c struct nfs4_client *clp = sop->so_client; sop 1258 fs/nfsd/nfs4state.c if (!atomic_dec_and_lock(&sop->so_count, &clp->cl_lock)) sop 1260 fs/nfsd/nfs4state.c sop->so_ops->so_unhash(sop); sop 1262 fs/nfsd/nfs4state.c nfs4_free_stateowner(sop); sop 4056 fs/nfsd/nfs4state.c struct nfs4_stateowner *sop; sop 4058 fs/nfsd/nfs4state.c sop = kmem_cache_alloc(slab, GFP_KERNEL); sop 4059 fs/nfsd/nfs4state.c if (!sop) sop 4062 fs/nfsd/nfs4state.c xdr_netobj_dup(&sop->so_owner, owner, GFP_KERNEL); sop 4063 fs/nfsd/nfs4state.c if (!sop->so_owner.data) { sop 4064 fs/nfsd/nfs4state.c kmem_cache_free(slab, sop); sop 4068 fs/nfsd/nfs4state.c INIT_LIST_HEAD(&sop->so_stateids); sop 4069 fs/nfsd/nfs4state.c sop->so_client = clp; sop 4070 fs/nfsd/nfs4state.c init_nfs4_replay(&sop->so_replay); sop 4071 fs/nfsd/nfs4state.c atomic_set(&sop->so_count, 1); sop 4072 fs/nfsd/nfs4state.c return sop; sop 5772 fs/nfsd/nfs4state.c struct nfs4_stateowner *sop = stp->st_stateowner; sop 5775 fs/nfsd/nfs4state.c status = nfsd4_check_seqid(cstate, sop, seqid); sop 6199 fs/nfsd/nfs4state.c static void nfs4_unhash_lockowner(struct nfs4_stateowner *sop) sop 6201 fs/nfsd/nfs4state.c unhash_lockowner_locked(lockowner(sop)); sop 6204 fs/nfsd/nfs4state.c static void nfs4_free_lockowner(struct nfs4_stateowner *sop) sop 6206 fs/nfsd/nfs4state.c struct nfs4_lockowner *lo = lockowner(sop); sop 6833 fs/nfsd/nfs4state.c struct nfs4_stateowner *sop; sop 6853 fs/nfsd/nfs4state.c list_for_each_entry(sop, &clp->cl_ownerstr_hashtbl[hashval], sop 6856 fs/nfsd/nfs4state.c if (sop->so_is_open_owner || !same_owner_str(sop, owner)) sop 6860 fs/nfsd/nfs4state.c lo = lockowner(sop); sop 6861 fs/nfsd/nfs4state.c list_for_each_entry(stp, &sop->so_stateids, st_perstateowner) { sop 6869 fs/nfsd/nfs4state.c nfs4_get_stateowner(sop); sop 444 fs/super.c const struct super_operations *sop = sb->s_op; sop 463 fs/super.c if (sop->put_super) sop 464 fs/super.c sop->put_super(sb); sop 629 ipc/sem.c struct sembuf *sop; sop 638 ipc/sem.c for (sop = sops; sop < sops + nsops; sop++) { sop 639 ipc/sem.c int idx = array_index_nospec(sop->sem_num, sma->sem_nsems); sop 641 ipc/sem.c sem_op = sop->sem_op; sop 653 ipc/sem.c if (sop->sem_flg & SEM_UNDO) { sop 654 ipc/sem.c int undo = un->semadj[sop->sem_num] - sem_op; sop 658 ipc/sem.c un->semadj[sop->sem_num] = undo; sop 664 ipc/sem.c sop--; sop 666 ipc/sem.c while (sop >= sops) { sop 667 ipc/sem.c ipc_update_pid(&sma->sems[sop->sem_num].sempid, pid); sop 668 ipc/sem.c sop--; sop 678 ipc/sem.c q->blocking = sop; sop 680 ipc/sem.c if (sop->sem_flg & IPC_NOWAIT) sop 686 ipc/sem.c sop--; sop 687 ipc/sem.c while (sop >= sops) { sop 688 ipc/sem.c sem_op = sop->sem_op; sop 689 ipc/sem.c sma->sems[sop->sem_num].semval -= sem_op; sop 690 ipc/sem.c if (sop->sem_flg & SEM_UNDO) sop 691 ipc/sem.c un->semadj[sop->sem_num] += sem_op; sop 692 ipc/sem.c sop--; sop 701 ipc/sem.c struct sembuf *sop; sop 719 ipc/sem.c for (sop = sops; sop < sops + nsops; sop++) { sop 720 ipc/sem.c int idx = array_index_nospec(sop->sem_num, sma->sem_nsems); sop 723 ipc/sem.c sem_op = sop->sem_op; sop 736 ipc/sem.c if (sop->sem_flg & SEM_UNDO) { sop 737 ipc/sem.c int undo = un->semadj[sop->sem_num] - sem_op; sop 745 ipc/sem.c for (sop = sops; sop < sops + nsops; sop++) { sop 746 ipc/sem.c curr = &sma->sems[sop->sem_num]; sop 747 ipc/sem.c sem_op = sop->sem_op; sop 750 ipc/sem.c if (sop->sem_flg & SEM_UNDO) { sop 751 ipc/sem.c int undo = un->semadj[sop->sem_num] - sem_op; sop 753 ipc/sem.c un->semadj[sop->sem_num] = undo; sop 762 ipc/sem.c q->blocking = sop; sop 763 ipc/sem.c return sop->sem_flg & IPC_NOWAIT ? -EAGAIN : 1; sop 1056 ipc/sem.c struct sembuf *sop = q->blocking; sop 1069 ipc/sem.c if (sop->sem_num != semnum) sop 1072 ipc/sem.c if (count_zero && sop->sem_op == 0) sop 1074 ipc/sem.c if (!count_zero && sop->sem_op < 0) sop 1972 ipc/sem.c struct sembuf *sops = fast_sops, *sop; sop 2007 ipc/sem.c for (sop = sops; sop < sops + nsops; sop++) { sop 2008 ipc/sem.c unsigned long mask = 1ULL << ((sop->sem_num) % BITS_PER_LONG); sop 2010 ipc/sem.c if (sop->sem_num >= max) sop 2011 ipc/sem.c max = sop->sem_num; sop 2012 ipc/sem.c if (sop->sem_flg & SEM_UNDO) sop 2023 ipc/sem.c if (sop->sem_op != 0) {