cps 1376 arch/mips/include/asm/octeon/cvmx-lmcx-defs.h uint64_t cps:3; cps 1384 arch/mips/include/asm/octeon/cvmx-lmcx-defs.h uint64_t cps:3; cps 729 drivers/atm/idt77252.c u32 cps = vc->estimator->maxcps; cps 731 drivers/atm/idt77252.c vc->estimator->cps = cps; cps 732 drivers/atm/idt77252.c vc->estimator->avcps = cps << 5; cps 2081 drivers/atm/idt77252.c u32 rate, cps; cps 2093 drivers/atm/idt77252.c est->cps = (est->avcps + 0x1f) >> 5; cps 2095 drivers/atm/idt77252.c cps = est->cps; cps 2096 drivers/atm/idt77252.c if (cps < (est->maxcps >> 4)) cps 2097 drivers/atm/idt77252.c cps = est->maxcps >> 4; cps 2099 drivers/atm/idt77252.c lacr = idt77252_rate_logindex(card, cps); cps 2124 drivers/atm/idt77252.c est->cps = est->maxcps; cps 2125 drivers/atm/idt77252.c est->avcps = est->cps << 5; cps 196 drivers/atm/idt77252.h u32 cps; cps 342 drivers/atm/iphase.c u32 exp, mantissa, cps; cps 349 drivers/atm/iphase.c cps = (1 << M_BITS) | mantissa; cps 351 drivers/atm/iphase.c cps = cps; cps 353 drivers/atm/iphase.c cps <<= (exp - M_BITS); cps 355 drivers/atm/iphase.c cps >>= (M_BITS - exp); cps 356 drivers/atm/iphase.c return cps; cps 506 drivers/block/drbd/drbd_worker.c int cps; /* correction per invocation of drbd_rs_controller() */ cps 527 drivers/block/drbd/drbd_worker.c cps = correction / steps; cps 528 drivers/block/drbd/drbd_worker.c fifo_add_val(plan, cps); cps 529 drivers/block/drbd/drbd_worker.c plan->total += cps * steps; cps 337 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.h #define amdgpu_dpm_check_state_equal(adev, cps, rps, equal) \ cps 338 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.h ((adev)->powerplay.pp_funcs->check_state_equal((adev)->powerplay.pp_handle, (cps), (rps), (equal))) cps 3245 drivers/gpu/drm/amd/amdgpu/kv_dpm.c struct amdgpu_ps *cps = (struct amdgpu_ps *)current_ps; cps 3249 drivers/gpu/drm/amd/amdgpu/kv_dpm.c if (adev == NULL || cps == NULL || rps == NULL || equal == NULL) cps 3252 drivers/gpu/drm/amd/amdgpu/kv_dpm.c kv_cps = kv_get_ps(cps); cps 3274 drivers/gpu/drm/amd/amdgpu/kv_dpm.c *equal = ((cps->vclk == rps->vclk) && (cps->dclk == rps->dclk)); cps 3275 drivers/gpu/drm/amd/amdgpu/kv_dpm.c *equal &= ((cps->evclk == rps->evclk) && (cps->ecclk == rps->ecclk)); cps 7947 drivers/gpu/drm/amd/amdgpu/si_dpm.c struct amdgpu_ps *cps = (struct amdgpu_ps *)current_ps; cps 7951 drivers/gpu/drm/amd/amdgpu/si_dpm.c if (adev == NULL || cps == NULL || rps == NULL || equal == NULL) cps 7954 drivers/gpu/drm/amd/amdgpu/si_dpm.c si_cps = si_get_ps((struct amdgpu_ps *)cps); cps 7977 drivers/gpu/drm/amd/amdgpu/si_dpm.c *equal = ((cps->vclk == rps->vclk) && (cps->dclk == rps->dclk)); cps 7978 drivers/gpu/drm/amd/amdgpu/si_dpm.c *equal &= ((cps->evclk == rps->evclk) && (cps->ecclk == rps->ecclk)); cps 238 drivers/gpu/drm/amd/include/kgd_pp_interface.h void *cps, cps 507 drivers/staging/fwserial/fwserial.c n = (elapsed * port->cps) / HZ + 1; cps 950 drivers/staging/fwserial/fwserial.c port->cps = (baud << 1) / frame; cps 250 drivers/staging/fwserial/fwserial.h unsigned int cps; cps 750 drivers/tty/rocket.c info->cps = baud / bits; cps 986 drivers/tty/rocket.c timeout = (sGetTxCnt(cp) + 1) * HZ / info->cps; cps 1429 drivers/tty/rocket.c printk(KERN_INFO "cps=%d...\n", info->cps); cps 1436 drivers/tty/rocket.c check_time = (HZ / info->cps) / 5; cps 1438 drivers/tty/rocket.c check_time = HZ * txcnt / info->cps; cps 1146 drivers/tty/rocket_int.h int cps; cps 57 fs/hpfs/map.c unsigned char *hpfs_load_code_page(struct super_block *s, secno cps) cps 66 fs/hpfs/map.c struct code_page_directory *cp = hpfs_map_sector(s, cps, &bh, 0); cps 119 fs/nfs/callback.h struct cb_process_state *cps); cps 138 fs/nfs/callback.h struct cb_process_state *cps); cps 144 fs/nfs/callback.h struct cb_process_state *cps); cps 161 fs/nfs/callback.h struct cb_process_state *cps); cps 176 fs/nfs/callback.h struct cb_process_state *cps); cps 185 fs/nfs/callback.h struct cb_process_state *cps); cps 197 fs/nfs/callback.h struct cb_process_state *cps); cps 201 fs/nfs/callback.h struct cb_process_state *cps); cps 203 fs/nfs/callback.h struct cb_process_state *cps); cps 24 fs/nfs/callback_proc.c struct cb_process_state *cps) cps 33 fs/nfs/callback_proc.c if (!cps->clp) /* Always set for v4.0. Set in cb_sequence for v4.1 */ cps 40 fs/nfs/callback_proc.c rpc_peeraddr2str(cps->clp->cl_rpcclient, RPC_DISPLAY_ADDR)); cps 42 fs/nfs/callback_proc.c inode = nfs_delegation_find_inode(cps->clp, &args->fh); cps 46 fs/nfs/callback_proc.c trace_nfs4_cb_getattr(cps->clp, &args->fh, NULL, cps 68 fs/nfs/callback_proc.c trace_nfs4_cb_getattr(cps->clp, &args->fh, inode, -ntohl(res->status)); cps 76 fs/nfs/callback_proc.c struct cb_process_state *cps) cps 83 fs/nfs/callback_proc.c if (!cps->clp) /* Always set for v4.0. Set in cb_sequence for v4.1 */ cps 87 fs/nfs/callback_proc.c rpc_peeraddr2str(cps->clp->cl_rpcclient, RPC_DISPLAY_ADDR)); cps 90 fs/nfs/callback_proc.c inode = nfs_delegation_find_inode(cps->clp, &args->fh); cps 94 fs/nfs/callback_proc.c trace_nfs4_cb_recall(cps->clp, &args->fh, NULL, cps 109 fs/nfs/callback_proc.c trace_nfs4_cb_recall(cps->clp, &args->fh, inode, cps 342 fs/nfs/callback_proc.c struct cb_process_state *cps) cps 347 fs/nfs/callback_proc.c if (cps->clp) cps 348 fs/nfs/callback_proc.c res = do_callback_layoutrecall(cps->clp, args); cps 364 fs/nfs/callback_proc.c struct cb_process_state *cps) cps 369 fs/nfs/callback_proc.c struct nfs_client *clp = cps->clp; cps 503 fs/nfs/callback_proc.c struct cb_process_state *cps) cps 513 fs/nfs/callback_proc.c clp = nfs4_find_client_sessionid(cps->net, args->csa_addr, cps 514 fs/nfs/callback_proc.c &args->csa_sessionid, cps->minorversion); cps 556 fs/nfs/callback_proc.c cps->slot = slot; cps 585 fs/nfs/callback_proc.c cps->clp = clp; /* put in nfs4_callback_compound */ cps 591 fs/nfs/callback_proc.c cps->drc_status = status; cps 607 fs/nfs/callback_proc.c struct cb_process_state *cps) cps 614 fs/nfs/callback_proc.c if (!cps->clp) /* set in cb_sequence */ cps 618 fs/nfs/callback_proc.c rpc_peeraddr2str(cps->clp->cl_rpcclient, RPC_DISPLAY_ADDR)); cps 630 fs/nfs/callback_proc.c nfs_expire_unused_delegation_types(cps->clp, flags); cps 633 fs/nfs/callback_proc.c pnfs_recall_all_layouts(cps->clp); cps 641 fs/nfs/callback_proc.c struct cb_process_state *cps) cps 648 fs/nfs/callback_proc.c if (!cps->clp) /* set in cb_sequence */ cps 652 fs/nfs/callback_proc.c rpc_peeraddr2str(cps->clp->cl_rpcclient, RPC_DISPLAY_ADDR), cps 655 fs/nfs/callback_proc.c fc_tbl = &cps->clp->cl_session->fc_slot_table; cps 660 fs/nfs/callback_proc.c nfs41_notify_server(cps->clp); cps 667 fs/nfs/callback_proc.c struct cb_process_state *cps) cps 671 fs/nfs/callback_proc.c if (!cps->clp) /* set in cb_sequence */ cps 675 fs/nfs/callback_proc.c rpc_peeraddr2str(cps->clp->cl_rpcclient, RPC_DISPLAY_ADDR)); cps 679 fs/nfs/callback_proc.c __wake_up(&cps->clp->cl_lock_waitq, TASK_NORMAL, 0, args); cps 699 fs/nfs/callback_proc.c struct cb_process_state *cps) cps 710 fs/nfs/callback_proc.c spin_lock(&cps->clp->cl_lock); cps 712 fs/nfs/callback_proc.c list_for_each_entry_rcu(server, &cps->clp->cl_superblocks, cps 730 fs/nfs/callback_proc.c list_add_tail(©->copies, &cps->clp->pending_cb_stateids); cps 733 fs/nfs/callback_proc.c spin_unlock(&cps->clp->cl_lock); cps 806 fs/nfs/callback_xdr.c static void nfs4_cb_free_slot(struct cb_process_state *cps) cps 808 fs/nfs/callback_xdr.c if (cps->slot) { cps 809 fs/nfs/callback_xdr.c nfs4_callback_free_slot(cps->clp->cl_session, cps->slot); cps 810 fs/nfs/callback_xdr.c cps->slot = NULL; cps 822 fs/nfs/callback_xdr.c static void nfs4_cb_free_slot(struct cb_process_state *cps) cps 868 fs/nfs/callback_xdr.c struct cb_process_state *cps) cps 880 fs/nfs/callback_xdr.c switch (cps->minorversion) { cps 899 fs/nfs/callback_xdr.c if (cps->drc_status) { cps 900 fs/nfs/callback_xdr.c status = cps->drc_status; cps 908 fs/nfs/callback_xdr.c status = op->process_op(argp, resp, cps); cps 930 fs/nfs/callback_xdr.c struct cb_process_state cps = { cps 948 fs/nfs/callback_xdr.c cps.clp = nfs4_find_client_ident(SVC_NET(rqstp), hdr_arg.cb_ident); cps 949 fs/nfs/callback_xdr.c if (!cps.clp || !check_gss_callback_principal(cps.clp, rqstp)) { cps 950 fs/nfs/callback_xdr.c if (cps.clp) cps 951 fs/nfs/callback_xdr.c nfs_put_client(cps.clp); cps 956 fs/nfs/callback_xdr.c cps.minorversion = hdr_arg.minorversion; cps 960 fs/nfs/callback_xdr.c if (cps.clp) cps 961 fs/nfs/callback_xdr.c nfs_put_client(cps.clp); cps 967 fs/nfs/callback_xdr.c &cps); cps 980 fs/nfs/callback_xdr.c nfs4_cb_free_slot(&cps); cps 981 fs/nfs/callback_xdr.c nfs_put_client(cps.clp); cps 378 include/net/ip_vs.h u64 cps; cps 395 include/net/ip_vs.h u64 cps; /* current connection rate */ cps 184 include/uapi/linux/ip_vs.h __u32 cps; /* current connection rate */ cps 827 net/netfilter/ipvs/ip_vs_ctl.c dst->cps = (u32)src->cps; cps 2225 net/netfilter/ipvs/ip_vs_ctl.c (unsigned long long)show.cps, cps 2281 net/netfilter/ipvs/ip_vs_ctl.c kstats.cps, cps 2976 net/netfilter/ipvs/ip_vs_ctl.c nla_put_u32(skb, IPVS_STATS_ATTR_CPS, (u32)kstats->cps) || cps 3009 net/netfilter/ipvs/ip_vs_ctl.c nla_put_u64_64bit(skb, IPVS_STATS_ATTR_CPS, kstats->cps, cps 113 net/netfilter/ipvs/ip_vs_est.c e->cps += ((s64)rate - (s64)e->cps) >> 2; cps 168 net/netfilter/ipvs/ip_vs_est.c est->cps = 0; cps 180 net/netfilter/ipvs/ip_vs_est.c dst->cps = (e->cps + 0x1FF) >> 10;