serv 190 fs/lockd/svc.c static int create_lockd_listener(struct svc_serv *serv, const char *name, serv 197 fs/lockd/svc.c xprt = svc_find_xprt(serv, name, net, family, 0); serv 199 fs/lockd/svc.c return svc_create_xprt(serv, name, net, family, port, serv 205 fs/lockd/svc.c static int create_lockd_family(struct svc_serv *serv, struct net *net, serv 210 fs/lockd/svc.c err = create_lockd_listener(serv, "udp", net, family, nlm_udpport, serv 215 fs/lockd/svc.c return create_lockd_listener(serv, "tcp", net, family, nlm_tcpport, serv 229 fs/lockd/svc.c static int make_socks(struct svc_serv *serv, struct net *net, serv 235 fs/lockd/svc.c err = create_lockd_family(serv, net, PF_INET, cred); serv 239 fs/lockd/svc.c err = create_lockd_family(serv, net, PF_INET6, cred); serv 250 fs/lockd/svc.c svc_shutdown_net(serv, net); serv 254 fs/lockd/svc.c static int lockd_up_net(struct svc_serv *serv, struct net *net, serv 263 fs/lockd/svc.c error = svc_bind(serv, net); serv 267 fs/lockd/svc.c error = make_socks(serv, net, cred); serv 279 fs/lockd/svc.c static void lockd_down_net(struct svc_serv *serv, struct net *net) serv 288 fs/lockd/svc.c svc_shutdown_net(serv, net); serv 376 fs/lockd/svc.c static int lockd_start_svc(struct svc_serv *serv) serv 386 fs/lockd/svc.c nlmsvc_rqst = svc_prepare_thread(serv, &serv->sv_pools[0], NUMA_NO_NODE); serv 397 fs/lockd/svc.c svc_sock_update_bufs(serv); serv 398 fs/lockd/svc.c serv->sv_maxconn = nlm_max_connections; serv 400 fs/lockd/svc.c nlmsvc_task = kthread_create(lockd, nlmsvc_rqst, "%s", serv->sv_name); serv 428 fs/lockd/svc.c struct svc_serv *serv; serv 454 fs/lockd/svc.c serv = svc_create(&nlmsvc_program, LOCKD_BUFSIZE, &lockd_sv_ops); serv 455 fs/lockd/svc.c if (!serv) { serv 464 fs/lockd/svc.c return serv; serv 472 fs/lockd/svc.c struct svc_serv *serv; serv 477 fs/lockd/svc.c serv = lockd_create_svc(); serv 478 fs/lockd/svc.c if (IS_ERR(serv)) { serv 479 fs/lockd/svc.c error = PTR_ERR(serv); serv 483 fs/lockd/svc.c error = lockd_up_net(serv, net, cred); serv 489 fs/lockd/svc.c error = lockd_start_svc(serv); serv 491 fs/lockd/svc.c lockd_down_net(serv, net); serv 500 fs/lockd/svc.c svc_destroy(serv); serv 35 fs/nfs/callback.c struct svc_serv *serv; serv 42 fs/nfs/callback.c static int nfs4_callback_up_net(struct svc_serv *serv, struct net *net) serv 48 fs/nfs/callback.c ret = svc_create_xprt(serv, "tcp", net, PF_INET, serv 57 fs/nfs/callback.c ret = svc_create_xprt(serv, "tcp", net, PF_INET6, serv 108 fs/nfs/callback.c struct svc_serv *serv = rqstp->rq_server; serv 120 fs/nfs/callback.c prepare_to_wait(&serv->sv_cb_waitq, &wq, TASK_INTERRUPTIBLE); serv 121 fs/nfs/callback.c spin_lock_bh(&serv->sv_cb_lock); serv 122 fs/nfs/callback.c if (!list_empty(&serv->sv_cb_list)) { serv 123 fs/nfs/callback.c req = list_first_entry(&serv->sv_cb_list, serv 126 fs/nfs/callback.c spin_unlock_bh(&serv->sv_cb_lock); serv 127 fs/nfs/callback.c finish_wait(&serv->sv_cb_waitq, &wq); serv 129 fs/nfs/callback.c error = bc_svc_process(serv, req, rqstp); serv 133 fs/nfs/callback.c spin_unlock_bh(&serv->sv_cb_lock); serv 136 fs/nfs/callback.c finish_wait(&serv->sv_cb_waitq, &wq); serv 145 fs/nfs/callback.c struct svc_serv *serv) serv 152 fs/nfs/callback.c xprt->bc_serv = serv; serv 156 fs/nfs/callback.c struct svc_serv *serv) serv 162 fs/nfs/callback.c struct svc_serv *serv) serv 167 fs/nfs/callback.c nfs_callback_bc_serv(minorversion, xprt, serv); serv 172 fs/nfs/callback.c if (serv->sv_nrthreads-1 == nrservs) serv 175 fs/nfs/callback.c ret = serv->sv_ops->svo_setup(serv, NULL, nrservs); serv 177 fs/nfs/callback.c serv->sv_ops->svo_setup(serv, NULL, 0); serv 184 fs/nfs/callback.c static void nfs_callback_down_net(u32 minorversion, struct svc_serv *serv, struct net *net) serv 192 fs/nfs/callback.c svc_shutdown_net(serv, net); serv 195 fs/nfs/callback.c static int nfs_callback_up_net(int minorversion, struct svc_serv *serv, serv 206 fs/nfs/callback.c ret = svc_bind(serv, net); serv 214 fs/nfs/callback.c ret = nfs4_callback_up_net(serv, net); serv 216 fs/nfs/callback.c set_bc_enabled(serv); serv 227 fs/nfs/callback.c svc_rpcb_cleanup(serv, net); serv 264 fs/nfs/callback.c struct svc_serv *serv; serv 269 fs/nfs/callback.c if (cb_info->serv) { serv 274 fs/nfs/callback.c svc_get(cb_info->serv); serv 275 fs/nfs/callback.c return cb_info->serv; serv 297 fs/nfs/callback.c serv = svc_create_pooled(&nfs4_callback_program, NFS4_CALLBACK_BUFSIZE, sv_ops); serv 298 fs/nfs/callback.c if (!serv) { serv 302 fs/nfs/callback.c cb_info->serv = serv; serv 306 fs/nfs/callback.c serv->sv_maxconn = 1024; serv 308 fs/nfs/callback.c return serv; serv 316 fs/nfs/callback.c struct svc_serv *serv; serv 323 fs/nfs/callback.c serv = nfs_callback_create_svc(minorversion); serv 324 fs/nfs/callback.c if (IS_ERR(serv)) { serv 325 fs/nfs/callback.c ret = PTR_ERR(serv); serv 329 fs/nfs/callback.c ret = nfs_callback_up_net(minorversion, serv, net, xprt); serv 333 fs/nfs/callback.c ret = nfs_callback_start_svc(minorversion, xprt, serv); serv 346 fs/nfs/callback.c cb_info->serv = NULL; serv 347 fs/nfs/callback.c svc_destroy(serv); serv 353 fs/nfs/callback.c nfs_callback_down_net(minorversion, serv, net); serv 364 fs/nfs/callback.c struct svc_serv *serv; serv 367 fs/nfs/callback.c serv = cb_info->serv; serv 368 fs/nfs/callback.c nfs_callback_down_net(minorversion, serv, net); serv 371 fs/nfs/callback.c svc_get(serv); serv 372 fs/nfs/callback.c serv->sv_ops->svo_setup(serv, NULL, 0); serv 373 fs/nfs/callback.c svc_destroy(serv); serv 375 fs/nfs/callback.c cb_info->serv = NULL; serv 494 fs/nfsd/nfssvc.c static void nfsd_last_thread(struct svc_serv *serv, struct net *net) serv 514 fs/nfsd/nfssvc.c svc_rpcb_cleanup(serv, net); serv 56 include/linux/sunrpc/bc_xprt.h static inline void set_bc_enabled(struct svc_serv *serv) serv 58 include/linux/sunrpc/bc_xprt.h serv->sv_bc_enabled = true; serv 77 include/linux/sunrpc/bc_xprt.h static inline void set_bc_enabled(struct svc_serv *serv) serv 122 include/linux/sunrpc/svc.h static inline void svc_get(struct svc_serv *serv) serv 124 include/linux/sunrpc/svc.h serv->sv_nrthreads++; serv 490 include/linux/sunrpc/svc.h int svc_rpcb_setup(struct svc_serv *serv, struct net *net); serv 491 include/linux/sunrpc/svc.h void svc_rpcb_cleanup(struct svc_serv *serv, struct net *net); serv 492 include/linux/sunrpc/svc.h int svc_bind(struct svc_serv *serv, struct net *net); serv 495 include/linux/sunrpc/svc.h struct svc_rqst *svc_rqst_alloc(struct svc_serv *serv, serv 497 include/linux/sunrpc/svc.h struct svc_rqst *svc_prepare_thread(struct svc_serv *serv, serv 507 include/linux/sunrpc/svc.h int svc_pool_stats_open(struct svc_serv *serv, struct file *file); serv 518 include/linux/sunrpc/svc.h struct svc_pool * svc_pool_for_cpu(struct svc_serv *serv, int cpu); serv 132 include/linux/sunrpc/svc_xprt.h struct svc_xprt *svc_find_xprt(struct svc_serv *serv, const char *xcl_name, serv 135 include/linux/sunrpc/svc_xprt.h int svc_xprt_names(struct svc_serv *serv, char *buf, const int buflen); serv 136 include/linux/sunrpc/svc_xprt.h void svc_add_new_perm_xprt(struct svc_serv *serv, struct svc_xprt *xprt); serv 59 include/linux/sunrpc/svcsock.h void svc_sock_update_bufs(struct svc_serv *serv); serv 61 include/linux/sunrpc/svcsock.h int svc_addsock(struct svc_serv *serv, const int fd, serv 66 include/linux/sunrpc/svcsock.h struct svc_xprt *svc_sock_create(struct svc_serv *serv, int prot); serv 44 include/net/caif/cfctrl.h enum cfctrl_srv serv, u8 phyid, serv 102 include/net/caif/cfctrl.h struct cfsrvl serv; serv 64 net/caif/cfcnfg.c enum cfctrl_srv serv, u8 phyid, serv 372 net/caif/cfcnfg.c cfcnfg_linkup_rsp(struct cflayer *layer, u8 channel_id, enum cfctrl_srv serv, serv 412 net/caif/cfcnfg.c switch (serv) { serv 17 net/caif/cfctrl.c #define container_obj(layr) container_of(layr, struct cfctrl, serv.layer) serv 42 net/caif/cfctrl.c caif_assert(offsetof(struct cfctrl, serv.layer) == 0); serv 45 net/caif/cfctrl.c cfsrvl_init(&this->serv, 0, &dev_info, false); serv 48 net/caif/cfctrl.c this->serv.layer.receive = cfctrl_recv; serv 49 net/caif/cfctrl.c sprintf(this->serv.layer.name, "ctrl"); serv 50 net/caif/cfctrl.c this->serv.layer.ctrlcmd = cfctrl_ctrlcmd; serv 57 net/caif/cfctrl.c return &this->serv.layer; serv 172 net/caif/cfctrl.c info->channel_id = cfctrl->serv.layer.id; serv 173 net/caif/cfctrl.c info->dev_info = &cfctrl->serv.dev_info; serv 180 net/caif/cfctrl.c struct cflayer *dn = cfctrl->serv.layer.dn; serv 189 net/caif/cfctrl.c caif_assert(offsetof(struct cfctrl, serv.layer) == 0); serv 192 net/caif/cfctrl.c cfctrl->serv.dev_info.id = physlinkid; serv 211 net/caif/cfctrl.c struct cflayer *dn = cfctrl->serv.layer.dn; serv 294 net/caif/cfctrl.c count = cfctrl_cancel_req(&cfctrl->serv.layer, serv 310 net/caif/cfctrl.c struct cflayer *dn = cfctrl->serv.layer.dn; serv 374 net/caif/cfctrl.c enum cfctrl_srv serv; serv 387 net/caif/cfctrl.c serv = tmp & CFCTRL_SRV_MASK; serv 388 net/caif/cfctrl.c linkparam.linktype = serv; serv 402 net/caif/cfctrl.c switch (serv) { serv 488 net/caif/cfctrl.c serv); serv 501 net/caif/cfctrl.c cfctrl->res.reject_rsp(cfctrl->serv.layer.up, serv 506 net/caif/cfctrl.c cfctrl->res.linksetup_rsp(cfctrl->serv. serv 508 net/caif/cfctrl.c serv, physlinkid, serv 520 net/caif/cfctrl.c cfctrl->res.linkdestroy_rsp(cfctrl->serv.layer.up, linkid); serv 17 net/caif/cfrfml.c #define container_obj(layr) container_of(layr, struct cfrfml, serv.layer) serv 25 net/caif/cfrfml.c struct cfsrvl serv; serv 54 net/caif/cfrfml.c cfsrvl_init(&this->serv, channel_id, dev_info, false); serv 55 net/caif/cfrfml.c this->serv.release = cfrfml_release; serv 56 net/caif/cfrfml.c this->serv.layer.receive = cfrfml_receive; serv 57 net/caif/cfrfml.c this->serv.layer.transmit = cfrfml_transmit; serv 65 net/caif/cfrfml.c snprintf(this->serv.layer.name, CAIF_LAYER_NAME_SZ, serv 68 net/caif/cfrfml.c return &this->serv.layer; serv 167 net/caif/cfrfml.c err = rfml->serv.layer.up->receive(rfml->serv.layer.up, pkt); serv 184 net/caif/cfrfml.c rfml->serv.dev_info.id); serv 201 net/caif/cfrfml.c cfpkt_info(pkt)->channel_id = rfml->serv.layer.id; serv 208 net/caif/cfrfml.c cfpkt_info(pkt)->dev_info = &rfml->serv.dev_info; serv 210 net/caif/cfrfml.c return rfml->serv.layer.dn->transmit(rfml->serv.layer.dn, pkt); serv 225 net/caif/cfrfml.c if (!cfsrvl_ready(&rfml->serv, &err)) serv 289 net/caif/cfrfml.c rfml->serv.dev_info.id); serv 36 net/sunrpc/svc.c static void svc_unregister(const struct svc_serv *serv, struct net *net); serv 38 net/sunrpc/svc.c #define svc_serv_is_pooled(serv) ((serv)->sv_ops->svo_function) serv 336 net/sunrpc/svc.c svc_pool_for_cpu(struct svc_serv *serv, int cpu) serv 346 net/sunrpc/svc.c if (svc_serv_is_pooled(serv)) { serv 356 net/sunrpc/svc.c return &serv->sv_pools[pidx % serv->sv_nrpools]; serv 359 net/sunrpc/svc.c int svc_rpcb_setup(struct svc_serv *serv, struct net *net) serv 368 net/sunrpc/svc.c svc_unregister(serv, net); serv 373 net/sunrpc/svc.c void svc_rpcb_cleanup(struct svc_serv *serv, struct net *net) serv 375 net/sunrpc/svc.c svc_unregister(serv, net); serv 380 net/sunrpc/svc.c static int svc_uses_rpcbind(struct svc_serv *serv) serv 385 net/sunrpc/svc.c for (progp = serv->sv_program; progp; progp = progp->pg_next) { serv 397 net/sunrpc/svc.c int svc_bind(struct svc_serv *serv, struct net *net) serv 399 net/sunrpc/svc.c if (!svc_uses_rpcbind(serv)) serv 401 net/sunrpc/svc.c return svc_rpcb_setup(serv, net); serv 407 net/sunrpc/svc.c __svc_init_bc(struct svc_serv *serv) serv 409 net/sunrpc/svc.c INIT_LIST_HEAD(&serv->sv_cb_list); serv 410 net/sunrpc/svc.c spin_lock_init(&serv->sv_cb_lock); serv 411 net/sunrpc/svc.c init_waitqueue_head(&serv->sv_cb_waitq); serv 415 net/sunrpc/svc.c __svc_init_bc(struct svc_serv *serv) serv 427 net/sunrpc/svc.c struct svc_serv *serv; serv 432 net/sunrpc/svc.c if (!(serv = kzalloc(sizeof(*serv), GFP_KERNEL))) serv 434 net/sunrpc/svc.c serv->sv_name = prog->pg_name; serv 435 net/sunrpc/svc.c serv->sv_program = prog; serv 436 net/sunrpc/svc.c serv->sv_nrthreads = 1; serv 437 net/sunrpc/svc.c serv->sv_stats = prog->pg_stats; serv 440 net/sunrpc/svc.c serv->sv_max_payload = bufsize? bufsize : 4096; serv 441 net/sunrpc/svc.c serv->sv_max_mesg = roundup(serv->sv_max_payload + PAGE_SIZE, PAGE_SIZE); serv 442 net/sunrpc/svc.c serv->sv_ops = ops; serv 456 net/sunrpc/svc.c serv->sv_xdrsize = xdrsize; serv 457 net/sunrpc/svc.c INIT_LIST_HEAD(&serv->sv_tempsocks); serv 458 net/sunrpc/svc.c INIT_LIST_HEAD(&serv->sv_permsocks); serv 459 net/sunrpc/svc.c timer_setup(&serv->sv_temptimer, NULL, 0); serv 460 net/sunrpc/svc.c spin_lock_init(&serv->sv_lock); serv 462 net/sunrpc/svc.c __svc_init_bc(serv); serv 464 net/sunrpc/svc.c serv->sv_nrpools = npools; serv 465 net/sunrpc/svc.c serv->sv_pools = serv 466 net/sunrpc/svc.c kcalloc(serv->sv_nrpools, sizeof(struct svc_pool), serv 468 net/sunrpc/svc.c if (!serv->sv_pools) { serv 469 net/sunrpc/svc.c kfree(serv); serv 473 net/sunrpc/svc.c for (i = 0; i < serv->sv_nrpools; i++) { serv 474 net/sunrpc/svc.c struct svc_pool *pool = &serv->sv_pools[i]; serv 477 net/sunrpc/svc.c i, serv->sv_name); serv 485 net/sunrpc/svc.c return serv; serv 500 net/sunrpc/svc.c struct svc_serv *serv; serv 503 net/sunrpc/svc.c serv = __svc_create(prog, bufsize, npools, ops); serv 504 net/sunrpc/svc.c if (!serv) serv 506 net/sunrpc/svc.c return serv; serv 513 net/sunrpc/svc.c void svc_shutdown_net(struct svc_serv *serv, struct net *net) serv 515 net/sunrpc/svc.c svc_close_net(serv, net); serv 517 net/sunrpc/svc.c if (serv->sv_ops->svo_shutdown) serv 518 net/sunrpc/svc.c serv->sv_ops->svo_shutdown(serv, net); serv 527 net/sunrpc/svc.c svc_destroy(struct svc_serv *serv) serv 530 net/sunrpc/svc.c serv->sv_program->pg_name, serv 531 net/sunrpc/svc.c serv->sv_nrthreads); serv 533 net/sunrpc/svc.c if (serv->sv_nrthreads) { serv 534 net/sunrpc/svc.c if (--(serv->sv_nrthreads) != 0) { serv 535 net/sunrpc/svc.c svc_sock_update_bufs(serv); serv 539 net/sunrpc/svc.c printk("svc_destroy: no threads for serv=%p!\n", serv); serv 541 net/sunrpc/svc.c del_timer_sync(&serv->sv_temptimer); serv 547 net/sunrpc/svc.c BUG_ON(!list_empty(&serv->sv_permsocks)); serv 548 net/sunrpc/svc.c BUG_ON(!list_empty(&serv->sv_tempsocks)); serv 550 net/sunrpc/svc.c cache_clean_deferred(serv); serv 552 net/sunrpc/svc.c if (svc_serv_is_pooled(serv)) serv 555 net/sunrpc/svc.c kfree(serv->sv_pools); serv 556 net/sunrpc/svc.c kfree(serv); serv 604 net/sunrpc/svc.c svc_rqst_alloc(struct svc_serv *serv, struct svc_pool *pool, int node) serv 614 net/sunrpc/svc.c rqstp->rq_server = serv; serv 617 net/sunrpc/svc.c rqstp->rq_argp = kmalloc_node(serv->sv_xdrsize, GFP_KERNEL, node); serv 621 net/sunrpc/svc.c rqstp->rq_resp = kmalloc_node(serv->sv_xdrsize, GFP_KERNEL, node); serv 625 net/sunrpc/svc.c if (!svc_init_buffer(rqstp, serv->sv_max_mesg, node)) serv 636 net/sunrpc/svc.c svc_prepare_thread(struct svc_serv *serv, struct svc_pool *pool, int node) serv 640 net/sunrpc/svc.c rqstp = svc_rqst_alloc(serv, pool, node); serv 644 net/sunrpc/svc.c serv->sv_nrthreads++; serv 657 net/sunrpc/svc.c choose_pool(struct svc_serv *serv, struct svc_pool *pool, unsigned int *state) serv 662 net/sunrpc/svc.c return &serv->sv_pools[(*state)++ % serv->sv_nrpools]; serv 669 net/sunrpc/svc.c choose_victim(struct svc_serv *serv, struct svc_pool *pool, unsigned int *state) serv 678 net/sunrpc/svc.c for (i = 0; i < serv->sv_nrpools; i++) { serv 679 net/sunrpc/svc.c pool = &serv->sv_pools[--(*state) % serv->sv_nrpools]; serv 708 net/sunrpc/svc.c svc_start_kthreads(struct svc_serv *serv, struct svc_pool *pool, int nrservs) serv 713 net/sunrpc/svc.c unsigned int state = serv->sv_nrthreads-1; serv 718 net/sunrpc/svc.c chosen_pool = choose_pool(serv, pool, &state); serv 721 net/sunrpc/svc.c rqstp = svc_prepare_thread(serv, chosen_pool, node); serv 725 net/sunrpc/svc.c __module_get(serv->sv_ops->svo_module); serv 726 net/sunrpc/svc.c task = kthread_create_on_node(serv->sv_ops->svo_function, rqstp, serv 727 net/sunrpc/svc.c node, "%s", serv->sv_name); serv 729 net/sunrpc/svc.c module_put(serv->sv_ops->svo_module); serv 735 net/sunrpc/svc.c if (serv->sv_nrpools > 1) serv 738 net/sunrpc/svc.c svc_sock_update_bufs(serv); serv 748 net/sunrpc/svc.c svc_signal_kthreads(struct svc_serv *serv, struct svc_pool *pool, int nrservs) serv 751 net/sunrpc/svc.c unsigned int state = serv->sv_nrthreads-1; serv 755 net/sunrpc/svc.c task = choose_victim(serv, pool, &state); serv 780 net/sunrpc/svc.c svc_set_num_threads(struct svc_serv *serv, struct svc_pool *pool, int nrservs) serv 784 net/sunrpc/svc.c nrservs -= (serv->sv_nrthreads-1); serv 792 net/sunrpc/svc.c return svc_start_kthreads(serv, pool, nrservs); serv 794 net/sunrpc/svc.c return svc_signal_kthreads(serv, pool, nrservs); serv 801 net/sunrpc/svc.c svc_stop_kthreads(struct svc_serv *serv, struct svc_pool *pool, int nrservs) serv 804 net/sunrpc/svc.c unsigned int state = serv->sv_nrthreads-1; serv 808 net/sunrpc/svc.c task = choose_victim(serv, pool, &state); serv 818 net/sunrpc/svc.c svc_set_num_threads_sync(struct svc_serv *serv, struct svc_pool *pool, int nrservs) serv 822 net/sunrpc/svc.c nrservs -= (serv->sv_nrthreads-1); serv 830 net/sunrpc/svc.c return svc_start_kthreads(serv, pool, nrservs); serv 832 net/sunrpc/svc.c return svc_stop_kthreads(serv, pool, nrservs); serv 855 net/sunrpc/svc.c struct svc_serv *serv = rqstp->rq_server; serv 867 net/sunrpc/svc.c if (serv) serv 868 net/sunrpc/svc.c svc_destroy(serv); serv 1059 net/sunrpc/svc.c int svc_register(const struct svc_serv *serv, struct net *net, serv 1071 net/sunrpc/svc.c for (progp = serv->sv_program; progp; progp = progp->pg_next) { serv 1121 net/sunrpc/svc.c static void svc_unregister(const struct svc_serv *serv, struct net *net) serv 1129 net/sunrpc/svc.c for (progp = serv->sv_program; progp; progp = progp->pg_next) { serv 1287 net/sunrpc/svc.c struct svc_serv *serv = rqstp->rq_server; serv 1325 net/sunrpc/svc.c for (progp = serv->sv_program; progp; progp = progp->pg_next) serv 1379 net/sunrpc/svc.c serv->sv_stats->rpccnt++; serv 1443 net/sunrpc/svc.c serv->sv_stats->rpcbadfmt++; serv 1455 net/sunrpc/svc.c serv->sv_stats->rpcbadauth++; serv 1465 net/sunrpc/svc.c serv->sv_stats->rpcbadfmt++; serv 1473 net/sunrpc/svc.c serv->sv_stats->rpcbadfmt++; serv 1482 net/sunrpc/svc.c serv->sv_stats->rpcbadfmt++; serv 1491 net/sunrpc/svc.c serv->sv_stats->rpcbadfmt++; serv 1504 net/sunrpc/svc.c struct svc_serv *serv = rqstp->rq_server; serv 1526 net/sunrpc/svc.c serv->sv_stats->rpcbadfmt++; serv 1550 net/sunrpc/svc.c bc_svc_process(struct svc_serv *serv, struct rpc_rqst *req, serv 1564 net/sunrpc/svc.c rqstp->rq_server = serv; serv 162 net/sunrpc/svc_xprt.c struct svc_xprt *xprt, struct svc_serv *serv) serv 168 net/sunrpc/svc_xprt.c xprt->xpt_server = serv; serv 182 net/sunrpc/svc_xprt.c struct svc_serv *serv, serv 218 net/sunrpc/svc_xprt.c return xcl->xcl_ops->xpo_create(serv, net, sap, len, flags); serv 246 net/sunrpc/svc_xprt.c void svc_add_new_perm_xprt(struct svc_serv *serv, struct svc_xprt *new) serv 249 net/sunrpc/svc_xprt.c spin_lock_bh(&serv->sv_lock); serv 250 net/sunrpc/svc_xprt.c list_add(&new->xpt_list, &serv->sv_permsocks); serv 251 net/sunrpc/svc_xprt.c spin_unlock_bh(&serv->sv_lock); serv 255 net/sunrpc/svc_xprt.c static int _svc_create_xprt(struct svc_serv *serv, const char *xprt_name, serv 274 net/sunrpc/svc_xprt.c newxprt = __svc_xpo_create(xcl, serv, net, family, port, flags); serv 280 net/sunrpc/svc_xprt.c svc_add_new_perm_xprt(serv, newxprt); serv 291 net/sunrpc/svc_xprt.c int svc_create_xprt(struct svc_serv *serv, const char *xprt_name, serv 299 net/sunrpc/svc_xprt.c err = _svc_create_xprt(serv, xprt_name, net, family, port, flags, cred); serv 302 net/sunrpc/svc_xprt.c err = _svc_create_xprt(serv, xprt_name, net, family, port, flags, cred); serv 540 net/sunrpc/svc_xprt.c void svc_wake_up(struct svc_serv *serv) serv 545 net/sunrpc/svc_xprt.c pool = &serv->sv_pools[0]; serv 598 net/sunrpc/svc_xprt.c static void svc_check_conn_limits(struct svc_serv *serv) serv 600 net/sunrpc/svc_xprt.c unsigned int limit = serv->sv_maxconn ? serv->sv_maxconn : serv 601 net/sunrpc/svc_xprt.c (serv->sv_nrthreads+3) * 20; serv 603 net/sunrpc/svc_xprt.c if (serv->sv_tmpcnt > limit) { serv 605 net/sunrpc/svc_xprt.c spin_lock_bh(&serv->sv_lock); serv 606 net/sunrpc/svc_xprt.c if (!list_empty(&serv->sv_tempsocks)) { serv 609 net/sunrpc/svc_xprt.c serv->sv_name, serv->sv_maxconn ? serv 616 net/sunrpc/svc_xprt.c xprt = list_entry(serv->sv_tempsocks.prev, serv 622 net/sunrpc/svc_xprt.c spin_unlock_bh(&serv->sv_lock); serv 633 net/sunrpc/svc_xprt.c struct svc_serv *serv = rqstp->rq_server; serv 639 net/sunrpc/svc_xprt.c pages = (serv->sv_max_mesg + 2 * PAGE_SIZE) >> PAGE_SHIFT; serv 752 net/sunrpc/svc_xprt.c static void svc_add_new_temp_xprt(struct svc_serv *serv, struct svc_xprt *newxpt) serv 754 net/sunrpc/svc_xprt.c spin_lock_bh(&serv->sv_lock); serv 756 net/sunrpc/svc_xprt.c list_add(&newxpt->xpt_list, &serv->sv_tempsocks); serv 757 net/sunrpc/svc_xprt.c serv->sv_tmpcnt++; serv 758 net/sunrpc/svc_xprt.c if (serv->sv_temptimer.function == NULL) { serv 760 net/sunrpc/svc_xprt.c serv->sv_temptimer.function = svc_age_temp_xprts; serv 761 net/sunrpc/svc_xprt.c mod_timer(&serv->sv_temptimer, serv 764 net/sunrpc/svc_xprt.c spin_unlock_bh(&serv->sv_lock); serv 770 net/sunrpc/svc_xprt.c struct svc_serv *serv = rqstp->rq_server; serv 792 net/sunrpc/svc_xprt.c svc_add_new_temp_xprt(serv, newxpt); serv 806 net/sunrpc/svc_xprt.c rqstp->rq_reserved = serv->sv_max_mesg; serv 824 net/sunrpc/svc_xprt.c struct svc_serv *serv = rqstp->rq_server; serv 864 net/sunrpc/svc_xprt.c if (serv->sv_stats) serv 865 net/sunrpc/svc_xprt.c serv->sv_stats->netcnt++; serv 930 net/sunrpc/svc_xprt.c struct svc_serv *serv = from_timer(serv, t, sv_temptimer); serv 936 net/sunrpc/svc_xprt.c if (!spin_trylock_bh(&serv->sv_lock)) { serv 939 net/sunrpc/svc_xprt.c mod_timer(&serv->sv_temptimer, jiffies + HZ); serv 943 net/sunrpc/svc_xprt.c list_for_each_safe(le, next, &serv->sv_tempsocks) { serv 960 net/sunrpc/svc_xprt.c spin_unlock_bh(&serv->sv_lock); serv 962 net/sunrpc/svc_xprt.c mod_timer(&serv->sv_temptimer, jiffies + svc_conn_age_period * HZ); serv 971 net/sunrpc/svc_xprt.c void svc_age_temp_xprts_now(struct svc_serv *serv, struct sockaddr *server_addr) serv 977 net/sunrpc/svc_xprt.c spin_lock_bh(&serv->sv_lock); serv 978 net/sunrpc/svc_xprt.c list_for_each_safe(le, next, &serv->sv_tempsocks) { serv 986 net/sunrpc/svc_xprt.c spin_unlock_bh(&serv->sv_lock); serv 1019 net/sunrpc/svc_xprt.c struct svc_serv *serv = xprt->xpt_server; serv 1031 net/sunrpc/svc_xprt.c spin_lock_bh(&serv->sv_lock); serv 1035 net/sunrpc/svc_xprt.c serv->sv_tmpcnt--; serv 1036 net/sunrpc/svc_xprt.c spin_unlock_bh(&serv->sv_lock); serv 1061 net/sunrpc/svc_xprt.c static int svc_close_list(struct svc_serv *serv, struct list_head *xprt_list, struct net *net) serv 1066 net/sunrpc/svc_xprt.c spin_lock(&serv->sv_lock); serv 1074 net/sunrpc/svc_xprt.c spin_unlock(&serv->sv_lock); serv 1078 net/sunrpc/svc_xprt.c static struct svc_xprt *svc_dequeue_net(struct svc_serv *serv, struct net *net) serv 1085 net/sunrpc/svc_xprt.c for (i = 0; i < serv->sv_nrpools; i++) { serv 1086 net/sunrpc/svc_xprt.c pool = &serv->sv_pools[i]; serv 1101 net/sunrpc/svc_xprt.c static void svc_clean_up_xprts(struct svc_serv *serv, struct net *net) serv 1105 net/sunrpc/svc_xprt.c while ((xprt = svc_dequeue_net(serv, net))) { serv 1123 net/sunrpc/svc_xprt.c void svc_close_net(struct svc_serv *serv, struct net *net) serv 1127 net/sunrpc/svc_xprt.c while (svc_close_list(serv, &serv->sv_permsocks, net) + serv 1128 net/sunrpc/svc_xprt.c svc_close_list(serv, &serv->sv_tempsocks, net)) { serv 1130 net/sunrpc/svc_xprt.c svc_clean_up_xprts(serv, net); serv 1273 net/sunrpc/svc_xprt.c struct svc_xprt *svc_find_xprt(struct svc_serv *serv, const char *xcl_name, serv 1281 net/sunrpc/svc_xprt.c if (serv == NULL || xcl_name == NULL) serv 1284 net/sunrpc/svc_xprt.c spin_lock_bh(&serv->sv_lock); serv 1285 net/sunrpc/svc_xprt.c list_for_each_entry(xprt, &serv->sv_permsocks, xpt_list) { serv 1298 net/sunrpc/svc_xprt.c spin_unlock_bh(&serv->sv_lock); serv 1328 net/sunrpc/svc_xprt.c int svc_xprt_names(struct svc_serv *serv, char *buf, const int buflen) serv 1335 net/sunrpc/svc_xprt.c if (!serv) serv 1338 net/sunrpc/svc_xprt.c spin_lock_bh(&serv->sv_lock); serv 1342 net/sunrpc/svc_xprt.c list_for_each_entry(xprt, &serv->sv_permsocks, xpt_list) { serv 1355 net/sunrpc/svc_xprt.c spin_unlock_bh(&serv->sv_lock); serv 1366 net/sunrpc/svc_xprt.c struct svc_serv *serv = m->private; serv 1372 net/sunrpc/svc_xprt.c return (pidx > serv->sv_nrpools ? NULL : &serv->sv_pools[pidx-1]); serv 1378 net/sunrpc/svc_xprt.c struct svc_serv *serv = m->private; serv 1383 net/sunrpc/svc_xprt.c pool = &serv->sv_pools[0]; serv 1385 net/sunrpc/svc_xprt.c unsigned int pidx = (pool - &serv->sv_pools[0]); serv 1386 net/sunrpc/svc_xprt.c if (pidx < serv->sv_nrpools-1) serv 1387 net/sunrpc/svc_xprt.c pool = &serv->sv_pools[pidx+1]; serv 1425 net/sunrpc/svc_xprt.c int svc_pool_stats_open(struct svc_serv *serv, struct file *file) serv 1431 net/sunrpc/svc_xprt.c ((struct seq_file *) file->private_data)->private = serv; serv 499 net/sunrpc/svcsock.c struct svc_serv *serv = svsk->sk_xprt.xpt_server; serv 524 net/sunrpc/svcsock.c svc_sock_setbufsize(svsk, serv->sv_nrthreads + 3); serv 594 net/sunrpc/svcsock.c if (serv->sv_stats) serv 595 net/sunrpc/svcsock.c serv->sv_stats->netudpcnt++; serv 621 net/sunrpc/svcsock.c struct svc_serv *serv = xprt->xpt_server; serv 629 net/sunrpc/svcsock.c required = atomic_read(&svsk->sk_xprt.xpt_reserved) + serv->sv_max_mesg; serv 646 net/sunrpc/svcsock.c static struct svc_xprt *svc_udp_create(struct svc_serv *serv, serv 651 net/sunrpc/svcsock.c return svc_create_socket(serv, IPPROTO_UDP, net, sa, salen, flags); serv 675 net/sunrpc/svcsock.c static void svc_udp_init(struct svc_sock *svsk, struct svc_serv *serv) serv 680 net/sunrpc/svcsock.c &svsk->sk_xprt, serv); serv 780 net/sunrpc/svcsock.c struct svc_serv *serv = svsk->sk_xprt.xpt_server; serv 796 net/sunrpc/svcsock.c serv->sv_name); serv 799 net/sunrpc/svcsock.c serv->sv_name, -err); serv 807 net/sunrpc/svcsock.c serv->sv_name, -err); serv 818 net/sunrpc/svcsock.c serv->sv_name, serv 821 net/sunrpc/svcsock.c dprintk("%s: connect from %s\n", serv->sv_name, serv 834 net/sunrpc/svcsock.c newsvsk = svc_setup_socket(serv, newsock, serv 851 net/sunrpc/svcsock.c if (serv->sv_stats) serv 852 net/sunrpc/svcsock.c serv->sv_stats->nettcpconn++; serv 921 net/sunrpc/svcsock.c struct svc_serv *serv = svsk->sk_xprt.xpt_server; serv 944 net/sunrpc/svcsock.c serv->sv_max_mesg) { serv 1036 net/sunrpc/svcsock.c struct svc_serv *serv = svsk->sk_xprt.xpt_server; serv 1115 net/sunrpc/svcsock.c if (serv->sv_stats) serv 1116 net/sunrpc/svcsock.c serv->sv_stats->nettcpcnt++; serv 1166 net/sunrpc/svcsock.c static struct svc_xprt *svc_tcp_create(struct svc_serv *serv, serv 1171 net/sunrpc/svcsock.c return svc_create_socket(serv, IPPROTO_TCP, net, sa, salen, flags); serv 1207 net/sunrpc/svcsock.c static void svc_tcp_init(struct svc_sock *svsk, struct svc_serv *serv) serv 1212 net/sunrpc/svcsock.c &svsk->sk_xprt, serv); serv 1245 net/sunrpc/svcsock.c void svc_sock_update_bufs(struct svc_serv *serv) serv 1253 net/sunrpc/svcsock.c spin_lock_bh(&serv->sv_lock); serv 1254 net/sunrpc/svcsock.c list_for_each_entry(svsk, &serv->sv_permsocks, sk_xprt.xpt_list) serv 1256 net/sunrpc/svcsock.c spin_unlock_bh(&serv->sv_lock); serv 1263 net/sunrpc/svcsock.c static struct svc_sock *svc_setup_socket(struct svc_serv *serv, serv 1281 net/sunrpc/svcsock.c err = svc_register(serv, sock_net(sock->sk), inet->sk_family, serv 1305 net/sunrpc/svcsock.c svc_udp_init(svsk, serv); serv 1307 net/sunrpc/svcsock.c svc_tcp_init(svsk, serv); serv 1346 net/sunrpc/svcsock.c int svc_addsock(struct svc_serv *serv, const int fd, char *name_return, serv 1371 net/sunrpc/svcsock.c svsk = svc_setup_socket(serv, so, SVC_SOCK_DEFAULTS); serv 1381 net/sunrpc/svcsock.c svc_add_new_perm_xprt(serv, &svsk->sk_xprt); serv 1392 net/sunrpc/svcsock.c static struct svc_xprt *svc_create_socket(struct svc_serv *serv, serv 1410 net/sunrpc/svcsock.c serv->sv_program->pg_name, protocol, serv 1463 net/sunrpc/svcsock.c svsk = svc_setup_socket(serv, sock, flags); serv 67 net/sunrpc/xprtrdma/svc_rdma_transport.c static struct svcxprt_rdma *svc_rdma_create_xprt(struct svc_serv *serv, serv 69 net/sunrpc/xprtrdma/svc_rdma_transport.c static struct svc_xprt *svc_rdma_create(struct svc_serv *serv, serv 128 net/sunrpc/xprtrdma/svc_rdma_transport.c static struct svcxprt_rdma *svc_rdma_create_xprt(struct svc_serv *serv, serv 137 net/sunrpc/xprtrdma/svc_rdma_transport.c svc_xprt_init(net, &svc_rdma_class, &cma_xprt->sc_xprt, serv); serv 309 net/sunrpc/xprtrdma/svc_rdma_transport.c static struct svc_xprt *svc_rdma_create(struct svc_serv *serv, serv 323 net/sunrpc/xprtrdma/svc_rdma_transport.c cma_xprt = svc_rdma_create_xprt(serv, net); serv 430 tools/usb/usbip/libsrc/vhci_driver.c char serv[NI_MAXSERV] = "unknown port"; serv 438 tools/usb/usbip/libsrc/vhci_driver.c ret = read_record(idev->port, host, sizeof(host), serv, sizeof(serv), serv 456 tools/usb/usbip/libsrc/vhci_driver.c host, serv, remote_busid);