Home
last modified time | relevance | path

Searched refs:dreq (Results 1 – 30 of 30) sorted by relevance

/linux-4.4.14/fs/nfs/
Ddirect.c107 static void nfs_direct_write_complete(struct nfs_direct_req *dreq, struct inode *inode);
110 static inline void get_dreq(struct nfs_direct_req *dreq) in get_dreq() argument
112 atomic_inc(&dreq->io_count); in get_dreq()
115 static inline int put_dreq(struct nfs_direct_req *dreq) in put_dreq() argument
117 return atomic_dec_and_test(&dreq->io_count); in put_dreq()
120 void nfs_direct_set_resched_writes(struct nfs_direct_req *dreq) in nfs_direct_set_resched_writes() argument
122 dreq->flags = NFS_ODIRECT_RESCHED_WRITES; in nfs_direct_set_resched_writes()
127 nfs_direct_good_bytes(struct nfs_direct_req *dreq, struct nfs_pgio_header *hdr) in nfs_direct_good_bytes() argument
132 if (dreq->mirror_count == 1) { in nfs_direct_good_bytes()
133 dreq->mirrors[hdr->pgio_mirror_idx].count += hdr->good_bytes; in nfs_direct_good_bytes()
[all …]
Dcache_lib.c67 void nfs_cache_defer_req_put(struct nfs_cache_defer_req *dreq) in nfs_cache_defer_req_put() argument
69 if (atomic_dec_and_test(&dreq->count)) in nfs_cache_defer_req_put()
70 kfree(dreq); in nfs_cache_defer_req_put()
75 struct nfs_cache_defer_req *dreq; in nfs_dns_cache_revisit() local
77 dreq = container_of(d, struct nfs_cache_defer_req, deferred_req); in nfs_dns_cache_revisit()
79 complete_all(&dreq->completion); in nfs_dns_cache_revisit()
80 nfs_cache_defer_req_put(dreq); in nfs_dns_cache_revisit()
85 struct nfs_cache_defer_req *dreq; in nfs_dns_cache_defer() local
87 dreq = container_of(req, struct nfs_cache_defer_req, req); in nfs_dns_cache_defer()
88 dreq->deferred_req.revisit = nfs_dns_cache_revisit; in nfs_dns_cache_defer()
[all …]
Ddns_resolve.c273 struct nfs_cache_defer_req *dreq) in do_cache_lookup() argument
279 ret = cache_check(cd, &(*item)->h, &dreq->req); in do_cache_lookup()
315 struct nfs_cache_defer_req *dreq; in do_cache_lookup_wait() local
318 dreq = nfs_cache_defer_req_alloc(); in do_cache_lookup_wait()
319 if (!dreq) in do_cache_lookup_wait()
321 ret = do_cache_lookup(cd, key, item, dreq); in do_cache_lookup_wait()
323 ret = nfs_cache_wait_for_upcall(dreq); in do_cache_lookup_wait()
327 nfs_cache_defer_req_put(dreq); in do_cache_lookup_wait()
Dcache_lib.h23 extern void nfs_cache_defer_req_put(struct nfs_cache_defer_req *dreq);
24 extern int nfs_cache_wait_for_upcall(struct nfs_cache_defer_req *dreq);
Dinternal.h495 struct nfs_direct_req *dreq);
516 struct nfs_direct_req *dreq);
521 extern ssize_t nfs_dreq_bytes_left(struct nfs_direct_req *dreq);
522 extern void nfs_direct_set_resched_writes(struct nfs_direct_req *dreq);
626 if (!cinfo->dreq) { in nfs_mark_page_unstable()
Dwrite.c844 cinfo->dreq = NULL; in nfs_init_cinfo_from_inode()
850 struct nfs_direct_req *dreq) in nfs_init_cinfo() argument
852 if (dreq) in nfs_init_cinfo()
853 nfs_init_cinfo_from_dreq(cinfo, dreq); in nfs_init_cinfo()
961 if ((ret == max) && !cinfo->dreq) in nfs_scan_commit_list()
1638 data->dreq = cinfo->dreq; in nfs_init_commit()
1662 if (!cinfo->dreq) in nfs_retry_commit()
1751 nfs_init_cinfo(&cinfo, data->inode, data->dreq); in nfs_commit_release_pages()
Dpagelist.c66 hdr->dreq = desc->pg_dreq; in nfs_pgheader_init()
1224 desc->pg_dreq = hdr->dreq; in nfs_pageio_resend()
Dpnfs_nfs.c107 if ((ret == max) && !cinfo->dreq) in pnfs_generic_transfer_commit_list()
Dnfs4proc.c4437 if (hdr->ds_clp != NULL || hdr->dreq != NULL) in nfs4_write_need_cache_consistency_data()
/linux-4.4.14/net/dccp/
Dminisocks.c88 struct dccp_request_sock *dreq = dccp_rsk(req); in dccp_create_openreq_child() local
95 newdp->dccps_service = dreq->dreq_service; in dccp_create_openreq_child()
96 newdp->dccps_timestamp_echo = dreq->dreq_timestamp_echo; in dccp_create_openreq_child()
97 newdp->dccps_timestamp_time = dreq->dreq_timestamp_time; in dccp_create_openreq_child()
112 newdp->dccps_iss = dreq->dreq_iss; in dccp_create_openreq_child()
113 newdp->dccps_gss = dreq->dreq_gss; in dccp_create_openreq_child()
115 newdp->dccps_isr = dreq->dreq_isr; in dccp_create_openreq_child()
116 newdp->dccps_gsr = dreq->dreq_gsr; in dccp_create_openreq_child()
121 if (dccp_feat_activate_values(newsk, &dreq->dreq_featneg)) { in dccp_create_openreq_child()
145 struct dccp_request_sock *dreq = dccp_rsk(req); in dccp_check_req() local
[all …]
Doptions.c51 int dccp_parse_options(struct sock *sk, struct dccp_request_sock *dreq, in dccp_parse_options() argument
102 if (dreq != NULL && (opt >= DCCPO_MIN_RX_CCID_SPECIFIC || in dccp_parse_options()
128 rc = dccp_feat_parse_options(sk, dreq, mandatory, opt, in dccp_parse_options()
147 if (dreq != NULL) { in dccp_parse_options()
148 dreq->dreq_timestamp_echo = ntohl(opt_val); in dccp_parse_options()
149 dreq->dreq_timestamp_time = dccp_timestamp(); in dccp_parse_options()
356 struct dccp_request_sock *dreq, in dccp_insert_option_timestamp_echo() argument
363 if (dreq != NULL) { in dccp_insert_option_timestamp_echo()
364 elapsed_time = dccp_timestamp() - dreq->dreq_timestamp_time; in dccp_insert_option_timestamp_echo()
365 tstamp_echo = htonl(dreq->dreq_timestamp_echo); in dccp_insert_option_timestamp_echo()
[all …]
Doutput.c397 struct dccp_request_sock *dreq; in dccp_make_response() local
416 dreq = dccp_rsk(req); in dccp_make_response()
418 dccp_inc_seqno(&dreq->dreq_gss); in dccp_make_response()
420 DCCP_SKB_CB(skb)->dccpd_seq = dreq->dreq_gss; in dccp_make_response()
423 if (dccp_feat_server_ccid_dependencies(dreq)) in dccp_make_response()
426 if (dccp_insert_options_rsk(dreq, skb)) in dccp_make_response()
438 dccp_hdr_set_seq(dh, dreq->dreq_gss); in dccp_make_response()
439 dccp_hdr_set_ack(dccp_hdr_ack_bits(skb), dreq->dreq_gsr); in dccp_make_response()
440 dccp_hdr_response(skb)->dccph_resp_service = dreq->dreq_service; in dccp_make_response()
Dipv4.c568 struct dccp_request_sock *dreq; in dccp_v4_conn_request() local
605 dreq = dccp_rsk(req); in dccp_v4_conn_request()
606 if (dccp_parse_options(sk, dreq, skb)) in dccp_v4_conn_request()
625 dreq->dreq_isr = dcb->dccpd_seq; in dccp_v4_conn_request()
626 dreq->dreq_gsr = dreq->dreq_isr; in dccp_v4_conn_request()
627 dreq->dreq_iss = dccp_v4_init_sequence(skb); in dccp_v4_conn_request()
628 dreq->dreq_gss = dreq->dreq_iss; in dccp_v4_conn_request()
629 dreq->dreq_service = service; in dccp_v4_conn_request()
Dipv6.c301 struct dccp_request_sock *dreq; in dccp_v6_conn_request() local
334 dreq = dccp_rsk(req); in dccp_v6_conn_request()
335 if (dccp_parse_options(sk, dreq, skb)) in dccp_v6_conn_request()
366 dreq->dreq_isr = dcb->dccpd_seq; in dccp_v6_conn_request()
367 dreq->dreq_gsr = dreq->dreq_isr; in dccp_v6_conn_request()
368 dreq->dreq_iss = dccp_v6_init_sequence(skb); in dccp_v6_conn_request()
369 dreq->dreq_gss = dreq->dreq_iss; in dccp_v6_conn_request()
370 dreq->dreq_service = service; in dccp_v6_conn_request()
Dfeat.c634 int dccp_feat_insert_opts(struct dccp_sock *dp, struct dccp_request_sock *dreq, in dccp_feat_insert_opts() argument
637 struct list_head *fn = dreq ? &dreq->dreq_featneg : &dp->dccps_featneg; in dccp_feat_insert_opts()
998 int dccp_feat_server_ccid_dependencies(struct dccp_request_sock *dreq) in dccp_feat_server_ccid_dependencies() argument
1000 struct list_head *fn = &dreq->dreq_featneg; in dccp_feat_server_ccid_dependencies()
1394 int dccp_feat_parse_options(struct sock *sk, struct dccp_request_sock *dreq, in dccp_feat_parse_options() argument
1398 struct list_head *fn = dreq ? &dreq->dreq_featneg : &dp->dccps_featneg; in dccp_feat_parse_options()
Ddccp.h474 int dccp_feat_server_ccid_dependencies(struct dccp_request_sock *dreq);
/linux-4.4.14/drivers/crypto/marvell/
Dtdma.c40 void mv_cesa_dma_step(struct mv_cesa_tdma_req *dreq) in mv_cesa_dma_step() argument
42 struct mv_cesa_engine *engine = dreq->base.engine; in mv_cesa_dma_step()
54 writel_relaxed(dreq->chain.first->cur_dma, in mv_cesa_dma_step()
59 void mv_cesa_dma_cleanup(struct mv_cesa_tdma_req *dreq) in mv_cesa_dma_cleanup() argument
63 for (tdma = dreq->chain.first; tdma;) { in mv_cesa_dma_cleanup()
75 dreq->chain.first = NULL; in mv_cesa_dma_cleanup()
76 dreq->chain.last = NULL; in mv_cesa_dma_cleanup()
79 void mv_cesa_dma_prepare(struct mv_cesa_tdma_req *dreq, in mv_cesa_dma_prepare() argument
84 for (tdma = dreq->chain.first; tdma; tdma = tdma->next) { in mv_cesa_dma_prepare()
Dhash.c51 struct mv_cesa_ahash_dma_req *dreq = &creq->req.dma; in mv_cesa_ahash_dma_alloc_cache() local
54 &dreq->cache_dma); in mv_cesa_ahash_dma_alloc_cache()
302 struct mv_cesa_tdma_req *dreq = &creq->req.dma.base; in mv_cesa_ahash_dma_prepare() local
304 mv_cesa_dma_prepare(dreq, dreq->base.engine); in mv_cesa_ahash_dma_prepare()
603 struct mv_cesa_tdma_req *dreq = &ahashdreq->base; in mv_cesa_ahash_dma_req_init() local
609 dreq->chain.first = NULL; in mv_cesa_ahash_dma_req_init()
610 dreq->chain.last = NULL; in mv_cesa_ahash_dma_req_init()
621 mv_cesa_tdma_desc_iter_init(&dreq->chain); in mv_cesa_ahash_dma_req_init()
628 ret = mv_cesa_ahash_dma_add_cache(&dreq->chain, &iter, creq, flags); in mv_cesa_ahash_dma_req_init()
639 ret = mv_cesa_dma_add_op_transfers(&dreq->chain, in mv_cesa_ahash_dma_req_init()
[all …]
Dcipher.c170 struct mv_cesa_tdma_req *dreq = &creq->req.dma; in mv_cesa_ablkcipher_dma_prepare() local
172 mv_cesa_dma_prepare(dreq, dreq->base.engine); in mv_cesa_ablkcipher_dma_prepare()
298 struct mv_cesa_tdma_req *dreq = &creq->req.dma; in mv_cesa_ablkcipher_dma_req_init() local
304 dreq->base.type = CESA_DMA_REQ; in mv_cesa_ablkcipher_dma_req_init()
305 dreq->chain.first = NULL; in mv_cesa_ablkcipher_dma_req_init()
306 dreq->chain.last = NULL; in mv_cesa_ablkcipher_dma_req_init()
361 dreq->chain = chain; in mv_cesa_ablkcipher_dma_req_init()
366 mv_cesa_dma_cleanup(dreq); in mv_cesa_ablkcipher_dma_req_init()
Dcesa.h767 void mv_cesa_dma_step(struct mv_cesa_tdma_req *dreq);
769 static inline int mv_cesa_dma_process(struct mv_cesa_tdma_req *dreq, in mv_cesa_dma_process() argument
781 void mv_cesa_dma_prepare(struct mv_cesa_tdma_req *dreq,
784 void mv_cesa_dma_cleanup(struct mv_cesa_tdma_req *dreq);
/linux-4.4.14/net/sunrpc/
Dcache.c535 static void __unhash_deferred_req(struct cache_deferred_req *dreq) in __unhash_deferred_req() argument
537 hlist_del_init(&dreq->hash); in __unhash_deferred_req()
538 if (!list_empty(&dreq->recent)) { in __unhash_deferred_req()
539 list_del_init(&dreq->recent); in __unhash_deferred_req()
544 static void __hash_deferred_req(struct cache_deferred_req *dreq, struct cache_head *item) in __hash_deferred_req() argument
548 INIT_LIST_HEAD(&dreq->recent); in __hash_deferred_req()
549 hlist_add_head(&dreq->hash, &cache_defer_hash[hash]); in __hash_deferred_req()
552 static void setup_deferral(struct cache_deferred_req *dreq, in setup_deferral() argument
557 dreq->item = item; in setup_deferral()
561 __hash_deferred_req(dreq, item); in setup_deferral()
[all …]
Dsvc_xprt.c1077 static void svc_revisit(struct cache_deferred_req *dreq, int too_many) in svc_revisit() argument
1080 container_of(dreq, struct svc_deferred_req, handle); in svc_revisit()
/linux-4.4.14/drivers/s390/block/
Ddasd_diag.c173 struct dasd_diag_req *dreq; in dasd_start_diag() local
184 dreq = (struct dasd_diag_req *) cqr->data; in dasd_start_diag()
189 private->iob.block_count = dreq->block_count; in dasd_start_diag()
191 private->iob.bio_list = dreq->bio; in dasd_start_diag()
511 struct dasd_diag_req *dreq; in dasd_diag_build_cp() local
550 dreq = (struct dasd_diag_req *) cqr->data; in dasd_diag_build_cp()
551 dreq->block_count = count; in dasd_diag_build_cp()
552 dbio = dreq->bio; in dasd_diag_build_cp()
/linux-4.4.14/fs/nilfs2/
Dbtree.c1726 union nilfs_bmap_ptr_req *dreq, in nilfs_btree_prepare_convert_and_insert() argument
1740 dreq->bpr_ptr = nilfs_btree_find_target_v(btree, NULL, key); in nilfs_btree_prepare_convert_and_insert()
1744 ret = nilfs_bmap_prepare_alloc_ptr(btree, dreq, dat); in nilfs_btree_prepare_convert_and_insert()
1751 nreq->bpr_ptr = dreq->bpr_ptr + 1; in nilfs_btree_prepare_convert_and_insert()
1771 nilfs_bmap_abort_alloc_ptr(btree, dreq, dat); in nilfs_btree_prepare_convert_and_insert()
1782 union nilfs_bmap_ptr_req *dreq, in nilfs_btree_commit_convert_and_insert() argument
1802 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat); in nilfs_btree_commit_convert_and_insert()
1809 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, ncblk); in nilfs_btree_commit_convert_and_insert()
1824 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat); in nilfs_btree_commit_convert_and_insert()
1831 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, in nilfs_btree_commit_convert_and_insert()
[all …]
/linux-4.4.14/drivers/dma/
Dbcm2835-dma.c77 unsigned int dreq; member
439 if (c->dreq != 0) in bcm2835_dma_prep_dma_cyclic()
441 BCM2835_DMA_PER_MAP(c->dreq); in bcm2835_dma_prep_dma_cyclic()
579 to_bcm2835_dma_chan(chan)->dreq = spec->args[0]; in bcm2835_dma_xlate()
/linux-4.4.14/include/linux/
Ddccp.h178 extern int dccp_parse_options(struct sock *sk, struct dccp_request_sock *dreq,
Dnfs_xdr.h1392 struct nfs_direct_req *dreq; member
1434 struct nfs_direct_req *dreq; /* O_DIRECT request */ member
1446 struct nfs_direct_req *dreq; /* O_DIRECT request */ member
/linux-4.4.14/fs/nfs/blocklayout/
Dblocklayout.c235 const bool is_dio = (header->dreq != NULL); in bl_read_pagelist()
/linux-4.4.14/fs/nfs/flexfilelayout/
Dflexfilelayout.c915 if (!hdr->dreq) { in ff_layout_reset_write()
922 nfs_direct_set_resched_writes(hdr->dreq); in ff_layout_reset_write()
/linux-4.4.14/drivers/infiniband/core/
Dcm.c137 static CM_COUNTER_ATTR(dreq, CM_DREQ_COUNTER);