/linux-4.1.27/fs/nfs/ |
D | direct.c | 107 static void nfs_direct_write_complete(struct nfs_direct_req *dreq, struct inode *inode); 110 static inline void get_dreq(struct nfs_direct_req *dreq) in get_dreq() argument 112 atomic_inc(&dreq->io_count); in get_dreq() 115 static inline int put_dreq(struct nfs_direct_req *dreq) in put_dreq() argument 117 return atomic_dec_and_test(&dreq->io_count); in put_dreq() 120 void nfs_direct_set_resched_writes(struct nfs_direct_req *dreq) in nfs_direct_set_resched_writes() argument 122 dreq->flags = NFS_ODIRECT_RESCHED_WRITES; in nfs_direct_set_resched_writes() 127 nfs_direct_good_bytes(struct nfs_direct_req *dreq, struct nfs_pgio_header *hdr) in nfs_direct_good_bytes() argument 132 if (dreq->mirror_count == 1) { in nfs_direct_good_bytes() 133 dreq->mirrors[hdr->pgio_mirror_idx].count += hdr->good_bytes; in nfs_direct_good_bytes() [all …]
|
D | cache_lib.c | 67 void nfs_cache_defer_req_put(struct nfs_cache_defer_req *dreq) in nfs_cache_defer_req_put() argument 69 if (atomic_dec_and_test(&dreq->count)) in nfs_cache_defer_req_put() 70 kfree(dreq); in nfs_cache_defer_req_put() 75 struct nfs_cache_defer_req *dreq; in nfs_dns_cache_revisit() local 77 dreq = container_of(d, struct nfs_cache_defer_req, deferred_req); in nfs_dns_cache_revisit() 79 complete_all(&dreq->completion); in nfs_dns_cache_revisit() 80 nfs_cache_defer_req_put(dreq); in nfs_dns_cache_revisit() 85 struct nfs_cache_defer_req *dreq; in nfs_dns_cache_defer() local 87 dreq = container_of(req, struct nfs_cache_defer_req, req); in nfs_dns_cache_defer() 88 dreq->deferred_req.revisit = nfs_dns_cache_revisit; in nfs_dns_cache_defer() [all …]
|
D | dns_resolve.c | 273 struct nfs_cache_defer_req *dreq) in do_cache_lookup() argument 279 ret = cache_check(cd, &(*item)->h, &dreq->req); in do_cache_lookup() 315 struct nfs_cache_defer_req *dreq; in do_cache_lookup_wait() local 318 dreq = nfs_cache_defer_req_alloc(); in do_cache_lookup_wait() 319 if (!dreq) in do_cache_lookup_wait() 321 ret = do_cache_lookup(cd, key, item, dreq); in do_cache_lookup_wait() 323 ret = nfs_cache_wait_for_upcall(dreq); in do_cache_lookup_wait() 327 nfs_cache_defer_req_put(dreq); in do_cache_lookup_wait()
|
D | cache_lib.h | 23 extern void nfs_cache_defer_req_put(struct nfs_cache_defer_req *dreq); 24 extern int nfs_cache_wait_for_upcall(struct nfs_cache_defer_req *dreq);
|
D | write.c | 792 if (!cinfo->dreq) in nfs_request_add_commit_list() 825 cinfo->dreq = NULL; in nfs_init_cinfo_from_inode() 831 struct nfs_direct_req *dreq) in nfs_init_cinfo() argument 833 if (dreq) in nfs_init_cinfo() 834 nfs_init_cinfo_from_dreq(cinfo, dreq); in nfs_init_cinfo() 941 if ((ret == max) && !cinfo->dreq) in nfs_scan_commit_list() 1616 data->dreq = cinfo->dreq; in nfs_init_commit() 1640 if (!cinfo->dreq) in nfs_retry_commit() 1729 nfs_init_cinfo(&cinfo, data->inode, data->dreq); in nfs_commit_release_pages()
|
D | internal.h | 476 struct nfs_direct_req *dreq); 497 struct nfs_direct_req *dreq); 502 extern ssize_t nfs_dreq_bytes_left(struct nfs_direct_req *dreq); 503 extern void nfs_direct_set_resched_writes(struct nfs_direct_req *dreq);
|
D | pagelist.c | 66 hdr->dreq = desc->pg_dreq; in nfs_pgheader_init() 1226 desc->pg_dreq = hdr->dreq; in nfs_pageio_resend()
|
D | pnfs_nfs.c | 107 if ((ret == max) && !cinfo->dreq) in pnfs_generic_transfer_commit_list()
|
D | nfs4proc.c | 4283 if (hdr->ds_clp != NULL || hdr->dreq != NULL) in nfs4_write_need_cache_consistency_data()
|
/linux-4.1.27/net/dccp/ |
D | minisocks.c | 88 struct dccp_request_sock *dreq = dccp_rsk(req); in dccp_create_openreq_child() local 95 newdp->dccps_service = dreq->dreq_service; in dccp_create_openreq_child() 96 newdp->dccps_timestamp_echo = dreq->dreq_timestamp_echo; in dccp_create_openreq_child() 97 newdp->dccps_timestamp_time = dreq->dreq_timestamp_time; in dccp_create_openreq_child() 112 newdp->dccps_iss = dreq->dreq_iss; in dccp_create_openreq_child() 113 newdp->dccps_gss = dreq->dreq_gss; in dccp_create_openreq_child() 115 newdp->dccps_isr = dreq->dreq_isr; in dccp_create_openreq_child() 116 newdp->dccps_gsr = dreq->dreq_gsr; in dccp_create_openreq_child() 121 if (dccp_feat_activate_values(newsk, &dreq->dreq_featneg)) { in dccp_create_openreq_child() 145 struct dccp_request_sock *dreq = dccp_rsk(req); in dccp_check_req() local [all …]
|
D | options.c | 51 int dccp_parse_options(struct sock *sk, struct dccp_request_sock *dreq, in dccp_parse_options() argument 102 if (dreq != NULL && (opt >= DCCPO_MIN_RX_CCID_SPECIFIC || in dccp_parse_options() 128 rc = dccp_feat_parse_options(sk, dreq, mandatory, opt, in dccp_parse_options() 147 if (dreq != NULL) { in dccp_parse_options() 148 dreq->dreq_timestamp_echo = ntohl(opt_val); in dccp_parse_options() 149 dreq->dreq_timestamp_time = dccp_timestamp(); in dccp_parse_options() 356 struct dccp_request_sock *dreq, in dccp_insert_option_timestamp_echo() argument 363 if (dreq != NULL) { in dccp_insert_option_timestamp_echo() 364 elapsed_time = dccp_timestamp() - dreq->dreq_timestamp_time; in dccp_insert_option_timestamp_echo() 365 tstamp_echo = htonl(dreq->dreq_timestamp_echo); in dccp_insert_option_timestamp_echo() [all …]
|
D | output.c | 397 struct dccp_request_sock *dreq; in dccp_make_response() local 411 dreq = dccp_rsk(req); in dccp_make_response() 413 dccp_inc_seqno(&dreq->dreq_gss); in dccp_make_response() 415 DCCP_SKB_CB(skb)->dccpd_seq = dreq->dreq_gss; in dccp_make_response() 418 if (dccp_feat_server_ccid_dependencies(dreq)) in dccp_make_response() 421 if (dccp_insert_options_rsk(dreq, skb)) in dccp_make_response() 433 dccp_hdr_set_seq(dh, dreq->dreq_gss); in dccp_make_response() 434 dccp_hdr_set_ack(dccp_hdr_ack_bits(skb), dreq->dreq_gsr); in dccp_make_response() 435 dccp_hdr_response(skb)->dccph_resp_service = dreq->dreq_service; in dccp_make_response()
|
D | ipv4.c | 597 struct dccp_request_sock *dreq; in dccp_v4_conn_request() local 634 dreq = dccp_rsk(req); in dccp_v4_conn_request() 635 if (dccp_parse_options(sk, dreq, skb)) in dccp_v4_conn_request() 654 dreq->dreq_isr = dcb->dccpd_seq; in dccp_v4_conn_request() 655 dreq->dreq_gsr = dreq->dreq_isr; in dccp_v4_conn_request() 656 dreq->dreq_iss = dccp_v4_init_sequence(skb); in dccp_v4_conn_request() 657 dreq->dreq_gss = dreq->dreq_iss; in dccp_v4_conn_request() 658 dreq->dreq_service = service; in dccp_v4_conn_request()
|
D | ipv6.c | 332 struct dccp_request_sock *dreq; in dccp_v6_conn_request() local 365 dreq = dccp_rsk(req); in dccp_v6_conn_request() 366 if (dccp_parse_options(sk, dreq, skb)) in dccp_v6_conn_request() 397 dreq->dreq_isr = dcb->dccpd_seq; in dccp_v6_conn_request() 398 dreq->dreq_gsr = dreq->dreq_isr; in dccp_v6_conn_request() 399 dreq->dreq_iss = dccp_v6_init_sequence(skb); in dccp_v6_conn_request() 400 dreq->dreq_gss = dreq->dreq_iss; in dccp_v6_conn_request() 401 dreq->dreq_service = service; in dccp_v6_conn_request()
|
D | feat.c | 634 int dccp_feat_insert_opts(struct dccp_sock *dp, struct dccp_request_sock *dreq, in dccp_feat_insert_opts() argument 637 struct list_head *fn = dreq ? &dreq->dreq_featneg : &dp->dccps_featneg; in dccp_feat_insert_opts() 998 int dccp_feat_server_ccid_dependencies(struct dccp_request_sock *dreq) in dccp_feat_server_ccid_dependencies() argument 1000 struct list_head *fn = &dreq->dreq_featneg; in dccp_feat_server_ccid_dependencies() 1394 int dccp_feat_parse_options(struct sock *sk, struct dccp_request_sock *dreq, in dccp_feat_parse_options() argument 1398 struct list_head *fn = dreq ? &dreq->dreq_featneg : &dp->dccps_featneg; in dccp_feat_parse_options()
|
D | dccp.h | 472 int dccp_feat_server_ccid_dependencies(struct dccp_request_sock *dreq);
|
/linux-4.1.27/net/sunrpc/ |
D | cache.c | 528 static void __unhash_deferred_req(struct cache_deferred_req *dreq) in __unhash_deferred_req() argument 530 hlist_del_init(&dreq->hash); in __unhash_deferred_req() 531 if (!list_empty(&dreq->recent)) { in __unhash_deferred_req() 532 list_del_init(&dreq->recent); in __unhash_deferred_req() 537 static void __hash_deferred_req(struct cache_deferred_req *dreq, struct cache_head *item) in __hash_deferred_req() argument 541 INIT_LIST_HEAD(&dreq->recent); in __hash_deferred_req() 542 hlist_add_head(&dreq->hash, &cache_defer_hash[hash]); in __hash_deferred_req() 545 static void setup_deferral(struct cache_deferred_req *dreq, in setup_deferral() argument 550 dreq->item = item; in setup_deferral() 554 __hash_deferred_req(dreq, item); in setup_deferral() [all …]
|
D | svc_xprt.c | 1077 static void svc_revisit(struct cache_deferred_req *dreq, int too_many) in svc_revisit() argument 1080 container_of(dreq, struct svc_deferred_req, handle); in svc_revisit()
|
/linux-4.1.27/drivers/staging/i2o/ |
D | i2o_block.c | 403 struct i2o_block_delayed_request *dreq = in i2o_block_delayed_request_fn() local 406 struct request_queue *q = dreq->queue; in i2o_block_delayed_request_fn() 412 kfree(dreq); in i2o_block_delayed_request_fn() 898 struct i2o_block_delayed_request *dreq; in i2o_block_request_fn() local 916 dreq = kmalloc(sizeof(*dreq), GFP_ATOMIC); in i2o_block_request_fn() 917 if (!dreq) in i2o_block_request_fn() 920 dreq->queue = q; in i2o_block_request_fn() 921 INIT_DELAYED_WORK(&dreq->work, in i2o_block_request_fn() 925 &dreq->work, in i2o_block_request_fn() 927 kfree(dreq); in i2o_block_request_fn()
|
/linux-4.1.27/drivers/s390/block/ |
D | dasd_diag.c | 166 struct dasd_diag_req *dreq; in dasd_start_diag() local 177 dreq = (struct dasd_diag_req *) cqr->data; in dasd_start_diag() 182 private->iob.block_count = dreq->block_count; in dasd_start_diag() 184 private->iob.bio_list = dreq->bio; in dasd_start_diag() 504 struct dasd_diag_req *dreq; in dasd_diag_build_cp() local 543 dreq = (struct dasd_diag_req *) cqr->data; in dasd_diag_build_cp() 544 dreq->block_count = count; in dasd_diag_build_cp() 545 dbio = dreq->bio; in dasd_diag_build_cp()
|
/linux-4.1.27/fs/nilfs2/ |
D | btree.c | 1731 union nilfs_bmap_ptr_req *dreq, in nilfs_btree_prepare_convert_and_insert() argument 1745 dreq->bpr_ptr = nilfs_btree_find_target_v(btree, NULL, key); in nilfs_btree_prepare_convert_and_insert() 1749 ret = nilfs_bmap_prepare_alloc_ptr(btree, dreq, dat); in nilfs_btree_prepare_convert_and_insert() 1756 nreq->bpr_ptr = dreq->bpr_ptr + 1; in nilfs_btree_prepare_convert_and_insert() 1776 nilfs_bmap_abort_alloc_ptr(btree, dreq, dat); in nilfs_btree_prepare_convert_and_insert() 1787 union nilfs_bmap_ptr_req *dreq, in nilfs_btree_commit_convert_and_insert() argument 1807 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat); in nilfs_btree_commit_convert_and_insert() 1814 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, ncblk); in nilfs_btree_commit_convert_and_insert() 1829 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat); in nilfs_btree_commit_convert_and_insert() 1836 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, in nilfs_btree_commit_convert_and_insert() [all …]
|
/linux-4.1.27/drivers/dma/ |
D | bcm2835-dma.c | 71 unsigned int dreq; member 418 if (c->dreq != 0) in bcm2835_dma_prep_dma_cyclic() 420 BCM2835_DMA_PER_MAP(c->dreq); in bcm2835_dma_prep_dma_cyclic() 549 to_bcm2835_dma_chan(chan)->dreq = spec->args[0]; in bcm2835_dma_xlate()
|
/linux-4.1.27/include/linux/ |
D | dccp.h | 178 extern int dccp_parse_options(struct sock *sk, struct dccp_request_sock *dreq,
|
D | nfs_xdr.h | 1334 struct nfs_direct_req *dreq; member 1376 struct nfs_direct_req *dreq; /* O_DIRECT request */ member 1388 struct nfs_direct_req *dreq; /* O_DIRECT request */ member
|
/linux-4.1.27/fs/nfs/flexfilelayout/ |
D | flexfilelayout.c | 624 if (!hdr->dreq) { in ff_layout_reset_write() 631 nfs_direct_set_resched_writes(hdr->dreq); in ff_layout_reset_write()
|
/linux-4.1.27/fs/nfs/blocklayout/ |
D | blocklayout.c | 236 const bool is_dio = (header->dreq != NULL); in bl_read_pagelist()
|
/linux-4.1.27/drivers/infiniband/core/ |
D | cm.c | 137 static CM_COUNTER_ATTR(dreq, CM_DREQ_COUNTER);
|