/linux-4.1.27/drivers/staging/lustre/lustre/ptlrpc/ |
D | client.c | 50 static int ptlrpc_send_new_req(struct ptlrpc_request *req); 51 static int ptlrpcd_check_work(struct ptlrpc_request *req); 134 struct ptlrpc_bulk_desc *ptlrpc_prep_bulk_imp(struct ptlrpc_request *req, in ptlrpc_prep_bulk_imp() argument 138 struct obd_import *imp = req->rq_import; in ptlrpc_prep_bulk_imp() 146 desc->bd_import_generation = req->rq_import_generation; in ptlrpc_prep_bulk_imp() 148 desc->bd_req = req; in ptlrpc_prep_bulk_imp() 154 req->rq_bulk = desc; in ptlrpc_prep_bulk_imp() 217 void ptlrpc_at_set_req_timeout(struct ptlrpc_request *req) in ptlrpc_at_set_req_timeout() argument 223 LASSERT(req->rq_import); in ptlrpc_at_set_req_timeout() 234 req->rq_timeout = req->rq_import->imp_server_timeout ? in ptlrpc_at_set_req_timeout() [all …]
|
D | sec_null.c | 79 int null_ctx_sign(struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req) in null_ctx_sign() argument 81 req->rq_reqbuf->lm_secflvr = SPTLRPC_FLVR_NULL; in null_ctx_sign() 83 if (!req->rq_import->imp_dlm_fake) { in null_ctx_sign() 84 struct obd_device *obd = req->rq_import->imp_obd; in null_ctx_sign() 85 null_encode_sec_part(req->rq_reqbuf, in null_ctx_sign() 88 req->rq_reqdata_len = req->rq_reqlen; in null_ctx_sign() 93 int null_ctx_verify(struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req) in null_ctx_verify() argument 97 LASSERT(req->rq_repdata); in null_ctx_verify() 99 req->rq_repmsg = req->rq_repdata; in null_ctx_verify() 100 req->rq_replen = req->rq_repdata_len; in null_ctx_verify() [all …]
|
D | sec.c | 318 struct ptlrpc_request *req, *next; in sptlrpc_cli_ctx_wakeup() local 321 list_for_each_entry_safe(req, next, &ctx->cc_req_list, in sptlrpc_cli_ctx_wakeup() 323 list_del_init(&req->rq_ctx_chain); in sptlrpc_cli_ctx_wakeup() 324 ptlrpc_client_wake_req(req); in sptlrpc_cli_ctx_wakeup() 393 int sptlrpc_req_get_ctx(struct ptlrpc_request *req) in sptlrpc_req_get_ctx() argument 395 struct obd_import *imp = req->rq_import; in sptlrpc_req_get_ctx() 399 LASSERT(!req->rq_cli_ctx); in sptlrpc_req_get_ctx() 406 req->rq_cli_ctx = get_my_ctx(sec); in sptlrpc_req_get_ctx() 410 if (!req->rq_cli_ctx) { in sptlrpc_req_get_ctx() 411 CERROR("req %p: fail to get context\n", req); in sptlrpc_req_get_ctx() [all …]
|
D | sec_plain.c | 191 int plain_ctx_sign(struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req) in plain_ctx_sign() argument 193 struct lustre_msg *msg = req->rq_reqbuf; in plain_ctx_sign() 196 msg->lm_secflvr = req->rq_flvr.sf_rpc; in plain_ctx_sign() 202 phdr->ph_bulk_hash_alg = req->rq_flvr.u_bulk.hash.hash_alg; in plain_ctx_sign() 204 if (req->rq_pack_udesc) in plain_ctx_sign() 206 if (req->rq_pack_bulk) in plain_ctx_sign() 209 req->rq_reqdata_len = lustre_msg_size_v2(msg->lm_bufcount, in plain_ctx_sign() 215 int plain_ctx_verify(struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req) in plain_ctx_verify() argument 217 struct lustre_msg *msg = req->rq_repdata; in plain_ctx_verify() 227 swabbed = ptlrpc_rep_need_swab(req); in plain_ctx_verify() [all …]
|
D | events.c | 57 struct ptlrpc_request *req = cbid->cbid_arg; in request_out_callback() local 63 DEBUG_REQ(D_NET, req, "type %d, status %d", ev->type, ev->status); in request_out_callback() 65 sptlrpc_request_out_callback(req); in request_out_callback() 66 spin_lock(&req->rq_lock); in request_out_callback() 67 req->rq_real_sent = get_seconds(); in request_out_callback() 69 req->rq_req_unlink = 0; in request_out_callback() 76 req->rq_net_err = 1; in request_out_callback() 77 ptlrpc_client_wake_req(req); in request_out_callback() 79 spin_unlock(&req->rq_lock); in request_out_callback() 81 ptlrpc_req_finished(req); in request_out_callback() [all …]
|
D | niobuf.c | 114 int ptlrpc_register_bulk(struct ptlrpc_request *req) in ptlrpc_register_bulk() argument 116 struct ptlrpc_bulk_desc *desc = req->rq_bulk; in ptlrpc_register_bulk() 139 if (req->rq_resend || req->rq_send_state == LUSTRE_IMP_REPLAY) in ptlrpc_register_bulk() 158 xid = req->rq_xid & ~((__u64)desc->bd_md_max_brw - 1); in ptlrpc_register_bulk() 160 req->rq_send_state != LUSTRE_IMP_REPLAY) || in ptlrpc_register_bulk() 208 req->rq_status = -ENOMEM; in ptlrpc_register_bulk() 214 req->rq_xid = --xid; in ptlrpc_register_bulk() 215 LASSERTF(desc->bd_last_xid == (req->rq_xid & PTLRPC_BULK_OPS_MASK), in ptlrpc_register_bulk() 217 desc->bd_last_xid, req->rq_xid); in ptlrpc_register_bulk() 231 desc->bd_last_xid, req->rq_xid, desc->bd_portal); in ptlrpc_register_bulk() [all …]
|
D | llog_client.c | 86 struct ptlrpc_request *req = NULL; in llog_client_open() local 95 req = ptlrpc_request_alloc(imp, &RQF_LLOG_ORIGIN_HANDLE_CREATE); in llog_client_open() 96 if (req == NULL) { in llog_client_open() 102 req_capsule_set_size(&req->rq_pill, &RMF_NAME, RCL_CLIENT, in llog_client_open() 105 rc = ptlrpc_request_pack(req, LUSTRE_LOG_VERSION, in llog_client_open() 108 ptlrpc_request_free(req); in llog_client_open() 109 req = NULL; in llog_client_open() 112 ptlrpc_request_set_replen(req); in llog_client_open() 114 body = req_capsule_client_get(&req->rq_pill, &RMF_LLOGD_BODY); in llog_client_open() 121 tmp = req_capsule_client_sized_get(&req->rq_pill, &RMF_NAME, in llog_client_open() [all …]
|
D | service.c | 64 static void ptlrpc_server_hpreq_fini(struct ptlrpc_request *req); 65 static void ptlrpc_at_remove_timed(struct ptlrpc_request *req); 183 ptlrpc_save_lock(struct ptlrpc_request *req, in ptlrpc_save_lock() argument 186 struct ptlrpc_reply_state *rs = req->rq_reply_state; in ptlrpc_save_lock() 192 if (req->rq_export->exp_disconnected) { in ptlrpc_save_lock() 830 static void ptlrpc_server_free_request(struct ptlrpc_request *req) in ptlrpc_server_free_request() argument 832 LASSERT(atomic_read(&req->rq_refcount) == 0); in ptlrpc_server_free_request() 833 LASSERT(list_empty(&req->rq_timed_list)); in ptlrpc_server_free_request() 837 ptlrpc_req_drop_rs(req); in ptlrpc_server_free_request() 839 sptlrpc_svc_ctx_decref(req); in ptlrpc_server_free_request() [all …]
|
D | recover.c | 73 struct ptlrpc_request *req = NULL; in ptlrpc_replay_next() local 110 req = list_entry(tmp, struct ptlrpc_request, in ptlrpc_replay_next() 114 if (req->rq_transno > last_transno) { in ptlrpc_replay_next() 122 req = list_entry(imp->imp_replay_cursor, in ptlrpc_replay_next() 125 if (req->rq_transno > last_transno) in ptlrpc_replay_next() 128 req = NULL; in ptlrpc_replay_next() 135 req = NULL; in ptlrpc_replay_next() 141 if (req == NULL) { in ptlrpc_replay_next() 143 req = list_entry(tmp, struct ptlrpc_request, in ptlrpc_replay_next() 146 if (req->rq_transno > last_transno) in ptlrpc_replay_next() [all …]
|
/linux-4.1.27/drivers/macintosh/ |
D | via-pmu68k.c | 106 static int pmu_send_request(struct adb_request *req, int sync); 114 static void pmu_done(struct adb_request *req); 192 volatile struct adb_request req; in pmu_init() local 197 pmu_request((struct adb_request *) &req, NULL, 2, PMU_SET_INTR_MASK, PMU_INT_ADB); in pmu_init() 199 while (!req.complete) { in pmu_init() 224 pmu_request((struct adb_request *) &req, NULL, 2, PMU_SET_INTR_MASK, in pmu_init() 227 while (!req.complete) { in pmu_init() 272 pmu_send_request(struct adb_request *req, int sync) in pmu_send_request() argument 278 req->complete = 1; in pmu_send_request() 284 switch (req->data[0]) { in pmu_send_request() [all …]
|
D | adb-iop.c | 66 static void adb_iop_end_req(struct adb_request *req, int state) in adb_iop_end_req() argument 68 req->complete = 1; in adb_iop_end_req() 69 current_req = req->next; in adb_iop_end_req() 70 if (req->done) (*req->done)(req); in adb_iop_end_req() 82 struct adb_request *req; in adb_iop_complete() local 87 req = current_req; in adb_iop_complete() 88 if ((adb_iop_state == sending) && req && req->reply_expected) { in adb_iop_complete() 105 struct adb_request *req; in adb_iop_listen() local 113 req = current_req; in adb_iop_listen() 116 printk("adb_iop_listen %p: rcvd packet, %d bytes: %02X %02X", req, in adb_iop_listen() [all …]
|
D | via-macii.c | 86 static int macii_send_request(struct adb_request *req, int sync); 87 static int macii_write(struct adb_request *req); 123 static int request_is_queued(struct adb_request *req) { in request_is_queued() argument 129 if (cur == req) { in request_is_queued() 205 static struct adb_request req; in macii_queue_poll() local 215 BUG_ON(request_is_queued(&req)); in macii_queue_poll() 217 adb_request(&req, NULL, ADBREQ_NOSEND, 1, in macii_queue_poll() 220 req.sent = 0; in macii_queue_poll() 221 req.complete = 0; in macii_queue_poll() 222 req.reply_len = 0; in macii_queue_poll() [all …]
|
D | via-pmu.c | 184 static int pmu_send_request(struct adb_request *req, int sync); 216 int pmu_polled_request(struct adb_request *req); 531 struct adb_request req; in init_pmu() local 536 pmu_request(&req, NULL, 2, PMU_SET_INTR_MASK, pmu_intr_mask); in init_pmu() 538 while (!req.complete) { in init_pmu() 563 pmu_request(&req, NULL, 2, PMU_SYSTEM_READY, 2); in init_pmu() 564 while (!req.complete) in init_pmu() 569 pmu_request(&req, NULL, 1, PMU_GET_VERSION); in init_pmu() 570 pmu_wait_complete(&req); in init_pmu() 571 if (req.reply_len > 0) in init_pmu() [all …]
|
D | via-maciisi.c | 82 static int maciisi_send_request(struct adb_request* req, int sync); 83 static void maciisi_sync(struct adb_request *req); 84 static int maciisi_write(struct adb_request* req); 229 maciisi_send_request(struct adb_request* req, int sync) in maciisi_send_request() argument 238 req->complete = 1; in maciisi_send_request() 245 for (i = 0; i < req->nbytes; i++) { in maciisi_send_request() 246 printk(" %.2x", req->data[i]); in maciisi_send_request() 252 req->reply_expected = 1; in maciisi_send_request() 254 i = maciisi_write(req); in maciisi_send_request() 271 maciisi_sync(req); in maciisi_send_request() [all …]
|
D | via-cuda.c | 91 static int cuda_send_request(struct adb_request *req, int sync); 101 static int cuda_write(struct adb_request *req); 103 int cuda_request(struct adb_request *req, 120 struct adb_request req; in find_via_cuda() local 137 cuda_request(&req, NULL, 3, CUDA_PACKET, CUDA_AUTOPOLL, 1); in find_via_cuda() 138 while (!req.complete) in find_via_cuda() 146 struct adb_request req; in find_via_cuda() local 190 cuda_request(&req, NULL, 3, CUDA_PACKET, CUDA_AUTOPOLL, 1); in find_via_cuda() 191 while (!req.complete) in find_via_cuda() 307 cuda_send_request(struct adb_request *req, int sync) in cuda_send_request() argument [all …]
|
D | adb.c | 116 static void printADBreply(struct adb_request *req) 120 printk("adb reply (%d)", req->reply_len); 121 for(i = 0; i < req->reply_len; i++) 122 printk(" %x", req->reply[i]); 132 struct adb_request req; in adb_scan_bus() local 137 adb_request(&req, NULL, ADBREQ_SYNC | ADBREQ_REPLY, 1, in adb_scan_bus() 139 if (req.reply_len > 1) in adb_scan_bus() 156 adb_request(&req, NULL, ADBREQ_SYNC | ADBREQ_REPLY, 1, in adb_scan_bus() 163 adb_request(&req, NULL, ADBREQ_SYNC, 3, in adb_scan_bus() 171 adb_request(&req, NULL, ADBREQ_SYNC | ADBREQ_REPLY, 1, in adb_scan_bus() [all …]
|
D | macio-adb.c | 66 static int macio_send_request(struct adb_request *req, int sync); 166 static int macio_send_request(struct adb_request *req, int sync) in macio_send_request() argument 171 if (req->data[0] != ADB_PACKET) in macio_send_request() 174 for (i = 0; i < req->nbytes - 1; ++i) in macio_send_request() 175 req->data[i] = req->data[i+1]; in macio_send_request() 176 --req->nbytes; in macio_send_request() 178 req->next = NULL; in macio_send_request() 179 req->sent = 0; in macio_send_request() 180 req->complete = 0; in macio_send_request() 181 req->reply_len = 0; in macio_send_request() [all …]
|
/linux-4.1.27/drivers/s390/scsi/ |
D | zfcp_fsf.c | 66 static void zfcp_fsf_class_not_supp(struct zfcp_fsf_req *req) in zfcp_fsf_class_not_supp() argument 68 dev_err(&req->adapter->ccw_device->dev, "FCP device not " in zfcp_fsf_class_not_supp() 70 zfcp_erp_adapter_shutdown(req->adapter, 0, "fscns_1"); in zfcp_fsf_class_not_supp() 71 req->status |= ZFCP_STATUS_FSFREQ_ERROR; in zfcp_fsf_class_not_supp() 78 void zfcp_fsf_req_free(struct zfcp_fsf_req *req) in zfcp_fsf_req_free() argument 80 if (likely(req->pool)) { in zfcp_fsf_req_free() 81 if (likely(req->qtcb)) in zfcp_fsf_req_free() 82 mempool_free(req->qtcb, req->adapter->pool.qtcb_pool); in zfcp_fsf_req_free() 83 mempool_free(req, req->pool); in zfcp_fsf_req_free() 87 if (likely(req->qtcb)) in zfcp_fsf_req_free() [all …]
|
D | zfcp_reqlist.h | 85 struct zfcp_fsf_req *req; in _zfcp_reqlist_find() local 89 list_for_each_entry(req, &rl->buckets[i], list) in _zfcp_reqlist_find() 90 if (req->req_id == req_id) in _zfcp_reqlist_find() 91 return req; in _zfcp_reqlist_find() 107 struct zfcp_fsf_req *req; in zfcp_reqlist_find() local 110 req = _zfcp_reqlist_find(rl, req_id); in zfcp_reqlist_find() 113 return req; in zfcp_reqlist_find() 132 struct zfcp_fsf_req *req; in zfcp_reqlist_find_rm() local 135 req = _zfcp_reqlist_find(rl, req_id); in zfcp_reqlist_find_rm() 136 if (req) in zfcp_reqlist_find_rm() [all …]
|
D | zfcp_dbf.h | 279 void zfcp_dbf_hba_fsf_resp(char *tag, int level, struct zfcp_fsf_req *req) in zfcp_dbf_hba_fsf_resp() argument 281 if (debug_level_enabled(req->adapter->dbf->hba, level)) in zfcp_dbf_hba_fsf_resp() 282 zfcp_dbf_hba_fsf_res(tag, req); in zfcp_dbf_hba_fsf_resp() 290 void zfcp_dbf_hba_fsf_response(struct zfcp_fsf_req *req) in zfcp_dbf_hba_fsf_response() argument 292 struct fsf_qtcb *qtcb = req->qtcb; in zfcp_dbf_hba_fsf_response() 296 zfcp_dbf_hba_fsf_resp("fs_perr", 1, req); in zfcp_dbf_hba_fsf_response() 299 zfcp_dbf_hba_fsf_resp("fs_ferr", 1, req); in zfcp_dbf_hba_fsf_response() 301 } else if ((req->fsf_command == FSF_QTCB_OPEN_PORT_WITH_DID) || in zfcp_dbf_hba_fsf_response() 302 (req->fsf_command == FSF_QTCB_OPEN_LUN)) { in zfcp_dbf_hba_fsf_response() 303 zfcp_dbf_hba_fsf_resp("fs_open", 4, req); in zfcp_dbf_hba_fsf_response() [all …]
|
/linux-4.1.27/drivers/block/drbd/ |
D | drbd_req.c | 37 static void _drbd_start_io_acct(struct drbd_device *device, struct drbd_request *req) in _drbd_start_io_acct() argument 39 generic_start_io_acct(bio_data_dir(req->master_bio), req->i.size >> 9, in _drbd_start_io_acct() 44 static void _drbd_end_io_acct(struct drbd_device *device, struct drbd_request *req) in _drbd_end_io_acct() argument 46 generic_end_io_acct(bio_data_dir(req->master_bio), in _drbd_end_io_acct() 47 &device->vdisk->part0, req->start_jif); in _drbd_end_io_acct() 53 struct drbd_request *req; in drbd_req_new() local 55 req = mempool_alloc(drbd_request_mempool, GFP_NOIO); in drbd_req_new() 56 if (!req) in drbd_req_new() 58 memset(req, 0, sizeof(*req)); in drbd_req_new() 60 drbd_req_make_private_bio(req, bio_src); in drbd_req_new() [all …]
|
/linux-4.1.27/net/tipc/ |
D | discover.c | 273 static void disc_update(struct tipc_link_req *req) in disc_update() argument 275 if (!req->num_nodes) { in disc_update() 276 if ((req->timer_intv == TIPC_LINK_REQ_INACTIVE) || in disc_update() 277 (req->timer_intv > TIPC_LINK_REQ_FAST)) { in disc_update() 278 req->timer_intv = TIPC_LINK_REQ_INIT; in disc_update() 279 mod_timer(&req->timer, jiffies + req->timer_intv); in disc_update() 288 void tipc_disc_add_dest(struct tipc_link_req *req) in tipc_disc_add_dest() argument 290 spin_lock_bh(&req->lock); in tipc_disc_add_dest() 291 req->num_nodes++; in tipc_disc_add_dest() 292 spin_unlock_bh(&req->lock); in tipc_disc_add_dest() [all …]
|
/linux-4.1.27/net/sunrpc/ |
D | backchannel_rqst.c | 57 static void xprt_free_allocation(struct rpc_rqst *req) in xprt_free_allocation() argument 61 dprintk("RPC: free allocations for req= %p\n", req); in xprt_free_allocation() 62 WARN_ON_ONCE(test_bit(RPC_BC_PA_IN_USE, &req->rq_bc_pa_state)); in xprt_free_allocation() 63 xbufp = &req->rq_rcv_buf; in xprt_free_allocation() 65 xbufp = &req->rq_snd_buf; in xprt_free_allocation() 67 kfree(req); in xprt_free_allocation() 92 struct rpc_rqst *req, *tmp; in xprt_setup_backchannel() local 109 req = kzalloc(sizeof(struct rpc_rqst), GFP_KERNEL); in xprt_setup_backchannel() 110 if (req == NULL) { in xprt_setup_backchannel() 116 dprintk("RPC: adding req= %p\n", req); in xprt_setup_backchannel() [all …]
|
D | xprt.c | 183 struct rpc_rqst *req = task->tk_rqstp; in xprt_reserve_xprt() local 192 if (req != NULL) in xprt_reserve_xprt() 193 req->rq_ntrans++; in xprt_reserve_xprt() 202 if (req == NULL) in xprt_reserve_xprt() 204 else if (!req->rq_ntrans) in xprt_reserve_xprt() 234 struct rpc_rqst *req = task->tk_rqstp; in xprt_reserve_xprt_cong() local 242 if (req == NULL) { in xprt_reserve_xprt_cong() 248 req->rq_ntrans++; in xprt_reserve_xprt_cong() 256 if (req == NULL) in xprt_reserve_xprt_cong() 258 else if (!req->rq_ntrans) in xprt_reserve_xprt_cong() [all …]
|
/linux-4.1.27/drivers/s390/cio/ |
D | ccwreq.c | 42 struct ccw_request *req = &cdev->private->req; in ccwreq_next_path() local 44 if (!req->singlepath) { in ccwreq_next_path() 45 req->mask = 0; in ccwreq_next_path() 48 req->retries = req->maxretries; in ccwreq_next_path() 49 req->mask = lpm_adjust(req->mask >> 1, req->lpm); in ccwreq_next_path() 51 return req->mask; in ccwreq_next_path() 59 struct ccw_request *req = &cdev->private->req; in ccwreq_stop() local 61 if (req->done) in ccwreq_stop() 63 req->done = 1; in ccwreq_stop() 66 if (rc && rc != -ENODEV && req->drc) in ccwreq_stop() [all …]
|
D | device_pgid.c | 57 struct ccw_request *req = &cdev->private->req; in nop_build_cp() local 64 req->cp = cp; in nop_build_cp() 73 struct ccw_request *req = &cdev->private->req; in nop_do() local 75 req->lpm = lpm_adjust(req->lpm, sch->schib.pmcw.pam & sch->opm & in nop_do() 77 if (!req->lpm) in nop_do() 105 struct ccw_request *req = &cdev->private->req; in nop_callback() local 109 sch->vpm |= req->lpm; in nop_callback() 112 cdev->private->path_noirq_mask |= req->lpm; in nop_callback() 115 cdev->private->path_notoper_mask |= req->lpm; in nop_callback() 121 req->lpm >>= 1; in nop_callback() [all …]
|
/linux-4.1.27/arch/powerpc/platforms/52xx/ |
D | mpc52xx_lpbfifo.c | 53 struct mpc52xx_lpbfifo_request *req; member 63 static void mpc52xx_lpbfifo_kick(struct mpc52xx_lpbfifo_request *req) in mpc52xx_lpbfifo_kick() argument 65 size_t transfer_size = req->size - req->pos; in mpc52xx_lpbfifo_kick() 71 int dma = !(req->flags & MPC52XX_LPBFIFO_FLAG_NO_DMA); in mpc52xx_lpbfifo_kick() 72 int write = req->flags & MPC52XX_LPBFIFO_FLAG_WRITE; in mpc52xx_lpbfifo_kick() 73 int poll_dma = req->flags & MPC52XX_LPBFIFO_FLAG_POLL_DMA; in mpc52xx_lpbfifo_kick() 96 data = req->data + req->pos; in mpc52xx_lpbfifo_kick() 150 bd->data[0] = req->data_phys + req->pos; in mpc52xx_lpbfifo_kick() 164 req->offset + req->pos); in mpc52xx_lpbfifo_kick() 167 bit_fields = req->cs << 24 | 0x000008; in mpc52xx_lpbfifo_kick() [all …]
|
/linux-4.1.27/drivers/video/fbdev/msm/ |
D | mdp_ppp.c | 92 static void rotate_dst_addr_x(struct mdp_blit_req *req, struct mdp_regs *regs) in rotate_dst_addr_x() argument 94 regs->dst0 += (req->dst_rect.w - in rotate_dst_addr_x() 95 min((uint32_t)16, req->dst_rect.w)) * regs->dst_bpp; in rotate_dst_addr_x() 96 regs->dst1 += (req->dst_rect.w - in rotate_dst_addr_x() 97 min((uint32_t)16, req->dst_rect.w)) * regs->dst_bpp; in rotate_dst_addr_x() 100 static void rotate_dst_addr_y(struct mdp_blit_req *req, struct mdp_regs *regs) in rotate_dst_addr_y() argument 102 regs->dst0 += (req->dst_rect.h - in rotate_dst_addr_y() 103 min((uint32_t)16, req->dst_rect.h)) * in rotate_dst_addr_y() 105 regs->dst1 += (req->dst_rect.h - in rotate_dst_addr_y() 106 min((uint32_t)16, req->dst_rect.h)) * in rotate_dst_addr_y() [all …]
|
/linux-4.1.27/net/ceph/ |
D | osd_client.c | 32 struct ceph_osd_request *req); 34 struct ceph_osd_request *req); 36 struct ceph_osd_request *req); 37 static void __enqueue_request(struct ceph_osd_request *req); 39 struct ceph_osd_request *req); 309 struct ceph_osd_request *req = container_of(kref, in ceph_osdc_release_request() local 313 dout("%s %p (r_request %p r_reply %p)\n", __func__, req, in ceph_osdc_release_request() 314 req->r_request, req->r_reply); in ceph_osdc_release_request() 315 WARN_ON(!RB_EMPTY_NODE(&req->r_node)); in ceph_osdc_release_request() 316 WARN_ON(!list_empty(&req->r_req_lru_item)); in ceph_osdc_release_request() [all …]
|
D | mon_client.c | 418 struct ceph_mon_generic_request *req; in __lookup_generic_req() local 422 req = rb_entry(n, struct ceph_mon_generic_request, node); in __lookup_generic_req() 423 if (tid < req->tid) in __lookup_generic_req() 425 else if (tid > req->tid) in __lookup_generic_req() 428 return req; in __lookup_generic_req() 438 struct ceph_mon_generic_request *req = NULL; in __insert_generic_request() local 442 req = rb_entry(parent, struct ceph_mon_generic_request, node); in __insert_generic_request() 443 if (new->tid < req->tid) in __insert_generic_request() 445 else if (new->tid > req->tid) in __insert_generic_request() 457 struct ceph_mon_generic_request *req = in release_generic_request() local [all …]
|
/linux-4.1.27/fs/ncpfs/ |
D | sock.c | 59 struct list_head req; member 76 struct ncp_request_reply *req; in ncp_alloc_req() local 78 req = kmalloc(sizeof(struct ncp_request_reply), GFP_KERNEL); in ncp_alloc_req() 79 if (!req) in ncp_alloc_req() 82 init_waitqueue_head(&req->wq); in ncp_alloc_req() 83 atomic_set(&req->refs, (1)); in ncp_alloc_req() 84 req->status = RQ_IDLE; in ncp_alloc_req() 86 return req; in ncp_alloc_req() 89 static void ncp_req_get(struct ncp_request_reply *req) in ncp_req_get() argument 91 atomic_inc(&req->refs); in ncp_req_get() [all …]
|
/linux-4.1.27/include/crypto/ |
D | skcipher.h | 34 struct skcipher_givcrypt_request *req) in skcipher_givcrypt_reqtfm() argument 36 return crypto_ablkcipher_reqtfm(&req->creq); in skcipher_givcrypt_reqtfm() 40 struct skcipher_givcrypt_request *req) in crypto_skcipher_givencrypt() argument 43 crypto_ablkcipher_crt(skcipher_givcrypt_reqtfm(req)); in crypto_skcipher_givencrypt() 44 return crt->givencrypt(req); in crypto_skcipher_givencrypt() 48 struct skcipher_givcrypt_request *req) in crypto_skcipher_givdecrypt() argument 51 crypto_ablkcipher_crt(skcipher_givcrypt_reqtfm(req)); in crypto_skcipher_givdecrypt() 52 return crt->givdecrypt(req); in crypto_skcipher_givdecrypt() 56 struct skcipher_givcrypt_request *req, struct crypto_ablkcipher *tfm) in skcipher_givcrypt_set_tfm() argument 58 req->creq.base.tfm = crypto_ablkcipher_tfm(tfm); in skcipher_givcrypt_set_tfm() [all …]
|
D | aead.h | 34 struct aead_givcrypt_request *req) in aead_givcrypt_reqtfm() argument 36 return crypto_aead_reqtfm(&req->areq); in aead_givcrypt_reqtfm() 39 static inline int crypto_aead_givencrypt(struct aead_givcrypt_request *req) in crypto_aead_givencrypt() argument 41 struct aead_tfm *crt = crypto_aead_crt(aead_givcrypt_reqtfm(req)); in crypto_aead_givencrypt() 42 return crt->givencrypt(req); in crypto_aead_givencrypt() 45 static inline int crypto_aead_givdecrypt(struct aead_givcrypt_request *req) in crypto_aead_givdecrypt() argument 47 struct aead_tfm *crt = crypto_aead_crt(aead_givcrypt_reqtfm(req)); in crypto_aead_givdecrypt() 48 return crt->givdecrypt(req); in crypto_aead_givdecrypt() 51 static inline void aead_givcrypt_set_tfm(struct aead_givcrypt_request *req, in aead_givcrypt_set_tfm() argument 54 req->areq.base.tfm = crypto_aead_tfm(tfm); in aead_givcrypt_set_tfm() [all …]
|
D | hash.h | 124 int (*init)(struct ahash_request *req); 125 int (*update)(struct ahash_request *req); 126 int (*final)(struct ahash_request *req); 127 int (*finup)(struct ahash_request *req); 128 int (*digest)(struct ahash_request *req); 129 int (*export)(struct ahash_request *req, void *out); 130 int (*import)(struct ahash_request *req, const void *in); 191 int (*init)(struct ahash_request *req); 192 int (*update)(struct ahash_request *req); 193 int (*final)(struct ahash_request *req); [all …]
|
D | compress.h | 62 struct comp_request *req); 64 struct comp_request *req); 69 struct comp_request *req); 71 struct comp_request *req); 111 struct comp_request *req) in crypto_compress_update() argument 113 return crypto_pcomp_alg(tfm)->compress_update(tfm, req); in crypto_compress_update() 117 struct comp_request *req) in crypto_compress_final() argument 119 return crypto_pcomp_alg(tfm)->compress_final(tfm, req); in crypto_compress_final() 134 struct comp_request *req) in crypto_decompress_update() argument 136 return crypto_pcomp_alg(tfm)->decompress_update(tfm, req); in crypto_decompress_update() [all …]
|
/linux-4.1.27/net/bluetooth/ |
D | hci_request.c | 30 void hci_req_init(struct hci_request *req, struct hci_dev *hdev) in hci_req_init() argument 32 skb_queue_head_init(&req->cmd_q); in hci_req_init() 33 req->hdev = hdev; in hci_req_init() 34 req->err = 0; in hci_req_init() 37 static int req_run(struct hci_request *req, hci_req_complete_t complete, in req_run() argument 40 struct hci_dev *hdev = req->hdev; in req_run() 44 BT_DBG("length %u", skb_queue_len(&req->cmd_q)); in req_run() 49 if (req->err) { in req_run() 50 skb_queue_purge(&req->cmd_q); in req_run() 51 return req->err; in req_run() [all …]
|
D | hci_request.h | 33 void hci_req_init(struct hci_request *req, struct hci_dev *hdev); 34 int hci_req_run(struct hci_request *req, hci_req_complete_t complete); 35 int hci_req_run_skb(struct hci_request *req, hci_req_complete_skb_t complete); 36 void hci_req_add(struct hci_request *req, u16 opcode, u32 plen, 38 void hci_req_add_ev(struct hci_request *req, u16 opcode, u32 plen, 47 void hci_req_add_le_scan_disable(struct hci_request *req); 48 void hci_req_add_le_passive_scan(struct hci_request *req); 51 void __hci_update_page_scan(struct hci_request *req); 53 int hci_update_random_address(struct hci_request *req, bool require_privacy, 57 void __hci_update_background_scan(struct hci_request *req);
|
/linux-4.1.27/drivers/staging/lustre/lustre/mdc/ |
D | mdc_reint.c | 108 struct ptlrpc_request *req; in mdc_setattr() local 124 req = ptlrpc_request_alloc(class_exp2cliimp(exp), in mdc_setattr() 126 if (req == NULL) { in mdc_setattr() 130 mdc_set_capa_size(req, &RMF_CAPA1, op_data->op_capa1); in mdc_setattr() 132 req_capsule_set_size(&req->rq_pill, &RMF_MDT_EPOCH, RCL_CLIENT, in mdc_setattr() 134 req_capsule_set_size(&req->rq_pill, &RMF_EADATA, RCL_CLIENT, ealen); in mdc_setattr() 135 req_capsule_set_size(&req->rq_pill, &RMF_LOGCOOKIES, RCL_CLIENT, in mdc_setattr() 138 rc = mdc_prep_elc_req(exp, req, MDS_REINT, &cancels, count); in mdc_setattr() 140 ptlrpc_request_free(req); in mdc_setattr() 151 mdc_setattr_pack(req, op_data, ea, ealen, ea2, ea2len); in mdc_setattr() [all …]
|
D | mdc_request.c | 63 static int mdc_unpack_capa(struct obd_export *exp, struct ptlrpc_request *req, in mdc_unpack_capa() argument 70 capa = req_capsule_server_get(&req->rq_pill, field); in mdc_unpack_capa() 85 static inline int mdc_queue_wait(struct ptlrpc_request *req) in mdc_queue_wait() argument 87 struct client_obd *cli = &req->rq_import->imp_obd->u.cli; in mdc_queue_wait() 97 rc = ptlrpc_queue_wait(req); in mdc_queue_wait() 108 struct ptlrpc_request *req; in send_getstatus() local 112 req = ptlrpc_request_alloc_pack(imp, &RQF_MDS_GETSTATUS, in send_getstatus() 114 if (req == NULL) in send_getstatus() 117 mdc_pack_body(req, NULL, NULL, 0, 0, -1, 0); in send_getstatus() 118 lustre_msg_add_flags(req->rq_reqmsg, msg_flags); in send_getstatus() [all …]
|
D | mdc_locks.c | 231 static inline void mdc_clear_replay_flag(struct ptlrpc_request *req, int rc) in mdc_clear_replay_flag() argument 234 if (req->rq_replay) { in mdc_clear_replay_flag() 235 spin_lock(&req->rq_lock); in mdc_clear_replay_flag() 236 req->rq_replay = 0; in mdc_clear_replay_flag() 237 spin_unlock(&req->rq_lock); in mdc_clear_replay_flag() 239 if (rc && req->rq_transno != 0) { in mdc_clear_replay_flag() 240 DEBUG_REQ(D_ERROR, req, "transno returned on error rc %d", rc); in mdc_clear_replay_flag() 256 static void mdc_realloc_openmsg(struct ptlrpc_request *req, in mdc_realloc_openmsg() argument 262 rc = sptlrpc_cli_enlarge_reqbuf(req, DLM_INTENT_REC_OFF + 4, in mdc_realloc_openmsg() 278 struct ptlrpc_request *req; in mdc_intent_open_pack() local [all …]
|
D | mdc_lib.c | 55 void mdc_pack_capa(struct ptlrpc_request *req, in mdc_pack_capa() argument 59 struct req_capsule *pill = &req->rq_pill; in mdc_pack_capa() 73 void mdc_is_subdir_pack(struct ptlrpc_request *req, const struct lu_fid *pfid, in mdc_is_subdir_pack() argument 76 struct mdt_body *b = req_capsule_client_get(&req->rq_pill, in mdc_is_subdir_pack() 88 void mdc_swap_layouts_pack(struct ptlrpc_request *req, in mdc_swap_layouts_pack() argument 91 struct mdt_body *b = req_capsule_client_get(&req->rq_pill, in mdc_swap_layouts_pack() 99 mdc_pack_capa(req, &RMF_CAPA1, op_data->op_capa1); in mdc_swap_layouts_pack() 100 mdc_pack_capa(req, &RMF_CAPA2, op_data->op_capa2); in mdc_swap_layouts_pack() 103 void mdc_pack_body(struct ptlrpc_request *req, in mdc_pack_body() argument 107 struct mdt_body *b = req_capsule_client_get(&req->rq_pill, in mdc_pack_body() [all …]
|
D | mdc_internal.h | 52 void mdc_pack_body(struct ptlrpc_request *req, const struct lu_fid *fid, 55 void mdc_pack_capa(struct ptlrpc_request *req, 57 int mdc_pack_req(struct ptlrpc_request *req, int version, int opc); 58 void mdc_is_subdir_pack(struct ptlrpc_request *req, const struct lu_fid *pfid, 60 void mdc_swap_layouts_pack(struct ptlrpc_request *req, 62 void mdc_readdir_pack(struct ptlrpc_request *req, __u64 pgoff, __u32 size, 64 void mdc_getattr_pack(struct ptlrpc_request *req, __u64 valid, int flags, 66 void mdc_setattr_pack(struct ptlrpc_request *req, struct md_op_data *op_data, 68 void mdc_create_pack(struct ptlrpc_request *req, struct md_op_data *op_data, 71 void mdc_open_pack(struct ptlrpc_request *req, struct md_op_data *op_data, [all …]
|
/linux-4.1.27/arch/um/drivers/ |
D | mconsole_user.c | 40 static int mconsole_reply_v0(struct mc_request *req, char *reply) in mconsole_reply_v0() argument 48 msg.msg_name = &(req->origin); in mconsole_reply_v0() 49 msg.msg_namelen = req->originlen; in mconsole_reply_v0() 56 return sendmsg(req->originating_fd, &msg, 0); in mconsole_reply_v0() 59 static struct mconsole_command *mconsole_parse(struct mc_request *req) in mconsole_parse() argument 66 if (!strncmp(req->request.data, cmd->command, in mconsole_parse() 79 int mconsole_get_request(int fd, struct mc_request *req) in mconsole_get_request() argument 83 req->originlen = sizeof(req->origin); in mconsole_get_request() 84 req->len = recvfrom(fd, &req->request, sizeof(req->request), 0, in mconsole_get_request() 85 (struct sockaddr *) req->origin, &req->originlen); in mconsole_get_request() [all …]
|
D | mconsole_kern.c | 59 struct mconsole_entry *req; in mc_work_proc() local 64 req = list_entry(mc_requests.next, struct mconsole_entry, list); in mc_work_proc() 65 list_del(&req->list); in mc_work_proc() 67 req->request.cmd->handler(&req->request); in mc_work_proc() 68 kfree(req); in mc_work_proc() 79 static struct mc_request req; /* that's OK */ in mconsole_interrupt() local 82 while (mconsole_get_request(fd, &req)) { in mconsole_interrupt() 83 if (req.cmd->context == MCONSOLE_INTR) in mconsole_interrupt() 84 (*req.cmd->handler)(&req); in mconsole_interrupt() 88 mconsole_reply(&req, "Out of memory", 1, 0); in mconsole_interrupt() [all …]
|
D | mconsole.h | 51 void (*handler)(struct mc_request *req); 72 extern int mconsole_reply_len(struct mc_request *req, const char *reply, 74 extern int mconsole_reply(struct mc_request *req, const char *str, int err, 77 extern void mconsole_version(struct mc_request *req); 78 extern void mconsole_help(struct mc_request *req); 79 extern void mconsole_halt(struct mc_request *req); 80 extern void mconsole_reboot(struct mc_request *req); 81 extern void mconsole_config(struct mc_request *req); 82 extern void mconsole_remove(struct mc_request *req); 83 extern void mconsole_sysrq(struct mc_request *req); [all …]
|
D | ubd_kern.c | 47 struct request *req; member 449 struct io_thread_req *req; in ubd_handler() local 456 n = os_read_file(thread_fd, &req, in ubd_handler() 458 if(n != sizeof(req)){ in ubd_handler() 466 blk_end_request(req->req, 0, req->length); in ubd_handler() 467 kfree(req); in ubd_handler() 1191 static void cowify_req(struct io_thread_req *req, unsigned long *bitmap, in cowify_req() argument 1194 __u64 sector = req->offset >> 9; in cowify_req() 1197 if(req->length > (sizeof(req->sector_mask) * 8) << 9) in cowify_req() 1200 if(req->op == UBD_READ) { in cowify_req() [all …]
|
/linux-4.1.27/crypto/ |
D | gcm.c | 63 void (*complete)(struct aead_request *req, int err); 86 struct aead_request *req) in crypto_gcm_reqctx() argument 88 unsigned long align = crypto_aead_alignmask(crypto_aead_reqtfm(req)); in crypto_gcm_reqctx() 90 return (void *)PTR_ALIGN((u8 *)aead_request_ctx(req), align + 1); in crypto_gcm_reqctx() 93 static void crypto_gcm_setkey_done(struct crypto_async_request *req, int err) in crypto_gcm_setkey_done() argument 95 struct crypto_gcm_setkey_result *result = req->data; in crypto_gcm_setkey_done() 117 struct ablkcipher_request req; in crypto_gcm_setkey() member 139 ablkcipher_request_set_tfm(&data->req, ctr); in crypto_gcm_setkey() 140 ablkcipher_request_set_callback(&data->req, CRYPTO_TFM_REQ_MAY_SLEEP | in crypto_gcm_setkey() 144 ablkcipher_request_set_crypt(&data->req, data->sg, data->sg, in crypto_gcm_setkey() [all …]
|
D | ahash.c | 131 int crypto_hash_walk_first(struct ahash_request *req, in crypto_hash_walk_first() argument 134 walk->total = req->nbytes; in crypto_hash_walk_first() 141 walk->alignmask = crypto_ahash_alignmask(crypto_ahash_reqtfm(req)); in crypto_hash_walk_first() 142 walk->sg = req->src; in crypto_hash_walk_first() 143 walk->flags = req->base.flags & CRYPTO_TFM_REQ_MASK; in crypto_hash_walk_first() 149 int crypto_ahash_walk_first(struct ahash_request *req, in crypto_ahash_walk_first() argument 152 walk->total = req->nbytes; in crypto_ahash_walk_first() 159 walk->alignmask = crypto_ahash_alignmask(crypto_ahash_reqtfm(req)); in crypto_ahash_walk_first() 160 walk->sg = req->src; in crypto_ahash_walk_first() 161 walk->flags = req->base.flags & CRYPTO_TFM_REQ_MASK; in crypto_ahash_walk_first() [all …]
|
D | zlib.c | 144 struct comp_request *req) in zlib_compress_update() argument 150 pr_debug("avail_in %u, avail_out %u\n", req->avail_in, req->avail_out); in zlib_compress_update() 151 stream->next_in = req->next_in; in zlib_compress_update() 152 stream->avail_in = req->avail_in; in zlib_compress_update() 153 stream->next_out = req->next_out; in zlib_compress_update() 154 stream->avail_out = req->avail_out; in zlib_compress_update() 170 ret = req->avail_out - stream->avail_out; in zlib_compress_update() 173 req->avail_in - stream->avail_in, ret); in zlib_compress_update() 174 req->next_in = stream->next_in; in zlib_compress_update() 175 req->avail_in = stream->avail_in; in zlib_compress_update() [all …]
|
D | authenc.c | 26 typedef u8 *(*authenc_ahash_t)(struct aead_request *req, unsigned int flags); 49 static void authenc_request_complete(struct aead_request *req, int err) in authenc_request_complete() argument 52 aead_request_complete(req, err); in authenc_request_complete() 125 struct aead_request *req = areq->data; in authenc_geniv_ahash_update_done() local 126 struct crypto_aead *authenc = crypto_aead_reqtfm(req); in authenc_geniv_ahash_update_done() 128 struct authenc_request_ctx *areq_ctx = aead_request_ctx(req); in authenc_geniv_ahash_update_done() 136 ahash_request_set_callback(ahreq, aead_request_flags(req) & in authenc_geniv_ahash_update_done() 138 areq_ctx->complete, req); in authenc_geniv_ahash_update_done() 149 authenc_request_complete(req, err); in authenc_geniv_ahash_update_done() 154 struct aead_request *req = areq->data; in authenc_geniv_ahash_done() local [all …]
|
D | authencesn.c | 53 static void authenc_esn_request_complete(struct aead_request *req, int err) in authenc_esn_request_complete() argument 56 aead_request_complete(req, err); in authenc_esn_request_complete() 99 struct aead_request *req = areq->data; in authenc_esn_geniv_ahash_update_done() local 100 struct crypto_aead *authenc_esn = crypto_aead_reqtfm(req); in authenc_esn_geniv_ahash_update_done() 102 struct authenc_esn_request_ctx *areq_ctx = aead_request_ctx(req); in authenc_esn_geniv_ahash_update_done() 110 ahash_request_set_callback(ahreq, aead_request_flags(req) & in authenc_esn_geniv_ahash_update_done() 112 areq_ctx->update_complete2, req); in authenc_esn_geniv_ahash_update_done() 120 ahash_request_set_callback(ahreq, aead_request_flags(req) & in authenc_esn_geniv_ahash_update_done() 122 areq_ctx->complete, req); in authenc_esn_geniv_ahash_update_done() 133 authenc_esn_request_complete(req, err); in authenc_esn_geniv_ahash_update_done() [all …]
|
D | chainiv.c | 48 static int chainiv_givencrypt(struct skcipher_givcrypt_request *req) in chainiv_givencrypt() argument 50 struct crypto_ablkcipher *geniv = skcipher_givcrypt_reqtfm(req); in chainiv_givencrypt() 52 struct ablkcipher_request *subreq = skcipher_givcrypt_reqctx(req); in chainiv_givencrypt() 57 ablkcipher_request_set_callback(subreq, req->creq.base.flags & in chainiv_givencrypt() 59 req->creq.base.complete, in chainiv_givencrypt() 60 req->creq.base.data); in chainiv_givencrypt() 61 ablkcipher_request_set_crypt(subreq, req->creq.src, req->creq.dst, in chainiv_givencrypt() 62 req->creq.nbytes, req->creq.info); in chainiv_givencrypt() 68 memcpy(req->giv, ctx->iv, ivsize); in chainiv_givencrypt() 83 static int chainiv_givencrypt_first(struct skcipher_givcrypt_request *req) in chainiv_givencrypt_first() argument [all …]
|
D | seqiv.c | 32 static void seqiv_complete2(struct skcipher_givcrypt_request *req, int err) in seqiv_complete2() argument 34 struct ablkcipher_request *subreq = skcipher_givcrypt_reqctx(req); in seqiv_complete2() 43 geniv = skcipher_givcrypt_reqtfm(req); in seqiv_complete2() 44 memcpy(req->creq.info, subreq->info, crypto_ablkcipher_ivsize(geniv)); in seqiv_complete2() 52 struct skcipher_givcrypt_request *req = base->data; in seqiv_complete() local 54 seqiv_complete2(req, err); in seqiv_complete() 55 skcipher_givcrypt_complete(req, err); in seqiv_complete() 58 static void seqiv_aead_complete2(struct aead_givcrypt_request *req, int err) in seqiv_aead_complete2() argument 60 struct aead_request *subreq = aead_givcrypt_reqctx(req); in seqiv_aead_complete2() 69 geniv = aead_givcrypt_reqtfm(req); in seqiv_aead_complete2() [all …]
|
D | cryptd.c | 137 struct crypto_async_request *req, *backlog; in cryptd_queue_worker() local 149 req = crypto_dequeue_request(&cpu_queue->queue); in cryptd_queue_worker() 153 if (!req) in cryptd_queue_worker() 158 req->complete(req, 0); in cryptd_queue_worker() 201 static void cryptd_blkcipher_crypt(struct ablkcipher_request *req, in cryptd_blkcipher_crypt() argument 212 rctx = ablkcipher_request_ctx(req); in cryptd_blkcipher_crypt() 218 desc.info = req->info; in cryptd_blkcipher_crypt() 221 err = crypt(&desc, req->dst, req->src, req->nbytes); in cryptd_blkcipher_crypt() 223 req->base.complete = rctx->complete; in cryptd_blkcipher_crypt() 227 rctx->complete(&req->base, err); in cryptd_blkcipher_crypt() [all …]
|
D | ablk_helper.c | 54 int __ablk_encrypt(struct ablkcipher_request *req) in __ablk_encrypt() argument 56 struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req); in __ablk_encrypt() 61 desc.info = req->info; in __ablk_encrypt() 65 &desc, req->dst, req->src, req->nbytes); in __ablk_encrypt() 69 int ablk_encrypt(struct ablkcipher_request *req) in ablk_encrypt() argument 71 struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req); in ablk_encrypt() 76 ablkcipher_request_ctx(req); in ablk_encrypt() 78 *cryptd_req = *req; in ablk_encrypt() 83 return __ablk_encrypt(req); in ablk_encrypt() 88 int ablk_decrypt(struct ablkcipher_request *req) in ablk_decrypt() argument [all …]
|
D | ccm.c | 51 struct aead_request *req) in crypto_ccm_reqctx() argument 53 unsigned long align = crypto_aead_alignmask(crypto_aead_reqtfm(req)); in crypto_ccm_reqctx() 55 return (void *)PTR_ALIGN((u8 *)aead_request_ctx(req), align + 1); in crypto_ccm_reqctx() 123 static int format_input(u8 *info, struct aead_request *req, in format_input() argument 126 struct crypto_aead *aead = crypto_aead_reqtfm(req); in format_input() 127 unsigned int lp = req->iv[0]; in format_input() 133 memcpy(info, req->iv, 16); in format_input() 139 if (req->assoclen) in format_input() 245 static int crypto_ccm_auth(struct aead_request *req, struct scatterlist *plain, in crypto_ccm_auth() argument 248 struct crypto_aead *aead = crypto_aead_reqtfm(req); in crypto_ccm_auth() [all …]
|
D | eseqiv.c | 42 static void eseqiv_complete2(struct skcipher_givcrypt_request *req) in eseqiv_complete2() argument 44 struct crypto_ablkcipher *geniv = skcipher_givcrypt_reqtfm(req); in eseqiv_complete2() 45 struct eseqiv_request_ctx *reqctx = skcipher_givcrypt_reqctx(req); in eseqiv_complete2() 47 memcpy(req->giv, PTR_ALIGN((u8 *)reqctx->tail, in eseqiv_complete2() 54 struct skcipher_givcrypt_request *req = base->data; in eseqiv_complete() local 59 eseqiv_complete2(req); in eseqiv_complete() 62 skcipher_givcrypt_complete(req, err); in eseqiv_complete() 65 static int eseqiv_givencrypt(struct skcipher_givcrypt_request *req) in eseqiv_givencrypt() argument 67 struct crypto_ablkcipher *geniv = skcipher_givcrypt_reqtfm(req); in eseqiv_givencrypt() 69 struct eseqiv_request_ctx *reqctx = skcipher_givcrypt_reqctx(req); in eseqiv_givencrypt() [all …]
|
D | mcryptd.c | 156 struct crypto_async_request *req, *backlog; in mcryptd_queue_worker() local 174 req = crypto_dequeue_request(&cpu_queue->queue); in mcryptd_queue_worker() 178 if (!req) { in mcryptd_queue_worker() 185 req->complete(req, 0); in mcryptd_queue_worker() 317 static int mcryptd_hash_enqueue(struct ahash_request *req, in mcryptd_hash_enqueue() argument 322 struct mcryptd_hash_request_ctx *rctx = ahash_request_ctx(req); in mcryptd_hash_enqueue() 323 struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); in mcryptd_hash_enqueue() 327 rctx->complete = req->base.complete; in mcryptd_hash_enqueue() 328 req->base.complete = complete; in mcryptd_hash_enqueue() 330 ret = mcryptd_enqueue_request(queue, &req->base, rctx); in mcryptd_hash_enqueue() [all …]
|
D | pcrypt.c | 120 struct aead_request *req = pcrypt_request_ctx(preq); in pcrypt_aead_serial() local 122 aead_request_complete(req->base.data, padata->info); in pcrypt_aead_serial() 128 struct aead_givcrypt_request *req = pcrypt_request_ctx(preq); in pcrypt_aead_giv_serial() local 130 aead_request_complete(req->areq.base.data, padata->info); in pcrypt_aead_giv_serial() 135 struct aead_request *req = areq->data; in pcrypt_aead_done() local 136 struct pcrypt_request *preq = aead_request_ctx(req); in pcrypt_aead_done() 140 req->base.flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP; in pcrypt_aead_done() 148 struct aead_request *req = pcrypt_request_ctx(preq); in pcrypt_aead_enc() local 150 padata->info = crypto_aead_encrypt(req); in pcrypt_aead_enc() 158 static int pcrypt_aead_encrypt(struct aead_request *req) in pcrypt_aead_encrypt() argument [all …]
|
/linux-4.1.27/net/9p/ |
D | client.c | 239 struct p9_req_t *req; in p9_tag_alloc() local 270 req = &c->reqs[row][col]; in p9_tag_alloc() 271 if (!req->wq) { in p9_tag_alloc() 272 req->wq = kmalloc(sizeof(wait_queue_head_t), GFP_NOFS); in p9_tag_alloc() 273 if (!req->wq) in p9_tag_alloc() 275 init_waitqueue_head(req->wq); in p9_tag_alloc() 278 if (!req->tc) in p9_tag_alloc() 279 req->tc = p9_fcall_alloc(alloc_msize); in p9_tag_alloc() 280 if (!req->rc) in p9_tag_alloc() 281 req->rc = p9_fcall_alloc(alloc_msize); in p9_tag_alloc() [all …]
|
/linux-4.1.27/fs/nfs/ |
D | pagelist.c | 61 hdr->req = nfs_list_entry(mirror->pg_list.next); in nfs_pgheader_init() 63 hdr->cred = hdr->req->wb_context->cred; in nfs_pgheader_init() 64 hdr->io_start = req_offset(hdr->req); in nfs_pgheader_init() 168 nfs_page_group_lock(struct nfs_page *req, bool nonblock) in nfs_page_group_lock() argument 170 struct nfs_page *head = req->wb_head; in nfs_page_group_lock() 191 nfs_page_group_lock_wait(struct nfs_page *req) in nfs_page_group_lock_wait() argument 193 struct nfs_page *head = req->wb_head; in nfs_page_group_lock_wait() 206 nfs_page_group_unlock(struct nfs_page *req) in nfs_page_group_unlock() argument 208 struct nfs_page *head = req->wb_head; in nfs_page_group_unlock() 224 nfs_page_group_sync_on_bit_locked(struct nfs_page *req, unsigned int bit) in nfs_page_group_sync_on_bit_locked() argument [all …]
|
D | write.c | 44 static void nfs_redirty_request(struct nfs_page *req); 49 static void nfs_clear_request_commit(struct nfs_page *req); 110 struct nfs_page *req = NULL; in nfs_page_find_head_request_locked() local 113 req = (struct nfs_page *)page_private(page); in nfs_page_find_head_request_locked() 115 req = nfs_page_search_commits_for_head_request_locked(nfsi, in nfs_page_find_head_request_locked() 118 if (req) { in nfs_page_find_head_request_locked() 119 WARN_ON_ONCE(req->wb_head != req); in nfs_page_find_head_request_locked() 120 kref_get(&req->wb_kref); in nfs_page_find_head_request_locked() 123 return req; in nfs_page_find_head_request_locked() 134 struct nfs_page *req = NULL; in nfs_page_find_head_request() local [all …]
|
D | direct.c | 290 struct nfs_page *req) in nfs_direct_setup_mirroring() argument 295 mirror_count = pgio->pg_ops->pg_get_mirror_count(pgio, req); in nfs_direct_setup_mirroring() 399 static void nfs_direct_readpage_release(struct nfs_page *req) in nfs_direct_readpage_release() argument 402 d_inode(req->wb_context->dentry)->i_sb->s_id, in nfs_direct_readpage_release() 403 (unsigned long long)NFS_FILEID(d_inode(req->wb_context->dentry)), in nfs_direct_readpage_release() 404 req->wb_bytes, in nfs_direct_readpage_release() 405 (long long)req_offset(req)); in nfs_direct_readpage_release() 406 nfs_release_request(req); in nfs_direct_readpage_release() 426 struct nfs_page *req = nfs_list_entry(hdr->pages.next); in nfs_direct_read_completion() local 427 struct page *page = req->wb_page; in nfs_direct_read_completion() [all …]
|
D | read.c | 121 static void nfs_readpage_release(struct nfs_page *req) in nfs_readpage_release() argument 123 struct inode *inode = d_inode(req->wb_context->dentry); in nfs_readpage_release() 126 (unsigned long long)NFS_FILEID(inode), req->wb_bytes, in nfs_readpage_release() 127 (long long)req_offset(req)); in nfs_readpage_release() 129 if (nfs_page_group_sync_on_bit(req, PG_UNLOCKPAGE)) { in nfs_readpage_release() 130 if (PageUptodate(req->wb_page)) in nfs_readpage_release() 131 nfs_readpage_to_fscache(inode, req->wb_page, 0); in nfs_readpage_release() 133 unlock_page(req->wb_page); in nfs_readpage_release() 135 nfs_release_request(req); in nfs_readpage_release() 138 static void nfs_page_group_set_uptodate(struct nfs_page *req) in nfs_page_group_set_uptodate() argument [all …]
|
/linux-4.1.27/include/net/ |
D | request_sock.h | 36 struct request_sock *req); 38 struct request_sock *req); 41 void (*destructor)(struct request_sock *req); 42 void (*syn_ack_timeout)(const struct request_sock *req); 45 int inet_rtx_syn_ack(struct sock *parent, struct request_sock *req); 74 struct request_sock *req = kmem_cache_alloc(ops->slab, GFP_ATOMIC); in reqsk_alloc() local 76 if (req) { in reqsk_alloc() 77 req->rsk_ops = ops; in reqsk_alloc() 79 req->rsk_listener = sk_listener; in reqsk_alloc() 84 atomic_set(&req->rsk_refcnt, 0); in reqsk_alloc() [all …]
|
/linux-4.1.27/drivers/media/usb/dvb-usb-v2/ |
D | ec168.c | 28 static int ec168_ctrl_msg(struct dvb_usb_device *d, struct ec168_req *req) in ec168_ctrl_msg() argument 35 switch (req->cmd) { in ec168_ctrl_msg() 41 request = req->cmd; in ec168_ctrl_msg() 45 request = req->cmd; in ec168_ctrl_msg() 65 KBUILD_MODNAME, req->cmd); in ec168_ctrl_msg() 70 buf = kmalloc(req->size, GFP_KERNEL); in ec168_ctrl_msg() 78 memcpy(buf, req->data, req->size); in ec168_ctrl_msg() 87 ret = usb_control_msg(d->udev, pipe, request, requesttype, req->value, in ec168_ctrl_msg() 88 req->index, buf, req->size, EC168_USB_TIMEOUT); in ec168_ctrl_msg() 90 dvb_usb_dbg_usb_control_msg(d->udev, request, requesttype, req->value, in ec168_ctrl_msg() [all …]
|
D | ce6230.c | 26 static int ce6230_ctrl_msg(struct dvb_usb_device *d, struct usb_req *req) in ce6230_ctrl_msg() argument 36 request = req->cmd; in ce6230_ctrl_msg() 37 value = req->value; in ce6230_ctrl_msg() 38 index = req->index; in ce6230_ctrl_msg() 40 switch (req->cmd) { in ce6230_ctrl_msg() 53 KBUILD_MODNAME, req->cmd); in ce6230_ctrl_msg() 58 buf = kmalloc(req->data_len, GFP_KERNEL); in ce6230_ctrl_msg() 66 memcpy(buf, req->data, req->data_len); in ce6230_ctrl_msg() 76 buf, req->data_len, CE6230_USB_TIMEOUT); in ce6230_ctrl_msg() 79 buf, req->data_len); in ce6230_ctrl_msg() [all …]
|
D | af9015.c | 31 static int af9015_ctrl_msg(struct dvb_usb_device *d, struct req_t *req) in af9015_ctrl_msg() argument 41 state->buf[0] = req->cmd; in af9015_ctrl_msg() 43 state->buf[2] = req->i2c_addr; in af9015_ctrl_msg() 44 state->buf[3] = req->addr >> 8; in af9015_ctrl_msg() 45 state->buf[4] = req->addr & 0xff; in af9015_ctrl_msg() 46 state->buf[5] = req->mbox; in af9015_ctrl_msg() 47 state->buf[6] = req->addr_len; in af9015_ctrl_msg() 48 state->buf[7] = req->data_len; in af9015_ctrl_msg() 50 switch (req->cmd) { in af9015_ctrl_msg() 63 if (((req->addr & 0xff00) == 0xff00) || in af9015_ctrl_msg() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lustre/lov/ |
D | lov_request.c | 64 struct lov_request *req = list_entry(pos, in lov_finish_set() local 67 list_del_init(&req->rq_link); in lov_finish_set() 69 if (req->rq_oi.oi_oa) in lov_finish_set() 70 OBDO_FREE(req->rq_oi.oi_oa); in lov_finish_set() 71 if (req->rq_oi.oi_md) in lov_finish_set() 72 OBD_FREE_LARGE(req->rq_oi.oi_md, req->rq_buflen); in lov_finish_set() 73 if (req->rq_oi.oi_osfs) in lov_finish_set() 74 OBD_FREE(req->rq_oi.oi_osfs, in lov_finish_set() 75 sizeof(*req->rq_oi.oi_osfs)); in lov_finish_set() 76 OBD_FREE(req, sizeof(*req)); in lov_finish_set() [all …]
|
/linux-4.1.27/drivers/base/power/ |
D | qos.c | 141 static int apply_constraint(struct dev_pm_qos_request *req, in apply_constraint() argument 144 struct dev_pm_qos *qos = req->dev->power.qos; in apply_constraint() 147 switch(req->type) { in apply_constraint() 150 &req->data.pnode, action, value); in apply_constraint() 155 req); in apply_constraint() 160 &req->data.pnode, action, value); in apply_constraint() 163 req->dev->power.set_latency_tolerance(req->dev, value); in apply_constraint() 167 ret = pm_qos_update_flags(&qos->flags, &req->data.flr, in apply_constraint() 237 struct dev_pm_qos_request *req, *tmp; in dev_pm_qos_constraints_destroy() local 261 plist_for_each_entry_safe(req, tmp, &c->list, data.pnode) { in dev_pm_qos_constraints_destroy() [all …]
|
/linux-4.1.27/fs/fuse/ |
D | dev.c | 37 static void fuse_request_init(struct fuse_req *req, struct page **pages, in fuse_request_init() argument 41 memset(req, 0, sizeof(*req)); in fuse_request_init() 44 INIT_LIST_HEAD(&req->list); in fuse_request_init() 45 INIT_LIST_HEAD(&req->intr_entry); in fuse_request_init() 46 init_waitqueue_head(&req->waitq); in fuse_request_init() 47 atomic_set(&req->count, 1); in fuse_request_init() 48 req->pages = pages; in fuse_request_init() 49 req->page_descs = page_descs; in fuse_request_init() 50 req->max_pages = npages; in fuse_request_init() 55 struct fuse_req *req = kmem_cache_alloc(fuse_req_cachep, flags); in __fuse_request_alloc() local [all …]
|
D | file.c | 84 static void fuse_release_end(struct fuse_conn *fc, struct fuse_req *req) in fuse_release_end() argument 86 iput(req->misc.release.inode); in fuse_release_end() 92 struct fuse_req *req = ff->reserved_req; in fuse_file_put() local 99 req->background = 0; in fuse_file_put() 100 iput(req->misc.release.inode); in fuse_file_put() 101 fuse_put_request(ff->fc, req); in fuse_file_put() 103 req->background = 0; in fuse_file_put() 104 fuse_request_send(ff->fc, req); in fuse_file_put() 105 iput(req->misc.release.inode); in fuse_file_put() 106 fuse_put_request(ff->fc, req); in fuse_file_put() [all …]
|
/linux-4.1.27/block/ |
D | bsg-lib.c | 55 struct request *req = job->req; in bsg_job_done() local 56 struct request *rsp = req->next_rq; in bsg_job_done() 59 err = job->req->errors = result; in bsg_job_done() 62 job->req->sense_len = sizeof(u32); in bsg_job_done() 64 job->req->sense_len = job->reply_len; in bsg_job_done() 66 req->resid_len = 0; in bsg_job_done() 74 blk_complete_request(req); in bsg_job_done() 90 static int bsg_map_buffer(struct bsg_buffer *buf, struct request *req) in bsg_map_buffer() argument 92 size_t sz = (sizeof(struct scatterlist) * req->nr_phys_segments); in bsg_map_buffer() 94 BUG_ON(!req->nr_phys_segments); in bsg_map_buffer() [all …]
|
D | blk-merge.c | 287 struct request *req, in ll_new_hw_segment() argument 292 if (req->nr_phys_segments + nr_phys_segs > queue_max_segments(q)) in ll_new_hw_segment() 295 if (blk_integrity_merge_bio(q, req, bio) == false) in ll_new_hw_segment() 302 req->nr_phys_segments += nr_phys_segs; in ll_new_hw_segment() 306 req->cmd_flags |= REQ_NOMERGE; in ll_new_hw_segment() 307 if (req == q->last_merge) in ll_new_hw_segment() 312 int ll_back_merge_fn(struct request_queue *q, struct request *req, in ll_back_merge_fn() argument 315 if (blk_rq_sectors(req) + bio_sectors(bio) > in ll_back_merge_fn() 316 blk_rq_get_max_sectors(req)) { in ll_back_merge_fn() 317 req->cmd_flags |= REQ_NOMERGE; in ll_back_merge_fn() [all …]
|
D | blk-timeout.c | 78 void blk_delete_timer(struct request *req) in blk_delete_timer() argument 80 list_del_init(&req->timeout_list); in blk_delete_timer() 83 static void blk_rq_timed_out(struct request *req) in blk_rq_timed_out() argument 85 struct request_queue *q = req->q; in blk_rq_timed_out() 89 ret = q->rq_timed_out_fn(req); in blk_rq_timed_out() 93 __blk_complete_request(req); in blk_rq_timed_out() 96 blk_add_timer(req); in blk_rq_timed_out() 97 blk_clear_rq_complete(req); in blk_rq_timed_out() 157 void blk_abort_request(struct request *req) in blk_abort_request() argument 159 if (blk_mark_rq_complete(req)) in blk_abort_request() [all …]
|
D | blk-core.c | 1380 void __blk_put_request(struct request_queue *q, struct request *req) in __blk_put_request() argument 1386 blk_mq_free_request(req); in __blk_put_request() 1390 blk_pm_put_request(req); in __blk_put_request() 1392 elv_completed_request(q, req); in __blk_put_request() 1395 WARN_ON(req->bio != NULL); in __blk_put_request() 1401 if (req->cmd_flags & REQ_ALLOCED) { in __blk_put_request() 1402 unsigned int flags = req->cmd_flags; in __blk_put_request() 1403 struct request_list *rl = blk_rq_rl(req); in __blk_put_request() 1405 BUG_ON(!list_empty(&req->queuelist)); in __blk_put_request() 1406 BUG_ON(ELV_ON_HASH(req)); in __blk_put_request() [all …]
|
/linux-4.1.27/drivers/staging/emxx_udc/ |
D | emxx_udc.c | 187 udc->ep0_req.req.buf = p_buf; in _nbu2ss_create_ep0_packet() 188 udc->ep0_req.req.length = length; in _nbu2ss_create_ep0_packet() 189 udc->ep0_req.req.dma = 0; in _nbu2ss_create_ep0_packet() 190 udc->ep0_req.req.zero = TRUE; in _nbu2ss_create_ep0_packet() 191 udc->ep0_req.req.complete = _nbu2ss_ep0_complete; in _nbu2ss_create_ep0_packet() 192 udc->ep0_req.req.status = -EINPROGRESS; in _nbu2ss_create_ep0_packet() 193 udc->ep0_req.req.context = udc; in _nbu2ss_create_ep0_packet() 194 udc->ep0_req.req.actual = 0; in _nbu2ss_create_ep0_packet() 483 struct nbu2ss_req *req, in _nbu2ss_dma_map_single() argument 487 if (req->req.dma == DMA_ADDR_INVALID) { in _nbu2ss_dma_map_single() [all …]
|
/linux-4.1.27/kernel/power/ |
D | qos.c | 192 struct pm_qos_request *req; in pm_qos_dbg_show_requests() local 229 plist_for_each_entry(req, &c->list, node) { in pm_qos_dbg_show_requests() 232 if ((req->node).prio != c->default_value) { in pm_qos_dbg_show_requests() 238 (req->node).prio, state); in pm_qos_dbg_show_requests() 331 struct pm_qos_flags_request *req) in pm_qos_flags_remove_req() argument 335 list_del(&req->node); in pm_qos_flags_remove_req() 336 list_for_each_entry(req, &pqf->list, node) in pm_qos_flags_remove_req() 337 val |= req->flags; in pm_qos_flags_remove_req() 354 struct pm_qos_flags_request *req, in pm_qos_update_flags() argument 366 pm_qos_flags_remove_req(pqf, req); in pm_qos_update_flags() [all …]
|
/linux-4.1.27/drivers/i2c/busses/ |
D | i2c-opal.c | 51 static int i2c_opal_send_request(u32 bus_id, struct opal_i2c_request *req) in i2c_opal_send_request() argument 64 rc = opal_i2c_request(token, bus_id, req); in i2c_opal_send_request() 89 struct opal_i2c_request req; in i2c_opal_master_xfer() local 95 memset(&req, 0, sizeof(req)); in i2c_opal_master_xfer() 100 req.type = (msgs[0].flags & I2C_M_RD) ? in i2c_opal_master_xfer() 102 req.addr = cpu_to_be16(msgs[0].addr); in i2c_opal_master_xfer() 103 req.size = cpu_to_be32(msgs[0].len); in i2c_opal_master_xfer() 104 req.buffer_ra = cpu_to_be64(__pa(msgs[0].buf)); in i2c_opal_master_xfer() 107 req.type = (msgs[1].flags & I2C_M_RD) ? in i2c_opal_master_xfer() 109 req.addr = cpu_to_be16(msgs[0].addr); in i2c_opal_master_xfer() [all …]
|
/linux-4.1.27/fs/ubifs/ |
D | budget.c | 375 const struct ubifs_budget_req *req) in calc_idx_growth() argument 379 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth() 380 req->new_dent; in calc_idx_growth() 391 const struct ubifs_budget_req *req) in calc_data_growth() argument 395 data_growth = req->new_ino ? c->bi.inode_budget : 0; in calc_data_growth() 396 if (req->new_page) in calc_data_growth() 398 if (req->new_dent) in calc_data_growth() 400 data_growth += req->new_ino_d; in calc_data_growth() 411 const struct ubifs_budget_req *req) in calc_dd_growth() argument 415 dd_growth = req->dirtied_page ? c->bi.page_budget : 0; in calc_dd_growth() [all …]
|
/linux-4.1.27/arch/m68k/mac/ |
D | misc.c | 39 struct adb_request req; in cuda_read_time() local 42 if (cuda_request(&req, NULL, 2, CUDA_PACKET, CUDA_GET_TIME) < 0) in cuda_read_time() 44 while (!req.complete) in cuda_read_time() 47 time = (req.reply[3] << 24) | (req.reply[4] << 16) in cuda_read_time() 48 | (req.reply[5] << 8) | req.reply[6]; in cuda_read_time() 54 struct adb_request req; in cuda_write_time() local 56 if (cuda_request(&req, NULL, 6, CUDA_PACKET, CUDA_SET_TIME, in cuda_write_time() 60 while (!req.complete) in cuda_write_time() 66 struct adb_request req; in cuda_read_pram() local 67 if (cuda_request(&req, NULL, 4, CUDA_PACKET, CUDA_GET_PRAM, in cuda_read_pram() [all …]
|
/linux-4.1.27/drivers/usb/isp1760/ |
D | isp1760-udc.c | 30 struct usb_request req; member 46 static inline struct isp1760_request *req_to_udc_req(struct usb_request *req) in req_to_udc_req() argument 48 return container_of(req, struct isp1760_request, req); in req_to_udc_req() 127 struct isp1760_request *req, in isp1760_udc_request_complete() argument 134 req, status); in isp1760_udc_request_complete() 136 req->ep = NULL; in isp1760_udc_request_complete() 137 req->req.status = status; in isp1760_udc_request_complete() 138 req->req.complete(&ep->ep, &req->req); in isp1760_udc_request_complete() 180 struct isp1760_request *req) in isp1760_udc_receive() argument 191 __func__, len, req->req.actual, req->req.length); in isp1760_udc_receive() [all …]
|
/linux-4.1.27/drivers/scsi/device_handler/ |
D | scsi_dh_hp_sw.c | 114 struct request *req; in hp_sw_tur() local 118 req = blk_get_request(sdev->request_queue, WRITE, GFP_NOIO); in hp_sw_tur() 119 if (IS_ERR(req)) in hp_sw_tur() 122 blk_rq_set_block_pc(req); in hp_sw_tur() 123 req->cmd_flags |= REQ_FAILFAST_DEV | REQ_FAILFAST_TRANSPORT | in hp_sw_tur() 125 req->cmd_len = COMMAND_SIZE(TEST_UNIT_READY); in hp_sw_tur() 126 req->cmd[0] = TEST_UNIT_READY; in hp_sw_tur() 127 req->timeout = HP_SW_TIMEOUT; in hp_sw_tur() 128 req->sense = h->sense; in hp_sw_tur() 129 memset(req->sense, 0, SCSI_SENSE_BUFFERSIZE); in hp_sw_tur() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lustre/include/ |
D | lustre_net.h | 270 #define ptlrpc_req_async_args(req) ((void *)&req->rq_async_args) argument 486 struct ptlrpc_request *req, 1525 struct ptlrpc_request *req, int rc) in ptlrpc_req_interpret() argument 1527 if (req->rq_interpret_reply != NULL) { in ptlrpc_req_interpret() 1528 req->rq_status = req->rq_interpret_reply(env, req, in ptlrpc_req_interpret() 1529 &req->rq_async_args, in ptlrpc_req_interpret() 1531 return req->rq_status; in ptlrpc_req_interpret() 1541 void ptlrpc_nrs_req_hp_move(struct ptlrpc_request *req); 1551 static inline bool ptlrpc_nrs_req_can_move(struct ptlrpc_request *req) in ptlrpc_nrs_req_can_move() argument 1553 struct ptlrpc_nrs_request *nrq = &req->rq_nrq; in ptlrpc_nrs_req_can_move() [all …]
|
D | lustre_sec.h | 403 struct ptlrpc_request *req); 416 struct ptlrpc_request *req); 429 struct ptlrpc_request *req); 442 struct ptlrpc_request *req); 465 struct ptlrpc_request *req, 482 struct ptlrpc_request *req, 629 struct ptlrpc_request *req, 640 struct ptlrpc_request *req); 653 struct ptlrpc_request *req, 666 struct ptlrpc_request *req); [all …]
|
/linux-4.1.27/fs/nilfs2/ |
D | dat.c | 54 struct nilfs_palloc_req *req, int create) in nilfs_dat_prepare_entry() argument 56 return nilfs_palloc_get_entry_block(dat, req->pr_entry_nr, in nilfs_dat_prepare_entry() 57 create, &req->pr_entry_bh); in nilfs_dat_prepare_entry() 61 struct nilfs_palloc_req *req) in nilfs_dat_commit_entry() argument 63 mark_buffer_dirty(req->pr_entry_bh); in nilfs_dat_commit_entry() 65 brelse(req->pr_entry_bh); in nilfs_dat_commit_entry() 69 struct nilfs_palloc_req *req) in nilfs_dat_abort_entry() argument 71 brelse(req->pr_entry_bh); in nilfs_dat_abort_entry() 74 int nilfs_dat_prepare_alloc(struct inode *dat, struct nilfs_palloc_req *req) in nilfs_dat_prepare_alloc() argument 78 ret = nilfs_palloc_prepare_alloc_entry(dat, req); in nilfs_dat_prepare_alloc() [all …]
|
D | ifile.c | 68 struct nilfs_palloc_req req; in nilfs_ifile_create_inode() local 71 req.pr_entry_nr = 0; /* 0 says find free inode from beginning of in nilfs_ifile_create_inode() 73 req.pr_entry_bh = NULL; in nilfs_ifile_create_inode() 75 ret = nilfs_palloc_prepare_alloc_entry(ifile, &req); in nilfs_ifile_create_inode() 77 ret = nilfs_palloc_get_entry_block(ifile, req.pr_entry_nr, 1, in nilfs_ifile_create_inode() 78 &req.pr_entry_bh); in nilfs_ifile_create_inode() 80 nilfs_palloc_abort_alloc_entry(ifile, &req); in nilfs_ifile_create_inode() 83 brelse(req.pr_entry_bh); in nilfs_ifile_create_inode() 86 nilfs_palloc_commit_alloc_entry(ifile, &req); in nilfs_ifile_create_inode() 87 mark_buffer_dirty(req.pr_entry_bh); in nilfs_ifile_create_inode() [all …]
|
D | alloc.c | 470 struct nilfs_palloc_req *req) in nilfs_palloc_prepare_alloc_entry() argument 484 group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset); in nilfs_palloc_prepare_alloc_entry() 492 maxgroup = nilfs_palloc_group(inode, req->pr_entry_nr, in nilfs_palloc_prepare_alloc_entry() 519 req->pr_entry_nr = in nilfs_palloc_prepare_alloc_entry() 524 req->pr_desc_bh = desc_bh; in nilfs_palloc_prepare_alloc_entry() 525 req->pr_bitmap_bh = bitmap_bh; in nilfs_palloc_prepare_alloc_entry() 554 struct nilfs_palloc_req *req) in nilfs_palloc_commit_alloc_entry() argument 556 mark_buffer_dirty(req->pr_bitmap_bh); in nilfs_palloc_commit_alloc_entry() 557 mark_buffer_dirty(req->pr_desc_bh); in nilfs_palloc_commit_alloc_entry() 560 brelse(req->pr_bitmap_bh); in nilfs_palloc_commit_alloc_entry() [all …]
|
/linux-4.1.27/drivers/net/ethernet/emulex/benet/ |
D | be_cmds.c | 900 struct be_cmd_req_eq_create *req; in be_cmd_eq_create() local 908 req = embedded_payload(wrb); in be_cmd_eq_create() 910 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, in be_cmd_eq_create() 911 OPCODE_COMMON_EQ_CREATE, sizeof(*req), wrb, in be_cmd_eq_create() 918 req->hdr.version = ver; in be_cmd_eq_create() 919 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in be_cmd_eq_create() 921 AMAP_SET_BITS(struct amap_eq_context, valid, req->context, 1); in be_cmd_eq_create() 923 AMAP_SET_BITS(struct amap_eq_context, size, req->context, 0); in be_cmd_eq_create() 924 AMAP_SET_BITS(struct amap_eq_context, count, req->context, in be_cmd_eq_create() 926 be_dws_cpu_to_le(req->context, sizeof(req->context)); in be_cmd_eq_create() [all …]
|
/linux-4.1.27/drivers/scsi/csiostor/ |
D | csio_scsi.c | 163 csio_scsi_fcp_cmnd(struct csio_ioreq *req, void *addr) in csio_scsi_fcp_cmnd() argument 166 struct scsi_cmnd *scmnd = csio_scsi_cmnd(req); in csio_scsi_fcp_cmnd() 178 if (req->nsge) in csio_scsi_fcp_cmnd() 179 if (req->datadir == DMA_TO_DEVICE) in csio_scsi_fcp_cmnd() 201 csio_scsi_init_cmd_wr(struct csio_ioreq *req, void *addr, uint32_t size) in csio_scsi_init_cmd_wr() argument 203 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_init_cmd_wr() 204 struct csio_rnode *rn = req->rnode; in csio_scsi_init_cmd_wr() 215 wr->cookie = (uintptr_t) req; in csio_scsi_init_cmd_wr() 216 wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx)); in csio_scsi_init_cmd_wr() 217 wr->tmo_val = (uint8_t) req->tmo; in csio_scsi_init_cmd_wr() [all …]
|
/linux-4.1.27/drivers/base/ |
D | devtmpfs.c | 39 static struct req { struct 40 struct req *next; argument 85 struct req req; in devtmpfs_create_node() local 90 req.mode = 0; in devtmpfs_create_node() 91 req.uid = GLOBAL_ROOT_UID; in devtmpfs_create_node() 92 req.gid = GLOBAL_ROOT_GID; in devtmpfs_create_node() 93 req.name = device_get_devnode(dev, &req.mode, &req.uid, &req.gid, &tmp); in devtmpfs_create_node() 94 if (!req.name) in devtmpfs_create_node() 97 if (req.mode == 0) in devtmpfs_create_node() 98 req.mode = 0600; in devtmpfs_create_node() [all …]
|
/linux-4.1.27/drivers/usb/gadget/udc/ |
D | gr_udc.c | 96 struct gr_request *req) in gr_dbgprint_request() argument 98 int buflen = ep->is_in ? req->req.length : req->req.actual; in gr_dbgprint_request() 102 dev_dbg(ep->dev->dev, "%s: 0x%p, %d bytes data%s:\n", str, req, buflen, in gr_dbgprint_request() 105 rowlen, 4, req->req.buf, plen, false); in gr_dbgprint_request() 117 struct gr_request *req) {} in gr_dbgprint_request() argument 134 struct gr_request *req; in gr_seq_ep_show() local 167 list_for_each_entry(req, &ep->queue, queue) { in gr_seq_ep_show() 171 seq_printf(seq, " 0x%p: 0x%p %d %d\n", req, in gr_seq_ep_show() 172 &req->req.buf, req->req.actual, req->req.length); in gr_seq_ep_show() 174 next = req->first_desc; in gr_seq_ep_show() [all …]
|
D | pxa25x_udc.c | 303 struct pxa25x_request *req; in pxa25x_ep_alloc_request() local 305 req = kzalloc(sizeof(*req), gfp_flags); in pxa25x_ep_alloc_request() 306 if (!req) in pxa25x_ep_alloc_request() 309 INIT_LIST_HEAD (&req->queue); in pxa25x_ep_alloc_request() 310 return &req->req; in pxa25x_ep_alloc_request() 320 struct pxa25x_request *req; in pxa25x_ep_free_request() local 322 req = container_of (_req, struct pxa25x_request, req); in pxa25x_ep_free_request() 323 WARN_ON(!list_empty (&req->queue)); in pxa25x_ep_free_request() 324 kfree(req); in pxa25x_ep_free_request() 332 static void done(struct pxa25x_ep *ep, struct pxa25x_request *req, int status) in done() argument [all …]
|
D | goku_udc.c | 268 struct goku_request *req; in goku_alloc_request() local 272 req = kzalloc(sizeof *req, gfp_flags); in goku_alloc_request() 273 if (!req) in goku_alloc_request() 276 INIT_LIST_HEAD(&req->queue); in goku_alloc_request() 277 return &req->req; in goku_alloc_request() 283 struct goku_request *req; in goku_free_request() local 288 req = container_of(_req, struct goku_request, req); in goku_free_request() 289 WARN_ON(!list_empty(&req->queue)); in goku_free_request() 290 kfree(req); in goku_free_request() 296 done(struct goku_ep *ep, struct goku_request *req, int status) in done() argument [all …]
|
D | pxa27x_udc.c | 142 struct pxa27x_request *req; in queues_dbg_show() local 160 list_for_each_entry(req, &ep->queue, queue) { in queues_dbg_show() 162 &req->req, req->req.actual, in queues_dbg_show() 163 req->req.length, req->req.buf); in queues_dbg_show() 647 struct pxa27x_request *req; in pxa_ep_alloc_request() local 649 req = kzalloc(sizeof *req, gfp_flags); in pxa_ep_alloc_request() 650 if (!req) in pxa_ep_alloc_request() 653 INIT_LIST_HEAD(&req->queue); in pxa_ep_alloc_request() 654 req->in_use = 0; in pxa_ep_alloc_request() 655 req->udc_usb_ep = container_of(_ep, struct udc_usb_ep, usb_ep); in pxa_ep_alloc_request() [all …]
|
D | amd5536udc.c | 74 static int udc_free_dma_chain(struct udc *dev, struct udc_request *req); 75 static int udc_create_dma_chain(struct udc_ep *ep, struct udc_request *req, 496 udc_free_request(&ep->ep, &ep->bna_dummy_req->req); in udc_ep_disable() 508 struct udc_request *req; in udc_alloc_request() local 518 req = kzalloc(sizeof(struct udc_request), gfp); in udc_alloc_request() 519 if (!req) in udc_alloc_request() 522 req->req.dma = DMA_DONT_USE; in udc_alloc_request() 523 INIT_LIST_HEAD(&req->queue); in udc_alloc_request() 528 &req->td_phys); in udc_alloc_request() 530 kfree(req); in udc_alloc_request() [all …]
|
D | atmel_usba_udc.c | 40 struct usba_request *req, *req_copy; in queue_dbg_open() local 49 list_for_each_entry(req, &ep->queue, queue) { in queue_dbg_open() 50 req_copy = kmemdup(req, sizeof(*req_copy), GFP_ATOMIC); in queue_dbg_open() 62 list_for_each_entry_safe(req, req_copy, queue_data, queue) { in queue_dbg_open() 63 list_del(&req->queue); in queue_dbg_open() 64 kfree(req); in queue_dbg_open() 88 struct usba_request *req, *tmp_req; in queue_dbg_read() local 96 list_for_each_entry_safe(req, tmp_req, queue, queue) { in queue_dbg_read() 99 req->req.buf, req->req.length, in queue_dbg_read() 100 req->req.no_interrupt ? 'i' : 'I', in queue_dbg_read() [all …]
|
D | fotg210-udc.c | 61 static void fotg210_done(struct fotg210_ep *ep, struct fotg210_request *req, in fotg210_done() argument 64 list_del_init(&req->queue); in fotg210_done() 68 req->req.status = -ESHUTDOWN; in fotg210_done() 70 req->req.status = status; in fotg210_done() 73 usb_gadget_giveback_request(&ep->ep, &req->req); in fotg210_done() 215 struct fotg210_request *req; in fotg210_ep_disable() local 223 req = list_entry(ep->queue.next, in fotg210_ep_disable() 226 fotg210_done(ep, req, -ECONNRESET); in fotg210_ep_disable() 236 struct fotg210_request *req; in fotg210_ep_alloc_request() local 238 req = kzalloc(sizeof(struct fotg210_request), gfp_flags); in fotg210_ep_alloc_request() [all …]
|
D | omap_udc.c | 270 struct omap_req *req; in omap_alloc_request() local 272 req = kzalloc(sizeof(*req), gfp_flags); in omap_alloc_request() 273 if (!req) in omap_alloc_request() 276 INIT_LIST_HEAD(&req->queue); in omap_alloc_request() 278 return &req->req; in omap_alloc_request() 284 struct omap_req *req = container_of(_req, struct omap_req, req); in omap_free_request() local 286 kfree(req); in omap_free_request() 292 done(struct omap_ep *ep, struct omap_req *req, int status) in done() argument 297 list_del_init(&req->queue); in done() 299 if (req->req.status == -EINPROGRESS) in done() [all …]
|
D | r8a66597-udc.c | 38 static void irq_ep0_write(struct r8a66597_ep *ep, struct r8a66597_request *req); 40 struct r8a66597_request *req); 45 struct r8a66597_request *req, int status); 572 struct r8a66597_request *req) in start_ep0_write() argument 579 if (req->req.length == 0) { in start_ep0_write() 582 transfer_complete(ep, req, 0); in start_ep0_write() 585 irq_ep0_write(ep, req); in start_ep0_write() 634 struct r8a66597_request *req) in sudmac_alloc_channel() argument 666 return usb_gadget_map_request(&r8a66597->gadget, &req->req, dma->dir); in sudmac_alloc_channel() 671 struct r8a66597_request *req) in sudmac_free_channel() argument [all …]
|
D | mv_u3d_core.c | 134 actual = curr_req->req.length; in mv_u3d_process_ep_req() 177 curr_req->req.actual = actual; in mv_u3d_process_ep_req() 187 void mv_u3d_done(struct mv_u3d_ep *ep, struct mv_u3d_req *req, int status) in mv_u3d_done() argument 195 list_del_init(&req->queue); in mv_u3d_done() 198 if (req->req.status == -EINPROGRESS) in mv_u3d_done() 199 req->req.status = status; in mv_u3d_done() 201 status = req->req.status; in mv_u3d_done() 204 if (!req->chain) in mv_u3d_done() 206 req->trb_head->trb_hw, req->trb_head->trb_dma); in mv_u3d_done() 209 (dma_addr_t)req->trb_head->trb_dma, in mv_u3d_done() [all …]
|
D | s3c2410_udc.c | 255 struct s3c2410_request *req, int status) in s3c2410_udc_done() argument 259 list_del_init(&req->queue); in s3c2410_udc_done() 261 if (likely(req->req.status == -EINPROGRESS)) in s3c2410_udc_done() 262 req->req.status = status; in s3c2410_udc_done() 264 status = req->req.status; in s3c2410_udc_done() 267 usb_gadget_giveback_request(&ep->ep, &req->req); in s3c2410_udc_done() 279 struct s3c2410_request *req; in s3c2410_udc_nuke() local 280 req = list_entry(ep->queue.next, struct s3c2410_request, in s3c2410_udc_nuke() 282 s3c2410_udc_done(ep, req, status); in s3c2410_udc_nuke() 299 struct s3c2410_request *req, in s3c2410_udc_write_packet() argument [all …]
|
D | fusb300_udc.c | 36 static void done(struct fusb300_ep *ep, struct fusb300_request *req, 254 struct fusb300_request *req; in fusb300_disable() local 262 req = list_entry(ep->queue.next, struct fusb300_request, queue); in fusb300_disable() 264 done(ep, req, -ECONNRESET); in fusb300_disable() 274 struct fusb300_request *req; in fusb300_alloc_request() local 276 req = kzalloc(sizeof(struct fusb300_request), gfp_flags); in fusb300_alloc_request() 277 if (!req) in fusb300_alloc_request() 279 INIT_LIST_HEAD(&req->queue); in fusb300_alloc_request() 281 return &req->req; in fusb300_alloc_request() 286 struct fusb300_request *req; in fusb300_free_request() local [all …]
|
D | udc-xilinx.c | 112 #define to_xusb_req(req) container_of((req), struct xusb_req, usb_req) argument 187 struct xusb_req *req; member 380 static int xudc_dma_send(struct xusb_ep *ep, struct xusb_req *req, in xudc_dma_send() argument 388 src = req->usb_req.dma + req->usb_req.actual; in xudc_dma_send() 389 if (req->usb_req.length) in xudc_dma_send() 434 static int xudc_dma_receive(struct xusb_ep *ep, struct xusb_req *req, in xudc_dma_receive() argument 442 dst = req->usb_req.dma + req->usb_req.actual; in xudc_dma_receive() 483 static int xudc_eptxrx(struct xusb_ep *ep, struct xusb_req *req, in xudc_eptxrx() argument 494 rc = xudc_dma_send(ep, req, bufferptr, bufferlen); in xudc_eptxrx() 496 rc = xudc_dma_receive(ep, req, bufferptr, bufferlen); in xudc_eptxrx() [all …]
|
D | m66592-udc.c | 38 static void irq_ep0_write(struct m66592_ep *ep, struct m66592_request *req); 39 static void irq_packet_write(struct m66592_ep *ep, struct m66592_request *req); 44 struct m66592_request *req, int status); 497 static void start_ep0_write(struct m66592_ep *ep, struct m66592_request *req) in start_ep0_write() argument 506 if (req->req.length == 0) { in start_ep0_write() 509 transfer_complete(ep, req, 0); in start_ep0_write() 512 irq_ep0_write(ep, req); in start_ep0_write() 516 static void start_packet_write(struct m66592_ep *ep, struct m66592_request *req) in start_packet_write() argument 529 irq_packet_write(ep, req); in start_packet_write() 532 static void start_packet_read(struct m66592_ep *ep, struct m66592_request *req) in start_packet_read() argument [all …]
|
D | fsl_udc_core.c | 161 static void done(struct fsl_ep *ep, struct fsl_req *req, int status) in done() argument 172 list_del_init(&req->queue); in done() 175 if (req->req.status == -EINPROGRESS) in done() 176 req->req.status = status; in done() 178 status = req->req.status; in done() 181 next_td = req->head; in done() 182 for (j = 0; j < req->dtd_count; j++) { in done() 184 if (j != req->dtd_count - 1) { in done() 190 usb_gadget_unmap_request(&ep->udc->gadget, &req->req, ep_is_in(ep)); in done() 194 ep->ep.name, &req->req, status, in done() [all …]
|
D | pch_udc.c | 410 struct usb_request req; member 1446 static void complete_req(struct pch_udc_ep *ep, struct pch_udc_request *req, in complete_req() argument 1454 list_del_init(&req->queue); in complete_req() 1457 if (req->req.status == -EINPROGRESS) in complete_req() 1458 req->req.status = status; in complete_req() 1460 status = req->req.status; in complete_req() 1463 if (req->dma_mapped) { in complete_req() 1464 if (req->dma == DMA_ADDR_INVALID) { in complete_req() 1466 dma_unmap_single(&dev->pdev->dev, req->req.dma, in complete_req() 1467 req->req.length, in complete_req() [all …]
|
D | mv_udc_core.c | 143 actual = curr_req->req.length; in process_ep_req() 203 curr_req->req.actual = actual; in process_ep_req() 213 static void done(struct mv_ep *ep, struct mv_req *req, int status) in done() argument 224 list_del_init(&req->queue); in done() 227 if (req->req.status == -EINPROGRESS) in done() 228 req->req.status = status; in done() 230 status = req->req.status; in done() 233 next_td = req->head; in done() 234 for (j = 0; j < req->dtd_count; j++) { in done() 236 if (j != req->dtd_count - 1) in done() [all …]
|
D | net2280.c | 485 struct net2280_request *req; in net2280_alloc_request() local 493 req = kzalloc(sizeof(*req), gfp_flags); in net2280_alloc_request() 494 if (!req) in net2280_alloc_request() 497 INIT_LIST_HEAD(&req->queue); in net2280_alloc_request() 504 &req->td_dma); in net2280_alloc_request() 506 kfree(req); in net2280_alloc_request() 511 req->td = td; in net2280_alloc_request() 513 return &req->req; in net2280_alloc_request() 519 struct net2280_request *req; in net2280_free_request() local 528 req = container_of(_req, struct net2280_request, req); in net2280_free_request() [all …]
|
D | fsl_qe_udc.c | 80 static void done(struct qe_ep *ep, struct qe_req *req, int status) in done() argument 89 list_del_init(&req->queue); in done() 92 if (req->req.status == -EINPROGRESS) in done() 93 req->req.status = status; in done() 95 status = req->req.status; in done() 97 if (req->mapped) { in done() 99 req->req.dma, req->req.length, in done() 103 req->req.dma = DMA_ADDR_INVALID; in done() 104 req->mapped = 0; in done() 107 req->req.dma, req->req.length, in done() [all …]
|
D | net2272.c | 333 struct net2272_request *req; in net2272_alloc_request() local 339 req = kzalloc(sizeof(*req), gfp_flags); in net2272_alloc_request() 340 if (!req) in net2272_alloc_request() 343 INIT_LIST_HEAD(&req->queue); in net2272_alloc_request() 345 return &req->req; in net2272_alloc_request() 352 struct net2272_request *req; in net2272_free_request() local 358 req = container_of(_req, struct net2272_request, req); in net2272_free_request() 359 WARN_ON(!list_empty(&req->queue)); in net2272_free_request() 360 kfree(req); in net2272_free_request() 364 net2272_done(struct net2272_ep *ep, struct net2272_request *req, int status) in net2272_done() argument [all …]
|
/linux-4.1.27/fs/ceph/ |
D | dir.c | 320 struct ceph_mds_request *req; in ceph_readdir() local 332 req = ceph_mdsc_create_request(mdsc, op, USE_AUTH_MDS); in ceph_readdir() 333 if (IS_ERR(req)) in ceph_readdir() 334 return PTR_ERR(req); in ceph_readdir() 335 err = ceph_alloc_readdir_reply_buffer(req, inode); in ceph_readdir() 337 ceph_mdsc_put_request(req); in ceph_readdir() 341 req->r_direct_mode = USE_AUTH_MDS; in ceph_readdir() 342 req->r_direct_hash = ceph_frag_value(frag); in ceph_readdir() 343 req->r_direct_is_hash = true; in ceph_readdir() 345 req->r_path2 = kstrdup(fi->last_name, GFP_NOFS); in ceph_readdir() [all …]
|
D | mds_client.c | 515 static void put_request_session(struct ceph_mds_request *req) in put_request_session() argument 517 if (req->r_session) { in put_request_session() 518 ceph_put_mds_session(req->r_session); in put_request_session() 519 req->r_session = NULL; in put_request_session() 525 struct ceph_mds_request *req = container_of(kref, in ceph_mdsc_release_request() local 528 destroy_reply_info(&req->r_reply_info); in ceph_mdsc_release_request() 529 if (req->r_request) in ceph_mdsc_release_request() 530 ceph_msg_put(req->r_request); in ceph_mdsc_release_request() 531 if (req->r_reply) in ceph_mdsc_release_request() 532 ceph_msg_put(req->r_reply); in ceph_mdsc_release_request() [all …]
|
D | export.c | 73 struct ceph_mds_request *req; in __fh_to_dentry() local 75 req = ceph_mdsc_create_request(mdsc, CEPH_MDS_OP_LOOKUPINO, in __fh_to_dentry() 77 if (IS_ERR(req)) in __fh_to_dentry() 78 return ERR_CAST(req); in __fh_to_dentry() 80 req->r_ino1 = vino; in __fh_to_dentry() 81 req->r_num_caps = 1; in __fh_to_dentry() 82 err = ceph_mdsc_do_request(mdsc, NULL, req); in __fh_to_dentry() 83 inode = req->r_target_inode; in __fh_to_dentry() 86 ceph_mdsc_put_request(req); in __fh_to_dentry() 128 struct ceph_mds_request *req; in __get_parent() local [all …]
|
D | debugfs.c | 53 struct ceph_mds_request *req; in mdsc_show() local 61 req = rb_entry(rp, struct ceph_mds_request, r_node); in mdsc_show() 63 if (req->r_request && req->r_session) in mdsc_show() 64 seq_printf(s, "%lld\tmds%d\t", req->r_tid, in mdsc_show() 65 req->r_session->s_mds); in mdsc_show() 66 else if (!req->r_request) in mdsc_show() 67 seq_printf(s, "%lld\t(no request)\t", req->r_tid); in mdsc_show() 69 seq_printf(s, "%lld\t(no session)\t", req->r_tid); in mdsc_show() 71 seq_printf(s, "%s", ceph_mds_op_name(req->r_op)); in mdsc_show() 73 if (req->r_got_unsafe) in mdsc_show() [all …]
|
D | locks.c | 13 struct ceph_mds_request *req); 40 struct ceph_mds_request *req; in ceph_lock_message() local 48 req = ceph_mdsc_create_request(mdsc, operation, USE_AUTH_MDS); in ceph_lock_message() 49 if (IS_ERR(req)) in ceph_lock_message() 50 return PTR_ERR(req); in ceph_lock_message() 51 req->r_inode = inode; in ceph_lock_message() 53 req->r_num_caps = 1; in ceph_lock_message() 68 req->r_args.filelock_change.rule = lock_type; in ceph_lock_message() 69 req->r_args.filelock_change.type = cmd; in ceph_lock_message() 70 req->r_args.filelock_change.owner = cpu_to_le64(owner); in ceph_lock_message() [all …]
|
D | ioctl.c | 67 struct ceph_mds_request *req; in ceph_ioctl_set_layout() local 106 req = ceph_mdsc_create_request(mdsc, CEPH_MDS_OP_SETLAYOUT, in ceph_ioctl_set_layout() 108 if (IS_ERR(req)) in ceph_ioctl_set_layout() 109 return PTR_ERR(req); in ceph_ioctl_set_layout() 110 req->r_inode = inode; in ceph_ioctl_set_layout() 112 req->r_num_caps = 1; in ceph_ioctl_set_layout() 114 req->r_inode_drop = CEPH_CAP_FILE_SHARED | CEPH_CAP_FILE_EXCL; in ceph_ioctl_set_layout() 116 req->r_args.setlayout.layout.fl_stripe_unit = in ceph_ioctl_set_layout() 118 req->r_args.setlayout.layout.fl_stripe_count = in ceph_ioctl_set_layout() 120 req->r_args.setlayout.layout.fl_object_size = in ceph_ioctl_set_layout() [all …]
|
/linux-4.1.27/drivers/crypto/qce/ |
D | sha.c | 40 struct ahash_request *req = ahash_request_cast(async_req); in qce_ahash_done() local 41 struct crypto_ahash *ahash = crypto_ahash_reqtfm(req); in qce_ahash_done() 42 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); in qce_ahash_done() 54 qce_unmapsg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE, in qce_ahash_done() 59 if (req->result) in qce_ahash_done() 60 memcpy(req->result, result->auth_iv, digestsize); in qce_ahash_done() 69 req->src = rctx->src_orig; in qce_ahash_done() 70 req->nbytes = rctx->nbytes_orig; in qce_ahash_done() 79 struct ahash_request *req = ahash_request_cast(async_req); in qce_ahash_async_req_handle() local 80 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); in qce_ahash_async_req_handle() [all …]
|
D | ablkcipher.c | 28 struct ablkcipher_request *req = ablkcipher_request_cast(async_req); in qce_ablkcipher_done() local 29 struct qce_cipher_reqctx *rctx = ablkcipher_request_ctx(req); in qce_ablkcipher_done() 37 diff_dst = (req->src != req->dst) ? true : false; in qce_ablkcipher_done() 64 struct ablkcipher_request *req = ablkcipher_request_cast(async_req); in qce_ablkcipher_async_req_handle() local 65 struct qce_cipher_reqctx *rctx = ablkcipher_request_ctx(req); in qce_ablkcipher_async_req_handle() 66 struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req); in qce_ablkcipher_async_req_handle() 75 rctx->iv = req->info; in qce_ablkcipher_async_req_handle() 77 rctx->cryptlen = req->nbytes; in qce_ablkcipher_async_req_handle() 79 diff_dst = (req->src != req->dst) ? true : false; in qce_ablkcipher_async_req_handle() 83 rctx->src_nents = qce_countsg(req->src, req->nbytes, in qce_ablkcipher_async_req_handle() [all …]
|
/linux-4.1.27/drivers/target/sbp/ |
D | sbp_target.c | 70 static int read_peer_guid(u64 *guid, const struct sbp_management_request *req) in read_peer_guid() argument 75 ret = sbp_run_transaction(req->card, TCODE_READ_QUADLET_REQUEST, in read_peer_guid() 76 req->node_addr, req->generation, req->speed, in read_peer_guid() 82 ret = sbp_run_transaction(req->card, TCODE_READ_QUADLET_REQUEST, in read_peer_guid() 83 req->node_addr, req->generation, req->speed, in read_peer_guid() 293 struct sbp_management_agent *agent, struct sbp_management_request *req, in sbp_management_request_login() argument 307 LOGIN_ORB_LUN(be32_to_cpu(req->orb.misc))); in sbp_management_request_login() 310 LOGIN_ORB_LUN(be32_to_cpu(req->orb.misc))); in sbp_management_request_login() 312 req->status.status = cpu_to_be32( in sbp_management_request_login() 318 ret = read_peer_guid(&guid, req); in sbp_management_request_login() [all …]
|
/linux-4.1.27/drivers/crypto/ |
D | picoxcell_crypto.c | 94 struct crypto_async_request *req; member 100 void (*complete)(struct spacc_req *req); 180 static int spacc_ablk_submit(struct spacc_req *req); 321 static int spacc_aead_make_ddts(struct spacc_req *req, u8 *giv) in spacc_aead_make_ddts() argument 323 struct aead_request *areq = container_of(req->req, struct aead_request, in spacc_aead_make_ddts() 325 struct spacc_engine *engine = req->engine; in spacc_aead_make_ddts() 334 src_ddt = dma_pool_alloc(engine->req_pool, GFP_ATOMIC, &req->src_addr); in spacc_aead_make_ddts() 338 dst_ddt = dma_pool_alloc(engine->req_pool, GFP_ATOMIC, &req->dst_addr); in spacc_aead_make_ddts() 340 dma_pool_free(engine->req_pool, src_ddt, req->src_addr); in spacc_aead_make_ddts() 344 req->src_ddt = src_ddt; in spacc_aead_make_ddts() [all …]
|
D | mv_cesa.c | 255 struct ablkcipher_request *req = ablkcipher_request_cast(cpg->cur_req); in mv_process_current_q() local 256 struct mv_ctx *ctx = crypto_tfm_ctx(req->base.tfm); in mv_process_current_q() 257 struct mv_req_ctx *req_ctx = ablkcipher_request_ctx(req); in mv_process_current_q() 270 memcpy(cpg->sram + SRAM_DATA_IV, req->info, 16); in mv_process_current_q() 310 struct ablkcipher_request *req = ablkcipher_request_cast(cpg->cur_req); in mv_crypto_algo_completion() local 311 struct mv_req_ctx *req_ctx = ablkcipher_request_ctx(req); in mv_crypto_algo_completion() 319 memcpy(req->info, cpg->sram + SRAM_DATA_IV_BUF, 16); in mv_crypto_algo_completion() 324 struct ahash_request *req = ahash_request_cast(cpg->cur_req); in mv_process_hash_current() local 325 const struct mv_tfm_hash_ctx *tfm_ctx = crypto_tfm_ctx(req->base.tfm); in mv_process_hash_current() 326 struct mv_req_hash_ctx *req_ctx = ahash_request_ctx(req); in mv_process_hash_current() [all …]
|
D | mxs-dcp.c | 197 struct ablkcipher_request *req, int init) in mxs_dcp_run_aes() argument 201 struct dcp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in mxs_dcp_run_aes() 253 struct ablkcipher_request *req = ablkcipher_request_cast(arq); in mxs_dcp_aes_block_crypt() local 255 struct dcp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in mxs_dcp_aes_block_crypt() 257 struct scatterlist *dst = req->dst; in mxs_dcp_aes_block_crypt() 258 struct scatterlist *src = req->src; in mxs_dcp_aes_block_crypt() 259 const int nents = sg_nents(req->src); in mxs_dcp_aes_block_crypt() 282 memcpy(key + AES_KEYSIZE_128, req->info, AES_KEYSIZE_128); in mxs_dcp_aes_block_crypt() 289 for_each_sg(req->src, src, nents, i) { in mxs_dcp_aes_block_crypt() 309 ret = mxs_dcp_run_aes(actx, req, init); in mxs_dcp_aes_block_crypt() [all …]
|
D | img-hash.c | 132 struct ahash_request *req; member 165 struct img_hash_request_ctx *ctx = ahash_request_ctx(hdev->req); in img_hash_start() 211 struct img_hash_request_ctx *ctx = ahash_request_ctx(hdev->req); in img_hash_dma_callback() 224 struct img_hash_request_ctx *ctx = ahash_request_ctx(hdev->req); in img_hash_xmit_dma() 254 struct img_hash_request_ctx *ctx = ahash_request_ctx(hdev->req); in img_hash_write_via_cpu() 256 ctx->bufcnt = sg_copy_to_buffer(hdev->req->src, sg_nents(ctx->sg), in img_hash_write_via_cpu() 257 ctx->buffer, hdev->req->nbytes); in img_hash_write_via_cpu() 259 ctx->total = hdev->req->nbytes; in img_hash_write_via_cpu() 269 static int img_hash_finish(struct ahash_request *req) in img_hash_finish() argument 271 struct img_hash_request_ctx *ctx = ahash_request_ctx(req); in img_hash_finish() [all …]
|
D | sahara.c | 577 static int sahara_aes_process(struct ablkcipher_request *req) in sahara_aes_process() argument 588 req->nbytes, req->src, req->dst); in sahara_aes_process() 591 dev->total = req->nbytes; in sahara_aes_process() 592 dev->in_sg = req->src; in sahara_aes_process() 593 dev->out_sg = req->dst; in sahara_aes_process() 595 rctx = ablkcipher_request_ctx(req); in sahara_aes_process() 596 ctx = crypto_ablkcipher_ctx(crypto_ablkcipher_reqtfm(req)); in sahara_aes_process() 600 if ((dev->flags & FLAGS_CBC) && req->info) in sahara_aes_process() 601 memcpy(dev->iv_base, req->info, AES_KEYSIZE_128); in sahara_aes_process() 664 static int sahara_aes_crypt(struct ablkcipher_request *req, unsigned long mode) in sahara_aes_crypt() argument [all …]
|
/linux-4.1.27/fs/cifs/ |
D | smb2pdu.c | 326 struct smb2_negotiate_req *req; in SMB2_negotiate() local 343 rc = small_smb2_init(SMB2_NEGOTIATE, NULL, (void **) &req); in SMB2_negotiate() 347 req->hdr.SessionId = 0; in SMB2_negotiate() 349 req->Dialects[0] = cpu_to_le16(ses->server->vals->protocol_id); in SMB2_negotiate() 351 req->DialectCount = cpu_to_le16(1); /* One vers= at a time for now */ in SMB2_negotiate() 352 inc_rfc1001_len(req, 2); in SMB2_negotiate() 356 req->SecurityMode = cpu_to_le16(SMB2_NEGOTIATE_SIGNING_REQUIRED); in SMB2_negotiate() 358 req->SecurityMode = cpu_to_le16(SMB2_NEGOTIATE_SIGNING_ENABLED); in SMB2_negotiate() 360 req->SecurityMode = 0; in SMB2_negotiate() 362 req->Capabilities = cpu_to_le32(ses->server->vals->req_capabilities); in SMB2_negotiate() [all …]
|
/linux-4.1.27/drivers/misc/genwqe/ |
D | card_ddcb.c | 204 struct ddcb_requ *req; in ddcb_requ_alloc() local 206 req = kzalloc(sizeof(*req), GFP_ATOMIC); in ddcb_requ_alloc() 207 if (!req) in ddcb_requ_alloc() 210 return &req->cmd; in ddcb_requ_alloc() 215 struct ddcb_requ *req = container_of(cmd, struct ddcb_requ, cmd); in ddcb_requ_free() local 217 kfree(req); in ddcb_requ_free() 220 static inline enum genwqe_requ_state ddcb_requ_get_state(struct ddcb_requ *req) in ddcb_requ_get_state() argument 222 return req->req_state; in ddcb_requ_get_state() 225 static inline void ddcb_requ_set_state(struct ddcb_requ *req, in ddcb_requ_set_state() argument 228 req->req_state = new_state; in ddcb_requ_set_state() [all …]
|
/linux-4.1.27/fs/lockd/ |
D | clntproc.c | 124 static void nlmclnt_setlockargs(struct nlm_rqst *req, struct file_lock *fl) in nlmclnt_setlockargs() argument 126 struct nlm_args *argp = &req->a_args; in nlmclnt_setlockargs() 128 char *nodename = req->a_host->h_rpcclnt->cl_nodename; in nlmclnt_setlockargs() 133 lock->oh.data = req->a_owner; in nlmclnt_setlockargs() 134 lock->oh.len = snprintf(req->a_owner, sizeof(req->a_owner), "%u@%s", in nlmclnt_setlockargs() 143 static void nlmclnt_release_lockargs(struct nlm_rqst *req) in nlmclnt_release_lockargs() argument 145 WARN_ON_ONCE(req->a_args.lock.fl.fl_ops != NULL); in nlmclnt_release_lockargs() 249 nlmclnt_call(struct rpc_cred *cred, struct nlm_rqst *req, u32 proc) in nlmclnt_call() argument 251 struct nlm_host *host = req->a_host; in nlmclnt_call() 253 struct nlm_args *argp = &req->a_args; in nlmclnt_call() [all …]
|
/linux-4.1.27/arch/powerpc/platforms/powermac/ |
D | time.c | 107 struct adb_request req; in cuda_get_time() local 110 if (cuda_request(&req, NULL, 2, CUDA_PACKET, CUDA_GET_TIME) < 0) in cuda_get_time() 112 while (!req.complete) in cuda_get_time() 114 if (req.reply_len != 7) in cuda_get_time() 116 req.reply_len); in cuda_get_time() 117 now = (req.reply[3] << 24) + (req.reply[4] << 16) in cuda_get_time() 118 + (req.reply[5] << 8) + req.reply[6]; in cuda_get_time() 127 struct adb_request req; in cuda_set_rtc_time() local 130 if (cuda_request(&req, NULL, 6, CUDA_PACKET, CUDA_SET_TIME, in cuda_set_rtc_time() 134 while (!req.complete) in cuda_set_rtc_time() [all …]
|
/linux-4.1.27/drivers/net/wireless/ti/wl18xx/ |
D | scan.c | 42 struct cfg80211_scan_request *req) in wl18xx_scan_send() argument 78 WARN_ON(req->n_ssids > 1); in wl18xx_scan_send() 86 wlcore_set_scan_chan_params(wl, cmd_channels, req->channels, in wl18xx_scan_send() 87 req->n_channels, req->n_ssids, in wl18xx_scan_send() 97 if (req->no_cck) in wl18xx_scan_send() 102 if (req->n_ssids) { in wl18xx_scan_send() 103 cmd->ssid_len = req->ssids[0].ssid_len; in wl18xx_scan_send() 104 memcpy(cmd->ssid, req->ssids[0].ssid, cmd->ssid_len); in wl18xx_scan_send() 112 req->ssids ? req->ssids[0].ssid : NULL, in wl18xx_scan_send() 113 req->ssids ? req->ssids[0].ssid_len : 0, in wl18xx_scan_send() [all …]
|
/linux-4.1.27/drivers/usb/gadget/legacy/ |
D | dbgp.c | 27 struct usb_request *req; member 94 static void dbgp_complete(struct usb_ep *ep, struct usb_request *req) in dbgp_complete() argument 98 int status = req->status; in dbgp_complete() 110 dbgp_consume(req->buf, req->actual); in dbgp_complete() 112 req->length = DBGP_REQ_LEN; in dbgp_complete() 113 err = usb_ep_queue(ep, req, GFP_ATOMIC); in dbgp_complete() 122 kfree(req->buf); in dbgp_complete() 123 usb_ep_free_request(dbgp.o_ep, req); in dbgp_complete() 133 struct usb_request *req; in dbgp_enable_ep_req() local 135 req = usb_ep_alloc_request(ep, GFP_KERNEL); in dbgp_enable_ep_req() [all …]
|
/linux-4.1.27/drivers/mtd/ubi/ |
D | cdev.c | 440 struct ubi_leb_change_req req; in vol_cdev_ioctl() local 442 err = copy_from_user(&req, argp, in vol_cdev_ioctl() 457 if (req.lnum < 0 || req.lnum >= vol->reserved_pebs || in vol_cdev_ioctl() 458 req.bytes < 0 || req.bytes > vol->usable_leb_size) in vol_cdev_ioctl() 465 err = ubi_start_leb_change(ubi, vol, &req); in vol_cdev_ioctl() 466 if (req.bytes == 0) in vol_cdev_ioctl() 505 struct ubi_map_req req; in vol_cdev_ioctl() local 507 err = copy_from_user(&req, argp, sizeof(struct ubi_map_req)); in vol_cdev_ioctl() 512 err = ubi_leb_map(desc, req.lnum); in vol_cdev_ioctl() 547 struct ubi_set_vol_prop_req req; in vol_cdev_ioctl() local [all …]
|
/linux-4.1.27/arch/s390/pci/ |
D | pci_insn.c | 16 static inline u8 __mpcifc(u64 req, struct zpci_fib *fib, u8 *status) in __mpcifc() argument 24 : [cc] "=d" (cc), [req] "+d" (req), [fib] "+Q" (*fib) in __mpcifc() 26 *status = req >> 24 & 0xff; in __mpcifc() 30 int zpci_mod_fc(u64 req, struct zpci_fib *fib) in zpci_mod_fc() argument 35 cc = __mpcifc(req, fib, &status); in zpci_mod_fc() 89 static inline int __pcilg(u64 *data, u64 req, u64 offset, u8 *status) in __pcilg() argument 91 register u64 __req asm("2") = req; in __pcilg() 102 : [cc] "+d" (cc), [data] "=d" (__data), [req] "+d" (__req) in __pcilg() 112 int zpci_load(u64 *data, u64 req, u64 offset) in zpci_load() argument 118 cc = __pcilg(data, req, offset, &status); in zpci_load() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lustre/osc/ |
D | osc_request.c | 99 struct ptlrpc_request *req, void *data, int rc); 193 static inline void osc_pack_capa(struct ptlrpc_request *req, in osc_pack_capa() argument 202 c = req_capsule_client_get(&req->rq_pill, &RMF_CAPA1); in osc_pack_capa() 209 static inline void osc_pack_req_body(struct ptlrpc_request *req, in osc_pack_req_body() argument 214 body = req_capsule_client_get(&req->rq_pill, &RMF_OST_BODY); in osc_pack_req_body() 217 lustre_set_wire_obdo(&req->rq_import->imp_connect_data, &body->oa, in osc_pack_req_body() 219 osc_pack_capa(req, body, oinfo->oi_capa); in osc_pack_req_body() 222 static inline void osc_set_capa_size(struct ptlrpc_request *req, in osc_set_capa_size() argument 227 req_capsule_set_size(&req->rq_pill, field, RCL_CLIENT, 0); in osc_set_capa_size() 234 struct ptlrpc_request *req, in osc_getattr_interpret() argument [all …]
|
/linux-4.1.27/drivers/scsi/qla2xxx/ |
D | qla_mid.c | 486 vha->req = base_vha->req; in qla24xx_create_vhost() 487 host->can_queue = base_vha->req->length + 128; in qla24xx_create_vhost() 517 qla25xx_free_req_que(struct scsi_qla_host *vha, struct req_que *req) in qla25xx_free_req_que() argument 520 uint16_t que_id = req->id; in qla25xx_free_req_que() 522 dma_free_coherent(&ha->pdev->dev, (req->length + 1) * in qla25xx_free_req_que() 523 sizeof(request_t), req->ring, req->dma); in qla25xx_free_req_que() 524 req->ring = NULL; in qla25xx_free_req_que() 525 req->dma = 0; in qla25xx_free_req_que() 532 kfree(req->outstanding_cmds); in qla25xx_free_req_que() 533 kfree(req); in qla25xx_free_req_que() [all …]
|
D | qla_iocb.c | 98 struct req_que *req = vha->req; in qla2x00_prep_cont_type0_iocb() local 100 req->ring_index++; in qla2x00_prep_cont_type0_iocb() 101 if (req->ring_index == req->length) { in qla2x00_prep_cont_type0_iocb() 102 req->ring_index = 0; in qla2x00_prep_cont_type0_iocb() 103 req->ring_ptr = req->ring; in qla2x00_prep_cont_type0_iocb() 105 req->ring_ptr++; in qla2x00_prep_cont_type0_iocb() 108 cont_pkt = (cont_entry_t *)req->ring_ptr; in qla2x00_prep_cont_type0_iocb() 124 qla2x00_prep_cont_type1_iocb(scsi_qla_host_t *vha, struct req_que *req) in qla2x00_prep_cont_type1_iocb() argument 129 req->ring_index++; in qla2x00_prep_cont_type1_iocb() 130 if (req->ring_index == req->length) { in qla2x00_prep_cont_type1_iocb() [all …]
|
/linux-4.1.27/drivers/infiniband/core/ |
D | addr.c | 185 static void queue_req(struct addr_req *req) in queue_req() argument 191 if (time_after_eq(req->timeout, temp_req->timeout)) in queue_req() 195 list_add(&req->list, &temp_req->list); in queue_req() 197 if (req_list.next == &req->list) in queue_req() 198 set_timeout(req->timeout); in queue_req() 336 struct addr_req *req, *temp_req; in process_req() local 343 list_for_each_entry_safe(req, temp_req, &req_list, list) { in process_req() 344 if (req->status == -ENODATA) { in process_req() 345 src_in = (struct sockaddr *) &req->src_addr; in process_req() 346 dst_in = (struct sockaddr *) &req->dst_addr; in process_req() [all …]
|
/linux-4.1.27/drivers/isdn/hardware/eicon/ |
D | divasfunc.c | 67 IDI_SYNC_REQ req; in diva_xdi_didd_register_adapter() local 81 req.didd_remove_adapter.e.Req = 0; in diva_xdi_didd_register_adapter() 82 req.didd_add_adapter.e.Rc = IDI_SYNC_REQ_DIDD_ADD_ADAPTER; in diva_xdi_didd_register_adapter() 83 req.didd_add_adapter.info.descriptor = (void *) &d; in diva_xdi_didd_register_adapter() 84 DAdapter.request((ENTITY *)&req); in diva_xdi_didd_register_adapter() 85 if (req.didd_add_adapter.e.Rc != 0xff) { in diva_xdi_didd_register_adapter() 97 IDI_SYNC_REQ req; in diva_xdi_didd_remove_adapter() local 102 req.didd_remove_adapter.e.Req = 0; in diva_xdi_didd_remove_adapter() 103 req.didd_remove_adapter.e.Rc = IDI_SYNC_REQ_DIDD_REMOVE_ADAPTER; in diva_xdi_didd_remove_adapter() 104 req.didd_remove_adapter.info.p_request = in diva_xdi_didd_remove_adapter() [all …]
|
D | diddfunc.c | 54 IDI_SYNC_REQ req; in connect_didd() local 63 req.didd_notify.e.Req = 0; in connect_didd() 64 req.didd_notify.e.Rc = in connect_didd() 66 req.didd_notify.info.callback = (void *)didd_callback; in connect_didd() 67 req.didd_notify.info.context = NULL; in connect_didd() 68 _DAdapter.request((ENTITY *)&req); in connect_didd() 69 if (req.didd_notify.e.Rc != 0xff) in connect_didd() 71 notify_handle = req.didd_notify.info.handle; in connect_didd() 84 IDI_SYNC_REQ req; in disconnect_didd() local 86 req.didd_notify.e.Req = 0; in disconnect_didd() [all …]
|
D | mntfunc.c | 79 IDI_SYNC_REQ req; in connect_didd() local 88 req.didd_notify.e.Req = 0; in connect_didd() 89 req.didd_notify.e.Rc = in connect_didd() 91 req.didd_notify.info.callback = (void *)didd_callback; in connect_didd() 92 req.didd_notify.info.context = NULL; in connect_didd() 93 DAdapter.request((ENTITY *)&req); in connect_didd() 94 if (req.didd_notify.e.Rc != 0xff) in connect_didd() 96 notify_handle = req.didd_notify.info.handle; in connect_didd() 98 req.didd_add_adapter.e.Req = 0; in connect_didd() 99 req.didd_add_adapter.e.Rc = in connect_didd() [all …]
|
/linux-4.1.27/drivers/scsi/be2iscsi/ |
D | be_mgmt.c | 163 struct be_cmd_req_modify_eq_delay *req; in be_cmd_modify_eq_delay() local 175 req = embedded_payload(wrb); in be_cmd_modify_eq_delay() 178 be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); in be_cmd_modify_eq_delay() 179 be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, in be_cmd_modify_eq_delay() 180 OPCODE_COMMON_MODIFY_EQ_DELAY, sizeof(*req)); in be_cmd_modify_eq_delay() 182 req->num_eq = cpu_to_le32(num); in be_cmd_modify_eq_delay() 184 req->delay[i].eq_id = cpu_to_le32(set_eqd[i].eq_id); in be_cmd_modify_eq_delay() 185 req->delay[i].phase = 0; in be_cmd_modify_eq_delay() 186 req->delay[i].delay_multiplier = in be_cmd_modify_eq_delay() 211 struct be_cmd_reopen_session_req *req; in mgmt_reopen_session() local [all …]
|
D | be_cmds.c | 827 struct be_cmd_req_eq_create *req = embedded_payload(wrb); in beiscsi_cmd_eq_create() local 835 be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); in beiscsi_cmd_eq_create() 837 be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, in beiscsi_cmd_eq_create() 838 OPCODE_COMMON_EQ_CREATE, sizeof(*req)); in beiscsi_cmd_eq_create() 840 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in beiscsi_cmd_eq_create() 842 AMAP_SET_BITS(struct amap_eq_context, func, req->context, in beiscsi_cmd_eq_create() 844 AMAP_SET_BITS(struct amap_eq_context, valid, req->context, 1); in beiscsi_cmd_eq_create() 845 AMAP_SET_BITS(struct amap_eq_context, size, req->context, 0); in beiscsi_cmd_eq_create() 846 AMAP_SET_BITS(struct amap_eq_context, count, req->context, in beiscsi_cmd_eq_create() 848 AMAP_SET_BITS(struct amap_eq_context, delaymult, req->context, in beiscsi_cmd_eq_create() [all …]
|
/linux-4.1.27/drivers/net/wireless/iwlwifi/mvm/ |
D | scan.c | 393 static void iwl_scan_offload_build_ssid(struct cfg80211_sched_scan_request *req, in iwl_scan_offload_build_ssid() argument 405 for (i = 0; i < req->n_match_sets && i < PROBE_OPTION_MAX; i++) { in iwl_scan_offload_build_ssid() 407 if (!req->match_sets[i].ssid.ssid_len) in iwl_scan_offload_build_ssid() 410 direct_scan[i].len = req->match_sets[i].ssid.ssid_len; in iwl_scan_offload_build_ssid() 411 memcpy(direct_scan[i].ssid, req->match_sets[i].ssid.ssid, in iwl_scan_offload_build_ssid() 417 for (j = 0; j < req->n_ssids && i < PROBE_OPTION_MAX; j++) { in iwl_scan_offload_build_ssid() 418 index = iwl_ssid_exist(req->ssids[j].ssid, in iwl_scan_offload_build_ssid() 419 req->ssids[j].ssid_len, in iwl_scan_offload_build_ssid() 422 if (!req->ssids[j].ssid_len && basic_ssid) in iwl_scan_offload_build_ssid() 425 direct_scan[i].len = req->ssids[j].ssid_len; in iwl_scan_offload_build_ssid() [all …]
|
/linux-4.1.27/drivers/crypto/nx/ |
D | nx-aes-ccm.c | 173 struct aead_request *req, in generate_pat() argument 203 if (!req->assoclen) { in generate_pat() 205 } else if (req->assoclen <= 14) { in generate_pat() 211 iauth_len = req->assoclen; in generate_pat() 212 } else if (req->assoclen <= 65280) { in generate_pat() 226 rc = generate_b0(iv, req->assoclen, authsize, nbytes, b0); in generate_pat() 236 if (req->assoclen <= 65280) { in generate_pat() 237 *(u16 *)b1 = (u16)req->assoclen; in generate_pat() 238 scatterwalk_map_and_copy(b1 + 2, req->assoc, 0, in generate_pat() 242 *(u32 *)&b1[2] = (u32)req->assoclen; in generate_pat() [all …]
|
D | nx-aes-gcm.c | 125 struct aead_request *req, in nx_gca() argument 132 unsigned int nbytes = req->assoclen; in nx_gca() 137 scatterwalk_start(&walk, req->assoc); in nx_gca() 162 req->assoc, processed, &to_process); in nx_gca() 173 req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP); in nx_gca() 183 atomic64_add(req->assoclen, &(nx_ctx->stats->aes_bytes)); in nx_gca() 193 static int gmac(struct aead_request *req, struct blkcipher_desc *desc) in gmac() argument 196 struct nx_crypto_ctx *nx_ctx = crypto_tfm_ctx(req->base.tfm); in gmac() 199 unsigned int nbytes = req->assoclen; in gmac() 228 req->assoc, processed, &to_process); in gmac() [all …]
|
/linux-4.1.27/net/ipv4/ |
D | inet_connection_sock.c | 298 struct request_sock *req; in inet_csk_accept() local 324 req = reqsk_queue_remove(queue); in inet_csk_accept() 325 newsk = req->sk; in inet_csk_accept() 329 tcp_rsk(req)->tfo_listener && in inet_csk_accept() 332 if (tcp_rsk(req)->tfo_listener) { in inet_csk_accept() 339 req->sk = NULL; in inet_csk_accept() 340 req = NULL; in inet_csk_accept() 346 if (req) in inet_csk_accept() 347 reqsk_put(req); in inet_csk_accept() 351 req = NULL; in inet_csk_accept() [all …]
|
D | tcp_minisocks.c | 366 void tcp_openreq_init_rwin(struct request_sock *req, in tcp_openreq_init_rwin() argument 369 struct inet_request_sock *ireq = inet_rsk(req); in tcp_openreq_init_rwin() 378 req->window_clamp = tp->window_clamp ? : dst_metric(dst, RTAX_WINDOW); in tcp_openreq_init_rwin() 382 (req->window_clamp > tcp_full_space(sk) || req->window_clamp == 0)) in tcp_openreq_init_rwin() 383 req->window_clamp = tcp_full_space(sk); in tcp_openreq_init_rwin() 388 &req->rcv_wnd, in tcp_openreq_init_rwin() 389 &req->window_clamp, in tcp_openreq_init_rwin() 398 const struct request_sock *req) in tcp_ecn_openreq_child() argument 400 tp->ecn_flags = inet_rsk(req)->ecn_ok ? TCP_ECN_OK : 0; in tcp_ecn_openreq_child() 438 struct sock *tcp_create_openreq_child(struct sock *sk, struct request_sock *req, struct sk_buff *sk… in tcp_create_openreq_child() argument [all …]
|
D | udp_diag.c | 22 const struct inet_diag_req_v2 *req, in sk_diag_dump() argument 28 return inet_sk_diag_fill(sk, NULL, skb, req, in sk_diag_dump() 36 const struct inet_diag_req_v2 *req) in udp_dump_one() argument 43 if (req->sdiag_family == AF_INET) in udp_dump_one() 45 req->id.idiag_src[0], req->id.idiag_sport, in udp_dump_one() 46 req->id.idiag_dst[0], req->id.idiag_dport, in udp_dump_one() 47 req->id.idiag_if, tbl); in udp_dump_one() 49 else if (req->sdiag_family == AF_INET6) in udp_dump_one() 51 (struct in6_addr *)req->id.idiag_src, in udp_dump_one() 52 req->id.idiag_sport, in udp_dump_one() [all …]
|
D | syncookies.c | 82 __u32 cookie_init_timestamp(struct request_sock *req) in cookie_init_timestamp() argument 88 ireq = inet_rsk(req); in cookie_init_timestamp() 223 struct request_sock *req, in get_cookie_sock() argument 229 child = icsk->icsk_af_ops->syn_recv_sock(sk, skb, req, dst); in get_cookie_sock() 231 atomic_set(&req->rsk_refcnt, 1); in get_cookie_sock() 232 inet_csk_reqsk_queue_add(sk, req, child); in get_cookie_sock() 234 reqsk_free(req); in get_cookie_sock() 301 struct request_sock *req; in cookie_v4_check() local 329 req = inet_reqsk_alloc(&tcp_request_sock_ops, sk); /* for safety */ in cookie_v4_check() 330 if (!req) in cookie_v4_check() [all …]
|
/linux-4.1.27/drivers/block/ |
D | nbd.c | 99 static void nbd_end_request(struct nbd_device *nbd, struct request *req) in nbd_end_request() argument 101 int error = req->errors ? -EIO : 0; in nbd_end_request() 102 struct request_queue *q = req->q; in nbd_end_request() 105 dev_dbg(nbd_to_dev(nbd), "request %p: %s\n", req, in nbd_end_request() 109 __blk_end_request_all(req, error); in nbd_end_request() 228 static int nbd_send_req(struct nbd_device *nbd, struct request *req) in nbd_send_req() argument 232 unsigned long size = blk_rq_bytes(req); in nbd_send_req() 236 request.type = htonl(nbd_cmd(req)); in nbd_send_req() 238 if (nbd_cmd(req) != NBD_CMD_FLUSH && nbd_cmd(req) != NBD_CMD_DISC) { in nbd_send_req() 239 request.from = cpu_to_be64((u64)blk_rq_pos(req) << 9); in nbd_send_req() [all …]
|
D | mg_disk.c | 112 struct request *req; member 151 if (__blk_end_request(host->req, err, nr_bytes)) in mg_end_request() 154 host->req = NULL; in mg_end_request() 160 return mg_end_request(host, err, blk_rq_cur_bytes(host->req)); in mg_end_request_cur() 168 if (host->req) in mg_dump_status() 169 name = host->req->rq_disk->disk_name; in mg_dump_status() 205 if (host->req) in mg_dump_status() 207 (unsigned int)blk_rq_pos(host->req)); in mg_dump_status() 445 if (host->req) in mg_bad_rw_intr() 446 if (++host->req->errors >= MG_MAX_ERRORS || in mg_bad_rw_intr() [all …]
|
/linux-4.1.27/fs/coda/ |
D | psdev.c | 101 struct upc_req *req = NULL; in coda_psdev_write() local 153 req = tmp; in coda_psdev_write() 154 list_del(&req->uc_chain); in coda_psdev_write() 160 if (!req) { in coda_psdev_write() 168 if (req->uc_outSize < nbytes) { in coda_psdev_write() 170 __func__, req->uc_outSize, (long)nbytes, in coda_psdev_write() 172 nbytes = req->uc_outSize; /* don't have more space! */ in coda_psdev_write() 174 if (copy_from_user(req->uc_data, buf, nbytes)) { in coda_psdev_write() 175 req->uc_flags |= CODA_REQ_ABORT; in coda_psdev_write() 176 wake_up(&req->uc_sleep); in coda_psdev_write() [all …]
|
/linux-4.1.27/arch/arm64/crypto/ |
D | aes-ce-ccm-glue.c | 68 static int ccm_init_mac(struct aead_request *req, u8 maciv[], u32 msglen) in ccm_init_mac() argument 70 struct crypto_aead *aead = crypto_aead_reqtfm(req); in ccm_init_mac() 72 u32 l = req->iv[0] + 1; in ccm_init_mac() 89 memcpy(maciv, req->iv, AES_BLOCK_SIZE - l); in ccm_init_mac() 99 if (req->assoclen) in ccm_init_mac() 102 memset(&req->iv[AES_BLOCK_SIZE - l], 0, l); in ccm_init_mac() 106 static void ccm_calculate_auth_mac(struct aead_request *req, u8 mac[]) in ccm_calculate_auth_mac() argument 108 struct crypto_aead *aead = crypto_aead_reqtfm(req); in ccm_calculate_auth_mac() 112 u32 len = req->assoclen; in ccm_calculate_auth_mac() 127 scatterwalk_start(&walk, req->assoc); in ccm_calculate_auth_mac() [all …]
|
/linux-4.1.27/arch/arm/crypto/ |
D | ghash-ce-core.S | 14 SHASH .req q0 15 SHASH2 .req q1 16 T1 .req q2 17 T2 .req q3 18 MASK .req q4 19 XL .req q5 20 XM .req q6 21 XH .req q7 22 IN1 .req q7 24 SHASH_L .req d0 [all …]
|
D | sha1-ce-core.S | 18 k0 .req q0 19 k1 .req q1 20 k2 .req q2 21 k3 .req q3 23 ta0 .req q4 24 ta1 .req q5 25 tb0 .req q5 26 tb1 .req q4 28 dga .req q6 29 dgb .req q7 [all …]
|
D | ghash-ce-glue.c | 151 static int ghash_async_init(struct ahash_request *req) in ghash_async_init() argument 153 struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); in ghash_async_init() 155 struct ahash_request *cryptd_req = ahash_request_ctx(req); in ghash_async_init() 159 memcpy(cryptd_req, req, sizeof(*req)); in ghash_async_init() 167 desc->flags = req->base.flags; in ghash_async_init() 172 static int ghash_async_update(struct ahash_request *req) in ghash_async_update() argument 174 struct ahash_request *cryptd_req = ahash_request_ctx(req); in ghash_async_update() 177 struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); in ghash_async_update() 181 memcpy(cryptd_req, req, sizeof(*req)); in ghash_async_update() 186 return shash_ahash_update(req, desc); in ghash_async_update() [all …]
|
/linux-4.1.27/drivers/mfd/ |
D | pcf50633-adc.c | 85 adc_enqueue_request(struct pcf50633 *pcf, struct pcf50633_adc_request *req) in adc_enqueue_request() argument 101 adc->queue[tail] = req; in adc_enqueue_request() 114 struct pcf50633_adc_sync_request *req = param; in pcf50633_adc_sync_read_callback() local 116 req->result = result; in pcf50633_adc_sync_read_callback() 117 complete(&req->completion); in pcf50633_adc_sync_read_callback() 122 struct pcf50633_adc_sync_request req; in pcf50633_adc_sync_read() local 125 init_completion(&req.completion); in pcf50633_adc_sync_read() 128 pcf50633_adc_sync_read_callback, &req); in pcf50633_adc_sync_read() 132 wait_for_completion(&req.completion); in pcf50633_adc_sync_read() 134 return req.result; in pcf50633_adc_sync_read() [all …]
|
/linux-4.1.27/drivers/input/touchscreen/ |
D | ad7877.c | 215 struct ser_req *req; in ad7877_read() local 218 req = kzalloc(sizeof *req, GFP_KERNEL); in ad7877_read() 219 if (!req) in ad7877_read() 222 spi_message_init(&req->msg); in ad7877_read() 224 req->command = (u16) (AD7877_WRITEADD(AD7877_REG_CTRL1) | in ad7877_read() 226 req->xfer[0].tx_buf = &req->command; in ad7877_read() 227 req->xfer[0].len = 2; in ad7877_read() 228 req->xfer[0].cs_change = 1; in ad7877_read() 230 req->xfer[1].rx_buf = &req->sample; in ad7877_read() 231 req->xfer[1].len = 2; in ad7877_read() [all …]
|
/linux-4.1.27/fs/ecryptfs/ |
D | kthread.c | 61 struct ecryptfs_open_req *req; in ecryptfs_threadfn() local 73 req = list_first_entry(&ecryptfs_kthread_ctl.req_list, in ecryptfs_threadfn() 76 list_del(&req->kthread_ctl_list); in ecryptfs_threadfn() 77 *req->lower_file = dentry_open(&req->path, in ecryptfs_threadfn() 79 complete(&req->done); in ecryptfs_threadfn() 106 struct ecryptfs_open_req *req, *tmp; in ecryptfs_destroy_kthread() local 110 list_for_each_entry_safe(req, tmp, &ecryptfs_kthread_ctl.req_list, in ecryptfs_destroy_kthread() 112 list_del(&req->kthread_ctl_list); in ecryptfs_destroy_kthread() 113 *req->lower_file = ERR_PTR(-EIO); in ecryptfs_destroy_kthread() 114 complete(&req->done); in ecryptfs_destroy_kthread() [all …]
|
/linux-4.1.27/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_vfpf.c | 53 first_tlv->resp_msg_offset = sizeof(bp->vf2pf_mbox->req); in bnx2x_vfpf_prep() 224 struct vfpf_acquire_tlv *req = &bp->vf2pf_mbox->req.acquire; in bnx2x_vfpf_acquire() local 232 bnx2x_vfpf_prep(bp, &req->first_tlv, CHANNEL_TLV_ACQUIRE, sizeof(*req)); in bnx2x_vfpf_acquire() 239 req->vfdev_info.vf_id = vf_id; in bnx2x_vfpf_acquire() 240 req->vfdev_info.vf_os = 0; in bnx2x_vfpf_acquire() 241 req->vfdev_info.fp_hsi_ver = ETH_FP_HSI_VERSION; in bnx2x_vfpf_acquire() 243 req->resc_request.num_rxqs = rx_count; in bnx2x_vfpf_acquire() 244 req->resc_request.num_txqs = tx_count; in bnx2x_vfpf_acquire() 245 req->resc_request.num_sbs = bp->igu_sb_cnt; in bnx2x_vfpf_acquire() 246 req->resc_request.num_mac_filters = VF_ACQUIRE_MAC_FILTERS; in bnx2x_vfpf_acquire() [all …]
|
/linux-4.1.27/drivers/usb/gadget/function/ |
D | f_printer.c | 237 struct usb_request *req; in printer_req_alloc() local 239 req = usb_ep_alloc_request(ep, gfp_flags); in printer_req_alloc() 241 if (req != NULL) { in printer_req_alloc() 242 req->length = len; in printer_req_alloc() 243 req->buf = kmalloc(len, gfp_flags); in printer_req_alloc() 244 if (req->buf == NULL) { in printer_req_alloc() 245 usb_ep_free_request(ep, req); in printer_req_alloc() 250 return req; in printer_req_alloc() 254 printer_req_free(struct usb_ep *ep, struct usb_request *req) in printer_req_free() argument 256 if (ep != NULL && req != NULL) { in printer_req_free() [all …]
|
D | uvc_video.c | 62 uvc_video_encode_bulk(struct usb_request *req, struct uvc_video *video, in uvc_video_encode_bulk() argument 65 void *mem = req->buf; in uvc_video_encode_bulk() 84 req->length = video->req_size - len; in uvc_video_encode_bulk() 85 req->zero = video->payload_size == video->max_payload_size; in uvc_video_encode_bulk() 102 uvc_video_encode_isoc(struct usb_request *req, struct uvc_video *video, in uvc_video_encode_isoc() argument 105 void *mem = req->buf; in uvc_video_encode_isoc() 118 req->length = video->req_size - len; in uvc_video_encode_isoc() 163 uvc_video_complete(struct usb_ep *ep, struct usb_request *req) in uvc_video_complete() argument 165 struct uvc_video *video = req->context; in uvc_video_complete() 171 switch (req->status) { in uvc_video_complete() [all …]
|
/linux-4.1.27/drivers/usb/gadget/ |
D | u_f.c | 19 struct usb_request *req; in alloc_ep_req() local 21 req = usb_ep_alloc_request(ep, GFP_ATOMIC); in alloc_ep_req() 22 if (req) { in alloc_ep_req() 23 req->length = len ?: default_len; in alloc_ep_req() 24 req->buf = kmalloc(req->length, GFP_ATOMIC); in alloc_ep_req() 25 if (!req->buf) { in alloc_ep_req() 26 usb_ep_free_request(ep, req); in alloc_ep_req() 27 req = NULL; in alloc_ep_req() 30 return req; in alloc_ep_req()
|
/linux-4.1.27/drivers/staging/lustre/lustre/ldlm/ |
D | ldlm_lockd.c | 180 static void ldlm_handle_cp_callback(struct ptlrpc_request *req, in ldlm_handle_cp_callback() argument 203 lvb_len = req_capsule_get_size(&req->rq_pill, &RMF_DLM_LVB, RCL_CLIENT); in ldlm_handle_cp_callback() 257 ldlm_convert_policy_to_local(req->rq_export, in ldlm_handle_cp_callback() 289 rc = ldlm_fill_lvb(lock, &req->rq_pill, RCL_CLIENT, in ldlm_handle_cp_callback() 329 static void ldlm_handle_gl_callback(struct ptlrpc_request *req, in ldlm_handle_gl_callback() argument 339 rc = lock->l_glimpse_ast(lock, req); in ldlm_handle_gl_callback() 341 if (req->rq_repmsg != NULL) { in ldlm_handle_gl_callback() 342 ptlrpc_reply(req); in ldlm_handle_gl_callback() 344 req->rq_status = rc; in ldlm_handle_gl_callback() 345 ptlrpc_error(req); in ldlm_handle_gl_callback() [all …]
|
D | ldlm_flock.c | 97 static inline void ldlm_flock_blocking_link(struct ldlm_lock *req, in ldlm_flock_blocking_link() argument 101 if (req->l_export == NULL) in ldlm_flock_blocking_link() 104 LASSERT(hlist_unhashed(&req->l_exp_flock_hash)); in ldlm_flock_blocking_link() 106 req->l_policy_data.l_flock.blocking_owner = in ldlm_flock_blocking_link() 108 req->l_policy_data.l_flock.blocking_export = in ldlm_flock_blocking_link() 110 req->l_policy_data.l_flock.blocking_refs = 0; in ldlm_flock_blocking_link() 112 cfs_hash_add(req->l_export->exp_flock_hash, in ldlm_flock_blocking_link() 113 &req->l_policy_data.l_flock.owner, in ldlm_flock_blocking_link() 114 &req->l_exp_flock_hash); in ldlm_flock_blocking_link() 117 static inline void ldlm_flock_blocking_unlink(struct ldlm_lock *req) in ldlm_flock_blocking_unlink() argument [all …]
|
D | ldlm_request.c | 506 int ldlm_cli_enqueue_fini(struct obd_export *exp, struct ptlrpc_request *req, in ldlm_cli_enqueue_fini() argument 538 reply = req_capsule_server_get(&req->rq_pill, &RMF_DLM_REP); in ldlm_cli_enqueue_fini() 547 size = req_capsule_get_size(&req->rq_pill, &RMF_DLM_LVB, in ldlm_cli_enqueue_fini() 563 rc = ldlm_fill_lvb(lock, &req->rq_pill, RCL_SERVER, in ldlm_cli_enqueue_fini() 660 rc = ldlm_fill_lvb(lock, &req->rq_pill, RCL_SERVER, in ldlm_cli_enqueue_fini() 744 int ldlm_prep_elc_req(struct obd_export *exp, struct ptlrpc_request *req, in ldlm_prep_elc_req() argument 749 struct req_capsule *pill = &req->rq_pill; in ldlm_prep_elc_req() 781 rc = ptlrpc_request_pack(req, version, opc); in ldlm_prep_elc_req() 798 ldlm_cli_cancel_list(cancels, pack, req, 0); in ldlm_prep_elc_req() 808 int ldlm_prep_enqueue_req(struct obd_export *exp, struct ptlrpc_request *req, in ldlm_prep_enqueue_req() argument [all …]
|
/linux-4.1.27/drivers/s390/char/ |
D | sclp.c | 60 static void sclp_suspend_req_cb(struct sclp_req *req, void *data) in sclp_suspend_req_cb() argument 228 struct sclp_req *req; in __sclp_req_queue_find_next_timeout() local 230 list_for_each_entry(req, &sclp_req_queue, list) { in __sclp_req_queue_find_next_timeout() 231 if (!req->queue_expires) in __sclp_req_queue_find_next_timeout() 234 (time_before(req->queue_expires, expires_next))) in __sclp_req_queue_find_next_timeout() 235 expires_next = req->queue_expires; in __sclp_req_queue_find_next_timeout() 246 struct sclp_req *req; in __sclp_req_queue_remove_expired_req() local 251 list_for_each_entry(req, &sclp_req_queue, list) { in __sclp_req_queue_remove_expired_req() 252 if (!req->queue_expires) in __sclp_req_queue_remove_expired_req() 254 if (time_before_eq(req->queue_expires, now)) { in __sclp_req_queue_remove_expired_req() [all …]
|
D | con3215.c | 111 struct raw3215_req *req; in raw3215_alloc_req() local 115 req = raw3215_freelist; in raw3215_alloc_req() 116 raw3215_freelist = req->next; in raw3215_alloc_req() 118 return req; in raw3215_alloc_req() 124 static inline void raw3215_free_req(struct raw3215_req *req) in raw3215_free_req() argument 128 if (req->type == RAW3215_FREE) in raw3215_free_req() 130 req->type = RAW3215_FREE; in raw3215_free_req() 132 req->next = raw3215_freelist; in raw3215_free_req() 133 raw3215_freelist = req; in raw3215_free_req() 145 struct raw3215_req *req; in raw3215_mk_read_req() local [all …]
|
/linux-4.1.27/drivers/net/usb/ |
D | cdc-phonet.c | 55 static void tx_complete(struct urb *req); 56 static void rx_complete(struct urb *req); 64 struct urb *req = NULL; in usbpn_xmit() local 71 req = usb_alloc_urb(0, GFP_ATOMIC); in usbpn_xmit() 72 if (!req) in usbpn_xmit() 74 usb_fill_bulk_urb(req, pnd->usb, pnd->tx_pipe, skb->data, skb->len, in usbpn_xmit() 76 req->transfer_flags = URB_ZERO_PACKET; in usbpn_xmit() 77 err = usb_submit_urb(req, GFP_ATOMIC); in usbpn_xmit() 79 usb_free_urb(req); in usbpn_xmit() 96 static void tx_complete(struct urb *req) in tx_complete() argument [all …]
|
/linux-4.1.27/drivers/crypto/amcc/ |
D | crypto4xx_alg.c | 73 int crypto4xx_encrypt(struct ablkcipher_request *req) in crypto4xx_encrypt() argument 75 struct crypto4xx_ctx *ctx = crypto_tfm_ctx(req->base.tfm); in crypto4xx_encrypt() 82 return crypto4xx_build_pd(&req->base, ctx, req->src, req->dst, in crypto4xx_encrypt() 83 req->nbytes, req->info, in crypto4xx_encrypt() 87 int crypto4xx_decrypt(struct ablkcipher_request *req) in crypto4xx_decrypt() argument 89 struct crypto4xx_ctx *ctx = crypto_tfm_ctx(req->base.tfm); in crypto4xx_decrypt() 96 return crypto4xx_build_pd(&req->base, ctx, req->src, req->dst, in crypto4xx_decrypt() 97 req->nbytes, req->info, in crypto4xx_decrypt() 237 int crypto4xx_hash_init(struct ahash_request *req) in crypto4xx_hash_init() argument 239 struct crypto4xx_ctx *ctx = crypto_tfm_ctx(req->base.tfm); in crypto4xx_hash_init() [all …]
|
/linux-4.1.27/drivers/crypto/ccp/ |
D | ccp-crypto-aes.c | 27 struct ablkcipher_request *req = ablkcipher_request_cast(async_req); in ccp_aes_complete() local 28 struct ccp_ctx *ctx = crypto_tfm_ctx(req->base.tfm); in ccp_aes_complete() 29 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_complete() 35 memcpy(req->info, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_complete() 70 static int ccp_aes_crypt(struct ablkcipher_request *req, bool encrypt) in ccp_aes_crypt() argument 72 struct ccp_ctx *ctx = crypto_tfm_ctx(req->base.tfm); in ccp_aes_crypt() 73 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_crypt() 84 (req->nbytes & (AES_BLOCK_SIZE - 1))) in ccp_aes_crypt() 88 if (!req->info) in ccp_aes_crypt() 91 memcpy(rctx->iv, req->info, AES_BLOCK_SIZE); in ccp_aes_crypt() [all …]
|
D | ccp-crypto-main.c | 66 struct crypto_async_request *req; member 140 struct crypto_async_request *req = crypto_cmd->req; in ccp_crypto_complete() local 141 struct ccp_ctx *ctx = crypto_tfm_ctx(req->tfm); in ccp_crypto_complete() 148 req->complete(req, -EINPROGRESS); in ccp_crypto_complete() 161 backlog->req->complete(backlog->req, -EINPROGRESS); in ccp_crypto_complete() 166 req->complete(req, -EINPROGRESS); in ccp_crypto_complete() 171 ret = ctx->complete(req, ret); in ccp_crypto_complete() 172 req->complete(req, ret); in ccp_crypto_complete() 185 ctx = crypto_tfm_ctx(held->req->tfm); in ccp_crypto_complete() 187 ret = ctx->complete(held->req, ret); in ccp_crypto_complete() [all …]
|
D | ccp-crypto-aes-xts.c | 86 struct ablkcipher_request *req = ablkcipher_request_cast(async_req); in ccp_aes_xts_complete() local 87 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_xts_complete() 92 memcpy(req->info, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_complete() 117 static int ccp_aes_xts_crypt(struct ablkcipher_request *req, in ccp_aes_xts_crypt() argument 121 crypto_ablkcipher_tfm(crypto_ablkcipher_reqtfm(req)); in ccp_aes_xts_crypt() 122 struct ccp_ctx *ctx = crypto_tfm_ctx(req->base.tfm); in ccp_aes_xts_crypt() 123 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_xts_crypt() 131 if (req->nbytes & (AES_BLOCK_SIZE - 1)) in ccp_aes_xts_crypt() 134 if (!req->info) in ccp_aes_xts_crypt() 138 if (req->nbytes <= unit_size_map[0].size) { in ccp_aes_xts_crypt() [all …]
|
/linux-4.1.27/include/scsi/ |
D | scsi_tcq.h | 17 struct request *req = NULL; in scsi_mq_find_tag() local 20 req = blk_mq_tag_to_rq(shost->tag_set.tags[hwq], in scsi_mq_find_tag() 22 return req ? (struct scsi_cmnd *)req->special : NULL; in scsi_mq_find_tag() 35 struct request *req; in scsi_find_tag() local 41 req = blk_queue_find_tag(sdev->request_queue, tag); in scsi_find_tag() 42 return req ? (struct scsi_cmnd *)req->special : NULL; in scsi_find_tag() 89 struct request *req; in scsi_host_find_tag() local 94 req = blk_map_queue_find_tag(shost->bqt, tag); in scsi_host_find_tag() 95 return req ? (struct scsi_cmnd *)req->special : NULL; in scsi_host_find_tag()
|
/linux-4.1.27/drivers/crypto/caam/ |
D | caamhash.c | 131 int (*update)(struct ahash_request *req); 132 int (*final)(struct ahash_request *req); 133 int (*finup)(struct ahash_request *req); 606 struct ahash_request *req, int dst_len) in ahash_unmap() argument 609 dma_unmap_sg_chained(dev, req->src, edesc->src_nents, in ahash_unmap() 621 struct ahash_request *req, int dst_len, u32 flag) in ahash_unmap_ctx() argument 623 struct crypto_ahash *ahash = crypto_ahash_reqtfm(req); in ahash_unmap_ctx() 625 struct caam_hash_state *state = ahash_request_ctx(req); in ahash_unmap_ctx() 629 ahash_unmap(dev, edesc, req, dst_len); in ahash_unmap_ctx() 635 struct ahash_request *req = context; in ahash_done() local [all …]
|
D | caamalg.c | 2155 struct aead_request *req) in aead_unmap() argument 2157 struct crypto_aead *aead = crypto_aead_reqtfm(req); in aead_unmap() 2160 dma_unmap_sg_chained(dev, req->assoc, edesc->assoc_nents, in aead_unmap() 2163 caam_unmap(dev, req->src, req->dst, in aead_unmap() 2171 struct ablkcipher_request *req) in ablkcipher_unmap() argument 2173 struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req); in ablkcipher_unmap() 2176 caam_unmap(dev, req->src, req->dst, in ablkcipher_unmap() 2185 struct aead_request *req = context; in aead_encrypt_done() local 2188 struct crypto_aead *aead = crypto_aead_reqtfm(req); in aead_encrypt_done() 2201 aead_unmap(jrdev, edesc, req); in aead_encrypt_done() [all …]
|
/linux-4.1.27/arch/um/os-Linux/ |
D | aio.c | 132 static int do_not_aio(struct aio_thread_req *req) in do_not_aio() argument 138 actual = lseek64(req->io_fd, req->offset, SEEK_SET); in do_not_aio() 139 if (actual != req->offset) in do_not_aio() 142 switch (req->type) { in do_not_aio() 144 n = read(req->io_fd, req->buf, req->len); in do_not_aio() 147 n = write(req->io_fd, req->buf, req->len); in do_not_aio() 150 n = read(req->io_fd, &c, sizeof(c)); in do_not_aio() 154 req->type); in do_not_aio() 171 struct aio_thread_req req; in not_aio_thread() local 177 err = read(aio_req_fd_r, &req, sizeof(req)); in not_aio_thread() [all …]
|
/linux-4.1.27/net/core/ |
D | request_sock.c | 101 struct request_sock *req; in reqsk_queue_destroy() local 104 while ((req = lopt->syn_table[i]) != NULL) { in reqsk_queue_destroy() 105 lopt->syn_table[i] = req->dl_next; in reqsk_queue_destroy() 112 if (del_timer_sync(&req->rsk_timer)) in reqsk_queue_destroy() 113 reqsk_put(req); in reqsk_queue_destroy() 114 reqsk_put(req); in reqsk_queue_destroy() 171 void reqsk_fastopen_remove(struct sock *sk, struct request_sock *req, in reqsk_fastopen_remove() argument 174 struct sock *lsk = req->rsk_listener; in reqsk_fastopen_remove() 182 tcp_rsk(req)->tfo_listener = false; in reqsk_fastopen_remove() 183 if (req->sk) /* the child socket hasn't been accepted yet */ in reqsk_fastopen_remove() [all …]
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
D | c2_intr.c | 121 struct c2_vq_req *req; in handle_vq() local 166 req = (struct c2_vq_req *) (unsigned long) host_msg->context; in handle_vq() 167 if (req == NULL) { in handle_vq() 183 if (!err) switch (req->event) { in handle_vq() 185 c2_set_qp_state(req->qp, in handle_vq() 198 cm_event.event = req->event; in handle_vq() 200 cm_event.local_addr = req->cm_id->local_addr; in handle_vq() 201 cm_event.remote_addr = req->cm_id->remote_addr; in handle_vq() 204 req->cm_id->event_handler(req->cm_id, &cm_event); in handle_vq() 210 req->reply_msg = (u64) (unsigned long) (reply_msg); in handle_vq() [all …]
|
/linux-4.1.27/include/linux/ |
D | crypto.h | 147 typedef void (*crypto_completion_t)(struct crypto_async_request *req, int err); 284 int (*encrypt)(struct ablkcipher_request *req); 285 int (*decrypt)(struct ablkcipher_request *req); 286 int (*givencrypt)(struct skcipher_givcrypt_request *req); 287 int (*givdecrypt)(struct skcipher_givcrypt_request *req); 326 int (*encrypt)(struct aead_request *req); 327 int (*decrypt)(struct aead_request *req); 328 int (*givencrypt)(struct aead_givcrypt_request *req); 329 int (*givdecrypt)(struct aead_givcrypt_request *req); 594 int (*encrypt)(struct ablkcipher_request *req); [all …]
|
D | nfs_page.h | 112 #define NFS_WBACK_BUSY(req) (test_bit(PG_BUSY,&(req)->wb_flags)) argument 137 struct nfs_page *req); 139 extern void nfs_unlock_request(struct nfs_page *req); 150 nfs_lock_request(struct nfs_page *req) in nfs_lock_request() argument 152 return !test_and_set_bit(PG_BUSY, &req->wb_flags); in nfs_lock_request() 161 nfs_list_add_request(struct nfs_page *req, struct list_head *head) in nfs_list_add_request() argument 163 list_add_tail(&req->wb_list, head); in nfs_list_add_request() 172 nfs_list_remove_request(struct nfs_page *req) in nfs_list_remove_request() argument 174 if (list_empty(&req->wb_list)) in nfs_list_remove_request() 176 list_del_init(&req->wb_list); in nfs_list_remove_request() [all …]
|
/linux-4.1.27/drivers/iio/adc/ |
D | twl4030-madc.c | 74 struct twl4030_madc_request req; in twl4030_madc_read() local 77 req.method = madc->use_second_irq ? TWL4030_MADC_SW2 : TWL4030_MADC_SW1; in twl4030_madc_read() 79 req.channels = BIT(chan->channel); in twl4030_madc_read() 80 req.active = false; in twl4030_madc_read() 81 req.func_cb = NULL; in twl4030_madc_read() 82 req.type = TWL4030_MADC_WAIT; in twl4030_madc_read() 83 req.raw = !(mask == IIO_CHAN_INFO_PROCESSED); in twl4030_madc_read() 84 req.do_avg = (mask == IIO_CHAN_INFO_AVERAGE_RAW); in twl4030_madc_read() 86 ret = twl4030_madc_conversion(&req); in twl4030_madc_read() 90 *val = req.rbuf[chan->channel]; in twl4030_madc_read() [all …]
|
/linux-4.1.27/drivers/infiniband/hw/mlx4/ |
D | mcg.c | 539 struct mcast_req *req = NULL; in mlx4_ib_mcg_timeout_handler() local 546 req = list_first_entry(&group->pending_list, struct mcast_req, group_list); in mlx4_ib_mcg_timeout_handler() 547 list_del(&req->group_list); in mlx4_ib_mcg_timeout_handler() 548 list_del(&req->func_list); in mlx4_ib_mcg_timeout_handler() 549 --group->func[req->func].num_pend_reqs; in mlx4_ib_mcg_timeout_handler() 551 kfree(req); in mlx4_ib_mcg_timeout_handler() 581 struct mcast_req *req) in handle_leave_req() argument 585 if (req->clean) in handle_leave_req() 586 leave_mask = group->func[req->func].join_state; in handle_leave_req() 588 status = check_leave(group, req->func, leave_mask); in handle_leave_req() [all …]
|
/linux-4.1.27/arch/arm64/lib/ |
D | strlen.S | 38 srcin .req x0 39 len .req x0 42 src .req x1 43 data1 .req x2 44 data2 .req x3 45 data2a .req x4 46 has_nul1 .req x5 47 has_nul2 .req x6 48 tmp1 .req x7 49 tmp2 .req x8 [all …]
|
D | strnlen.S | 39 srcin .req x0 40 len .req x0 41 limit .req x1 44 src .req x2 45 data1 .req x3 46 data2 .req x4 47 data2a .req x5 48 has_nul1 .req x6 49 has_nul2 .req x7 50 tmp1 .req x8 [all …]
|
D | memcpy.S | 39 dstin .req x0 40 src .req x1 41 count .req x2 42 tmp1 .req x3 43 tmp1w .req w3 44 tmp2 .req x4 45 tmp2w .req w4 46 tmp3 .req x5 47 tmp3w .req w5 48 dst .req x6 [all …]
|
D | memmove.S | 40 dstin .req x0 41 src .req x1 42 count .req x2 43 tmp1 .req x3 44 tmp1w .req w3 45 tmp2 .req x4 46 tmp2w .req w4 47 tmp3 .req x5 48 tmp3w .req w5 49 dst .req x6 [all …]
|
/linux-4.1.27/drivers/scsi/ |
D | scsi_lib.c | 220 struct request *req; in scsi_execute() local 224 req = blk_get_request(sdev->request_queue, write, __GFP_WAIT); in scsi_execute() 225 if (IS_ERR(req)) in scsi_execute() 227 blk_rq_set_block_pc(req); in scsi_execute() 229 if (bufflen && blk_rq_map_kern(sdev->request_queue, req, in scsi_execute() 233 req->cmd_len = COMMAND_SIZE(cmd[0]); in scsi_execute() 234 memcpy(req->cmd, cmd, req->cmd_len); in scsi_execute() 235 req->sense = sense; in scsi_execute() 236 req->sense_len = 0; in scsi_execute() 237 req->retries = retries; in scsi_execute() [all …]
|
/linux-4.1.27/drivers/memstick/host/ |
D | tifm_ms.c | 74 struct memstick_request *req; member 192 if (host->req->long_data) { in tifm_ms_transfer_data() 193 length = host->req->sg.length - host->block_pos; in tifm_ms_transfer_data() 194 off = host->req->sg.offset + host->block_pos; in tifm_ms_transfer_data() 196 length = host->req->data_len - host->block_pos; in tifm_ms_transfer_data() 205 if (host->req->long_data) { in tifm_ms_transfer_data() 206 pg = nth_page(sg_page(&host->req->sg), in tifm_ms_transfer_data() 215 buf = host->req->data + host->block_pos; in tifm_ms_transfer_data() 216 p_cnt = host->req->data_len - host->block_pos; in tifm_ms_transfer_data() 219 t_size = host->req->data_dir == WRITE in tifm_ms_transfer_data() [all …]
|
/linux-4.1.27/drivers/net/fddi/skfp/ |
D | ess.c | 555 struct smt_sba_alc_req *req ; in ess_send_alc_req() local 586 req = smtod(mb,struct smt_sba_alc_req *) ; in ess_send_alc_req() 587 req->smt.smt_tid = smc->ess.alloc_trans_id = smt_get_tid(smc) ; in ess_send_alc_req() 588 req->smt.smt_dest = smt_sba_da ; in ess_send_alc_req() 591 req->s_type.para.p_type = SMT_P0015 ; in ess_send_alc_req() 592 req->s_type.para.p_len = sizeof(struct smt_p_0015) - PARA_LEN ; in ess_send_alc_req() 593 req->s_type.res_type = SYNC_BW ; in ess_send_alc_req() 596 req->cmd.para.p_type = SMT_P0016 ; in ess_send_alc_req() 597 req->cmd.para.p_len = sizeof(struct smt_p_0016) - PARA_LEN ; in ess_send_alc_req() 598 req->cmd.sba_cmd = REQUEST_ALLOCATION ; in ess_send_alc_req() [all …]
|
/linux-4.1.27/include/trace/events/ |
D | hswadsp.h | 245 struct sst_hsw_audio_data_format_ipc *req), 247 TP_ARGS(stream, req), 262 __entry->frequency = req->frequency; 263 __entry->bitdepth = req->bitdepth; 264 __entry->map = req->map; 265 __entry->config = req->config; 266 __entry->style = req->style; 267 __entry->ch_num = req->ch_num; 268 __entry->valid_bit = req->valid_bit; 281 struct sst_hsw_ipc_stream_alloc_req *req), [all …]
|
/linux-4.1.27/drivers/infiniband/hw/cxgb3/ |
D | iwch_cm.c | 173 struct cpl_tid_release *req; in release_tid() local 175 skb = get_skb(skb, sizeof *req, GFP_KERNEL); in release_tid() 178 req = (struct cpl_tid_release *) skb_put(skb, sizeof(*req)); in release_tid() 179 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in release_tid() 180 OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_TID_RELEASE, hwtid)); in release_tid() 188 struct cpl_set_tcb_field *req; in iwch_quiesce_tid() local 189 struct sk_buff *skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); in iwch_quiesce_tid() 193 req = (struct cpl_set_tcb_field *) skb_put(skb, sizeof(*req)); in iwch_quiesce_tid() 194 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in iwch_quiesce_tid() 195 req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid)); in iwch_quiesce_tid() [all …]
|
/linux-4.1.27/drivers/media/usb/as102/ |
D | as10x_cmd_stream.c | 38 sizeof(pcmd->body.add_pid_filter.req)); in as10x_cmd_add_PID_filter() 41 pcmd->body.add_pid_filter.req.proc_id = in as10x_cmd_add_PID_filter() 43 pcmd->body.add_pid_filter.req.pid = cpu_to_le16(filter->pid); in as10x_cmd_add_PID_filter() 44 pcmd->body.add_pid_filter.req.stream_type = filter->type; in as10x_cmd_add_PID_filter() 47 pcmd->body.add_pid_filter.req.idx = filter->idx; in as10x_cmd_add_PID_filter() 49 pcmd->body.add_pid_filter.req.idx = 0xFF; in as10x_cmd_add_PID_filter() 54 sizeof(pcmd->body.add_pid_filter.req) in as10x_cmd_add_PID_filter() 95 sizeof(pcmd->body.del_pid_filter.req)); in as10x_cmd_del_PID_filter() 98 pcmd->body.del_pid_filter.req.proc_id = in as10x_cmd_del_PID_filter() 100 pcmd->body.del_pid_filter.req.pid = cpu_to_le16(pid_value); in as10x_cmd_del_PID_filter() [all …]
|
D | as10x_cmd_cfg.c | 43 sizeof(pcmd->body.context.req)); in as10x_cmd_get_context() 46 pcmd->body.context.req.proc_id = cpu_to_le16(CONTROL_PROC_CONTEXT); in as10x_cmd_get_context() 47 pcmd->body.context.req.tag = cpu_to_le16(tag); in as10x_cmd_get_context() 48 pcmd->body.context.req.type = cpu_to_le16(GET_CONTEXT_DATA); in as10x_cmd_get_context() 54 sizeof(pcmd->body.context.req) in as10x_cmd_get_context() 99 sizeof(pcmd->body.context.req)); in as10x_cmd_set_context() 102 pcmd->body.context.req.proc_id = cpu_to_le16(CONTROL_PROC_CONTEXT); in as10x_cmd_set_context() 104 pcmd->body.context.req.reg_val.u.value32 = (__force u32)cpu_to_le32(value); in as10x_cmd_set_context() 105 pcmd->body.context.req.tag = cpu_to_le16(tag); in as10x_cmd_set_context() 106 pcmd->body.context.req.type = cpu_to_le16(SET_CONTEXT_DATA); in as10x_cmd_set_context() [all …]
|
/linux-4.1.27/arch/x86/crypto/ |
D | ghash-clmulni-intel_glue.c | 165 static int ghash_async_init(struct ahash_request *req) in ghash_async_init() argument 167 struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); in ghash_async_init() 169 struct ahash_request *cryptd_req = ahash_request_ctx(req); in ghash_async_init() 173 memcpy(cryptd_req, req, sizeof(*req)); in ghash_async_init() 181 desc->flags = req->base.flags; in ghash_async_init() 186 static int ghash_async_update(struct ahash_request *req) in ghash_async_update() argument 188 struct ahash_request *cryptd_req = ahash_request_ctx(req); in ghash_async_update() 191 struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); in ghash_async_update() 195 memcpy(cryptd_req, req, sizeof(*req)); in ghash_async_update() 200 return shash_ahash_update(req, desc); in ghash_async_update() [all …]
|
/linux-4.1.27/drivers/usb/chipidea/ |
D | udc.c | 364 u32 mul = hwreq->req.length / hwep->ep.maxpacket; in add_td_to_list() 366 if (hwreq->req.length == 0 in add_td_to_list() 367 || hwreq->req.length % hwep->ep.maxpacket) in add_td_to_list() 372 temp = (u32) (hwreq->req.dma + hwreq->req.actual); in add_td_to_list() 382 hwreq->req.actual += length; in add_td_to_list() 417 unsigned rest = hwreq->req.length; in _hardware_enqueue() 422 if (hwreq->req.status == -EALREADY) in _hardware_enqueue() 425 hwreq->req.status = -EALREADY; in _hardware_enqueue() 427 ret = usb_gadget_map_request(&ci->gadget, &hwreq->req, hwep->dir); in _hardware_enqueue() 435 if (hwreq->req.dma % PAGE_SIZE) in _hardware_enqueue() [all …]
|
/linux-4.1.27/drivers/media/usb/dvb-usb/ |
D | az6027.c | 301 static int az6027_usb_in_op(struct dvb_usb_device *d, u8 req, in az6027_usb_in_op() argument 310 req, in az6027_usb_in_op() 324 deb_xfer("in: req. %02x, val: %04x, ind: %04x, buffer: ", req, value, index); in az6027_usb_in_op() 332 u8 req, in az6027_usb_out_op() argument 340 deb_xfer("out: req. %02x, val: %04x, ind: %04x, buffer: ", req, value, index); in az6027_usb_out_op() 348 req, in az6027_usb_out_op() 369 u8 req; in az6027_streaming_ctrl() local 376 req = 0xBC; in az6027_streaming_ctrl() 381 ret = az6027_usb_out_op(adap->dev, req, value, index, NULL, blen); in az6027_streaming_ctrl() 416 u8 req; in az6027_ci_read_attribute_mem() local [all …]
|
/linux-4.1.27/drivers/net/wireless/ath/wil6210/ |
D | rx_reorder.c | 257 struct wil_back_rx *req = kzalloc(sizeof(*req), GFP_KERNEL); in wil_addba_rx_request() local 259 if (!req) in wil_addba_rx_request() 262 req->cidxtid = cidxtid; in wil_addba_rx_request() 263 req->dialog_token = dialog_token; in wil_addba_rx_request() 264 req->ba_param_set = le16_to_cpu(ba_param_set); in wil_addba_rx_request() 265 req->ba_timeout = le16_to_cpu(ba_timeout); in wil_addba_rx_request() 266 req->ba_seq_ctrl = le16_to_cpu(ba_seq_ctrl); in wil_addba_rx_request() 269 list_add_tail(&req->list, &wil->back_rx_pending); in wil_addba_rx_request() 278 struct wil_back_rx *req) in wil_back_rx_handle() argument 289 u16 req_agg_wsize = WIL_GET_BITS(req->ba_param_set, 6, 15); in wil_back_rx_handle() [all …]
|
/linux-4.1.27/include/crypto/internal/ |
D | skcipher.h | 58 int skcipher_null_givencrypt(struct skcipher_givcrypt_request *req); 59 int skcipher_null_givdecrypt(struct skcipher_givcrypt_request *req); 88 struct skcipher_givcrypt_request *req) in skcipher_givcrypt_reqctx() argument 90 return ablkcipher_request_ctx(&req->creq); in skcipher_givcrypt_reqctx() 93 static inline void ablkcipher_request_complete(struct ablkcipher_request *req, in ablkcipher_request_complete() argument 96 req->base.complete(&req->base, err); in ablkcipher_request_complete() 100 struct skcipher_givcrypt_request *req, int err) in skcipher_givcrypt_complete() argument 102 ablkcipher_request_complete(&req->creq, err); in skcipher_givcrypt_complete() 105 static inline u32 ablkcipher_request_flags(struct ablkcipher_request *req) in ablkcipher_request_flags() argument 107 return req->base.flags; in ablkcipher_request_flags()
|
/linux-4.1.27/net/unix/ |
D | diag.c | 74 struct sock *req, *peer; in sk_diag_dump_icons() local 76 req = skb->sk; in sk_diag_dump_icons() 82 unix_state_lock_nested(req); in sk_diag_dump_icons() 83 peer = unix_sk(req)->peer; in sk_diag_dump_icons() 85 unix_state_unlock(req); in sk_diag_dump_icons() 112 static int sk_diag_fill(struct sock *sk, struct sk_buff *skb, struct unix_diag_req *req, in sk_diag_fill() argument 131 if ((req->udiag_show & UDIAG_SHOW_NAME) && in sk_diag_fill() 135 if ((req->udiag_show & UDIAG_SHOW_VFS) && in sk_diag_fill() 139 if ((req->udiag_show & UDIAG_SHOW_PEER) && in sk_diag_fill() 143 if ((req->udiag_show & UDIAG_SHOW_ICONS) && in sk_diag_fill() [all …]
|