/linux-4.1.27/drivers/dma/ |
H A D | sirf-dma.c | 117 struct sirfsoc_dma_desc *sdesc = NULL; sirfsoc_dma_execute() local 124 sdesc = list_first_entry(&schan->queued, struct sirfsoc_dma_desc, sirfsoc_dma_execute() 127 list_move_tail(&sdesc->node, &schan->active); sirfsoc_dma_execute() 130 writel_relaxed(sdesc->width, sdma->base + SIRFSOC_DMA_WIDTH_0 + sirfsoc_dma_execute() 133 (sdesc->dir << SIRFSOC_DMA_DIR_CTRL_BIT), sirfsoc_dma_execute() 135 writel_relaxed(sdesc->xlen, sdma->base + cid * 0x10 + sirfsoc_dma_execute() 137 writel_relaxed(sdesc->ylen, sdma->base + cid * 0x10 + sirfsoc_dma_execute() 146 writel(sdesc->addr >> 2, sdma->base + cid * 0x10 + SIRFSOC_DMA_CH_ADDR); sirfsoc_dma_execute() 148 if (sdesc->cyclic) { sirfsoc_dma_execute() 161 struct sirfsoc_dma_desc *sdesc = NULL; sirfsoc_dma_irq() local 173 sdesc = list_first_entry(&schan->active, struct sirfsoc_dma_desc, sirfsoc_dma_irq() 175 if (!sdesc->cyclic) { sirfsoc_dma_irq() 197 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_process_completed() local 214 list_for_each_entry(sdesc, &list, node) { sirfsoc_dma_process_completed() 215 desc = &sdesc->desc; sirfsoc_dma_process_completed() 231 sdesc = list_first_entry(&schan->active, struct sirfsoc_dma_desc, sirfsoc_dma_process_completed() 234 if (!sdesc || (sdesc && !sdesc->cyclic)) { sirfsoc_dma_process_completed() 244 desc = &sdesc->desc; sirfsoc_dma_process_completed() 266 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_tx_submit() local 270 sdesc = container_of(txd, struct sirfsoc_dma_desc, desc); sirfsoc_dma_tx_submit() 275 list_move_tail(&sdesc->node, &schan->queued); sirfsoc_dma_tx_submit() 381 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_alloc_chan_resources() local 390 sdesc = kzalloc(sizeof(*sdesc), GFP_KERNEL); sirfsoc_dma_alloc_chan_resources() 391 if (!sdesc) { sirfsoc_dma_alloc_chan_resources() 397 dma_async_tx_descriptor_init(&sdesc->desc, chan); sirfsoc_dma_alloc_chan_resources() 398 sdesc->desc.flags = DMA_CTRL_ACK; sirfsoc_dma_alloc_chan_resources() 399 sdesc->desc.tx_submit = sirfsoc_dma_tx_submit; sirfsoc_dma_alloc_chan_resources() 401 list_add_tail(&sdesc->node, &descs); sirfsoc_dma_alloc_chan_resources() 421 struct sirfsoc_dma_desc *sdesc, *tmp; sirfsoc_dma_free_chan_resources() local 439 list_for_each_entry_safe(sdesc, tmp, &descs, node) sirfsoc_dma_free_chan_resources() 440 kfree(sdesc); sirfsoc_dma_free_chan_resources() 468 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_tx_status() local 476 sdesc = list_first_entry(&schan->active, struct sirfsoc_dma_desc, sirfsoc_dma_tx_status() 478 dma_request_bytes = (sdesc->xlen + 1) * (sdesc->ylen + 1) * sirfsoc_dma_tx_status() 479 (sdesc->width * SIRFSOC_DMA_WORD_LEN); sirfsoc_dma_tx_status() 484 residue = dma_request_bytes - (dma_pos - sdesc->addr); sirfsoc_dma_tx_status() 498 struct sirfsoc_dma_desc *sdesc = NULL; sirfsoc_dma_prep_interleaved() local 510 sdesc = list_first_entry(&schan->free, struct sirfsoc_dma_desc, sirfsoc_dma_prep_interleaved() 512 list_del(&sdesc->node); sirfsoc_dma_prep_interleaved() 516 if (!sdesc) { sirfsoc_dma_prep_interleaved() 531 sdesc->cyclic = 0; sirfsoc_dma_prep_interleaved() 532 sdesc->xlen = xt->sgl[0].size / SIRFSOC_DMA_WORD_LEN; sirfsoc_dma_prep_interleaved() 533 sdesc->width = (xt->sgl[0].size + xt->sgl[0].icg) / sirfsoc_dma_prep_interleaved() 535 sdesc->ylen = xt->numf - 1; sirfsoc_dma_prep_interleaved() 537 sdesc->addr = xt->src_start; sirfsoc_dma_prep_interleaved() 538 sdesc->dir = 1; sirfsoc_dma_prep_interleaved() 540 sdesc->addr = xt->dst_start; sirfsoc_dma_prep_interleaved() 541 sdesc->dir = 0; sirfsoc_dma_prep_interleaved() 544 list_add_tail(&sdesc->node, &schan->prepared); sirfsoc_dma_prep_interleaved() 552 return &sdesc->desc; sirfsoc_dma_prep_interleaved() 566 struct sirfsoc_dma_desc *sdesc = NULL; sirfsoc_dma_prep_cyclic() local 586 sdesc = list_first_entry(&schan->free, struct sirfsoc_dma_desc, sirfsoc_dma_prep_cyclic() 588 list_del(&sdesc->node); sirfsoc_dma_prep_cyclic() 592 if (!sdesc) sirfsoc_dma_prep_cyclic() 597 sdesc->addr = addr; sirfsoc_dma_prep_cyclic() 598 sdesc->cyclic = 1; sirfsoc_dma_prep_cyclic() 599 sdesc->xlen = 0; sirfsoc_dma_prep_cyclic() 600 sdesc->ylen = buf_len / SIRFSOC_DMA_WORD_LEN - 1; sirfsoc_dma_prep_cyclic() 601 sdesc->width = 1; sirfsoc_dma_prep_cyclic() 602 list_add_tail(&sdesc->node, &schan->prepared); sirfsoc_dma_prep_cyclic() 605 return &sdesc->desc; sirfsoc_dma_prep_cyclic() 815 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_pm_suspend() local 838 sdesc = list_first_entry(&schan->active, sirfsoc_dma_pm_suspend() 856 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_pm_resume() local 871 sdesc = list_first_entry(&schan->active, sirfsoc_dma_pm_resume() 874 writel_relaxed(sdesc->width, sirfsoc_dma_pm_resume() 876 writel_relaxed(sdesc->xlen, sirfsoc_dma_pm_resume() 878 writel_relaxed(sdesc->ylen, sirfsoc_dma_pm_resume() 882 writel_relaxed(sdesc->addr >> 2, sirfsoc_dma_pm_resume()
|
/linux-4.1.27/security/keys/ |
H A D | trusted.c | 37 struct sdesc { struct 45 static struct sdesc *init_sdesc(struct crypto_shash *alg) init_sdesc() 47 struct sdesc *sdesc; init_sdesc() local 51 sdesc = kmalloc(size, GFP_KERNEL); init_sdesc() 52 if (!sdesc) init_sdesc() 54 sdesc->shash.tfm = alg; init_sdesc() 55 sdesc->shash.flags = 0x0; init_sdesc() 56 return sdesc; init_sdesc() 62 struct sdesc *sdesc; TSS_sha1() local 65 sdesc = init_sdesc(hashalg); TSS_sha1() 66 if (IS_ERR(sdesc)) { TSS_sha1() 68 return PTR_ERR(sdesc); TSS_sha1() 71 ret = crypto_shash_digest(&sdesc->shash, data, datalen, digest); TSS_sha1() 72 kfree(sdesc); TSS_sha1() 79 struct sdesc *sdesc; TSS_rawhmac() local 85 sdesc = init_sdesc(hmacalg); TSS_rawhmac() 86 if (IS_ERR(sdesc)) { TSS_rawhmac() 88 return PTR_ERR(sdesc); TSS_rawhmac() 94 ret = crypto_shash_init(&sdesc->shash); TSS_rawhmac() 108 ret = crypto_shash_update(&sdesc->shash, data, dlen); TSS_rawhmac() 114 ret = crypto_shash_final(&sdesc->shash, digest); TSS_rawhmac() 116 kfree(sdesc); TSS_rawhmac() 128 struct sdesc *sdesc; TSS_authhmac() local 135 sdesc = init_sdesc(hashalg); TSS_authhmac() 136 if (IS_ERR(sdesc)) { TSS_authhmac() 138 return PTR_ERR(sdesc); TSS_authhmac() 142 ret = crypto_shash_init(&sdesc->shash); TSS_authhmac() 155 ret = crypto_shash_update(&sdesc->shash, data, dlen); TSS_authhmac() 161 ret = crypto_shash_final(&sdesc->shash, paramdigest); TSS_authhmac() 167 kfree(sdesc); TSS_authhmac() 189 struct sdesc *sdesc; TSS_checkhmac1() local 207 sdesc = init_sdesc(hashalg); TSS_checkhmac1() 208 if (IS_ERR(sdesc)) { TSS_checkhmac1() 210 return PTR_ERR(sdesc); TSS_checkhmac1() 212 ret = crypto_shash_init(&sdesc->shash); TSS_checkhmac1() 215 ret = crypto_shash_update(&sdesc->shash, (const u8 *)&result, TSS_checkhmac1() 219 ret = crypto_shash_update(&sdesc->shash, (const u8 *)&ordinal, TSS_checkhmac1() 229 ret = crypto_shash_update(&sdesc->shash, buffer + dpos, dlen); TSS_checkhmac1() 235 ret = crypto_shash_final(&sdesc->shash, paramdigest); TSS_checkhmac1() 248 kfree(sdesc); TSS_checkhmac1() 276 struct sdesc *sdesc; TSS_checkhmac2() local 299 sdesc = init_sdesc(hashalg); TSS_checkhmac2() 300 if (IS_ERR(sdesc)) { TSS_checkhmac2() 302 return PTR_ERR(sdesc); TSS_checkhmac2() 304 ret = crypto_shash_init(&sdesc->shash); TSS_checkhmac2() 307 ret = crypto_shash_update(&sdesc->shash, (const u8 *)&result, TSS_checkhmac2() 311 ret = crypto_shash_update(&sdesc->shash, (const u8 *)&ordinal, TSS_checkhmac2() 322 ret = crypto_shash_update(&sdesc->shash, buffer + dpos, dlen); TSS_checkhmac2() 328 ret = crypto_shash_final(&sdesc->shash, paramdigest); TSS_checkhmac2() 349 kfree(sdesc); TSS_checkhmac2()
|
/linux-4.1.27/drivers/dma/sh/ |
H A D | sudmac.c | 53 #define to_desc(sdesc) container_of(sdesc, struct sudmac_desc, shdma_desc) 103 struct shdma_desc *sdesc) sudmac_set_reg() 119 struct shdma_desc *sdesc) sudmac_start_xfer() 122 struct sudmac_desc *sd = to_desc(sdesc); sudmac_start_xfer() 124 sudmac_set_reg(sc, &sd->hw, sdesc); sudmac_start_xfer() 176 struct shdma_desc *sdesc, sudmac_desc_setup() 180 struct sudmac_desc *sd = to_desc(sdesc); sudmac_desc_setup() 219 struct shdma_desc *sdesc) sudmac_get_partial() 222 struct sudmac_desc *sd = to_desc(sdesc); sudmac_get_partial() 229 struct shdma_desc *sdesc) sudmac_desc_completed() 232 struct sudmac_desc *sd = to_desc(sdesc); sudmac_desc_completed() 102 sudmac_set_reg(struct sudmac_chan *sc, struct sudmac_regs *hw, struct shdma_desc *sdesc) sudmac_set_reg() argument 118 sudmac_start_xfer(struct shdma_chan *schan, struct shdma_desc *sdesc) sudmac_start_xfer() argument 175 sudmac_desc_setup(struct shdma_chan *schan, struct shdma_desc *sdesc, dma_addr_t src, dma_addr_t dst, size_t *len) sudmac_desc_setup() argument 218 sudmac_get_partial(struct shdma_chan *schan, struct shdma_desc *sdesc) sudmac_get_partial() argument 228 sudmac_desc_completed(struct shdma_chan *schan, struct shdma_desc *sdesc) sudmac_desc_completed() argument
|
H A D | shdma-base.c | 59 struct shdma_desc *sdesc; shdma_chan_xfer_ld_queue() local 66 list_for_each_entry(sdesc, &schan->ld_queue, node) shdma_chan_xfer_ld_queue() 67 if (sdesc->mark == DESC_SUBMITTED) { shdma_chan_xfer_ld_queue() 68 ops->start_xfer(schan, sdesc); shdma_chan_xfer_ld_queue() 162 struct shdma_desc *sdesc; shdma_get_desc() local 164 list_for_each_entry(sdesc, &schan->ld_free, node) shdma_get_desc() 165 if (sdesc->mark != DESC_PREPARED) { shdma_get_desc() 166 BUG_ON(sdesc->mark != DESC_IDLE); shdma_get_desc() 167 list_del(&sdesc->node); shdma_get_desc() 168 return sdesc; shdma_get_desc() 840 struct shdma_desc *sdesc; shdma_tx_status() local 842 list_for_each_entry(sdesc, &schan->ld_queue, node) shdma_tx_status() 843 if (sdesc->cookie == cookie) { shdma_tx_status() 864 struct shdma_desc *sdesc; shdma_for_each_chan() local 886 list_for_each_entry(sdesc, &dl, node) { shdma_for_each_chan() 887 struct dma_async_tx_descriptor *tx = &sdesc->async_tx; shdma_for_each_chan() 888 sdesc->mark = DESC_IDLE; shdma_for_each_chan() 925 struct shdma_desc *sdesc; chan_irqt() local 928 list_for_each_entry(sdesc, &schan->ld_queue, node) { chan_irqt() 929 if (sdesc->mark == DESC_SUBMITTED && chan_irqt() 930 ops->desc_completed(schan, sdesc)) { chan_irqt() 932 sdesc->async_tx.cookie, &sdesc->async_tx); chan_irqt() 933 sdesc->mark = DESC_COMPLETED; chan_irqt()
|
H A D | rcar-hpbdma.c | 124 #define to_desc(sdesc) container_of(sdesc, struct hpb_desc, shdma_desc) 314 struct shdma_desc *sdesc) hpb_dmae_start_xfer() 318 struct hpb_desc *desc = to_desc(sdesc); hpb_dmae_start_xfer() 334 struct shdma_desc *sdesc) hpb_dmae_desc_completed() 362 struct shdma_desc *sdesc, hpb_dmae_desc_setup() 365 struct hpb_desc *desc = to_desc(sdesc); hpb_dmae_desc_setup() 378 struct shdma_desc *sdesc) hpb_dmae_get_partial() 380 struct hpb_desc *desc = to_desc(sdesc); hpb_dmae_get_partial() 313 hpb_dmae_start_xfer(struct shdma_chan *schan, struct shdma_desc *sdesc) hpb_dmae_start_xfer() argument 333 hpb_dmae_desc_completed(struct shdma_chan *schan, struct shdma_desc *sdesc) hpb_dmae_desc_completed() argument 361 hpb_dmae_desc_setup(struct shdma_chan *schan, struct shdma_desc *sdesc, dma_addr_t src, dma_addr_t dst, size_t *len) hpb_dmae_desc_setup() argument 377 hpb_dmae_get_partial(struct shdma_chan *schan, struct shdma_desc *sdesc) hpb_dmae_get_partial() argument
|
H A D | shdmac.c | 287 struct shdma_desc *sdesc) sh_dmae_start_xfer() 291 struct sh_dmae_desc *sh_desc = container_of(sdesc, sh_dmae_start_xfer() 294 sdesc->async_tx.cookie, sh_chan->shdma_chan.id, sh_dmae_start_xfer() 382 struct shdma_desc *sdesc, sh_dmae_desc_setup() 385 struct sh_dmae_desc *sh_desc = container_of(sdesc, sh_dmae_desc_setup() 420 struct shdma_desc *sdesc) sh_dmae_get_partial() 424 struct sh_dmae_desc *sh_desc = container_of(sdesc, sh_dmae_get_partial() 460 struct shdma_desc *sdesc) sh_dmae_desc_completed() 464 struct sh_dmae_desc *sh_desc = container_of(sdesc, sh_dmae_desc_completed() 469 return (sdesc->direction == DMA_DEV_TO_MEM && sh_dmae_desc_completed() 471 (sdesc->direction != DMA_DEV_TO_MEM && sh_dmae_desc_completed() 286 sh_dmae_start_xfer(struct shdma_chan *schan, struct shdma_desc *sdesc) sh_dmae_start_xfer() argument 381 sh_dmae_desc_setup(struct shdma_chan *schan, struct shdma_desc *sdesc, dma_addr_t src, dma_addr_t dst, size_t *len) sh_dmae_desc_setup() argument 419 sh_dmae_get_partial(struct shdma_chan *schan, struct shdma_desc *sdesc) sh_dmae_get_partial() argument 459 sh_dmae_desc_completed(struct shdma_chan *schan, struct shdma_desc *sdesc) sh_dmae_desc_completed() argument
|
/linux-4.1.27/include/media/ |
H A D | soc_camera_platform.h | 46 if (&icd->sdesc->subdev_desc != ssdd) soc_camera_platform_add() 77 if (&icd->sdesc->subdev_desc != ssdd || !pdev) soc_camera_platform_del()
|
H A D | soc_camera.h | 32 struct soc_camera_desc *sdesc; member in struct:soc_camera_device 266 return icd->sdesc; to_soc_camera_desc()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | sge.c | 190 * @sdesc: pointer to the software buffer descriptor 195 static inline dma_addr_t get_buf_addr(const struct rx_sw_desc *sdesc) get_buf_addr() argument 197 return sdesc->dma_addr & ~(dma_addr_t)(RX_LARGE_BUF | RX_UNMAPPED_BUF); get_buf_addr() 202 * @sdesc: pointer to the software buffer descriptor 207 static inline bool is_buf_mapped(const struct rx_sw_desc *sdesc) is_buf_mapped() argument 209 return !(sdesc->dma_addr & RX_UNMAPPED_BUF); is_buf_mapped() 380 struct tx_sw_desc *sdesc; free_tx_desc() local 386 sdesc = &tq->sdesc[cidx]; free_tx_desc() 392 if (sdesc->skb) { free_tx_desc() 394 unmap_sgl(dev, sdesc->skb, sdesc->sgl, tq); free_tx_desc() 395 dev_consume_skb_any(sdesc->skb); free_tx_desc() 396 sdesc->skb = NULL; free_tx_desc() 399 sdesc++; free_tx_desc() 402 sdesc = tq->sdesc; free_tx_desc() 452 * @sdesc: pointer to the software buffer descriptor 455 const struct rx_sw_desc *sdesc) get_buf_size() 459 return (s->fl_pg_order > 0 && (sdesc->dma_addr & RX_LARGE_BUF) get_buf_size() 476 struct rx_sw_desc *sdesc = &fl->sdesc[fl->cidx]; free_rx_bufs() local 478 if (is_buf_mapped(sdesc)) free_rx_bufs() 479 dma_unmap_page(adapter->pdev_dev, get_buf_addr(sdesc), free_rx_bufs() 480 get_buf_size(adapter, sdesc), free_rx_bufs() 482 put_page(sdesc->page); free_rx_bufs() 483 sdesc->page = NULL; free_rx_bufs() 505 struct rx_sw_desc *sdesc = &fl->sdesc[fl->cidx]; unmap_rx_buf() local 507 if (is_buf_mapped(sdesc)) unmap_rx_buf() 508 dma_unmap_page(adapter->pdev_dev, get_buf_addr(sdesc), unmap_rx_buf() 509 get_buf_size(adapter, sdesc), unmap_rx_buf() 511 sdesc->page = NULL; unmap_rx_buf() 569 * @sdesc: pointer to the softwore RX buffer descriptor 573 static inline void set_rx_sw_desc(struct rx_sw_desc *sdesc, struct page *page, set_rx_sw_desc() argument 576 sdesc->page = page; set_rx_sw_desc() 577 sdesc->dma_addr = dma_addr; set_rx_sw_desc() 614 struct rx_sw_desc *sdesc = &fl->sdesc[fl->pidx]; refill_fl() local 665 set_rx_sw_desc(sdesc, page, dma_addr); refill_fl() 666 sdesc++; refill_fl() 671 sdesc = fl->sdesc; refill_fl() 694 set_rx_sw_desc(sdesc, page, dma_addr); refill_fl() 695 sdesc++; refill_fl() 700 sdesc = fl->sdesc; refill_fl() 1425 tq->sdesc[last_desc].skb = skb; t4vf_eth_xmit() 1426 tq->sdesc[last_desc].sgl = sgl; t4vf_eth_xmit() 1692 struct rx_sw_desc *sdesc; restore_rx_bufs() local 1699 sdesc = &fl->sdesc[fl->cidx]; restore_rx_bufs() 1700 sdesc->page = gl->frags[frags].page; restore_rx_bufs() 1701 sdesc->dma_addr |= RX_UNMAPPED_BUF; restore_rx_bufs() 1759 const struct rx_sw_desc *sdesc; process_responses() local 1788 sdesc = &rxq->fl.sdesc[rxq->fl.cidx]; process_responses() 1789 bufsz = get_buf_size(adapter, sdesc); process_responses() 1790 fp->page = sdesc->page; process_responses() 1806 get_buf_addr(sdesc), process_responses() 2251 &fl->addr, &fl->sdesc, s->stat_len); t4vf_sge_alloc_rxq() 2343 kfree(fl->sdesc); t4vf_sge_alloc_rxq() 2344 fl->sdesc = NULL; t4vf_sge_alloc_rxq() 2382 &txq->q.phys_addr, &txq->q.sdesc, s->stat_len); t4vf_sge_alloc_eth_txq() 2425 kfree(txq->q.sdesc); t4vf_sge_alloc_eth_txq() 2426 txq->q.sdesc = NULL; t4vf_sge_alloc_eth_txq() 2466 tq->sdesc = NULL; free_txq() 2495 kfree(fl->sdesc); free_rspq_fl() 2496 fl->sdesc = NULL; free_rspq_fl() 2523 kfree(txq->q.sdesc); t4vf_free_sge_resources() 454 get_buf_size(const struct adapter *adapter, const struct rx_sw_desc *sdesc) get_buf_size() argument
|
H A D | adapter.h | 141 struct rx_sw_desc *sdesc; /* address of SW RX descriptor ring */ member in struct:sge_fl 247 struct tx_sw_desc *sdesc; /* address of SW TX descriptor ring */ member in struct:sge_txq
|
/linux-4.1.27/security/keys/encrypted-keys/ |
H A D | encrypted.c | 57 struct sdesc { struct 324 static struct sdesc *alloc_sdesc(struct crypto_shash *alg) alloc_sdesc() 326 struct sdesc *sdesc; alloc_sdesc() local 330 sdesc = kmalloc(size, GFP_KERNEL); alloc_sdesc() 331 if (!sdesc) alloc_sdesc() 333 sdesc->shash.tfm = alg; alloc_sdesc() 334 sdesc->shash.flags = 0x0; alloc_sdesc() 335 return sdesc; alloc_sdesc() 341 struct sdesc *sdesc; calc_hmac() local 344 sdesc = alloc_sdesc(hmacalg); calc_hmac() 345 if (IS_ERR(sdesc)) { calc_hmac() 347 return PTR_ERR(sdesc); calc_hmac() 352 ret = crypto_shash_digest(&sdesc->shash, buf, buflen, digest); calc_hmac() 353 kfree(sdesc); calc_hmac() 359 struct sdesc *sdesc; calc_hash() local 362 sdesc = alloc_sdesc(hashalg); calc_hash() 363 if (IS_ERR(sdesc)) { calc_hash() 365 return PTR_ERR(sdesc); calc_hash() 368 ret = crypto_shash_digest(&sdesc->shash, buf, buflen, digest); calc_hash() 369 kfree(sdesc); calc_hash()
|
/linux-4.1.27/drivers/media/platform/soc_camera/ |
H A D | soc_camera.c | 691 struct soc_camera_desc *sdesc = to_soc_camera_desc(icd); soc_camera_open() local 706 if (sdesc->subdev_desc.reset) soc_camera_open() 708 sdesc->subdev_desc.reset(icd->control); soc_camera_open() 1174 struct soc_camera_desc *sdesc = to_soc_camera_desc(icd); scan_add_host() local 1175 struct soc_camera_subdev_desc *ssdd = &sdesc->subdev_desc; scan_add_host() 1242 static int soc_camera_dyn_pdev(struct soc_camera_desc *sdesc, soc_camera_dyn_pdev() argument 1260 ret = platform_device_add_data(pdev, sdesc, sizeof(*sdesc)); soc_camera_dyn_pdev() 1337 struct soc_camera_desc *sdesc) soc_camera_i2c_init() 1342 struct soc_camera_host_desc *shd = &sdesc->host_desc; soc_camera_i2c_init() 1362 ssdd = kmemdup(&sdesc->subdev_desc, sizeof(*ssdd), GFP_KERNEL); soc_camera_i2c_init() 1453 struct soc_camera_desc *sdesc = to_soc_camera_desc(icd); soc_camera_async_bound() local 1457 memcpy(&sdesc->subdev_desc, ssdd, soc_camera_async_bound() 1458 sizeof(sdesc->subdev_desc)); soc_camera_async_bound() 1510 struct soc_camera_desc sdesc = {.host_desc.bus_id = ici->nr,}; scan_async_group() local 1534 sdesc.host_desc.board_info = ERR_PTR(-ENODATA); scan_async_group() 1536 ret = soc_camera_dyn_pdev(&sdesc, sasc); scan_async_group() 1594 #define soc_camera_i2c_init(icd, sdesc) (-ENODEV) 1612 struct soc_camera_desc sdesc = {.host_desc.bus_id = ici->nr,}; soc_of_bind() local 1633 sdesc.host_desc.board_info = ERR_PTR(-ENODATA); soc_of_bind() 1635 ret = soc_camera_dyn_pdev(&sdesc, sasc); soc_of_bind() 1730 struct soc_camera_desc *sdesc = to_soc_camera_desc(icd); soc_camera_probe() local 1731 struct soc_camera_host_desc *shd = &sdesc->host_desc; soc_camera_probe() 1761 ret = soc_camera_i2c_init(icd, sdesc); soc_camera_probe() 1828 struct soc_camera_desc *sdesc = to_soc_camera_desc(icd); soc_camera_remove() local 1837 if (sdesc->host_desc.board_info) { soc_camera_remove() 1843 sdesc->host_desc.del_device(icd); soc_camera_remove() 2171 struct soc_camera_desc *sdesc = pdev->dev.platform_data; soc_camera_pdrv_probe() local 2172 struct soc_camera_subdev_desc *ssdd = &sdesc->subdev_desc; soc_camera_pdrv_probe() 2176 if (!sdesc) soc_camera_pdrv_probe() 2195 icd->iface = sdesc->host_desc.bus_id; soc_camera_pdrv_probe() 2196 icd->sdesc = sdesc; soc_camera_pdrv_probe() 1336 soc_camera_i2c_init(struct soc_camera_device *icd, struct soc_camera_desc *sdesc) soc_camera_i2c_init() argument
|
H A D | soc_camera_platform.c | 57 return soc_camera_set_power(p->icd->control, &p->icd->sdesc->subdev_desc, NULL, on); soc_camera_platform_s_power()
|
/linux-4.1.27/drivers/net/wireless/ath/ath10k/ |
H A D | ce.c | 277 struct ce_desc *desc, *sdesc; ath10k_ce_send_nolock() local 297 sdesc = CE_SRC_RING_TO_DESC(src_ring->shadow_base, write_index); ath10k_ce_send_nolock() 306 sdesc->addr = __cpu_to_le32(buffer); ath10k_ce_send_nolock() 307 sdesc->nbytes = __cpu_to_le16(nbytes); ath10k_ce_send_nolock() 308 sdesc->flags = __cpu_to_le16(desc_flags); ath10k_ce_send_nolock() 310 *desc = *sdesc; ath10k_ce_send_nolock() 459 struct ce_desc sdesc; ath10k_ce_completed_recv_next_nolock() local 463 sdesc = *desc; ath10k_ce_completed_recv_next_nolock() 465 nbytes = __le16_to_cpu(sdesc.nbytes); ath10k_ce_completed_recv_next_nolock() 479 *bufferp = __le32_to_cpu(sdesc.addr); ath10k_ce_completed_recv_next_nolock() 481 *transfer_idp = MS(__le16_to_cpu(sdesc.flags), CE_DESC_FLAGS_META_DATA); ath10k_ce_completed_recv_next_nolock() 483 if (__le16_to_cpu(sdesc.flags) & CE_DESC_FLAGS_BYTE_SWAP) ath10k_ce_completed_recv_next_nolock() 591 struct ce_desc *sdesc, *sbase; ath10k_ce_completed_send_next_nolock() local 617 sdesc = CE_SRC_RING_TO_DESC(sbase, sw_index); ath10k_ce_completed_send_next_nolock() 620 *bufferp = __le32_to_cpu(sdesc->addr); ath10k_ce_completed_send_next_nolock() 621 *nbytesp = __le16_to_cpu(sdesc->nbytes); ath10k_ce_completed_send_next_nolock() 622 *transfer_idp = MS(__le16_to_cpu(sdesc->flags), ath10k_ce_completed_send_next_nolock()
|
/linux-4.1.27/drivers/crypto/ccp/ |
H A D | ccp-crypto-sha.c | 243 SHASH_DESC_ON_STACK(sdesc, shash); ccp_sha_setkey() 259 sdesc->tfm = shash; ccp_sha_setkey() 260 sdesc->flags = crypto_ahash_get_flags(tfm) & ccp_sha_setkey() 263 ret = crypto_shash_digest(sdesc, key, key_len, ccp_sha_setkey()
|
/linux-4.1.27/crypto/ |
H A D | drbg.c | 1575 struct sdesc { struct 1582 struct sdesc *sdesc; drbg_init_hash_kernel() local 1591 sdesc = kzalloc(sizeof(struct shash_desc) + crypto_shash_descsize(tfm), drbg_init_hash_kernel() 1593 if (!sdesc) { drbg_init_hash_kernel() 1598 sdesc->shash.tfm = tfm; drbg_init_hash_kernel() 1599 sdesc->shash.flags = 0; drbg_init_hash_kernel() 1600 drbg->priv_data = sdesc; drbg_init_hash_kernel() 1606 struct sdesc *sdesc = (struct sdesc *)drbg->priv_data; drbg_fini_hash_kernel() local 1607 if (sdesc) { drbg_fini_hash_kernel() 1608 crypto_free_shash(sdesc->shash.tfm); drbg_fini_hash_kernel() 1609 kzfree(sdesc); drbg_fini_hash_kernel() 1618 struct sdesc *sdesc = (struct sdesc *)drbg->priv_data; drbg_kcapi_hash() local 1622 crypto_shash_setkey(sdesc->shash.tfm, key, drbg_statelen(drbg)); drbg_kcapi_hash() 1623 crypto_shash_init(&sdesc->shash); drbg_kcapi_hash() 1625 crypto_shash_update(&sdesc->shash, input->buf, input->len); drbg_kcapi_hash() 1626 return crypto_shash_final(&sdesc->shash, outval); drbg_kcapi_hash()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 397 d = &q->sdesc[cidx]; free_tx_desc() 408 d = q->sdesc; free_tx_desc() 495 struct rx_sw_desc *d = &q->sdesc[q->cidx]; free_rx_bufs() 522 struct rx_sw_desc *d = &q->sdesc[q->cidx]; unmap_rx_buf() 594 struct rx_sw_desc *sd = &q->sdesc[q->pidx]; refill_fl() 629 sd = q->sdesc; refill_fl() 657 sd = q->sdesc; refill_fl() 1275 q->q.sdesc[last_desc].skb = skb; t4_eth_xmit() 1276 q->q.sdesc[last_desc].sgl = (struct ulptx_sgl *)(cpl + 1); t4_eth_xmit() 1570 q->q.sdesc[last_desc].skb = skb; service_ofldq() 1949 d = &q->sdesc[q->cidx]; restore_rx_bufs() 2032 rsd = &rxq->fl.sdesc[rxq->fl.cidx]; process_responses() 2479 &fl->sdesc, s->stat_len, NUMA_NO_NODE); t4_sge_alloc_rxq() 2546 kfree(fl->sdesc); t4_sge_alloc_rxq() 2547 fl->sdesc = NULL; t4_sge_alloc_rxq() 2584 &txq->q.phys_addr, &txq->q.sdesc, s->stat_len, t4_sge_alloc_eth_txq() 2610 kfree(txq->q.sdesc); t4_sge_alloc_eth_txq() 2611 txq->q.sdesc = NULL; t4_sge_alloc_eth_txq() 2692 &txq->q.phys_addr, &txq->q.sdesc, s->stat_len, t4_sge_alloc_ofld_txq() 2716 kfree(txq->q.sdesc); t4_sge_alloc_ofld_txq() 2717 txq->q.sdesc = NULL; t4_sge_alloc_ofld_txq() 2742 q->sdesc = NULL; free_txq() 2767 kfree(fl->sdesc); free_rspq_fl() 2768 fl->sdesc = NULL; free_rspq_fl() 2811 kfree(etq->q.sdesc); t4_free_sge_resources() 2830 kfree(q->q.sdesc); t4_free_sge_resources()
|
H A D | cxgb4.h | 460 struct rx_sw_desc *sdesc; /* address of SW Rx descriptor ring */ member in struct:sge_fl 562 struct tx_sw_desc *sdesc; /* address of SW Tx descriptor ring */ member in struct:sge_txq
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | adapter.h | 116 struct rx_sw_desc *sdesc; /* address of SW Rx descriptor ring */ member in struct:sge_fl 177 struct tx_sw_desc *sdesc; /* address of SW Tx descriptor ring */ member in struct:sge_txq
|
H A D | sge.c | 240 struct tx_sw_desc *d = &q->sdesc[cidx]; unmap_skb() 269 d = cidx + 1 == q->size ? q->sdesc : d + 1; unmap_skb() 295 d = &q->sdesc[cidx]; free_tx_desc() 308 d = q->sdesc; free_tx_desc() 385 struct rx_sw_desc *d = &q->sdesc[cidx]; free_rx_bufs() 502 struct rx_sw_desc *sd = &q->sdesc[q->pidx]; refill_fl() 545 sd = q->sdesc; refill_fl() 579 q->sdesc[q->pidx] = q->sdesc[idx]; recycle_rx_buf() 683 kfree(q->fl[i].sdesc); t3_free_qset() 695 if (q->txq[i].sdesc) { t3_free_qset() 698 kfree(q->txq[i].sdesc); t3_free_qset() 783 struct rx_sw_desc *sd = &fl->sdesc[fl->cidx]; get_packet() 843 struct rx_sw_desc *sd = &fl->sdesc[fl->cidx]; get_packet_pg() 1058 struct tx_sw_desc *sd = &q->sdesc[pidx]; write_wr_hdr_sgl() 1102 sd = q->sdesc; write_wr_hdr_sgl() 1176 q->sdesc[pidx].skb = NULL; write_tx_pkt_wr() 1588 q->sdesc[pidx].skb = NULL; write_ofld_wr() 2066 struct rx_sw_desc *sd = &fl->sdesc[fl->cidx]; lro_add_page() 2308 void *addr = fl->sdesc[fl->cidx].pg_chunk.va; process_responses() 2985 &q->fl[0].phys_addr, &q->fl[0].sdesc); t3_sge_alloc_qset() 2992 &q->fl[1].phys_addr, &q->fl[1].sdesc); t3_sge_alloc_qset() 3012 &q->txq[i].sdesc); t3_sge_alloc_qset()
|
/linux-4.1.27/fs/cifs/ |
H A D | smbencrypt.c | 138 struct sdesc *sdescmd4; mdfour()
|
H A D | cifsglob.h | 110 struct sdesc { struct 121 struct sdesc *sdeschmacmd5; /* ctxt to generate ntlmv2 hash, CR1 */ 122 struct sdesc *sdescmd5; /* ctxt to generate cifs/smb signature */ 123 struct sdesc *sdeschmacsha256; /* ctxt to generate smb2 signature */ 124 struct sdesc *sdesccmacaes; /* ctxt to generate smb3 signature */
|
H A D | link.c | 62 struct sdesc *sdescmd5; symlink_hash()
|
/linux-4.1.27/drivers/infiniband/ulp/iser/ |
H A D | iser_initiator.c | 178 iser_dbg("sdesc %p lkey mismatch, fixing\n", tx_desc); iser_create_send_desc()
|
/linux-4.1.27/drivers/net/ethernet/nvidia/ |
H A D | forcedeth.c | 6007 dev_info(&pci_dev->dev, "%s%s%s%s%s%s%s%s%s%s%sdesc-v%u\n", nv_probe()
|