Searched refs:sdesc (Results 1 - 25 of 25) sorted by relevance

/linux-4.1.27/drivers/dma/
H A Dsirf-dma.c117 struct sirfsoc_dma_desc *sdesc = NULL; sirfsoc_dma_execute() local
124 sdesc = list_first_entry(&schan->queued, struct sirfsoc_dma_desc, sirfsoc_dma_execute()
127 list_move_tail(&sdesc->node, &schan->active); sirfsoc_dma_execute()
130 writel_relaxed(sdesc->width, sdma->base + SIRFSOC_DMA_WIDTH_0 + sirfsoc_dma_execute()
133 (sdesc->dir << SIRFSOC_DMA_DIR_CTRL_BIT), sirfsoc_dma_execute()
135 writel_relaxed(sdesc->xlen, sdma->base + cid * 0x10 + sirfsoc_dma_execute()
137 writel_relaxed(sdesc->ylen, sdma->base + cid * 0x10 + sirfsoc_dma_execute()
146 writel(sdesc->addr >> 2, sdma->base + cid * 0x10 + SIRFSOC_DMA_CH_ADDR); sirfsoc_dma_execute()
148 if (sdesc->cyclic) { sirfsoc_dma_execute()
161 struct sirfsoc_dma_desc *sdesc = NULL; sirfsoc_dma_irq() local
173 sdesc = list_first_entry(&schan->active, struct sirfsoc_dma_desc, sirfsoc_dma_irq()
175 if (!sdesc->cyclic) { sirfsoc_dma_irq()
197 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_process_completed() local
214 list_for_each_entry(sdesc, &list, node) { sirfsoc_dma_process_completed()
215 desc = &sdesc->desc; sirfsoc_dma_process_completed()
231 sdesc = list_first_entry(&schan->active, struct sirfsoc_dma_desc, sirfsoc_dma_process_completed()
234 if (!sdesc || (sdesc && !sdesc->cyclic)) { sirfsoc_dma_process_completed()
244 desc = &sdesc->desc; sirfsoc_dma_process_completed()
266 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_tx_submit() local
270 sdesc = container_of(txd, struct sirfsoc_dma_desc, desc); sirfsoc_dma_tx_submit()
275 list_move_tail(&sdesc->node, &schan->queued); sirfsoc_dma_tx_submit()
381 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_alloc_chan_resources() local
390 sdesc = kzalloc(sizeof(*sdesc), GFP_KERNEL); sirfsoc_dma_alloc_chan_resources()
391 if (!sdesc) { sirfsoc_dma_alloc_chan_resources()
397 dma_async_tx_descriptor_init(&sdesc->desc, chan); sirfsoc_dma_alloc_chan_resources()
398 sdesc->desc.flags = DMA_CTRL_ACK; sirfsoc_dma_alloc_chan_resources()
399 sdesc->desc.tx_submit = sirfsoc_dma_tx_submit; sirfsoc_dma_alloc_chan_resources()
401 list_add_tail(&sdesc->node, &descs); sirfsoc_dma_alloc_chan_resources()
421 struct sirfsoc_dma_desc *sdesc, *tmp; sirfsoc_dma_free_chan_resources() local
439 list_for_each_entry_safe(sdesc, tmp, &descs, node) sirfsoc_dma_free_chan_resources()
440 kfree(sdesc); sirfsoc_dma_free_chan_resources()
468 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_tx_status() local
476 sdesc = list_first_entry(&schan->active, struct sirfsoc_dma_desc, sirfsoc_dma_tx_status()
478 dma_request_bytes = (sdesc->xlen + 1) * (sdesc->ylen + 1) * sirfsoc_dma_tx_status()
479 (sdesc->width * SIRFSOC_DMA_WORD_LEN); sirfsoc_dma_tx_status()
484 residue = dma_request_bytes - (dma_pos - sdesc->addr); sirfsoc_dma_tx_status()
498 struct sirfsoc_dma_desc *sdesc = NULL; sirfsoc_dma_prep_interleaved() local
510 sdesc = list_first_entry(&schan->free, struct sirfsoc_dma_desc, sirfsoc_dma_prep_interleaved()
512 list_del(&sdesc->node); sirfsoc_dma_prep_interleaved()
516 if (!sdesc) { sirfsoc_dma_prep_interleaved()
531 sdesc->cyclic = 0; sirfsoc_dma_prep_interleaved()
532 sdesc->xlen = xt->sgl[0].size / SIRFSOC_DMA_WORD_LEN; sirfsoc_dma_prep_interleaved()
533 sdesc->width = (xt->sgl[0].size + xt->sgl[0].icg) / sirfsoc_dma_prep_interleaved()
535 sdesc->ylen = xt->numf - 1; sirfsoc_dma_prep_interleaved()
537 sdesc->addr = xt->src_start; sirfsoc_dma_prep_interleaved()
538 sdesc->dir = 1; sirfsoc_dma_prep_interleaved()
540 sdesc->addr = xt->dst_start; sirfsoc_dma_prep_interleaved()
541 sdesc->dir = 0; sirfsoc_dma_prep_interleaved()
544 list_add_tail(&sdesc->node, &schan->prepared); sirfsoc_dma_prep_interleaved()
552 return &sdesc->desc; sirfsoc_dma_prep_interleaved()
566 struct sirfsoc_dma_desc *sdesc = NULL; sirfsoc_dma_prep_cyclic() local
586 sdesc = list_first_entry(&schan->free, struct sirfsoc_dma_desc, sirfsoc_dma_prep_cyclic()
588 list_del(&sdesc->node); sirfsoc_dma_prep_cyclic()
592 if (!sdesc) sirfsoc_dma_prep_cyclic()
597 sdesc->addr = addr; sirfsoc_dma_prep_cyclic()
598 sdesc->cyclic = 1; sirfsoc_dma_prep_cyclic()
599 sdesc->xlen = 0; sirfsoc_dma_prep_cyclic()
600 sdesc->ylen = buf_len / SIRFSOC_DMA_WORD_LEN - 1; sirfsoc_dma_prep_cyclic()
601 sdesc->width = 1; sirfsoc_dma_prep_cyclic()
602 list_add_tail(&sdesc->node, &schan->prepared); sirfsoc_dma_prep_cyclic()
605 return &sdesc->desc; sirfsoc_dma_prep_cyclic()
815 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_pm_suspend() local
838 sdesc = list_first_entry(&schan->active, sirfsoc_dma_pm_suspend()
856 struct sirfsoc_dma_desc *sdesc; sirfsoc_dma_pm_resume() local
871 sdesc = list_first_entry(&schan->active, sirfsoc_dma_pm_resume()
874 writel_relaxed(sdesc->width, sirfsoc_dma_pm_resume()
876 writel_relaxed(sdesc->xlen, sirfsoc_dma_pm_resume()
878 writel_relaxed(sdesc->ylen, sirfsoc_dma_pm_resume()
882 writel_relaxed(sdesc->addr >> 2, sirfsoc_dma_pm_resume()
/linux-4.1.27/security/keys/
H A Dtrusted.c37 struct sdesc { struct
45 static struct sdesc *init_sdesc(struct crypto_shash *alg) init_sdesc()
47 struct sdesc *sdesc; init_sdesc() local
51 sdesc = kmalloc(size, GFP_KERNEL); init_sdesc()
52 if (!sdesc) init_sdesc()
54 sdesc->shash.tfm = alg; init_sdesc()
55 sdesc->shash.flags = 0x0; init_sdesc()
56 return sdesc; init_sdesc()
62 struct sdesc *sdesc; TSS_sha1() local
65 sdesc = init_sdesc(hashalg); TSS_sha1()
66 if (IS_ERR(sdesc)) { TSS_sha1()
68 return PTR_ERR(sdesc); TSS_sha1()
71 ret = crypto_shash_digest(&sdesc->shash, data, datalen, digest); TSS_sha1()
72 kfree(sdesc); TSS_sha1()
79 struct sdesc *sdesc; TSS_rawhmac() local
85 sdesc = init_sdesc(hmacalg); TSS_rawhmac()
86 if (IS_ERR(sdesc)) { TSS_rawhmac()
88 return PTR_ERR(sdesc); TSS_rawhmac()
94 ret = crypto_shash_init(&sdesc->shash); TSS_rawhmac()
108 ret = crypto_shash_update(&sdesc->shash, data, dlen); TSS_rawhmac()
114 ret = crypto_shash_final(&sdesc->shash, digest); TSS_rawhmac()
116 kfree(sdesc); TSS_rawhmac()
128 struct sdesc *sdesc; TSS_authhmac() local
135 sdesc = init_sdesc(hashalg); TSS_authhmac()
136 if (IS_ERR(sdesc)) { TSS_authhmac()
138 return PTR_ERR(sdesc); TSS_authhmac()
142 ret = crypto_shash_init(&sdesc->shash); TSS_authhmac()
155 ret = crypto_shash_update(&sdesc->shash, data, dlen); TSS_authhmac()
161 ret = crypto_shash_final(&sdesc->shash, paramdigest); TSS_authhmac()
167 kfree(sdesc); TSS_authhmac()
189 struct sdesc *sdesc; TSS_checkhmac1() local
207 sdesc = init_sdesc(hashalg); TSS_checkhmac1()
208 if (IS_ERR(sdesc)) { TSS_checkhmac1()
210 return PTR_ERR(sdesc); TSS_checkhmac1()
212 ret = crypto_shash_init(&sdesc->shash); TSS_checkhmac1()
215 ret = crypto_shash_update(&sdesc->shash, (const u8 *)&result, TSS_checkhmac1()
219 ret = crypto_shash_update(&sdesc->shash, (const u8 *)&ordinal, TSS_checkhmac1()
229 ret = crypto_shash_update(&sdesc->shash, buffer + dpos, dlen); TSS_checkhmac1()
235 ret = crypto_shash_final(&sdesc->shash, paramdigest); TSS_checkhmac1()
248 kfree(sdesc); TSS_checkhmac1()
276 struct sdesc *sdesc; TSS_checkhmac2() local
299 sdesc = init_sdesc(hashalg); TSS_checkhmac2()
300 if (IS_ERR(sdesc)) { TSS_checkhmac2()
302 return PTR_ERR(sdesc); TSS_checkhmac2()
304 ret = crypto_shash_init(&sdesc->shash); TSS_checkhmac2()
307 ret = crypto_shash_update(&sdesc->shash, (const u8 *)&result, TSS_checkhmac2()
311 ret = crypto_shash_update(&sdesc->shash, (const u8 *)&ordinal, TSS_checkhmac2()
322 ret = crypto_shash_update(&sdesc->shash, buffer + dpos, dlen); TSS_checkhmac2()
328 ret = crypto_shash_final(&sdesc->shash, paramdigest); TSS_checkhmac2()
349 kfree(sdesc); TSS_checkhmac2()
/linux-4.1.27/drivers/dma/sh/
H A Dsudmac.c53 #define to_desc(sdesc) container_of(sdesc, struct sudmac_desc, shdma_desc)
103 struct shdma_desc *sdesc) sudmac_set_reg()
119 struct shdma_desc *sdesc) sudmac_start_xfer()
122 struct sudmac_desc *sd = to_desc(sdesc); sudmac_start_xfer()
124 sudmac_set_reg(sc, &sd->hw, sdesc); sudmac_start_xfer()
176 struct shdma_desc *sdesc, sudmac_desc_setup()
180 struct sudmac_desc *sd = to_desc(sdesc); sudmac_desc_setup()
219 struct shdma_desc *sdesc) sudmac_get_partial()
222 struct sudmac_desc *sd = to_desc(sdesc); sudmac_get_partial()
229 struct shdma_desc *sdesc) sudmac_desc_completed()
232 struct sudmac_desc *sd = to_desc(sdesc); sudmac_desc_completed()
102 sudmac_set_reg(struct sudmac_chan *sc, struct sudmac_regs *hw, struct shdma_desc *sdesc) sudmac_set_reg() argument
118 sudmac_start_xfer(struct shdma_chan *schan, struct shdma_desc *sdesc) sudmac_start_xfer() argument
175 sudmac_desc_setup(struct shdma_chan *schan, struct shdma_desc *sdesc, dma_addr_t src, dma_addr_t dst, size_t *len) sudmac_desc_setup() argument
218 sudmac_get_partial(struct shdma_chan *schan, struct shdma_desc *sdesc) sudmac_get_partial() argument
228 sudmac_desc_completed(struct shdma_chan *schan, struct shdma_desc *sdesc) sudmac_desc_completed() argument
H A Dshdma-base.c59 struct shdma_desc *sdesc; shdma_chan_xfer_ld_queue() local
66 list_for_each_entry(sdesc, &schan->ld_queue, node) shdma_chan_xfer_ld_queue()
67 if (sdesc->mark == DESC_SUBMITTED) { shdma_chan_xfer_ld_queue()
68 ops->start_xfer(schan, sdesc); shdma_chan_xfer_ld_queue()
162 struct shdma_desc *sdesc; shdma_get_desc() local
164 list_for_each_entry(sdesc, &schan->ld_free, node) shdma_get_desc()
165 if (sdesc->mark != DESC_PREPARED) { shdma_get_desc()
166 BUG_ON(sdesc->mark != DESC_IDLE); shdma_get_desc()
167 list_del(&sdesc->node); shdma_get_desc()
168 return sdesc; shdma_get_desc()
840 struct shdma_desc *sdesc; shdma_tx_status() local
842 list_for_each_entry(sdesc, &schan->ld_queue, node) shdma_tx_status()
843 if (sdesc->cookie == cookie) { shdma_tx_status()
864 struct shdma_desc *sdesc; shdma_for_each_chan() local
886 list_for_each_entry(sdesc, &dl, node) { shdma_for_each_chan()
887 struct dma_async_tx_descriptor *tx = &sdesc->async_tx; shdma_for_each_chan()
888 sdesc->mark = DESC_IDLE; shdma_for_each_chan()
925 struct shdma_desc *sdesc; chan_irqt() local
928 list_for_each_entry(sdesc, &schan->ld_queue, node) { chan_irqt()
929 if (sdesc->mark == DESC_SUBMITTED && chan_irqt()
930 ops->desc_completed(schan, sdesc)) { chan_irqt()
932 sdesc->async_tx.cookie, &sdesc->async_tx); chan_irqt()
933 sdesc->mark = DESC_COMPLETED; chan_irqt()
H A Drcar-hpbdma.c124 #define to_desc(sdesc) container_of(sdesc, struct hpb_desc, shdma_desc)
314 struct shdma_desc *sdesc) hpb_dmae_start_xfer()
318 struct hpb_desc *desc = to_desc(sdesc); hpb_dmae_start_xfer()
334 struct shdma_desc *sdesc) hpb_dmae_desc_completed()
362 struct shdma_desc *sdesc, hpb_dmae_desc_setup()
365 struct hpb_desc *desc = to_desc(sdesc); hpb_dmae_desc_setup()
378 struct shdma_desc *sdesc) hpb_dmae_get_partial()
380 struct hpb_desc *desc = to_desc(sdesc); hpb_dmae_get_partial()
313 hpb_dmae_start_xfer(struct shdma_chan *schan, struct shdma_desc *sdesc) hpb_dmae_start_xfer() argument
333 hpb_dmae_desc_completed(struct shdma_chan *schan, struct shdma_desc *sdesc) hpb_dmae_desc_completed() argument
361 hpb_dmae_desc_setup(struct shdma_chan *schan, struct shdma_desc *sdesc, dma_addr_t src, dma_addr_t dst, size_t *len) hpb_dmae_desc_setup() argument
377 hpb_dmae_get_partial(struct shdma_chan *schan, struct shdma_desc *sdesc) hpb_dmae_get_partial() argument
H A Dshdmac.c287 struct shdma_desc *sdesc) sh_dmae_start_xfer()
291 struct sh_dmae_desc *sh_desc = container_of(sdesc, sh_dmae_start_xfer()
294 sdesc->async_tx.cookie, sh_chan->shdma_chan.id, sh_dmae_start_xfer()
382 struct shdma_desc *sdesc, sh_dmae_desc_setup()
385 struct sh_dmae_desc *sh_desc = container_of(sdesc, sh_dmae_desc_setup()
420 struct shdma_desc *sdesc) sh_dmae_get_partial()
424 struct sh_dmae_desc *sh_desc = container_of(sdesc, sh_dmae_get_partial()
460 struct shdma_desc *sdesc) sh_dmae_desc_completed()
464 struct sh_dmae_desc *sh_desc = container_of(sdesc, sh_dmae_desc_completed()
469 return (sdesc->direction == DMA_DEV_TO_MEM && sh_dmae_desc_completed()
471 (sdesc->direction != DMA_DEV_TO_MEM && sh_dmae_desc_completed()
286 sh_dmae_start_xfer(struct shdma_chan *schan, struct shdma_desc *sdesc) sh_dmae_start_xfer() argument
381 sh_dmae_desc_setup(struct shdma_chan *schan, struct shdma_desc *sdesc, dma_addr_t src, dma_addr_t dst, size_t *len) sh_dmae_desc_setup() argument
419 sh_dmae_get_partial(struct shdma_chan *schan, struct shdma_desc *sdesc) sh_dmae_get_partial() argument
459 sh_dmae_desc_completed(struct shdma_chan *schan, struct shdma_desc *sdesc) sh_dmae_desc_completed() argument
/linux-4.1.27/include/media/
H A Dsoc_camera_platform.h46 if (&icd->sdesc->subdev_desc != ssdd) soc_camera_platform_add()
77 if (&icd->sdesc->subdev_desc != ssdd || !pdev) soc_camera_platform_del()
H A Dsoc_camera.h32 struct soc_camera_desc *sdesc; member in struct:soc_camera_device
266 return icd->sdesc; to_soc_camera_desc()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dsge.c190 * @sdesc: pointer to the software buffer descriptor
195 static inline dma_addr_t get_buf_addr(const struct rx_sw_desc *sdesc) get_buf_addr() argument
197 return sdesc->dma_addr & ~(dma_addr_t)(RX_LARGE_BUF | RX_UNMAPPED_BUF); get_buf_addr()
202 * @sdesc: pointer to the software buffer descriptor
207 static inline bool is_buf_mapped(const struct rx_sw_desc *sdesc) is_buf_mapped() argument
209 return !(sdesc->dma_addr & RX_UNMAPPED_BUF); is_buf_mapped()
380 struct tx_sw_desc *sdesc; free_tx_desc() local
386 sdesc = &tq->sdesc[cidx]; free_tx_desc()
392 if (sdesc->skb) { free_tx_desc()
394 unmap_sgl(dev, sdesc->skb, sdesc->sgl, tq); free_tx_desc()
395 dev_consume_skb_any(sdesc->skb); free_tx_desc()
396 sdesc->skb = NULL; free_tx_desc()
399 sdesc++; free_tx_desc()
402 sdesc = tq->sdesc; free_tx_desc()
452 * @sdesc: pointer to the software buffer descriptor
455 const struct rx_sw_desc *sdesc) get_buf_size()
459 return (s->fl_pg_order > 0 && (sdesc->dma_addr & RX_LARGE_BUF) get_buf_size()
476 struct rx_sw_desc *sdesc = &fl->sdesc[fl->cidx]; free_rx_bufs() local
478 if (is_buf_mapped(sdesc)) free_rx_bufs()
479 dma_unmap_page(adapter->pdev_dev, get_buf_addr(sdesc), free_rx_bufs()
480 get_buf_size(adapter, sdesc), free_rx_bufs()
482 put_page(sdesc->page); free_rx_bufs()
483 sdesc->page = NULL; free_rx_bufs()
505 struct rx_sw_desc *sdesc = &fl->sdesc[fl->cidx]; unmap_rx_buf() local
507 if (is_buf_mapped(sdesc)) unmap_rx_buf()
508 dma_unmap_page(adapter->pdev_dev, get_buf_addr(sdesc), unmap_rx_buf()
509 get_buf_size(adapter, sdesc), unmap_rx_buf()
511 sdesc->page = NULL; unmap_rx_buf()
569 * @sdesc: pointer to the softwore RX buffer descriptor
573 static inline void set_rx_sw_desc(struct rx_sw_desc *sdesc, struct page *page, set_rx_sw_desc() argument
576 sdesc->page = page; set_rx_sw_desc()
577 sdesc->dma_addr = dma_addr; set_rx_sw_desc()
614 struct rx_sw_desc *sdesc = &fl->sdesc[fl->pidx]; refill_fl() local
665 set_rx_sw_desc(sdesc, page, dma_addr); refill_fl()
666 sdesc++; refill_fl()
671 sdesc = fl->sdesc; refill_fl()
694 set_rx_sw_desc(sdesc, page, dma_addr); refill_fl()
695 sdesc++; refill_fl()
700 sdesc = fl->sdesc; refill_fl()
1425 tq->sdesc[last_desc].skb = skb; t4vf_eth_xmit()
1426 tq->sdesc[last_desc].sgl = sgl; t4vf_eth_xmit()
1692 struct rx_sw_desc *sdesc; restore_rx_bufs() local
1699 sdesc = &fl->sdesc[fl->cidx]; restore_rx_bufs()
1700 sdesc->page = gl->frags[frags].page; restore_rx_bufs()
1701 sdesc->dma_addr |= RX_UNMAPPED_BUF; restore_rx_bufs()
1759 const struct rx_sw_desc *sdesc; process_responses() local
1788 sdesc = &rxq->fl.sdesc[rxq->fl.cidx]; process_responses()
1789 bufsz = get_buf_size(adapter, sdesc); process_responses()
1790 fp->page = sdesc->page; process_responses()
1806 get_buf_addr(sdesc), process_responses()
2251 &fl->addr, &fl->sdesc, s->stat_len); t4vf_sge_alloc_rxq()
2343 kfree(fl->sdesc); t4vf_sge_alloc_rxq()
2344 fl->sdesc = NULL; t4vf_sge_alloc_rxq()
2382 &txq->q.phys_addr, &txq->q.sdesc, s->stat_len); t4vf_sge_alloc_eth_txq()
2425 kfree(txq->q.sdesc); t4vf_sge_alloc_eth_txq()
2426 txq->q.sdesc = NULL; t4vf_sge_alloc_eth_txq()
2466 tq->sdesc = NULL; free_txq()
2495 kfree(fl->sdesc); free_rspq_fl()
2496 fl->sdesc = NULL; free_rspq_fl()
2523 kfree(txq->q.sdesc); t4vf_free_sge_resources()
454 get_buf_size(const struct adapter *adapter, const struct rx_sw_desc *sdesc) get_buf_size() argument
H A Dadapter.h141 struct rx_sw_desc *sdesc; /* address of SW RX descriptor ring */ member in struct:sge_fl
247 struct tx_sw_desc *sdesc; /* address of SW TX descriptor ring */ member in struct:sge_txq
/linux-4.1.27/security/keys/encrypted-keys/
H A Dencrypted.c57 struct sdesc { struct
324 static struct sdesc *alloc_sdesc(struct crypto_shash *alg) alloc_sdesc()
326 struct sdesc *sdesc; alloc_sdesc() local
330 sdesc = kmalloc(size, GFP_KERNEL); alloc_sdesc()
331 if (!sdesc) alloc_sdesc()
333 sdesc->shash.tfm = alg; alloc_sdesc()
334 sdesc->shash.flags = 0x0; alloc_sdesc()
335 return sdesc; alloc_sdesc()
341 struct sdesc *sdesc; calc_hmac() local
344 sdesc = alloc_sdesc(hmacalg); calc_hmac()
345 if (IS_ERR(sdesc)) { calc_hmac()
347 return PTR_ERR(sdesc); calc_hmac()
352 ret = crypto_shash_digest(&sdesc->shash, buf, buflen, digest); calc_hmac()
353 kfree(sdesc); calc_hmac()
359 struct sdesc *sdesc; calc_hash() local
362 sdesc = alloc_sdesc(hashalg); calc_hash()
363 if (IS_ERR(sdesc)) { calc_hash()
365 return PTR_ERR(sdesc); calc_hash()
368 ret = crypto_shash_digest(&sdesc->shash, buf, buflen, digest); calc_hash()
369 kfree(sdesc); calc_hash()
/linux-4.1.27/drivers/media/platform/soc_camera/
H A Dsoc_camera.c691 struct soc_camera_desc *sdesc = to_soc_camera_desc(icd); soc_camera_open() local
706 if (sdesc->subdev_desc.reset) soc_camera_open()
708 sdesc->subdev_desc.reset(icd->control); soc_camera_open()
1174 struct soc_camera_desc *sdesc = to_soc_camera_desc(icd); scan_add_host() local
1175 struct soc_camera_subdev_desc *ssdd = &sdesc->subdev_desc; scan_add_host()
1242 static int soc_camera_dyn_pdev(struct soc_camera_desc *sdesc, soc_camera_dyn_pdev() argument
1260 ret = platform_device_add_data(pdev, sdesc, sizeof(*sdesc)); soc_camera_dyn_pdev()
1337 struct soc_camera_desc *sdesc) soc_camera_i2c_init()
1342 struct soc_camera_host_desc *shd = &sdesc->host_desc; soc_camera_i2c_init()
1362 ssdd = kmemdup(&sdesc->subdev_desc, sizeof(*ssdd), GFP_KERNEL); soc_camera_i2c_init()
1453 struct soc_camera_desc *sdesc = to_soc_camera_desc(icd); soc_camera_async_bound() local
1457 memcpy(&sdesc->subdev_desc, ssdd, soc_camera_async_bound()
1458 sizeof(sdesc->subdev_desc)); soc_camera_async_bound()
1510 struct soc_camera_desc sdesc = {.host_desc.bus_id = ici->nr,}; scan_async_group() local
1534 sdesc.host_desc.board_info = ERR_PTR(-ENODATA); scan_async_group()
1536 ret = soc_camera_dyn_pdev(&sdesc, sasc); scan_async_group()
1594 #define soc_camera_i2c_init(icd, sdesc) (-ENODEV)
1612 struct soc_camera_desc sdesc = {.host_desc.bus_id = ici->nr,}; soc_of_bind() local
1633 sdesc.host_desc.board_info = ERR_PTR(-ENODATA); soc_of_bind()
1635 ret = soc_camera_dyn_pdev(&sdesc, sasc); soc_of_bind()
1730 struct soc_camera_desc *sdesc = to_soc_camera_desc(icd); soc_camera_probe() local
1731 struct soc_camera_host_desc *shd = &sdesc->host_desc; soc_camera_probe()
1761 ret = soc_camera_i2c_init(icd, sdesc); soc_camera_probe()
1828 struct soc_camera_desc *sdesc = to_soc_camera_desc(icd); soc_camera_remove() local
1837 if (sdesc->host_desc.board_info) { soc_camera_remove()
1843 sdesc->host_desc.del_device(icd); soc_camera_remove()
2171 struct soc_camera_desc *sdesc = pdev->dev.platform_data; soc_camera_pdrv_probe() local
2172 struct soc_camera_subdev_desc *ssdd = &sdesc->subdev_desc; soc_camera_pdrv_probe()
2176 if (!sdesc) soc_camera_pdrv_probe()
2195 icd->iface = sdesc->host_desc.bus_id; soc_camera_pdrv_probe()
2196 icd->sdesc = sdesc; soc_camera_pdrv_probe()
1336 soc_camera_i2c_init(struct soc_camera_device *icd, struct soc_camera_desc *sdesc) soc_camera_i2c_init() argument
H A Dsoc_camera_platform.c57 return soc_camera_set_power(p->icd->control, &p->icd->sdesc->subdev_desc, NULL, on); soc_camera_platform_s_power()
/linux-4.1.27/drivers/net/wireless/ath/ath10k/
H A Dce.c277 struct ce_desc *desc, *sdesc; ath10k_ce_send_nolock() local
297 sdesc = CE_SRC_RING_TO_DESC(src_ring->shadow_base, write_index); ath10k_ce_send_nolock()
306 sdesc->addr = __cpu_to_le32(buffer); ath10k_ce_send_nolock()
307 sdesc->nbytes = __cpu_to_le16(nbytes); ath10k_ce_send_nolock()
308 sdesc->flags = __cpu_to_le16(desc_flags); ath10k_ce_send_nolock()
310 *desc = *sdesc; ath10k_ce_send_nolock()
459 struct ce_desc sdesc; ath10k_ce_completed_recv_next_nolock() local
463 sdesc = *desc; ath10k_ce_completed_recv_next_nolock()
465 nbytes = __le16_to_cpu(sdesc.nbytes); ath10k_ce_completed_recv_next_nolock()
479 *bufferp = __le32_to_cpu(sdesc.addr); ath10k_ce_completed_recv_next_nolock()
481 *transfer_idp = MS(__le16_to_cpu(sdesc.flags), CE_DESC_FLAGS_META_DATA); ath10k_ce_completed_recv_next_nolock()
483 if (__le16_to_cpu(sdesc.flags) & CE_DESC_FLAGS_BYTE_SWAP) ath10k_ce_completed_recv_next_nolock()
591 struct ce_desc *sdesc, *sbase; ath10k_ce_completed_send_next_nolock() local
617 sdesc = CE_SRC_RING_TO_DESC(sbase, sw_index); ath10k_ce_completed_send_next_nolock()
620 *bufferp = __le32_to_cpu(sdesc->addr); ath10k_ce_completed_send_next_nolock()
621 *nbytesp = __le16_to_cpu(sdesc->nbytes); ath10k_ce_completed_send_next_nolock()
622 *transfer_idp = MS(__le16_to_cpu(sdesc->flags), ath10k_ce_completed_send_next_nolock()
/linux-4.1.27/drivers/crypto/ccp/
H A Dccp-crypto-sha.c243 SHASH_DESC_ON_STACK(sdesc, shash); ccp_sha_setkey()
259 sdesc->tfm = shash; ccp_sha_setkey()
260 sdesc->flags = crypto_ahash_get_flags(tfm) & ccp_sha_setkey()
263 ret = crypto_shash_digest(sdesc, key, key_len, ccp_sha_setkey()
/linux-4.1.27/crypto/
H A Ddrbg.c1575 struct sdesc { struct
1582 struct sdesc *sdesc; drbg_init_hash_kernel() local
1591 sdesc = kzalloc(sizeof(struct shash_desc) + crypto_shash_descsize(tfm), drbg_init_hash_kernel()
1593 if (!sdesc) { drbg_init_hash_kernel()
1598 sdesc->shash.tfm = tfm; drbg_init_hash_kernel()
1599 sdesc->shash.flags = 0; drbg_init_hash_kernel()
1600 drbg->priv_data = sdesc; drbg_init_hash_kernel()
1606 struct sdesc *sdesc = (struct sdesc *)drbg->priv_data; drbg_fini_hash_kernel() local
1607 if (sdesc) { drbg_fini_hash_kernel()
1608 crypto_free_shash(sdesc->shash.tfm); drbg_fini_hash_kernel()
1609 kzfree(sdesc); drbg_fini_hash_kernel()
1618 struct sdesc *sdesc = (struct sdesc *)drbg->priv_data; drbg_kcapi_hash() local
1622 crypto_shash_setkey(sdesc->shash.tfm, key, drbg_statelen(drbg)); drbg_kcapi_hash()
1623 crypto_shash_init(&sdesc->shash); drbg_kcapi_hash()
1625 crypto_shash_update(&sdesc->shash, input->buf, input->len); drbg_kcapi_hash()
1626 return crypto_shash_final(&sdesc->shash, outval); drbg_kcapi_hash()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/
H A Dsge.c397 d = &q->sdesc[cidx]; free_tx_desc()
408 d = q->sdesc; free_tx_desc()
495 struct rx_sw_desc *d = &q->sdesc[q->cidx]; free_rx_bufs()
522 struct rx_sw_desc *d = &q->sdesc[q->cidx]; unmap_rx_buf()
594 struct rx_sw_desc *sd = &q->sdesc[q->pidx]; refill_fl()
629 sd = q->sdesc; refill_fl()
657 sd = q->sdesc; refill_fl()
1275 q->q.sdesc[last_desc].skb = skb; t4_eth_xmit()
1276 q->q.sdesc[last_desc].sgl = (struct ulptx_sgl *)(cpl + 1); t4_eth_xmit()
1570 q->q.sdesc[last_desc].skb = skb; service_ofldq()
1949 d = &q->sdesc[q->cidx]; restore_rx_bufs()
2032 rsd = &rxq->fl.sdesc[rxq->fl.cidx]; process_responses()
2479 &fl->sdesc, s->stat_len, NUMA_NO_NODE); t4_sge_alloc_rxq()
2546 kfree(fl->sdesc); t4_sge_alloc_rxq()
2547 fl->sdesc = NULL; t4_sge_alloc_rxq()
2584 &txq->q.phys_addr, &txq->q.sdesc, s->stat_len, t4_sge_alloc_eth_txq()
2610 kfree(txq->q.sdesc); t4_sge_alloc_eth_txq()
2611 txq->q.sdesc = NULL; t4_sge_alloc_eth_txq()
2692 &txq->q.phys_addr, &txq->q.sdesc, s->stat_len, t4_sge_alloc_ofld_txq()
2716 kfree(txq->q.sdesc); t4_sge_alloc_ofld_txq()
2717 txq->q.sdesc = NULL; t4_sge_alloc_ofld_txq()
2742 q->sdesc = NULL; free_txq()
2767 kfree(fl->sdesc); free_rspq_fl()
2768 fl->sdesc = NULL; free_rspq_fl()
2811 kfree(etq->q.sdesc); t4_free_sge_resources()
2830 kfree(q->q.sdesc); t4_free_sge_resources()
H A Dcxgb4.h460 struct rx_sw_desc *sdesc; /* address of SW Rx descriptor ring */ member in struct:sge_fl
562 struct tx_sw_desc *sdesc; /* address of SW Tx descriptor ring */ member in struct:sge_txq
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/
H A Dadapter.h116 struct rx_sw_desc *sdesc; /* address of SW Rx descriptor ring */ member in struct:sge_fl
177 struct tx_sw_desc *sdesc; /* address of SW Tx descriptor ring */ member in struct:sge_txq
H A Dsge.c240 struct tx_sw_desc *d = &q->sdesc[cidx]; unmap_skb()
269 d = cidx + 1 == q->size ? q->sdesc : d + 1; unmap_skb()
295 d = &q->sdesc[cidx]; free_tx_desc()
308 d = q->sdesc; free_tx_desc()
385 struct rx_sw_desc *d = &q->sdesc[cidx]; free_rx_bufs()
502 struct rx_sw_desc *sd = &q->sdesc[q->pidx]; refill_fl()
545 sd = q->sdesc; refill_fl()
579 q->sdesc[q->pidx] = q->sdesc[idx]; recycle_rx_buf()
683 kfree(q->fl[i].sdesc); t3_free_qset()
695 if (q->txq[i].sdesc) { t3_free_qset()
698 kfree(q->txq[i].sdesc); t3_free_qset()
783 struct rx_sw_desc *sd = &fl->sdesc[fl->cidx]; get_packet()
843 struct rx_sw_desc *sd = &fl->sdesc[fl->cidx]; get_packet_pg()
1058 struct tx_sw_desc *sd = &q->sdesc[pidx]; write_wr_hdr_sgl()
1102 sd = q->sdesc; write_wr_hdr_sgl()
1176 q->sdesc[pidx].skb = NULL; write_tx_pkt_wr()
1588 q->sdesc[pidx].skb = NULL; write_ofld_wr()
2066 struct rx_sw_desc *sd = &fl->sdesc[fl->cidx]; lro_add_page()
2308 void *addr = fl->sdesc[fl->cidx].pg_chunk.va; process_responses()
2985 &q->fl[0].phys_addr, &q->fl[0].sdesc); t3_sge_alloc_qset()
2992 &q->fl[1].phys_addr, &q->fl[1].sdesc); t3_sge_alloc_qset()
3012 &q->txq[i].sdesc); t3_sge_alloc_qset()
/linux-4.1.27/fs/cifs/
H A Dsmbencrypt.c138 struct sdesc *sdescmd4; mdfour()
H A Dcifsglob.h110 struct sdesc { struct
121 struct sdesc *sdeschmacmd5; /* ctxt to generate ntlmv2 hash, CR1 */
122 struct sdesc *sdescmd5; /* ctxt to generate cifs/smb signature */
123 struct sdesc *sdeschmacsha256; /* ctxt to generate smb2 signature */
124 struct sdesc *sdesccmacaes; /* ctxt to generate smb3 signature */
H A Dlink.c62 struct sdesc *sdescmd5; symlink_hash()
/linux-4.1.27/drivers/infiniband/ulp/iser/
H A Diser_initiator.c178 iser_dbg("sdesc %p lkey mismatch, fixing\n", tx_desc); iser_create_send_desc()
/linux-4.1.27/drivers/net/ethernet/nvidia/
H A Dforcedeth.c6007 dev_info(&pci_dev->dev, "%s%s%s%s%s%s%s%s%s%s%sdesc-v%u\n", nv_probe()

Completed in 652 milliseconds