ring_cmd 682 drivers/crypto/hisilicon/sec/sec_drv.c struct sec_queue_ring_cmd *msg_ring = &queue->ring_cmd; ring_cmd 844 drivers/crypto/hisilicon/sec/sec_drv.c struct sec_queue_ring_cmd *msg_ring = &queue->ring_cmd; ring_cmd 859 drivers/crypto/hisilicon/sec/sec_drv.c struct sec_queue_ring_cmd *msg_ring = &queue->ring_cmd; ring_cmd 886 drivers/crypto/hisilicon/sec/sec_drv.c struct sec_queue_ring_cmd *msg_ring = &queue->ring_cmd; ring_cmd 906 drivers/crypto/hisilicon/sec/sec_drv.c sec_queue_cmdbase_addr(queue, queue->ring_cmd.paddr); ring_cmd 1080 drivers/crypto/hisilicon/sec/sec_drv.c struct sec_queue_ring_cmd *ring_cmd = &queue->ring_cmd; ring_cmd 1085 drivers/crypto/hisilicon/sec/sec_drv.c ring_cmd->vaddr = dma_alloc_coherent(dev, SEC_Q_CMD_SIZE, ring_cmd 1086 drivers/crypto/hisilicon/sec/sec_drv.c &ring_cmd->paddr, GFP_KERNEL); ring_cmd 1087 drivers/crypto/hisilicon/sec/sec_drv.c if (!ring_cmd->vaddr) ring_cmd 1090 drivers/crypto/hisilicon/sec/sec_drv.c atomic_set(&ring_cmd->used, 0); ring_cmd 1091 drivers/crypto/hisilicon/sec/sec_drv.c mutex_init(&ring_cmd->lock); ring_cmd 1092 drivers/crypto/hisilicon/sec/sec_drv.c ring_cmd->callback = sec_alg_callback; ring_cmd 1123 drivers/crypto/hisilicon/sec/sec_drv.c dma_free_coherent(dev, SEC_Q_CMD_SIZE, queue->ring_cmd.vaddr, ring_cmd 1124 drivers/crypto/hisilicon/sec/sec_drv.c queue->ring_cmd.paddr); ring_cmd 1137 drivers/crypto/hisilicon/sec/sec_drv.c dma_free_coherent(dev, SEC_Q_CMD_SIZE, queue->ring_cmd.vaddr, ring_cmd 1138 drivers/crypto/hisilicon/sec/sec_drv.c queue->ring_cmd.paddr); ring_cmd 339 drivers/crypto/hisilicon/sec/sec_drv.h struct sec_queue_ring_cmd ring_cmd; ring_cmd 289 drivers/net/fddi/defza.c struct fza_ring_cmd __iomem *ring = fp->ring_cmd + fp->ring_cmd_index; ring_cmd 1346 drivers/net/fddi/defza.c fp->ring_cmd = mmio + FZA_RING_CMD; ring_cmd 677 drivers/net/fddi/defza.h struct fza_ring_cmd __iomem *ring_cmd;