ccdf 47 arch/s390/pci/pci_event.c static void __zpci_event_error(struct zpci_ccdf_err *ccdf) ccdf 49 arch/s390/pci/pci_event.c struct zpci_dev *zdev = get_zdev_by_fid(ccdf->fid); ccdf 53 arch/s390/pci/pci_event.c zpci_err_hex(ccdf, sizeof(*ccdf)); ccdf 59 arch/s390/pci/pci_event.c pdev ? pci_name(pdev) : "n/a", ccdf->pec, ccdf->fid); ccdf 74 arch/s390/pci/pci_event.c static void __zpci_event_availability(struct zpci_ccdf_avail *ccdf) ccdf 76 arch/s390/pci/pci_event.c struct zpci_dev *zdev = get_zdev_by_fid(ccdf->fid); ccdf 85 arch/s390/pci/pci_event.c pdev ? pci_name(pdev) : "n/a", ccdf->pec, ccdf->fid); ccdf 87 arch/s390/pci/pci_event.c zpci_err_hex(ccdf, sizeof(*ccdf)); ccdf 89 arch/s390/pci/pci_event.c switch (ccdf->pec) { ccdf 92 arch/s390/pci/pci_event.c ret = clp_add_pci_device(ccdf->fid, ccdf->fh, 0); ccdf 95 arch/s390/pci/pci_event.c zdev = get_zdev_by_fid(ccdf->fid); ccdf 100 arch/s390/pci/pci_event.c zdev->fh = ccdf->fh; ccdf 110 arch/s390/pci/pci_event.c clp_add_pci_device(ccdf->fid, ccdf->fh, 0); ccdf 138 arch/s390/pci/pci_event.c zdev->fh = ccdf->fh; ccdf 141 arch/s390/pci/pci_event.c if (!clp_get_state(ccdf->fid, &state) && ccdf 219 drivers/dma/fsl-qdma.c qdma_ccdf_addr_get64(const struct fsl_qdma_format *ccdf) ccdf 221 drivers/dma/fsl-qdma.c return le64_to_cpu(ccdf->data) & (U64_MAX >> 24); ccdf 225 drivers/dma/fsl-qdma.c qdma_desc_addr_set64(struct fsl_qdma_format *ccdf, u64 addr) ccdf 227 drivers/dma/fsl-qdma.c ccdf->addr_hi = upper_32_bits(addr); ccdf 228 drivers/dma/fsl-qdma.c ccdf->addr_lo = cpu_to_le32(lower_32_bits(addr)); ccdf 232 drivers/dma/fsl-qdma.c qdma_ccdf_get_queue(const struct fsl_qdma_format *ccdf) ccdf 234 drivers/dma/fsl-qdma.c return ccdf->cfg8b_w1 & U8_MAX; ccdf 238 drivers/dma/fsl-qdma.c qdma_ccdf_get_offset(const struct fsl_qdma_format *ccdf) ccdf 240 drivers/dma/fsl-qdma.c return (le32_to_cpu(ccdf->cfg) & QDMA_CCDF_MASK) >> QDMA_CCDF_OFFSET; ccdf 244 drivers/dma/fsl-qdma.c qdma_ccdf_set_format(struct fsl_qdma_format *ccdf, int offset) ccdf 246 drivers/dma/fsl-qdma.c ccdf->cfg = cpu_to_le32(QDMA_CCDF_FOTMAT | offset); ccdf 250 drivers/dma/fsl-qdma.c qdma_ccdf_get_status(const struct fsl_qdma_format *ccdf) ccdf 252 drivers/dma/fsl-qdma.c return (le32_to_cpu(ccdf->status) & QDMA_CCDF_MASK) >> QDMA_CCDF_STATUS; ccdf 256 drivers/dma/fsl-qdma.c qdma_ccdf_set_ser(struct fsl_qdma_format *ccdf, int status) ccdf 258 drivers/dma/fsl-qdma.c ccdf->status = cpu_to_le32(QDMA_CCDF_SER | status); ccdf 347 drivers/dma/fsl-qdma.c struct fsl_qdma_format *ccdf, *csgf_desc, *csgf_src, *csgf_dest; ccdf 349 drivers/dma/fsl-qdma.c ccdf = fsl_comp->virt_addr; ccdf 359 drivers/dma/fsl-qdma.c qdma_desc_addr_set64(ccdf, fsl_comp->bus_addr + 16); ccdf 360 drivers/dma/fsl-qdma.c qdma_ccdf_set_format(ccdf, qdma_ccdf_get_offset(ccdf)); ccdf 361 drivers/dma/fsl-qdma.c qdma_ccdf_set_ser(ccdf, qdma_ccdf_get_status(ccdf)); ccdf 296 drivers/s390/cio/chsc.c u8 ccdf[PAGE_SIZE - 24 - 16]; /* content-code dependent field */ ccdf 305 drivers/s390/cio/chsc.c u8 ccdf[PAGE_SIZE - 24 - 56]; /* content-code dependent field */ ccdf 388 drivers/s390/cio/chsc.c struct lir *lir = (struct lir *) &sei_area->ccdf; ccdf 393 drivers/s390/cio/chsc.c sei_area->rs, sei_area->rsid, sei_area->ccdf[0]); ccdf 474 drivers/s390/cio/chsc.c data = sei_area->ccdf; ccdf 510 drivers/s390/cio/chsc.c data = (struct chp_config_data *) &(sei_area->ccdf); ccdf 573 drivers/s390/cio/chsc.c zpci_event_error(sei_area->ccdf); ccdf 576 drivers/s390/cio/chsc.c zpci_event_availability(sei_area->ccdf);