sba 13 arch/alpha/include/asm/agp_backend.h u32 sba : 1; sba 1015 arch/alpha/kernel/core_marvel.c agp->mode.bits.sba ? " - SBA" : "", agp->mode.bits.rq); sba 649 arch/alpha/kernel/core_titan.c pctl.pctl_r_bits.apctl_v_agp_sba_en = agp->mode.bits.sba; sba 789 arch/alpha/kernel/core_titan.c agp->capability.bits.sba = 1; sba 798 arch/alpha/kernel/core_titan.c agp->mode.bits.sba = pctl.pctl_r_bits.apctl_v_agp_sba_en; sba 151 arch/s390/include/asm/qdio.h u64 sba[QDIO_MAX_ELEMENTS_PER_BUFFER]; sba 646 drivers/atm/fore200e.c u32 hcr = fore200e->bus->read(fore200e->regs.sba.hcr) & SBA200E_HCR_STICKY; sba 647 drivers/atm/fore200e.c fore200e->bus->write(hcr | SBA200E_HCR_INTR_ENA, fore200e->regs.sba.hcr); sba 652 drivers/atm/fore200e.c return fore200e->bus->read(fore200e->regs.sba.hcr) & SBA200E_HCR_INTR_REQ; sba 657 drivers/atm/fore200e.c u32 hcr = fore200e->bus->read(fore200e->regs.sba.hcr) & SBA200E_HCR_STICKY; sba 658 drivers/atm/fore200e.c fore200e->bus->write(hcr | SBA200E_HCR_INTR_CLR, fore200e->regs.sba.hcr); sba 663 drivers/atm/fore200e.c fore200e->bus->write(SBA200E_HCR_RESET, fore200e->regs.sba.hcr); sba 665 drivers/atm/fore200e.c fore200e->bus->write(0, fore200e->regs.sba.hcr); sba 674 drivers/atm/fore200e.c fore200e->regs.sba.hcr = of_ioremap(&op->resource[0], 0, SBA200E_HCR_LENGTH, "SBA HCR"); sba 675 drivers/atm/fore200e.c fore200e->regs.sba.bsr = of_ioremap(&op->resource[1], 0, SBA200E_BSR_LENGTH, "SBA BSR"); sba 676 drivers/atm/fore200e.c fore200e->regs.sba.isr = of_ioremap(&op->resource[2], 0, SBA200E_ISR_LENGTH, "SBA ISR"); sba 686 drivers/atm/fore200e.c fore200e->bus->write(0x02, fore200e->regs.sba.isr); /* XXX hardwired interrupt level */ sba 702 drivers/atm/fore200e.c of_iounmap(&op->resource[0], fore200e->regs.sba.hcr, SBA200E_HCR_LENGTH); sba 703 drivers/atm/fore200e.c of_iounmap(&op->resource[1], fore200e->regs.sba.bsr, SBA200E_BSR_LENGTH); sba 704 drivers/atm/fore200e.c of_iounmap(&op->resource[2], fore200e->regs.sba.isr, SBA200E_ISR_LENGTH); sba 792 drivers/atm/fore200e.h struct fore200e_sba_regs sba; /* SBA-200E registers */ sba 387 drivers/char/agp/parisc-agp.c struct parisc_device *sba = NULL, *lba = NULL; sba 394 drivers/char/agp/parisc-agp.c sba = sba_list->dev; sba 395 drivers/char/agp/parisc-agp.c if (!IS_PLUTO(sba)) { sba 401 drivers/char/agp/parisc-agp.c device_for_each_child(&sba->dev, &lba, find_quicksilver); sba 113 drivers/dma/bcm-sba-raid.c struct sba_device *sba; sba 203 drivers/dma/bcm-sba-raid.c static struct sba_request *sba_alloc_request(struct sba_device *sba) sba 209 drivers/dma/bcm-sba-raid.c spin_lock_irqsave(&sba->reqs_lock, flags); sba 210 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_free_list, node) { sba 212 drivers/dma/bcm-sba-raid.c list_move_tail(&req->node, &sba->reqs_alloc_list); sba 217 drivers/dma/bcm-sba-raid.c spin_unlock_irqrestore(&sba->reqs_lock, flags); sba 226 drivers/dma/bcm-sba-raid.c mbox_client_peek_data(sba->mchan); sba 235 drivers/dma/bcm-sba-raid.c dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan); sba 242 drivers/dma/bcm-sba-raid.c static void _sba_pending_request(struct sba_device *sba, sba 245 drivers/dma/bcm-sba-raid.c lockdep_assert_held(&sba->reqs_lock); sba 248 drivers/dma/bcm-sba-raid.c list_move_tail(&req->node, &sba->reqs_pending_list); sba 249 drivers/dma/bcm-sba-raid.c if (list_empty(&sba->reqs_active_list)) sba 250 drivers/dma/bcm-sba-raid.c sba->reqs_fence = false; sba 254 drivers/dma/bcm-sba-raid.c static bool _sba_active_request(struct sba_device *sba, sba 257 drivers/dma/bcm-sba-raid.c lockdep_assert_held(&sba->reqs_lock); sba 258 drivers/dma/bcm-sba-raid.c if (list_empty(&sba->reqs_active_list)) sba 259 drivers/dma/bcm-sba-raid.c sba->reqs_fence = false; sba 260 drivers/dma/bcm-sba-raid.c if (sba->reqs_fence) sba 264 drivers/dma/bcm-sba-raid.c list_move_tail(&req->node, &sba->reqs_active_list); sba 266 drivers/dma/bcm-sba-raid.c sba->reqs_fence = true; sba 271 drivers/dma/bcm-sba-raid.c static void _sba_abort_request(struct sba_device *sba, sba 274 drivers/dma/bcm-sba-raid.c lockdep_assert_held(&sba->reqs_lock); sba 277 drivers/dma/bcm-sba-raid.c list_move_tail(&req->node, &sba->reqs_aborted_list); sba 278 drivers/dma/bcm-sba-raid.c if (list_empty(&sba->reqs_active_list)) sba 279 drivers/dma/bcm-sba-raid.c sba->reqs_fence = false; sba 283 drivers/dma/bcm-sba-raid.c static void _sba_free_request(struct sba_device *sba, sba 286 drivers/dma/bcm-sba-raid.c lockdep_assert_held(&sba->reqs_lock); sba 289 drivers/dma/bcm-sba-raid.c list_move_tail(&req->node, &sba->reqs_free_list); sba 290 drivers/dma/bcm-sba-raid.c if (list_empty(&sba->reqs_active_list)) sba 291 drivers/dma/bcm-sba-raid.c sba->reqs_fence = false; sba 298 drivers/dma/bcm-sba-raid.c struct sba_device *sba = req->sba; sba 300 drivers/dma/bcm-sba-raid.c spin_lock_irqsave(&sba->reqs_lock, flags); sba 302 drivers/dma/bcm-sba-raid.c _sba_free_request(sba, req); sba 304 drivers/dma/bcm-sba-raid.c _sba_free_request(sba, nreq); sba 306 drivers/dma/bcm-sba-raid.c spin_unlock_irqrestore(&sba->reqs_lock, flags); sba 313 drivers/dma/bcm-sba-raid.c struct sba_device *sba = req->sba; sba 315 drivers/dma/bcm-sba-raid.c spin_lock_irqsave(&sba->reqs_lock, flags); sba 321 drivers/dma/bcm-sba-raid.c spin_unlock_irqrestore(&sba->reqs_lock, flags); sba 324 drivers/dma/bcm-sba-raid.c static void sba_cleanup_nonpending_requests(struct sba_device *sba) sba 329 drivers/dma/bcm-sba-raid.c spin_lock_irqsave(&sba->reqs_lock, flags); sba 332 drivers/dma/bcm-sba-raid.c list_for_each_entry_safe(req, req1, &sba->reqs_alloc_list, node) sba 333 drivers/dma/bcm-sba-raid.c _sba_free_request(sba, req); sba 336 drivers/dma/bcm-sba-raid.c list_for_each_entry_safe(req, req1, &sba->reqs_active_list, node) sba 337 drivers/dma/bcm-sba-raid.c _sba_abort_request(sba, req); sba 344 drivers/dma/bcm-sba-raid.c spin_unlock_irqrestore(&sba->reqs_lock, flags); sba 347 drivers/dma/bcm-sba-raid.c static void sba_cleanup_pending_requests(struct sba_device *sba) sba 352 drivers/dma/bcm-sba-raid.c spin_lock_irqsave(&sba->reqs_lock, flags); sba 355 drivers/dma/bcm-sba-raid.c list_for_each_entry_safe(req, req1, &sba->reqs_pending_list, node) sba 356 drivers/dma/bcm-sba-raid.c _sba_free_request(sba, req); sba 358 drivers/dma/bcm-sba-raid.c spin_unlock_irqrestore(&sba->reqs_lock, flags); sba 361 drivers/dma/bcm-sba-raid.c static int sba_send_mbox_request(struct sba_device *sba, sba 368 drivers/dma/bcm-sba-raid.c ret = mbox_send_message(sba->mchan, &req->msg); sba 370 drivers/dma/bcm-sba-raid.c dev_err(sba->dev, "send message failed with error %d", ret); sba 377 drivers/dma/bcm-sba-raid.c dev_err(sba->dev, "message error %d", ret); sba 381 drivers/dma/bcm-sba-raid.c mbox_client_txdone(sba->mchan, ret); sba 387 drivers/dma/bcm-sba-raid.c static void _sba_process_pending_requests(struct sba_device *sba) sba 395 drivers/dma/bcm-sba-raid.c while (!list_empty(&sba->reqs_pending_list) && count) { sba 397 drivers/dma/bcm-sba-raid.c req = list_first_entry(&sba->reqs_pending_list, sba 401 drivers/dma/bcm-sba-raid.c if (!_sba_active_request(sba, req)) sba 405 drivers/dma/bcm-sba-raid.c ret = sba_send_mbox_request(sba, req); sba 407 drivers/dma/bcm-sba-raid.c _sba_pending_request(sba, req); sba 415 drivers/dma/bcm-sba-raid.c static void sba_process_received_request(struct sba_device *sba, sba 428 drivers/dma/bcm-sba-raid.c spin_lock_irqsave(&sba->reqs_lock, flags); sba 430 drivers/dma/bcm-sba-raid.c spin_unlock_irqrestore(&sba->reqs_lock, flags); sba 439 drivers/dma/bcm-sba-raid.c spin_lock_irqsave(&sba->reqs_lock, flags); sba 443 drivers/dma/bcm-sba-raid.c _sba_free_request(sba, nreq); sba 447 drivers/dma/bcm-sba-raid.c _sba_free_request(sba, first); sba 450 drivers/dma/bcm-sba-raid.c _sba_process_pending_requests(sba); sba 452 drivers/dma/bcm-sba-raid.c spin_unlock_irqrestore(&sba->reqs_lock, flags); sba 456 drivers/dma/bcm-sba-raid.c static void sba_write_stats_in_seqfile(struct sba_device *sba, sba 464 drivers/dma/bcm-sba-raid.c spin_lock_irqsave(&sba->reqs_lock, flags); sba 466 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_free_list, node) sba 470 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_alloc_list, node) sba 473 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_pending_list, node) sba 476 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_active_list, node) sba 479 drivers/dma/bcm-sba-raid.c list_for_each_entry(req, &sba->reqs_aborted_list, node) sba 482 drivers/dma/bcm-sba-raid.c spin_unlock_irqrestore(&sba->reqs_lock, flags); sba 484 drivers/dma/bcm-sba-raid.c seq_printf(file, "maximum requests = %d\n", sba->max_req); sba 515 drivers/dma/bcm-sba-raid.c struct sba_device *sba = to_sba_device(dchan); sba 518 drivers/dma/bcm-sba-raid.c spin_lock_irqsave(&sba->reqs_lock, flags); sba 519 drivers/dma/bcm-sba-raid.c _sba_process_pending_requests(sba); sba 520 drivers/dma/bcm-sba-raid.c spin_unlock_irqrestore(&sba->reqs_lock, flags); sba 527 drivers/dma/bcm-sba-raid.c struct sba_device *sba; sba 533 drivers/dma/bcm-sba-raid.c sba = to_sba_device(tx->chan); sba 537 drivers/dma/bcm-sba-raid.c spin_lock_irqsave(&sba->reqs_lock, flags); sba 539 drivers/dma/bcm-sba-raid.c _sba_pending_request(sba, req); sba 541 drivers/dma/bcm-sba-raid.c _sba_pending_request(sba, nreq); sba 542 drivers/dma/bcm-sba-raid.c spin_unlock_irqrestore(&sba->reqs_lock, flags); sba 552 drivers/dma/bcm-sba-raid.c struct sba_device *sba = to_sba_device(dchan); sba 558 drivers/dma/bcm-sba-raid.c mbox_client_peek_data(sba->mchan); sba 575 drivers/dma/bcm-sba-raid.c cmd = sba_cmd_enc(cmd, req->sba->hw_resp_size, sba 586 drivers/dma/bcm-sba-raid.c cmdsp->data_len = req->sba->hw_resp_size; sba 592 drivers/dma/bcm-sba-raid.c cmd = sba_cmd_enc(cmd, req->sba->hw_resp_size, sba 604 drivers/dma/bcm-sba-raid.c if (req->sba->hw_resp_size) { sba 607 drivers/dma/bcm-sba-raid.c cmdsp->resp_len = req->sba->hw_resp_size; sba 611 drivers/dma/bcm-sba-raid.c cmdsp->data_len = req->sba->hw_resp_size; sba 616 drivers/dma/bcm-sba-raid.c msg->sba.cmds = cmds; sba 617 drivers/dma/bcm-sba-raid.c msg->sba.cmds_count = cmdsp - cmds; sba 626 drivers/dma/bcm-sba-raid.c struct sba_device *sba = to_sba_device(dchan); sba 629 drivers/dma/bcm-sba-raid.c req = sba_alloc_request(sba); sba 692 drivers/dma/bcm-sba-raid.c if (req->sba->hw_resp_size) { sba 695 drivers/dma/bcm-sba-raid.c cmdsp->resp_len = req->sba->hw_resp_size; sba 704 drivers/dma/bcm-sba-raid.c msg->sba.cmds = cmds; sba 705 drivers/dma/bcm-sba-raid.c msg->sba.cmds_count = cmdsp - cmds; sba 711 drivers/dma/bcm-sba-raid.c sba_prep_dma_memcpy_req(struct sba_device *sba, sba 718 drivers/dma/bcm-sba-raid.c req = sba_alloc_request(sba); sba 741 drivers/dma/bcm-sba-raid.c struct sba_device *sba = to_sba_device(dchan); sba 746 drivers/dma/bcm-sba-raid.c req_len = (len < sba->hw_buf_size) ? len : sba->hw_buf_size; sba 748 drivers/dma/bcm-sba-raid.c req = sba_prep_dma_memcpy_req(sba, off, dst, src, sba 831 drivers/dma/bcm-sba-raid.c if (req->sba->hw_resp_size) { sba 834 drivers/dma/bcm-sba-raid.c cmdsp->resp_len = req->sba->hw_resp_size; sba 843 drivers/dma/bcm-sba-raid.c msg->sba.cmds = cmds; sba 844 drivers/dma/bcm-sba-raid.c msg->sba.cmds_count = cmdsp - cmds; sba 850 drivers/dma/bcm-sba-raid.c sba_prep_dma_xor_req(struct sba_device *sba, sba 857 drivers/dma/bcm-sba-raid.c req = sba_alloc_request(sba); sba 880 drivers/dma/bcm-sba-raid.c struct sba_device *sba = to_sba_device(dchan); sba 884 drivers/dma/bcm-sba-raid.c if (unlikely(src_cnt > sba->max_xor_srcs)) sba 889 drivers/dma/bcm-sba-raid.c req_len = (len < sba->hw_buf_size) ? len : sba->hw_buf_size; sba 891 drivers/dma/bcm-sba-raid.c req = sba_prep_dma_xor_req(sba, off, dst, src, src_cnt, sba 1014 drivers/dma/bcm-sba-raid.c if (req->sba->hw_resp_size) { sba 1017 drivers/dma/bcm-sba-raid.c cmdsp->resp_len = req->sba->hw_resp_size; sba 1041 drivers/dma/bcm-sba-raid.c if (req->sba->hw_resp_size) { sba 1044 drivers/dma/bcm-sba-raid.c cmdsp->resp_len = req->sba->hw_resp_size; sba 1054 drivers/dma/bcm-sba-raid.c msg->sba.cmds = cmds; sba 1055 drivers/dma/bcm-sba-raid.c msg->sba.cmds_count = cmdsp - cmds; sba 1061 drivers/dma/bcm-sba-raid.c sba_prep_dma_pq_req(struct sba_device *sba, dma_addr_t off, sba 1068 drivers/dma/bcm-sba-raid.c req = sba_alloc_request(sba); sba 1174 drivers/dma/bcm-sba-raid.c if (req->sba->hw_resp_size) { sba 1177 drivers/dma/bcm-sba-raid.c cmdsp->resp_len = req->sba->hw_resp_size; sba 1202 drivers/dma/bcm-sba-raid.c pos = (dpos < req->sba->max_pq_coefs) ? sba 1203 drivers/dma/bcm-sba-raid.c dpos : (req->sba->max_pq_coefs - 1); sba 1231 drivers/dma/bcm-sba-raid.c pos = (dpos < req->sba->max_pq_coefs) ? sba 1232 drivers/dma/bcm-sba-raid.c dpos : (req->sba->max_pq_coefs - 1); sba 1295 drivers/dma/bcm-sba-raid.c if (req->sba->hw_resp_size) { sba 1298 drivers/dma/bcm-sba-raid.c cmdsp->resp_len = req->sba->hw_resp_size; sba 1308 drivers/dma/bcm-sba-raid.c msg->sba.cmds = cmds; sba 1309 drivers/dma/bcm-sba-raid.c msg->sba.cmds_count = cmdsp - cmds; sba 1315 drivers/dma/bcm-sba-raid.c sba_prep_dma_pq_single_req(struct sba_device *sba, dma_addr_t off, sba 1323 drivers/dma/bcm-sba-raid.c req = sba_alloc_request(sba); sba 1350 drivers/dma/bcm-sba-raid.c struct sba_device *sba = to_sba_device(dchan); sba 1354 drivers/dma/bcm-sba-raid.c if (unlikely(src_cnt > sba->max_pq_srcs)) sba 1357 drivers/dma/bcm-sba-raid.c if (sba->max_pq_coefs <= raid6_gflog[scf[i]]) sba 1368 drivers/dma/bcm-sba-raid.c req_len = (len < sba->hw_buf_size) ? len : sba->hw_buf_size; sba 1384 drivers/dma/bcm-sba-raid.c req = sba_prep_dma_pq_single_req(sba, sba 1402 drivers/dma/bcm-sba-raid.c req = sba_prep_dma_pq_single_req(sba, sba 1416 drivers/dma/bcm-sba-raid.c req = sba_prep_dma_pq_req(sba, off, sba 1446 drivers/dma/bcm-sba-raid.c struct sba_device *sba = req->sba; sba 1450 drivers/dma/bcm-sba-raid.c dev_err(sba->dev, "%s got message with error %d", sba 1451 drivers/dma/bcm-sba-raid.c dma_chan_name(&sba->dma_chan), m->error); sba 1454 drivers/dma/bcm-sba-raid.c sba_process_received_request(sba, req); sba 1461 drivers/dma/bcm-sba-raid.c struct sba_device *sba = dev_get_drvdata(file->private); sba 1464 drivers/dma/bcm-sba-raid.c sba_write_stats_in_seqfile(sba, file); sba 1471 drivers/dma/bcm-sba-raid.c static int sba_prealloc_channel_resources(struct sba_device *sba) sba 1476 drivers/dma/bcm-sba-raid.c sba->resp_base = dma_alloc_coherent(sba->mbox_dev, sba 1477 drivers/dma/bcm-sba-raid.c sba->max_resp_pool_size, sba 1478 drivers/dma/bcm-sba-raid.c &sba->resp_dma_base, GFP_KERNEL); sba 1479 drivers/dma/bcm-sba-raid.c if (!sba->resp_base) sba 1482 drivers/dma/bcm-sba-raid.c sba->cmds_base = dma_alloc_coherent(sba->mbox_dev, sba 1483 drivers/dma/bcm-sba-raid.c sba->max_cmds_pool_size, sba 1484 drivers/dma/bcm-sba-raid.c &sba->cmds_dma_base, GFP_KERNEL); sba 1485 drivers/dma/bcm-sba-raid.c if (!sba->cmds_base) { sba 1490 drivers/dma/bcm-sba-raid.c spin_lock_init(&sba->reqs_lock); sba 1491 drivers/dma/bcm-sba-raid.c sba->reqs_fence = false; sba 1492 drivers/dma/bcm-sba-raid.c INIT_LIST_HEAD(&sba->reqs_alloc_list); sba 1493 drivers/dma/bcm-sba-raid.c INIT_LIST_HEAD(&sba->reqs_pending_list); sba 1494 drivers/dma/bcm-sba-raid.c INIT_LIST_HEAD(&sba->reqs_active_list); sba 1495 drivers/dma/bcm-sba-raid.c INIT_LIST_HEAD(&sba->reqs_aborted_list); sba 1496 drivers/dma/bcm-sba-raid.c INIT_LIST_HEAD(&sba->reqs_free_list); sba 1498 drivers/dma/bcm-sba-raid.c for (i = 0; i < sba->max_req; i++) { sba 1499 drivers/dma/bcm-sba-raid.c req = devm_kzalloc(sba->dev, sba 1500 drivers/dma/bcm-sba-raid.c struct_size(req, cmds, sba->max_cmd_per_req), sba 1507 drivers/dma/bcm-sba-raid.c req->sba = sba; sba 1511 drivers/dma/bcm-sba-raid.c for (j = 0; j < sba->max_cmd_per_req; j++) { sba 1513 drivers/dma/bcm-sba-raid.c req->cmds[j].cmd_dma = sba->cmds_base + sba 1514 drivers/dma/bcm-sba-raid.c (i * sba->max_cmd_per_req + j) * sizeof(u64); sba 1515 drivers/dma/bcm-sba-raid.c req->cmds[j].cmd_dma_addr = sba->cmds_dma_base + sba 1516 drivers/dma/bcm-sba-raid.c (i * sba->max_cmd_per_req + j) * sizeof(u64); sba 1520 drivers/dma/bcm-sba-raid.c dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan); sba 1523 drivers/dma/bcm-sba-raid.c req->tx.phys = sba->resp_dma_base + i * sba->hw_resp_size; sba 1524 drivers/dma/bcm-sba-raid.c list_add_tail(&req->node, &sba->reqs_free_list); sba 1530 drivers/dma/bcm-sba-raid.c dma_free_coherent(sba->mbox_dev, sba 1531 drivers/dma/bcm-sba-raid.c sba->max_cmds_pool_size, sba 1532 drivers/dma/bcm-sba-raid.c sba->cmds_base, sba->cmds_dma_base); sba 1534 drivers/dma/bcm-sba-raid.c dma_free_coherent(sba->mbox_dev, sba 1535 drivers/dma/bcm-sba-raid.c sba->max_resp_pool_size, sba 1536 drivers/dma/bcm-sba-raid.c sba->resp_base, sba->resp_dma_base); sba 1540 drivers/dma/bcm-sba-raid.c static void sba_freeup_channel_resources(struct sba_device *sba) sba 1542 drivers/dma/bcm-sba-raid.c dmaengine_terminate_all(&sba->dma_chan); sba 1543 drivers/dma/bcm-sba-raid.c dma_free_coherent(sba->mbox_dev, sba->max_cmds_pool_size, sba 1544 drivers/dma/bcm-sba-raid.c sba->cmds_base, sba->cmds_dma_base); sba 1545 drivers/dma/bcm-sba-raid.c dma_free_coherent(sba->mbox_dev, sba->max_resp_pool_size, sba 1546 drivers/dma/bcm-sba-raid.c sba->resp_base, sba->resp_dma_base); sba 1547 drivers/dma/bcm-sba-raid.c sba->resp_base = NULL; sba 1548 drivers/dma/bcm-sba-raid.c sba->resp_dma_base = 0; sba 1551 drivers/dma/bcm-sba-raid.c static int sba_async_register(struct sba_device *sba) sba 1554 drivers/dma/bcm-sba-raid.c struct dma_device *dma_dev = &sba->dma_dev; sba 1557 drivers/dma/bcm-sba-raid.c sba->dma_chan.device = dma_dev; sba 1558 drivers/dma/bcm-sba-raid.c dma_cookie_init(&sba->dma_chan); sba 1572 drivers/dma/bcm-sba-raid.c dma_dev->dev = sba->mbox_dev; sba 1591 drivers/dma/bcm-sba-raid.c dma_dev->max_xor = sba->max_xor_srcs; sba 1597 drivers/dma/bcm-sba-raid.c dma_set_maxpq(dma_dev, sba->max_pq_srcs, 0); sba 1602 drivers/dma/bcm-sba-raid.c list_add_tail(&sba->dma_chan.device_node, &dma_dev->channels); sba 1607 drivers/dma/bcm-sba-raid.c dev_err(sba->dev, "async device register error %d", ret); sba 1611 drivers/dma/bcm-sba-raid.c dev_info(sba->dev, "%s capabilities: %s%s%s%s\n", sba 1612 drivers/dma/bcm-sba-raid.c dma_chan_name(&sba->dma_chan), sba 1624 drivers/dma/bcm-sba-raid.c struct sba_device *sba; sba 1629 drivers/dma/bcm-sba-raid.c sba = devm_kzalloc(&pdev->dev, sizeof(*sba), GFP_KERNEL); sba 1630 drivers/dma/bcm-sba-raid.c if (!sba) sba 1633 drivers/dma/bcm-sba-raid.c sba->dev = &pdev->dev; sba 1634 drivers/dma/bcm-sba-raid.c platform_set_drvdata(pdev, sba); sba 1643 drivers/dma/bcm-sba-raid.c if (of_device_is_compatible(sba->dev->of_node, "brcm,iproc-sba")) sba 1644 drivers/dma/bcm-sba-raid.c sba->ver = SBA_VER_1; sba 1645 drivers/dma/bcm-sba-raid.c else if (of_device_is_compatible(sba->dev->of_node, sba 1647 drivers/dma/bcm-sba-raid.c sba->ver = SBA_VER_2; sba 1652 drivers/dma/bcm-sba-raid.c switch (sba->ver) { sba 1654 drivers/dma/bcm-sba-raid.c sba->hw_buf_size = 4096; sba 1655 drivers/dma/bcm-sba-raid.c sba->hw_resp_size = 8; sba 1656 drivers/dma/bcm-sba-raid.c sba->max_pq_coefs = 6; sba 1657 drivers/dma/bcm-sba-raid.c sba->max_pq_srcs = 6; sba 1660 drivers/dma/bcm-sba-raid.c sba->hw_buf_size = 4096; sba 1661 drivers/dma/bcm-sba-raid.c sba->hw_resp_size = 8; sba 1662 drivers/dma/bcm-sba-raid.c sba->max_pq_coefs = 30; sba 1668 drivers/dma/bcm-sba-raid.c sba->max_pq_srcs = 12; sba 1673 drivers/dma/bcm-sba-raid.c sba->max_req = SBA_MAX_REQ_PER_MBOX_CHANNEL; sba 1674 drivers/dma/bcm-sba-raid.c sba->max_cmd_per_req = sba->max_pq_srcs + 3; sba 1675 drivers/dma/bcm-sba-raid.c sba->max_xor_srcs = sba->max_cmd_per_req - 1; sba 1676 drivers/dma/bcm-sba-raid.c sba->max_resp_pool_size = sba->max_req * sba->hw_resp_size; sba 1677 drivers/dma/bcm-sba-raid.c sba->max_cmds_pool_size = sba->max_req * sba 1678 drivers/dma/bcm-sba-raid.c sba->max_cmd_per_req * sizeof(u64); sba 1681 drivers/dma/bcm-sba-raid.c sba->client.dev = &pdev->dev; sba 1682 drivers/dma/bcm-sba-raid.c sba->client.rx_callback = sba_receive_message; sba 1683 drivers/dma/bcm-sba-raid.c sba->client.tx_block = false; sba 1684 drivers/dma/bcm-sba-raid.c sba->client.knows_txdone = true; sba 1685 drivers/dma/bcm-sba-raid.c sba->client.tx_tout = 0; sba 1688 drivers/dma/bcm-sba-raid.c sba->mchan = mbox_request_channel(&sba->client, 0); sba 1689 drivers/dma/bcm-sba-raid.c if (IS_ERR(sba->mchan)) { sba 1690 drivers/dma/bcm-sba-raid.c ret = PTR_ERR(sba->mchan); sba 1705 drivers/dma/bcm-sba-raid.c sba->mbox_dev = &mbox_pdev->dev; sba 1708 drivers/dma/bcm-sba-raid.c ret = sba_prealloc_channel_resources(sba); sba 1717 drivers/dma/bcm-sba-raid.c sba->root = debugfs_create_dir(dev_name(sba->dev), NULL); sba 1720 drivers/dma/bcm-sba-raid.c debugfs_create_devm_seqfile(sba->dev, "stats", sba->root, sba 1726 drivers/dma/bcm-sba-raid.c ret = sba_async_register(sba); sba 1731 drivers/dma/bcm-sba-raid.c dev_info(sba->dev, "%s using SBAv%d mailbox channel from %s", sba 1732 drivers/dma/bcm-sba-raid.c dma_chan_name(&sba->dma_chan), sba->ver+1, sba 1733 drivers/dma/bcm-sba-raid.c dev_name(sba->mbox_dev)); sba 1738 drivers/dma/bcm-sba-raid.c debugfs_remove_recursive(sba->root); sba 1739 drivers/dma/bcm-sba-raid.c sba_freeup_channel_resources(sba); sba 1741 drivers/dma/bcm-sba-raid.c mbox_free_channel(sba->mchan); sba 1747 drivers/dma/bcm-sba-raid.c struct sba_device *sba = platform_get_drvdata(pdev); sba 1749 drivers/dma/bcm-sba-raid.c dma_async_device_unregister(&sba->dma_dev); sba 1751 drivers/dma/bcm-sba-raid.c debugfs_remove_recursive(sba->root); sba 1753 drivers/dma/bcm-sba-raid.c sba_freeup_channel_resources(sba); sba 1755 drivers/dma/bcm-sba-raid.c mbox_free_channel(sba->mchan); sba 719 drivers/mailbox/bcm-flexrm-mailbox.c if (!msg->sba.cmds || !msg->sba.cmds_count) sba 722 drivers/mailbox/bcm-flexrm-mailbox.c for (i = 0; i < msg->sba.cmds_count; i++) { sba 723 drivers/mailbox/bcm-flexrm-mailbox.c if (((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_B) || sba 724 drivers/mailbox/bcm-flexrm-mailbox.c (msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_C)) && sba 725 drivers/mailbox/bcm-flexrm-mailbox.c (msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_OUTPUT)) sba 727 drivers/mailbox/bcm-flexrm-mailbox.c if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_B) && sba 728 drivers/mailbox/bcm-flexrm-mailbox.c (msg->sba.cmds[i].data_len > SRCT_LENGTH_MASK)) sba 730 drivers/mailbox/bcm-flexrm-mailbox.c if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_C) && sba 731 drivers/mailbox/bcm-flexrm-mailbox.c (msg->sba.cmds[i].data_len > SRCT_LENGTH_MASK)) sba 733 drivers/mailbox/bcm-flexrm-mailbox.c if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_RESP) && sba 734 drivers/mailbox/bcm-flexrm-mailbox.c (msg->sba.cmds[i].resp_len > DSTT_LENGTH_MASK)) sba 736 drivers/mailbox/bcm-flexrm-mailbox.c if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_OUTPUT) && sba 737 drivers/mailbox/bcm-flexrm-mailbox.c (msg->sba.cmds[i].data_len > DSTT_LENGTH_MASK)) sba 749 drivers/mailbox/bcm-flexrm-mailbox.c for (i = 0; i < msg->sba.cmds_count; i++) { sba 752 drivers/mailbox/bcm-flexrm-mailbox.c if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_B) || sba 753 drivers/mailbox/bcm-flexrm-mailbox.c (msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_C)) sba 756 drivers/mailbox/bcm-flexrm-mailbox.c if (msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_RESP) sba 759 drivers/mailbox/bcm-flexrm-mailbox.c if (msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_OUTPUT) sba 776 drivers/mailbox/bcm-flexrm-mailbox.c for (i = 0; i < msg->sba.cmds_count; i++) { sba 777 drivers/mailbox/bcm-flexrm-mailbox.c c = &msg->sba.cmds[i]; sba 611 drivers/net/fddi/skfp/h/cmtdef.h void sba(); sba 455 drivers/net/fddi/skfp/h/smc.h struct s_sba sba ; /* SBA variables */ sba 514 drivers/net/fddi/skfp/smt.c smc->sba.sm = sm ; sba 515 drivers/net/fddi/skfp/smt.c sba(smc,NIF) ; sba 1544 drivers/net/wireless/ath/ath6kl/debug.c pstream.sba = cpu_to_le32(val32); sba 1178 drivers/net/wireless/ath/ath6kl/wmi.h __le32 sba; sba 1235 drivers/parisc/sba_iommu.c setup_ibase_imask(struct parisc_device *sba, struct ioc *ioc, int ioc_num) sba 1242 drivers/parisc/sba_iommu.c device_for_each_child(&sba->dev, &ibase_data, sba 1260 drivers/parisc/sba_iommu.c sba_ioc_init_pluto(struct parisc_device *sba, struct ioc *ioc, int ioc_num) sba 1322 drivers/parisc/sba_iommu.c setup_ibase_imask(sba, ioc, ioc_num); sba 1369 drivers/parisc/sba_iommu.c device_for_each_child(&sba->dev, &agp_found, sba_ioc_find_quicksilver); sba 1381 drivers/parisc/sba_iommu.c sba_ioc_init(struct parisc_device *sba, struct ioc *ioc, int ioc_num) sba 1468 drivers/parisc/sba_iommu.c setup_ibase_imask(sba, ioc, ioc_num); sba 1994 drivers/parisc/sba_iommu.c struct sba_device *sba = dev_get_drvdata(&sba_dev->dev); sba 2000 drivers/parisc/sba_iommu.c return &(sba->ioc[iocnum]); sba 2015 drivers/parisc/sba_iommu.c struct sba_device *sba = dev_get_drvdata(&sba_dev->dev); sba 2027 drivers/parisc/sba_iommu.c void __iomem *reg = sba->sba_hpa + i*0x18; sba 2058 drivers/parisc/sba_iommu.c struct sba_device *sba = dev_get_drvdata(&sba_dev->dev); sba 2067 drivers/parisc/sba_iommu.c base = READ_REG32(sba->sba_hpa + LMMIO_DIST_BASE); sba 2075 drivers/parisc/sba_iommu.c size = (~READ_REG32(sba->sba_hpa + LMMIO_DIST_MASK)) / ROPES_PER_IOC; sba 372 drivers/s390/char/tty3270.c char *sba, *str; sba 415 drivers/s390/char/tty3270.c sba = invalid_sba; sba 427 drivers/s390/char/tty3270.c if (s->string[1] == sba[0] && s->string[2] == sba[1]) sba 433 drivers/s390/char/tty3270.c sba = s->string + s->len - 3; sba 435 drivers/s390/char/tty3270.c sba = invalid_sba; sba 52 drivers/s390/net/ism.h u64 sba; sba 90 drivers/s390/net/ism.h u64 sba; sba 198 drivers/s390/net/ism.h struct ism_sba *sba; sba 89 drivers/s390/net/ism_drv.c struct ism_sba *sba; sba 91 drivers/s390/net/ism_drv.c sba = dma_alloc_coherent(&ism->pdev->dev, PAGE_SIZE, &dma_handle, sba 93 drivers/s390/net/ism_drv.c if (!sba) sba 99 drivers/s390/net/ism_drv.c cmd.request.sba = dma_handle; sba 102 drivers/s390/net/ism_drv.c dma_free_coherent(&ism->pdev->dev, PAGE_SIZE, sba, dma_handle); sba 106 drivers/s390/net/ism_drv.c ism->sba = sba; sba 145 drivers/s390/net/ism_drv.c if (!ism->sba) sba 153 drivers/s390/net/ism_drv.c ism->sba, ism->sba_dma_addr); sba 155 drivers/s390/net/ism_drv.c ism->sba = NULL; sba 410 drivers/s390/net/ism_drv.c bv = (void *) &ism->sba->dmb_bits[ISM_DMB_WORD_OFFSET]; sba 411 drivers/s390/net/ism_drv.c end = sizeof(ism->sba->dmb_bits) * BITS_PER_BYTE - ISM_DMB_BIT_OFFSET; sba 414 drivers/s390/net/ism_drv.c ism->sba->s = 0; sba 422 drivers/s390/net/ism_drv.c ism->sba->dmbe_mask[bit + ISM_DMB_BIT_OFFSET] = 0; sba 427 drivers/s390/net/ism_drv.c if (ism->sba->e) { sba 428 drivers/s390/net/ism_drv.c ism->sba->e = 0; sba 473 drivers/s390/net/qeth_core_main.c if (aob->sba[i] && buffer->is_header[i]) sba 475 drivers/s390/net/qeth_core_main.c (void *) aob->sba[i]); sba 3096 include/linux/ieee80211.h __le16 sba; sba 53 include/linux/mailbox/brcm-message.h } sba;