sba                13 arch/alpha/include/asm/agp_backend.h 		u32 sba : 1;
sba              1015 arch/alpha/kernel/core_marvel.c 	       agp->mode.bits.sba ? " - SBA" : "", agp->mode.bits.rq);
sba               649 arch/alpha/kernel/core_titan.c 	pctl.pctl_r_bits.apctl_v_agp_sba_en = agp->mode.bits.sba;
sba               789 arch/alpha/kernel/core_titan.c 	agp->capability.bits.sba = 1;
sba               798 arch/alpha/kernel/core_titan.c 	agp->mode.bits.sba = pctl.pctl_r_bits.apctl_v_agp_sba_en;
sba               151 arch/s390/include/asm/qdio.h 	u64 sba[QDIO_MAX_ELEMENTS_PER_BUFFER];
sba               646 drivers/atm/fore200e.c 	u32 hcr = fore200e->bus->read(fore200e->regs.sba.hcr) & SBA200E_HCR_STICKY;
sba               647 drivers/atm/fore200e.c 	fore200e->bus->write(hcr | SBA200E_HCR_INTR_ENA, fore200e->regs.sba.hcr);
sba               652 drivers/atm/fore200e.c 	return fore200e->bus->read(fore200e->regs.sba.hcr) & SBA200E_HCR_INTR_REQ;
sba               657 drivers/atm/fore200e.c 	u32 hcr = fore200e->bus->read(fore200e->regs.sba.hcr) & SBA200E_HCR_STICKY;
sba               658 drivers/atm/fore200e.c 	fore200e->bus->write(hcr | SBA200E_HCR_INTR_CLR, fore200e->regs.sba.hcr);
sba               663 drivers/atm/fore200e.c 	fore200e->bus->write(SBA200E_HCR_RESET, fore200e->regs.sba.hcr);
sba               665 drivers/atm/fore200e.c 	fore200e->bus->write(0, fore200e->regs.sba.hcr);
sba               674 drivers/atm/fore200e.c 	fore200e->regs.sba.hcr = of_ioremap(&op->resource[0], 0, SBA200E_HCR_LENGTH, "SBA HCR");
sba               675 drivers/atm/fore200e.c 	fore200e->regs.sba.bsr = of_ioremap(&op->resource[1], 0, SBA200E_BSR_LENGTH, "SBA BSR");
sba               676 drivers/atm/fore200e.c 	fore200e->regs.sba.isr = of_ioremap(&op->resource[2], 0, SBA200E_ISR_LENGTH, "SBA ISR");
sba               686 drivers/atm/fore200e.c 	fore200e->bus->write(0x02, fore200e->regs.sba.isr); /* XXX hardwired interrupt level */
sba               702 drivers/atm/fore200e.c 	of_iounmap(&op->resource[0], fore200e->regs.sba.hcr, SBA200E_HCR_LENGTH);
sba               703 drivers/atm/fore200e.c 	of_iounmap(&op->resource[1], fore200e->regs.sba.bsr, SBA200E_BSR_LENGTH);
sba               704 drivers/atm/fore200e.c 	of_iounmap(&op->resource[2], fore200e->regs.sba.isr, SBA200E_ISR_LENGTH);
sba               792 drivers/atm/fore200e.h     struct fore200e_sba_regs sba;    /* SBA-200E registers */
sba               387 drivers/char/agp/parisc-agp.c 	struct parisc_device *sba = NULL, *lba = NULL;
sba               394 drivers/char/agp/parisc-agp.c 	sba = sba_list->dev;
sba               395 drivers/char/agp/parisc-agp.c 	if (!IS_PLUTO(sba)) {
sba               401 drivers/char/agp/parisc-agp.c 	device_for_each_child(&sba->dev, &lba, find_quicksilver);
sba               113 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba;
sba               203 drivers/dma/bcm-sba-raid.c static struct sba_request *sba_alloc_request(struct sba_device *sba)
sba               209 drivers/dma/bcm-sba-raid.c 	spin_lock_irqsave(&sba->reqs_lock, flags);
sba               210 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_free_list, node) {
sba               212 drivers/dma/bcm-sba-raid.c 			list_move_tail(&req->node, &sba->reqs_alloc_list);
sba               217 drivers/dma/bcm-sba-raid.c 	spin_unlock_irqrestore(&sba->reqs_lock, flags);
sba               226 drivers/dma/bcm-sba-raid.c 		mbox_client_peek_data(sba->mchan);
sba               235 drivers/dma/bcm-sba-raid.c 	dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan);
sba               242 drivers/dma/bcm-sba-raid.c static void _sba_pending_request(struct sba_device *sba,
sba               245 drivers/dma/bcm-sba-raid.c 	lockdep_assert_held(&sba->reqs_lock);
sba               248 drivers/dma/bcm-sba-raid.c 	list_move_tail(&req->node, &sba->reqs_pending_list);
sba               249 drivers/dma/bcm-sba-raid.c 	if (list_empty(&sba->reqs_active_list))
sba               250 drivers/dma/bcm-sba-raid.c 		sba->reqs_fence = false;
sba               254 drivers/dma/bcm-sba-raid.c static bool _sba_active_request(struct sba_device *sba,
sba               257 drivers/dma/bcm-sba-raid.c 	lockdep_assert_held(&sba->reqs_lock);
sba               258 drivers/dma/bcm-sba-raid.c 	if (list_empty(&sba->reqs_active_list))
sba               259 drivers/dma/bcm-sba-raid.c 		sba->reqs_fence = false;
sba               260 drivers/dma/bcm-sba-raid.c 	if (sba->reqs_fence)
sba               264 drivers/dma/bcm-sba-raid.c 	list_move_tail(&req->node, &sba->reqs_active_list);
sba               266 drivers/dma/bcm-sba-raid.c 		sba->reqs_fence = true;
sba               271 drivers/dma/bcm-sba-raid.c static void _sba_abort_request(struct sba_device *sba,
sba               274 drivers/dma/bcm-sba-raid.c 	lockdep_assert_held(&sba->reqs_lock);
sba               277 drivers/dma/bcm-sba-raid.c 	list_move_tail(&req->node, &sba->reqs_aborted_list);
sba               278 drivers/dma/bcm-sba-raid.c 	if (list_empty(&sba->reqs_active_list))
sba               279 drivers/dma/bcm-sba-raid.c 		sba->reqs_fence = false;
sba               283 drivers/dma/bcm-sba-raid.c static void _sba_free_request(struct sba_device *sba,
sba               286 drivers/dma/bcm-sba-raid.c 	lockdep_assert_held(&sba->reqs_lock);
sba               289 drivers/dma/bcm-sba-raid.c 	list_move_tail(&req->node, &sba->reqs_free_list);
sba               290 drivers/dma/bcm-sba-raid.c 	if (list_empty(&sba->reqs_active_list))
sba               291 drivers/dma/bcm-sba-raid.c 		sba->reqs_fence = false;
sba               298 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = req->sba;
sba               300 drivers/dma/bcm-sba-raid.c 	spin_lock_irqsave(&sba->reqs_lock, flags);
sba               302 drivers/dma/bcm-sba-raid.c 	_sba_free_request(sba, req);
sba               304 drivers/dma/bcm-sba-raid.c 		_sba_free_request(sba, nreq);
sba               306 drivers/dma/bcm-sba-raid.c 	spin_unlock_irqrestore(&sba->reqs_lock, flags);
sba               313 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = req->sba;
sba               315 drivers/dma/bcm-sba-raid.c 	spin_lock_irqsave(&sba->reqs_lock, flags);
sba               321 drivers/dma/bcm-sba-raid.c 	spin_unlock_irqrestore(&sba->reqs_lock, flags);
sba               324 drivers/dma/bcm-sba-raid.c static void sba_cleanup_nonpending_requests(struct sba_device *sba)
sba               329 drivers/dma/bcm-sba-raid.c 	spin_lock_irqsave(&sba->reqs_lock, flags);
sba               332 drivers/dma/bcm-sba-raid.c 	list_for_each_entry_safe(req, req1, &sba->reqs_alloc_list, node)
sba               333 drivers/dma/bcm-sba-raid.c 		_sba_free_request(sba, req);
sba               336 drivers/dma/bcm-sba-raid.c 	list_for_each_entry_safe(req, req1, &sba->reqs_active_list, node)
sba               337 drivers/dma/bcm-sba-raid.c 		_sba_abort_request(sba, req);
sba               344 drivers/dma/bcm-sba-raid.c 	spin_unlock_irqrestore(&sba->reqs_lock, flags);
sba               347 drivers/dma/bcm-sba-raid.c static void sba_cleanup_pending_requests(struct sba_device *sba)
sba               352 drivers/dma/bcm-sba-raid.c 	spin_lock_irqsave(&sba->reqs_lock, flags);
sba               355 drivers/dma/bcm-sba-raid.c 	list_for_each_entry_safe(req, req1, &sba->reqs_pending_list, node)
sba               356 drivers/dma/bcm-sba-raid.c 		_sba_free_request(sba, req);
sba               358 drivers/dma/bcm-sba-raid.c 	spin_unlock_irqrestore(&sba->reqs_lock, flags);
sba               361 drivers/dma/bcm-sba-raid.c static int sba_send_mbox_request(struct sba_device *sba,
sba               368 drivers/dma/bcm-sba-raid.c 	ret = mbox_send_message(sba->mchan, &req->msg);
sba               370 drivers/dma/bcm-sba-raid.c 		dev_err(sba->dev, "send message failed with error %d", ret);
sba               377 drivers/dma/bcm-sba-raid.c 		dev_err(sba->dev, "message error %d", ret);
sba               381 drivers/dma/bcm-sba-raid.c 	mbox_client_txdone(sba->mchan, ret);
sba               387 drivers/dma/bcm-sba-raid.c static void _sba_process_pending_requests(struct sba_device *sba)
sba               395 drivers/dma/bcm-sba-raid.c 	while (!list_empty(&sba->reqs_pending_list) && count) {
sba               397 drivers/dma/bcm-sba-raid.c 		req = list_first_entry(&sba->reqs_pending_list,
sba               401 drivers/dma/bcm-sba-raid.c 		if (!_sba_active_request(sba, req))
sba               405 drivers/dma/bcm-sba-raid.c 		ret = sba_send_mbox_request(sba, req);
sba               407 drivers/dma/bcm-sba-raid.c 			_sba_pending_request(sba, req);
sba               415 drivers/dma/bcm-sba-raid.c static void sba_process_received_request(struct sba_device *sba,
sba               428 drivers/dma/bcm-sba-raid.c 			spin_lock_irqsave(&sba->reqs_lock, flags);
sba               430 drivers/dma/bcm-sba-raid.c 			spin_unlock_irqrestore(&sba->reqs_lock, flags);
sba               439 drivers/dma/bcm-sba-raid.c 		spin_lock_irqsave(&sba->reqs_lock, flags);
sba               443 drivers/dma/bcm-sba-raid.c 			_sba_free_request(sba, nreq);
sba               447 drivers/dma/bcm-sba-raid.c 		_sba_free_request(sba, first);
sba               450 drivers/dma/bcm-sba-raid.c 		_sba_process_pending_requests(sba);
sba               452 drivers/dma/bcm-sba-raid.c 		spin_unlock_irqrestore(&sba->reqs_lock, flags);
sba               456 drivers/dma/bcm-sba-raid.c static void sba_write_stats_in_seqfile(struct sba_device *sba,
sba               464 drivers/dma/bcm-sba-raid.c 	spin_lock_irqsave(&sba->reqs_lock, flags);
sba               466 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_free_list, node)
sba               470 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_alloc_list, node)
sba               473 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_pending_list, node)
sba               476 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_active_list, node)
sba               479 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_aborted_list, node)
sba               482 drivers/dma/bcm-sba-raid.c 	spin_unlock_irqrestore(&sba->reqs_lock, flags);
sba               484 drivers/dma/bcm-sba-raid.c 	seq_printf(file, "maximum requests   = %d\n", sba->max_req);
sba               515 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = to_sba_device(dchan);
sba               518 drivers/dma/bcm-sba-raid.c 	spin_lock_irqsave(&sba->reqs_lock, flags);
sba               519 drivers/dma/bcm-sba-raid.c 	_sba_process_pending_requests(sba);
sba               520 drivers/dma/bcm-sba-raid.c 	spin_unlock_irqrestore(&sba->reqs_lock, flags);
sba               527 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba;
sba               533 drivers/dma/bcm-sba-raid.c 	sba = to_sba_device(tx->chan);
sba               537 drivers/dma/bcm-sba-raid.c 	spin_lock_irqsave(&sba->reqs_lock, flags);
sba               539 drivers/dma/bcm-sba-raid.c 	_sba_pending_request(sba, req);
sba               541 drivers/dma/bcm-sba-raid.c 		_sba_pending_request(sba, nreq);
sba               542 drivers/dma/bcm-sba-raid.c 	spin_unlock_irqrestore(&sba->reqs_lock, flags);
sba               552 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = to_sba_device(dchan);
sba               558 drivers/dma/bcm-sba-raid.c 	mbox_client_peek_data(sba->mchan);
sba               575 drivers/dma/bcm-sba-raid.c 	cmd = sba_cmd_enc(cmd, req->sba->hw_resp_size,
sba               586 drivers/dma/bcm-sba-raid.c 	cmdsp->data_len = req->sba->hw_resp_size;
sba               592 drivers/dma/bcm-sba-raid.c 	cmd = sba_cmd_enc(cmd, req->sba->hw_resp_size,
sba               604 drivers/dma/bcm-sba-raid.c 	if (req->sba->hw_resp_size) {
sba               607 drivers/dma/bcm-sba-raid.c 		cmdsp->resp_len = req->sba->hw_resp_size;
sba               611 drivers/dma/bcm-sba-raid.c 	cmdsp->data_len = req->sba->hw_resp_size;
sba               616 drivers/dma/bcm-sba-raid.c 	msg->sba.cmds = cmds;
sba               617 drivers/dma/bcm-sba-raid.c 	msg->sba.cmds_count = cmdsp - cmds;
sba               626 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = to_sba_device(dchan);
sba               629 drivers/dma/bcm-sba-raid.c 	req = sba_alloc_request(sba);
sba               692 drivers/dma/bcm-sba-raid.c 	if (req->sba->hw_resp_size) {
sba               695 drivers/dma/bcm-sba-raid.c 		cmdsp->resp_len = req->sba->hw_resp_size;
sba               704 drivers/dma/bcm-sba-raid.c 	msg->sba.cmds = cmds;
sba               705 drivers/dma/bcm-sba-raid.c 	msg->sba.cmds_count = cmdsp - cmds;
sba               711 drivers/dma/bcm-sba-raid.c sba_prep_dma_memcpy_req(struct sba_device *sba,
sba               718 drivers/dma/bcm-sba-raid.c 	req = sba_alloc_request(sba);
sba               741 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = to_sba_device(dchan);
sba               746 drivers/dma/bcm-sba-raid.c 		req_len = (len < sba->hw_buf_size) ? len : sba->hw_buf_size;
sba               748 drivers/dma/bcm-sba-raid.c 		req = sba_prep_dma_memcpy_req(sba, off, dst, src,
sba               831 drivers/dma/bcm-sba-raid.c 	if (req->sba->hw_resp_size) {
sba               834 drivers/dma/bcm-sba-raid.c 		cmdsp->resp_len = req->sba->hw_resp_size;
sba               843 drivers/dma/bcm-sba-raid.c 	msg->sba.cmds = cmds;
sba               844 drivers/dma/bcm-sba-raid.c 	msg->sba.cmds_count = cmdsp - cmds;
sba               850 drivers/dma/bcm-sba-raid.c sba_prep_dma_xor_req(struct sba_device *sba,
sba               857 drivers/dma/bcm-sba-raid.c 	req = sba_alloc_request(sba);
sba               880 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = to_sba_device(dchan);
sba               884 drivers/dma/bcm-sba-raid.c 	if (unlikely(src_cnt > sba->max_xor_srcs))
sba               889 drivers/dma/bcm-sba-raid.c 		req_len = (len < sba->hw_buf_size) ? len : sba->hw_buf_size;
sba               891 drivers/dma/bcm-sba-raid.c 		req = sba_prep_dma_xor_req(sba, off, dst, src, src_cnt,
sba              1014 drivers/dma/bcm-sba-raid.c 		if (req->sba->hw_resp_size) {
sba              1017 drivers/dma/bcm-sba-raid.c 			cmdsp->resp_len = req->sba->hw_resp_size;
sba              1041 drivers/dma/bcm-sba-raid.c 		if (req->sba->hw_resp_size) {
sba              1044 drivers/dma/bcm-sba-raid.c 			cmdsp->resp_len = req->sba->hw_resp_size;
sba              1054 drivers/dma/bcm-sba-raid.c 	msg->sba.cmds = cmds;
sba              1055 drivers/dma/bcm-sba-raid.c 	msg->sba.cmds_count = cmdsp - cmds;
sba              1061 drivers/dma/bcm-sba-raid.c sba_prep_dma_pq_req(struct sba_device *sba, dma_addr_t off,
sba              1068 drivers/dma/bcm-sba-raid.c 	req = sba_alloc_request(sba);
sba              1174 drivers/dma/bcm-sba-raid.c 	if (req->sba->hw_resp_size) {
sba              1177 drivers/dma/bcm-sba-raid.c 		cmdsp->resp_len = req->sba->hw_resp_size;
sba              1202 drivers/dma/bcm-sba-raid.c 	pos = (dpos < req->sba->max_pq_coefs) ?
sba              1203 drivers/dma/bcm-sba-raid.c 		dpos : (req->sba->max_pq_coefs - 1);
sba              1231 drivers/dma/bcm-sba-raid.c 		pos = (dpos < req->sba->max_pq_coefs) ?
sba              1232 drivers/dma/bcm-sba-raid.c 			dpos : (req->sba->max_pq_coefs - 1);
sba              1295 drivers/dma/bcm-sba-raid.c 	if (req->sba->hw_resp_size) {
sba              1298 drivers/dma/bcm-sba-raid.c 		cmdsp->resp_len = req->sba->hw_resp_size;
sba              1308 drivers/dma/bcm-sba-raid.c 	msg->sba.cmds = cmds;
sba              1309 drivers/dma/bcm-sba-raid.c 	msg->sba.cmds_count = cmdsp - cmds;
sba              1315 drivers/dma/bcm-sba-raid.c sba_prep_dma_pq_single_req(struct sba_device *sba, dma_addr_t off,
sba              1323 drivers/dma/bcm-sba-raid.c 	req = sba_alloc_request(sba);
sba              1350 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = to_sba_device(dchan);
sba              1354 drivers/dma/bcm-sba-raid.c 	if (unlikely(src_cnt > sba->max_pq_srcs))
sba              1357 drivers/dma/bcm-sba-raid.c 		if (sba->max_pq_coefs <= raid6_gflog[scf[i]])
sba              1368 drivers/dma/bcm-sba-raid.c 		req_len = (len < sba->hw_buf_size) ? len : sba->hw_buf_size;
sba              1384 drivers/dma/bcm-sba-raid.c 				req = sba_prep_dma_pq_single_req(sba,
sba              1402 drivers/dma/bcm-sba-raid.c 				req = sba_prep_dma_pq_single_req(sba,
sba              1416 drivers/dma/bcm-sba-raid.c 			req = sba_prep_dma_pq_req(sba, off,
sba              1446 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = req->sba;
sba              1450 drivers/dma/bcm-sba-raid.c 		dev_err(sba->dev, "%s got message with error %d",
sba              1451 drivers/dma/bcm-sba-raid.c 			dma_chan_name(&sba->dma_chan), m->error);
sba              1454 drivers/dma/bcm-sba-raid.c 	sba_process_received_request(sba, req);
sba              1461 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = dev_get_drvdata(file->private);
sba              1464 drivers/dma/bcm-sba-raid.c 	sba_write_stats_in_seqfile(sba, file);
sba              1471 drivers/dma/bcm-sba-raid.c static int sba_prealloc_channel_resources(struct sba_device *sba)
sba              1476 drivers/dma/bcm-sba-raid.c 	sba->resp_base = dma_alloc_coherent(sba->mbox_dev,
sba              1477 drivers/dma/bcm-sba-raid.c 					    sba->max_resp_pool_size,
sba              1478 drivers/dma/bcm-sba-raid.c 					    &sba->resp_dma_base, GFP_KERNEL);
sba              1479 drivers/dma/bcm-sba-raid.c 	if (!sba->resp_base)
sba              1482 drivers/dma/bcm-sba-raid.c 	sba->cmds_base = dma_alloc_coherent(sba->mbox_dev,
sba              1483 drivers/dma/bcm-sba-raid.c 					    sba->max_cmds_pool_size,
sba              1484 drivers/dma/bcm-sba-raid.c 					    &sba->cmds_dma_base, GFP_KERNEL);
sba              1485 drivers/dma/bcm-sba-raid.c 	if (!sba->cmds_base) {
sba              1490 drivers/dma/bcm-sba-raid.c 	spin_lock_init(&sba->reqs_lock);
sba              1491 drivers/dma/bcm-sba-raid.c 	sba->reqs_fence = false;
sba              1492 drivers/dma/bcm-sba-raid.c 	INIT_LIST_HEAD(&sba->reqs_alloc_list);
sba              1493 drivers/dma/bcm-sba-raid.c 	INIT_LIST_HEAD(&sba->reqs_pending_list);
sba              1494 drivers/dma/bcm-sba-raid.c 	INIT_LIST_HEAD(&sba->reqs_active_list);
sba              1495 drivers/dma/bcm-sba-raid.c 	INIT_LIST_HEAD(&sba->reqs_aborted_list);
sba              1496 drivers/dma/bcm-sba-raid.c 	INIT_LIST_HEAD(&sba->reqs_free_list);
sba              1498 drivers/dma/bcm-sba-raid.c 	for (i = 0; i < sba->max_req; i++) {
sba              1499 drivers/dma/bcm-sba-raid.c 		req = devm_kzalloc(sba->dev,
sba              1500 drivers/dma/bcm-sba-raid.c 				   struct_size(req, cmds, sba->max_cmd_per_req),
sba              1507 drivers/dma/bcm-sba-raid.c 		req->sba = sba;
sba              1511 drivers/dma/bcm-sba-raid.c 		for (j = 0; j < sba->max_cmd_per_req; j++) {
sba              1513 drivers/dma/bcm-sba-raid.c 			req->cmds[j].cmd_dma = sba->cmds_base +
sba              1514 drivers/dma/bcm-sba-raid.c 				(i * sba->max_cmd_per_req + j) * sizeof(u64);
sba              1515 drivers/dma/bcm-sba-raid.c 			req->cmds[j].cmd_dma_addr = sba->cmds_dma_base +
sba              1516 drivers/dma/bcm-sba-raid.c 				(i * sba->max_cmd_per_req + j) * sizeof(u64);
sba              1520 drivers/dma/bcm-sba-raid.c 		dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan);
sba              1523 drivers/dma/bcm-sba-raid.c 		req->tx.phys = sba->resp_dma_base + i * sba->hw_resp_size;
sba              1524 drivers/dma/bcm-sba-raid.c 		list_add_tail(&req->node, &sba->reqs_free_list);
sba              1530 drivers/dma/bcm-sba-raid.c 	dma_free_coherent(sba->mbox_dev,
sba              1531 drivers/dma/bcm-sba-raid.c 			  sba->max_cmds_pool_size,
sba              1532 drivers/dma/bcm-sba-raid.c 			  sba->cmds_base, sba->cmds_dma_base);
sba              1534 drivers/dma/bcm-sba-raid.c 	dma_free_coherent(sba->mbox_dev,
sba              1535 drivers/dma/bcm-sba-raid.c 			  sba->max_resp_pool_size,
sba              1536 drivers/dma/bcm-sba-raid.c 			  sba->resp_base, sba->resp_dma_base);
sba              1540 drivers/dma/bcm-sba-raid.c static void sba_freeup_channel_resources(struct sba_device *sba)
sba              1542 drivers/dma/bcm-sba-raid.c 	dmaengine_terminate_all(&sba->dma_chan);
sba              1543 drivers/dma/bcm-sba-raid.c 	dma_free_coherent(sba->mbox_dev, sba->max_cmds_pool_size,
sba              1544 drivers/dma/bcm-sba-raid.c 			  sba->cmds_base, sba->cmds_dma_base);
sba              1545 drivers/dma/bcm-sba-raid.c 	dma_free_coherent(sba->mbox_dev, sba->max_resp_pool_size,
sba              1546 drivers/dma/bcm-sba-raid.c 			  sba->resp_base, sba->resp_dma_base);
sba              1547 drivers/dma/bcm-sba-raid.c 	sba->resp_base = NULL;
sba              1548 drivers/dma/bcm-sba-raid.c 	sba->resp_dma_base = 0;
sba              1551 drivers/dma/bcm-sba-raid.c static int sba_async_register(struct sba_device *sba)
sba              1554 drivers/dma/bcm-sba-raid.c 	struct dma_device *dma_dev = &sba->dma_dev;
sba              1557 drivers/dma/bcm-sba-raid.c 	sba->dma_chan.device = dma_dev;
sba              1558 drivers/dma/bcm-sba-raid.c 	dma_cookie_init(&sba->dma_chan);
sba              1572 drivers/dma/bcm-sba-raid.c 	dma_dev->dev = sba->mbox_dev;
sba              1591 drivers/dma/bcm-sba-raid.c 		dma_dev->max_xor = sba->max_xor_srcs;
sba              1597 drivers/dma/bcm-sba-raid.c 		dma_set_maxpq(dma_dev, sba->max_pq_srcs, 0);
sba              1602 drivers/dma/bcm-sba-raid.c 	list_add_tail(&sba->dma_chan.device_node, &dma_dev->channels);
sba              1607 drivers/dma/bcm-sba-raid.c 		dev_err(sba->dev, "async device register error %d", ret);
sba              1611 drivers/dma/bcm-sba-raid.c 	dev_info(sba->dev, "%s capabilities: %s%s%s%s\n",
sba              1612 drivers/dma/bcm-sba-raid.c 	dma_chan_name(&sba->dma_chan),
sba              1624 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba;
sba              1629 drivers/dma/bcm-sba-raid.c 	sba = devm_kzalloc(&pdev->dev, sizeof(*sba), GFP_KERNEL);
sba              1630 drivers/dma/bcm-sba-raid.c 	if (!sba)
sba              1633 drivers/dma/bcm-sba-raid.c 	sba->dev = &pdev->dev;
sba              1634 drivers/dma/bcm-sba-raid.c 	platform_set_drvdata(pdev, sba);
sba              1643 drivers/dma/bcm-sba-raid.c 	if (of_device_is_compatible(sba->dev->of_node, "brcm,iproc-sba"))
sba              1644 drivers/dma/bcm-sba-raid.c 		sba->ver = SBA_VER_1;
sba              1645 drivers/dma/bcm-sba-raid.c 	else if (of_device_is_compatible(sba->dev->of_node,
sba              1647 drivers/dma/bcm-sba-raid.c 		sba->ver = SBA_VER_2;
sba              1652 drivers/dma/bcm-sba-raid.c 	switch (sba->ver) {
sba              1654 drivers/dma/bcm-sba-raid.c 		sba->hw_buf_size = 4096;
sba              1655 drivers/dma/bcm-sba-raid.c 		sba->hw_resp_size = 8;
sba              1656 drivers/dma/bcm-sba-raid.c 		sba->max_pq_coefs = 6;
sba              1657 drivers/dma/bcm-sba-raid.c 		sba->max_pq_srcs = 6;
sba              1660 drivers/dma/bcm-sba-raid.c 		sba->hw_buf_size = 4096;
sba              1661 drivers/dma/bcm-sba-raid.c 		sba->hw_resp_size = 8;
sba              1662 drivers/dma/bcm-sba-raid.c 		sba->max_pq_coefs = 30;
sba              1668 drivers/dma/bcm-sba-raid.c 		sba->max_pq_srcs = 12;
sba              1673 drivers/dma/bcm-sba-raid.c 	sba->max_req = SBA_MAX_REQ_PER_MBOX_CHANNEL;
sba              1674 drivers/dma/bcm-sba-raid.c 	sba->max_cmd_per_req = sba->max_pq_srcs + 3;
sba              1675 drivers/dma/bcm-sba-raid.c 	sba->max_xor_srcs = sba->max_cmd_per_req - 1;
sba              1676 drivers/dma/bcm-sba-raid.c 	sba->max_resp_pool_size = sba->max_req * sba->hw_resp_size;
sba              1677 drivers/dma/bcm-sba-raid.c 	sba->max_cmds_pool_size = sba->max_req *
sba              1678 drivers/dma/bcm-sba-raid.c 				  sba->max_cmd_per_req * sizeof(u64);
sba              1681 drivers/dma/bcm-sba-raid.c 	sba->client.dev			= &pdev->dev;
sba              1682 drivers/dma/bcm-sba-raid.c 	sba->client.rx_callback		= sba_receive_message;
sba              1683 drivers/dma/bcm-sba-raid.c 	sba->client.tx_block		= false;
sba              1684 drivers/dma/bcm-sba-raid.c 	sba->client.knows_txdone	= true;
sba              1685 drivers/dma/bcm-sba-raid.c 	sba->client.tx_tout		= 0;
sba              1688 drivers/dma/bcm-sba-raid.c 	sba->mchan = mbox_request_channel(&sba->client, 0);
sba              1689 drivers/dma/bcm-sba-raid.c 	if (IS_ERR(sba->mchan)) {
sba              1690 drivers/dma/bcm-sba-raid.c 		ret = PTR_ERR(sba->mchan);
sba              1705 drivers/dma/bcm-sba-raid.c 	sba->mbox_dev = &mbox_pdev->dev;
sba              1708 drivers/dma/bcm-sba-raid.c 	ret = sba_prealloc_channel_resources(sba);
sba              1717 drivers/dma/bcm-sba-raid.c 	sba->root = debugfs_create_dir(dev_name(sba->dev), NULL);
sba              1720 drivers/dma/bcm-sba-raid.c 	debugfs_create_devm_seqfile(sba->dev, "stats", sba->root,
sba              1726 drivers/dma/bcm-sba-raid.c 	ret = sba_async_register(sba);
sba              1731 drivers/dma/bcm-sba-raid.c 	dev_info(sba->dev, "%s using SBAv%d mailbox channel from %s",
sba              1732 drivers/dma/bcm-sba-raid.c 		 dma_chan_name(&sba->dma_chan), sba->ver+1,
sba              1733 drivers/dma/bcm-sba-raid.c 		 dev_name(sba->mbox_dev));
sba              1738 drivers/dma/bcm-sba-raid.c 	debugfs_remove_recursive(sba->root);
sba              1739 drivers/dma/bcm-sba-raid.c 	sba_freeup_channel_resources(sba);
sba              1741 drivers/dma/bcm-sba-raid.c 	mbox_free_channel(sba->mchan);
sba              1747 drivers/dma/bcm-sba-raid.c 	struct sba_device *sba = platform_get_drvdata(pdev);
sba              1749 drivers/dma/bcm-sba-raid.c 	dma_async_device_unregister(&sba->dma_dev);
sba              1751 drivers/dma/bcm-sba-raid.c 	debugfs_remove_recursive(sba->root);
sba              1753 drivers/dma/bcm-sba-raid.c 	sba_freeup_channel_resources(sba);
sba              1755 drivers/dma/bcm-sba-raid.c 	mbox_free_channel(sba->mchan);
sba               719 drivers/mailbox/bcm-flexrm-mailbox.c 	if (!msg->sba.cmds || !msg->sba.cmds_count)
sba               722 drivers/mailbox/bcm-flexrm-mailbox.c 	for (i = 0; i < msg->sba.cmds_count; i++) {
sba               723 drivers/mailbox/bcm-flexrm-mailbox.c 		if (((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_B) ||
sba               724 drivers/mailbox/bcm-flexrm-mailbox.c 		     (msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_C)) &&
sba               725 drivers/mailbox/bcm-flexrm-mailbox.c 		    (msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_OUTPUT))
sba               727 drivers/mailbox/bcm-flexrm-mailbox.c 		if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_B) &&
sba               728 drivers/mailbox/bcm-flexrm-mailbox.c 		    (msg->sba.cmds[i].data_len > SRCT_LENGTH_MASK))
sba               730 drivers/mailbox/bcm-flexrm-mailbox.c 		if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_C) &&
sba               731 drivers/mailbox/bcm-flexrm-mailbox.c 		    (msg->sba.cmds[i].data_len > SRCT_LENGTH_MASK))
sba               733 drivers/mailbox/bcm-flexrm-mailbox.c 		if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_RESP) &&
sba               734 drivers/mailbox/bcm-flexrm-mailbox.c 		    (msg->sba.cmds[i].resp_len > DSTT_LENGTH_MASK))
sba               736 drivers/mailbox/bcm-flexrm-mailbox.c 		if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_OUTPUT) &&
sba               737 drivers/mailbox/bcm-flexrm-mailbox.c 		    (msg->sba.cmds[i].data_len > DSTT_LENGTH_MASK))
sba               749 drivers/mailbox/bcm-flexrm-mailbox.c 	for (i = 0; i < msg->sba.cmds_count; i++) {
sba               752 drivers/mailbox/bcm-flexrm-mailbox.c 		if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_B) ||
sba               753 drivers/mailbox/bcm-flexrm-mailbox.c 		    (msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_C))
sba               756 drivers/mailbox/bcm-flexrm-mailbox.c 		if (msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_RESP)
sba               759 drivers/mailbox/bcm-flexrm-mailbox.c 		if (msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_OUTPUT)
sba               776 drivers/mailbox/bcm-flexrm-mailbox.c 	for (i = 0; i < msg->sba.cmds_count; i++) {
sba               777 drivers/mailbox/bcm-flexrm-mailbox.c 		c = &msg->sba.cmds[i];
sba               611 drivers/net/fddi/skfp/h/cmtdef.h void sba();
sba               455 drivers/net/fddi/skfp/h/smc.h 	struct s_sba	sba ;		/* SBA variables */
sba               514 drivers/net/fddi/skfp/smt.c 			smc->sba.sm = sm ;
sba               515 drivers/net/fddi/skfp/smt.c 			sba(smc,NIF) ;
sba              1544 drivers/net/wireless/ath/ath6kl/debug.c 	pstream.sba = cpu_to_le32(val32);
sba              1178 drivers/net/wireless/ath/ath6kl/wmi.h 	__le32 sba;
sba              1235 drivers/parisc/sba_iommu.c setup_ibase_imask(struct parisc_device *sba, struct ioc *ioc, int ioc_num)
sba              1242 drivers/parisc/sba_iommu.c 	device_for_each_child(&sba->dev, &ibase_data,
sba              1260 drivers/parisc/sba_iommu.c sba_ioc_init_pluto(struct parisc_device *sba, struct ioc *ioc, int ioc_num)
sba              1322 drivers/parisc/sba_iommu.c 	setup_ibase_imask(sba, ioc, ioc_num);
sba              1369 drivers/parisc/sba_iommu.c 	device_for_each_child(&sba->dev, &agp_found, sba_ioc_find_quicksilver);
sba              1381 drivers/parisc/sba_iommu.c sba_ioc_init(struct parisc_device *sba, struct ioc *ioc, int ioc_num)
sba              1468 drivers/parisc/sba_iommu.c 	setup_ibase_imask(sba, ioc, ioc_num);
sba              1994 drivers/parisc/sba_iommu.c 	struct sba_device *sba = dev_get_drvdata(&sba_dev->dev);
sba              2000 drivers/parisc/sba_iommu.c 	return &(sba->ioc[iocnum]);
sba              2015 drivers/parisc/sba_iommu.c 	struct sba_device *sba = dev_get_drvdata(&sba_dev->dev);
sba              2027 drivers/parisc/sba_iommu.c 		void __iomem *reg = sba->sba_hpa + i*0x18;
sba              2058 drivers/parisc/sba_iommu.c 	struct sba_device *sba = dev_get_drvdata(&sba_dev->dev);
sba              2067 drivers/parisc/sba_iommu.c 	base = READ_REG32(sba->sba_hpa + LMMIO_DIST_BASE);
sba              2075 drivers/parisc/sba_iommu.c 	size = (~READ_REG32(sba->sba_hpa + LMMIO_DIST_MASK)) / ROPES_PER_IOC;
sba               372 drivers/s390/char/tty3270.c 	char *sba, *str;
sba               415 drivers/s390/char/tty3270.c 	sba = invalid_sba;
sba               427 drivers/s390/char/tty3270.c 			if (s->string[1] == sba[0] && s->string[2] == sba[1])
sba               433 drivers/s390/char/tty3270.c 				sba = s->string + s->len - 3;
sba               435 drivers/s390/char/tty3270.c 				sba = invalid_sba;
sba                52 drivers/s390/net/ism.h 		u64 sba;
sba                90 drivers/s390/net/ism.h 		u64 sba;
sba               198 drivers/s390/net/ism.h 	struct ism_sba *sba;
sba                89 drivers/s390/net/ism_drv.c 	struct ism_sba *sba;
sba                91 drivers/s390/net/ism_drv.c 	sba = dma_alloc_coherent(&ism->pdev->dev, PAGE_SIZE, &dma_handle,
sba                93 drivers/s390/net/ism_drv.c 	if (!sba)
sba                99 drivers/s390/net/ism_drv.c 	cmd.request.sba = dma_handle;
sba               102 drivers/s390/net/ism_drv.c 		dma_free_coherent(&ism->pdev->dev, PAGE_SIZE, sba, dma_handle);
sba               106 drivers/s390/net/ism_drv.c 	ism->sba = sba;
sba               145 drivers/s390/net/ism_drv.c 	if (!ism->sba)
sba               153 drivers/s390/net/ism_drv.c 			  ism->sba, ism->sba_dma_addr);
sba               155 drivers/s390/net/ism_drv.c 	ism->sba = NULL;
sba               410 drivers/s390/net/ism_drv.c 	bv = (void *) &ism->sba->dmb_bits[ISM_DMB_WORD_OFFSET];
sba               411 drivers/s390/net/ism_drv.c 	end = sizeof(ism->sba->dmb_bits) * BITS_PER_BYTE - ISM_DMB_BIT_OFFSET;
sba               414 drivers/s390/net/ism_drv.c 	ism->sba->s = 0;
sba               422 drivers/s390/net/ism_drv.c 		ism->sba->dmbe_mask[bit + ISM_DMB_BIT_OFFSET] = 0;
sba               427 drivers/s390/net/ism_drv.c 	if (ism->sba->e) {
sba               428 drivers/s390/net/ism_drv.c 		ism->sba->e = 0;
sba               473 drivers/s390/net/qeth_core_main.c 		if (aob->sba[i] && buffer->is_header[i])
sba               475 drivers/s390/net/qeth_core_main.c 					(void *) aob->sba[i]);
sba              3096 include/linux/ieee80211.h 	__le16 sba;
sba                53 include/linux/mailbox/brcm-message.h 		} sba;