dma_pages 33 arch/c6x/mm/dma-coherent.c static u32 dma_pages; dma_pages 49 arch/c6x/mm/dma-coherent.c pos = bitmap_find_free_region(dma_bitmap, dma_pages, order); dma_pages 60 arch/c6x/mm/dma-coherent.c if (addr < dma_base || (pos + (1 << order)) >= dma_pages) { dma_pages 131 arch/c6x/mm/dma-coherent.c dma_pages = dma_size >> PAGE_SHIFT; dma_pages 133 arch/c6x/mm/dma-coherent.c ++dma_pages; dma_pages 135 arch/c6x/mm/dma-coherent.c dma_bitmap = memblock_alloc(BITS_TO_LONGS(dma_pages) * sizeof(long), dma_pages 139 arch/c6x/mm/dma-coherent.c __func__, BITS_TO_LONGS(dma_pages) * sizeof(long), dma_pages 310 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages = kmalloc_array(dev_priv->nr_dma_pages, dma_pages 313 drivers/gpu/drm/savage/savage_bci.c if (dev_priv->dma_pages == NULL) dma_pages 317 drivers/gpu/drm/savage/savage_bci.c SET_AGE(&dev_priv->dma_pages[i].age, 0, 0); dma_pages 318 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].used = 0; dma_pages 319 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].flushed = 0; dma_pages 336 drivers/gpu/drm/savage/savage_bci.c SET_AGE(&dev_priv->dma_pages[i].age, event, wrap); dma_pages 337 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].used = 0; dma_pages 338 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].flushed = 0; dma_pages 362 drivers/gpu/drm/savage/savage_bci.c if (dev_priv->dma_pages[page].age.wrap > wrap || dma_pages 363 drivers/gpu/drm/savage/savage_bci.c (dev_priv->dma_pages[page].age.wrap == wrap && dma_pages 364 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[page].age.event > event)) { dma_pages 366 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[page].age.event) dma_pages 376 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used; dma_pages 383 drivers/gpu/drm/savage/savage_bci.c cur, dev_priv->dma_pages[cur].used, n, rest, nr_pages); dma_pages 387 drivers/gpu/drm/savage/savage_bci.c cur * SAVAGE_DMA_PAGE_SIZE + dev_priv->dma_pages[cur].used; dma_pages 390 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used += rest; dma_pages 398 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].age = dev_priv->last_dma_age; dma_pages 399 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].used = 0; dma_pages 400 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].flushed = 0; dma_pages 407 drivers/gpu/drm/savage/savage_bci.c if (dev_priv->dma_pages[i].used) { dma_pages 409 drivers/gpu/drm/savage/savage_bci.c i, dev_priv->dma_pages[i].used); dma_pages 413 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].used = SAVAGE_DMA_PAGE_SIZE; dma_pages 415 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].used = n; dma_pages 421 drivers/gpu/drm/savage/savage_bci.c i, dev_priv->dma_pages[i].used, n); dma_pages 438 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used == dev_priv->dma_pages[cur].flushed) dma_pages 443 drivers/gpu/drm/savage/savage_bci.c pad = -dev_priv->dma_pages[cur].used & 1; dma_pages 444 drivers/gpu/drm/savage/savage_bci.c align = -(dev_priv->dma_pages[cur].used + pad) & 7; dma_pages 448 drivers/gpu/drm/savage/savage_bci.c first, cur, dev_priv->dma_pages[first].flushed, dma_pages 449 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used, pad, align); dma_pages 454 drivers/gpu/drm/savage/savage_bci.c cur * SAVAGE_DMA_PAGE_SIZE + dev_priv->dma_pages[cur].used; dma_pages 455 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used += pad; dma_pages 467 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[first].flushed) * 4; dma_pages 469 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used - dev_priv->dma_pages[first].flushed; dma_pages 480 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used += align; dma_pages 486 drivers/gpu/drm/savage/savage_bci.c SET_AGE(&dev_priv->dma_pages[i].age, event, wrap); dma_pages 487 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].used = 0; dma_pages 488 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].flushed = 0; dma_pages 491 drivers/gpu/drm/savage/savage_bci.c if (dev_priv->dma_pages[cur].used == SAVAGE_DMA_PAGE_SIZE) { dma_pages 492 drivers/gpu/drm/savage/savage_bci.c SET_AGE(&dev_priv->dma_pages[cur].age, event, wrap); dma_pages 493 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used = 0; dma_pages 494 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].flushed = 0; dma_pages 502 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].flushed = dev_priv->dma_pages[i].used; dma_pages 507 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used, dma_pages 508 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].flushed); dma_pages 517 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[dev_priv->current_dma_page].used == 0) dma_pages 522 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[dev_priv->current_dma_page].used); dma_pages 525 drivers/gpu/drm/savage/savage_bci.c i <= dev_priv->current_dma_page && dev_priv->dma_pages[i].used; dma_pages 532 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].used != SAVAGE_DMA_PAGE_SIZE) { dma_pages 534 drivers/gpu/drm/savage/savage_bci.c i, dev_priv->dma_pages[i].used); dma_pages 537 drivers/gpu/drm/savage/savage_bci.c BEGIN_BCI(dev_priv->dma_pages[i].used); dma_pages 538 drivers/gpu/drm/savage/savage_bci.c for (j = 0; j < dev_priv->dma_pages[i].used; ++j) { dma_pages 541 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[i].used = 0; dma_pages 917 drivers/gpu/drm/savage/savage_bci.c kfree(dev_priv->dma_pages); dma_pages 178 drivers/gpu/drm/savage/savage_drv.h drm_savage_dma_page_t *dma_pages; dma_pages 520 drivers/gpu/drm/savage/savage_drv.h dev_priv->dma_pages[cur].used; \ dma_pages 526 drivers/gpu/drm/savage/savage_drv.h dev_priv->dma_pages[cur].used; \ dma_pages 527 drivers/gpu/drm/savage/savage_drv.h if (dev_priv->dma_pages[cur].used == 0) \ dma_pages 529 drivers/gpu/drm/savage/savage_drv.h dev_priv->dma_pages[cur].used += (n); \ dma_pages 545 drivers/gpu/drm/savage/savage_drv.h dev_priv->dma_pages[cur].used; \ dma_pages 106 net/rds/ib_fmr.c u64 *dma_pages; dma_pages 153 net/rds/ib_fmr.c dma_pages = kmalloc_array_node(sizeof(u64), page_cnt, GFP_ATOMIC, dma_pages 155 net/rds/ib_fmr.c if (!dma_pages) { dma_pages 166 net/rds/ib_fmr.c dma_pages[page_cnt++] = dma_pages 170 net/rds/ib_fmr.c ret = ib_map_phys_fmr(fmr->fmr, dma_pages, page_cnt, io_addr); dma_pages 193 net/rds/ib_fmr.c kfree(dma_pages);