Lines Matching refs:fb
30 gf100_fb_memtype_valid(struct nvkm_fb *fb, u32 tile_flags) in gf100_fb_memtype_valid() argument
39 struct gf100_fb *fb = gf100_fb(base); in gf100_fb_intr() local
40 struct nvkm_subdev *subdev = &fb->base.subdev; in gf100_fb_intr()
52 struct gf100_fb *fb = gf100_fb(base); in gf100_fb_init() local
53 struct nvkm_device *device = fb->base.subdev.device; in gf100_fb_init()
55 if (fb->r100c10_page) in gf100_fb_init()
56 nvkm_wr32(device, 0x100c10, fb->r100c10 >> 8); in gf100_fb_init()
64 struct gf100_fb *fb = gf100_fb(base); in gf100_fb_dtor() local
65 struct nvkm_device *device = fb->base.subdev.device; in gf100_fb_dtor()
67 if (fb->r100c10_page) { in gf100_fb_dtor()
68 dma_unmap_page(device->dev, fb->r100c10, PAGE_SIZE, in gf100_fb_dtor()
70 __free_page(fb->r100c10_page); in gf100_fb_dtor()
73 return fb; in gf100_fb_dtor()
80 struct gf100_fb *fb; in gf100_fb_new_() local
82 if (!(fb = kzalloc(sizeof(*fb), GFP_KERNEL))) in gf100_fb_new_()
84 nvkm_fb_ctor(func, device, index, &fb->base); in gf100_fb_new_()
85 *pfb = &fb->base; in gf100_fb_new_()
87 fb->r100c10_page = alloc_page(GFP_KERNEL | __GFP_ZERO); in gf100_fb_new_()
88 if (fb->r100c10_page) { in gf100_fb_new_()
89 fb->r100c10 = dma_map_page(device->dev, fb->r100c10_page, 0, in gf100_fb_new_()
91 if (dma_mapping_error(device->dev, fb->r100c10)) in gf100_fb_new_()