Lines Matching refs:rd_gl

727 		gl = rgd->rd_gl;  in gfs2_clear_rgrpd()
931 &gfs2_rgrp_glops, CREATE, &rgd->rd_gl); in read_rindex_entry()
935 rgd->rd_gl->gl_object = rgd; in read_rindex_entry()
936 rgd->rd_gl->gl_vm.start = rgd->rd_addr * bsize; in read_rindex_entry()
937 rgd->rd_gl->gl_vm.end = rgd->rd_gl->gl_vm.start + (rgd->rd_length * bsize) - 1; in read_rindex_entry()
938 rgd->rd_rgl = (struct gfs2_rgrp_lvb *)rgd->rd_gl->gl_lksb.sb_lvbptr; in read_rindex_entry()
949 gfs2_glock_put(rgd->rd_gl); in read_rindex_entry()
1149 struct gfs2_glock *gl = rgd->rd_gl; in gfs2_rgrp_bh_get()
1389 ret = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_EXCLUSIVE, 0, &gh); in gfs2_fitrim()
1412 gfs2_trans_add_meta(rgd->rd_gl, bh); in gfs2_fitrim()
1849 const struct gfs2_glock *gl = rgd->rd_gl; in gfs2_rgrp_congested()
1897 rs->rs_rbm.rgd->rd_gl->gl_dstamp)); in gfs2_rgrp_used_recently()
1933 struct gfs2_glock *gl = rgd->rd_gl; in fast_to_acquire()
1990 if (!gfs2_glock_is_locked_by_me(rs->rs_rbm.rgd->rd_gl)) { in gfs2_inplace_reserve()
2003 error = gfs2_glock_nq_init(rs->rs_rbm.rgd->rd_gl, in gfs2_inplace_reserve()
2138 gfs2_trans_add_meta(rbm->rgd->rd_gl, rbm_bi(rbm)->bi_bh); in gfs2_alloc_extent()
2145 gfs2_trans_add_meta(pos.rgd->rd_gl, rbm_bi(&pos)->bi_bh); in gfs2_alloc_extent()
2186 gfs2_trans_add_meta(rbm.rgd->rd_gl, bi->bi_bh); in rgblk_free()
2229 gfs2_rgrp_dump(NULL, rgd->rd_gl); in gfs2_rgrp_error()
2376 gfs2_trans_add_meta(rbm.rgd->rd_gl, rbm.rgd->rd_bits[0].bi_bh); in gfs2_alloc_blocks()
2417 gfs2_trans_add_meta(rgd->rd_gl, rgd->rd_bits[0].bi_bh); in __gfs2_free_blocks()
2454 gfs2_trans_add_meta(rgd->rd_gl, rgd->rd_bits[0].bi_bh); in gfs2_unlink_di()
2475 gfs2_trans_add_meta(rgd->rd_gl, rgd->rd_bits[0].bi_bh); in gfs2_free_uninit_di()
2513 error = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_SHARED, 0, &rgd_gh); in gfs2_check_blk_type()
2599 gfs2_holder_init(rlist->rl_rgd[x]->rd_gl, in gfs2_rlist_alloc()