Home
last modified time | relevance | path

Searched refs:i_gl (Results 1 – 19 of 19) sorted by relevance

/linux-4.4.14/fs/gfs2/
Dinode.c158 error = gfs2_glock_get(sdp, no_addr, &gfs2_inode_glops, CREATE, &ip->i_gl); in gfs2_inode_lookup()
161 ip->i_gl->gl_object = ip; in gfs2_inode_lookup()
199 ip->i_gl->gl_object = NULL; in gfs2_inode_lookup()
200 gfs2_glock_put(ip->i_gl); in gfs2_inode_lookup()
303 if (gfs2_glock_is_locked_by_me(dip->i_gl) == NULL) { in gfs2_lookupi()
304 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); in gfs2_lookupi()
443 bh = gfs2_meta_new(ip->i_gl, ip->i_eattr); in gfs2_init_xattr()
444 gfs2_trans_add_meta(ip->i_gl, bh); in gfs2_init_xattr()
471 dibh = gfs2_meta_new(ip->i_gl, ip->i_no_addr); in init_dinode()
472 gfs2_trans_add_meta(ip->i_gl, dibh); in init_dinode()
[all …]
Dxattr.c122 error = gfs2_meta_read(ip->i_gl, ip->i_eattr, DIO_WAIT, &bh); in ea_foreach()
146 error = gfs2_meta_read(ip->i_gl, bn, DIO_WAIT, &eabh); in ea_foreach()
274 gfs2_trans_add_meta(ip->i_gl, bh); in ea_dealloc_unstuffed()
313 gfs2_trans_add_meta(ip->i_gl, dibh); in ea_dealloc_unstuffed()
434 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); in gfs2_listxattr()
480 error = gfs2_meta_read(ip->i_gl, be64_to_cpu(*dataptrs), 0, in gfs2_iter_unstuffed()
513 gfs2_trans_add_meta(ip->i_gl, bh[x]); in gfs2_iter_unstuffed()
634 *bhp = gfs2_meta_new(ip->i_gl, block); in ea_alloc_blk()
635 gfs2_trans_add_meta(ip->i_gl, *bhp); in ea_alloc_blk()
696 bh = gfs2_meta_new(ip->i_gl, block); in ea_write()
[all …]
Dsuper.c402 struct gfs2_glock *j_gl = ip->i_gl; in gfs2_make_fs_rw()
479 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE, in gfs2_statfs_init()
530 gfs2_trans_add_meta(l_ip->i_gl, l_bh); in gfs2_statfs_change()
558 gfs2_trans_add_meta(l_ip->i_gl, l_bh); in update_statfs()
569 gfs2_trans_add_meta(m_ip->i_gl, m_bh); in update_statfs()
585 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE, in gfs2_statfs_sync()
659 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &lfcc->gh); in gfs2_lock_fs_check_clean()
745 struct address_space *metamapping = gfs2_glock2aspace(ip->i_gl); in gfs2_write_inode()
750 gfs2_log_flush(GFS2_SB(inode), ip->i_gl, NORMAL_FLUSH); in gfs2_write_inode()
788 if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { in gfs2_dirty_inode()
[all …]
Dfile.c66 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, in gfs2_llseek()
99 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); in gfs2_readdir()
160 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_get_flags()
226 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); in do_gfs2_set_flags()
259 gfs2_log_flush(sdp, ip->i_gl, NORMAL_FLUSH); in do_gfs2_set_flags()
273 gfs2_trans_add_meta(ip->i_gl, bh); in do_gfs2_set_flags()
410 gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); in gfs2_page_mkwrite()
415 set_bit(GLF_DIRTY, &ip->i_gl->gl_flags); in gfs2_page_mkwrite()
523 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, in gfs2_mmap()
593 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, in gfs2_open()
[all …]
Ddir.c97 bh = gfs2_meta_new(ip->i_gl, block); in gfs2_dir_get_new_buffer()
98 gfs2_trans_add_meta(ip->i_gl, bh); in gfs2_dir_get_new_buffer()
111 error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, &bh); in gfs2_dir_get_existing_buffer()
132 gfs2_trans_add_meta(ip->i_gl, dibh); in gfs2_dir_write_stuffed()
214 gfs2_trans_add_meta(ip->i_gl, bh); in gfs2_dir_write_data()
236 gfs2_trans_add_meta(ip->i_gl, dibh); in gfs2_dir_write_data()
306 bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); in gfs2_dir_read_data()
308 error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, &bh); in gfs2_dir_read_data()
659 gfs2_trans_add_meta(dip->i_gl, bh); in dirent_del()
702 gfs2_trans_add_meta(ip->i_gl, bh); in gfs2_init_dirent()
[all …]
Dbmap.c97 gfs2_trans_add_data(ip->i_gl, bh); in gfs2_unstuffer_page()
157 gfs2_trans_add_meta(ip->i_gl, dibh); in gfs2_unstuff_dinode()
472 gfs2_trans_add_meta(ip->i_gl, dibh); in gfs2_bmap_alloc()
523 gfs2_indirect_init(mp, ip->i_gl, i, 0, bn++); in gfs2_bmap_alloc()
549 gfs2_trans_add_meta(ip->i_gl, mp->mp_bh[i-1]); in gfs2_bmap_alloc()
551 gfs2_indirect_init(mp, ip->i_gl, i, in gfs2_bmap_alloc()
561 gfs2_trans_add_meta(ip->i_gl, mp->mp_bh[end_of_metadata]); in gfs2_bmap_alloc()
801 gfs2_trans_add_meta(ip->i_gl, dibh); in do_strip()
802 gfs2_trans_add_meta(ip->i_gl, bh); in do_strip()
906 gfs2_metapath_ra(ip->i_gl, bh, top); in recursive_scan()
[all …]
Ddentry.c64 had_lock = (gfs2_glock_is_locked_by_me(dip->i_gl) != NULL); in gfs2_drevalidate()
66 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); in gfs2_drevalidate()
Daops.c56 gfs2_trans_add_data(ip->i_gl, bh); in gfs2_page_add_databufs()
107 if (gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(ip->i_gl))) in gfs2_writepage_common()
434 gfs2_log_flush(sdp, ip->i_gl, NORMAL_FLUSH); in gfs2_jdata_writepages()
532 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_readpage()
616 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_readpages()
658 gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &ip->i_gh); in gfs2_write_begin()
663 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, in gfs2_write_begin()
886 BUG_ON(gfs2_glock_is_locked_by_me(ip->i_gl) == NULL); in gfs2_write_end()
905 gfs2_trans_add_meta(ip->i_gl, dibh); in gfs2_write_end()
956 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); in gfs2_bmap()
[all …]
Dlops.c529 struct gfs2_glock *gl = ip->i_gl; in buf_lo_scan_elements()
600 gfs2_meta_sync(ip->i_gl); in buf_lo_after_scan()
606 gfs2_meta_sync(ip->i_gl); in buf_lo_after_scan()
766 struct gfs2_glock *gl = ip->i_gl; in databuf_lo_scan_elements()
817 gfs2_meta_sync(ip->i_gl); in databuf_lo_after_scan()
824 gfs2_meta_sync(ip->i_gl); in databuf_lo_after_scan()
Dquota.c391 error = gfs2_meta_read(ip->i_gl, bh_map.b_blocknr, DIO_WAIT, &bh); in bh_get()
623 gfs2_trans_add_meta(ip->i_gl, qd->qd_bh); in do_qc()
705 gfs2_trans_add_data(ip->i_gl, bh); in gfs2_write_buf_to_page()
866 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); in do_sync()
925 gfs2_log_flush(ip->i_gl->gl_name.ln_sbd, ip->i_gl, NORMAL_FLUSH); in do_sync()
979 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &i_gh); in do_glock()
1339 bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); in gfs2_quota_init()
1646 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); in gfs2_set_dqblk()
Dmeta_io.c311 bh = gfs2_getbuf(ip->i_gl, bstart, NO_CREATE); in gfs2_meta_wipe()
340 struct gfs2_glock *gl = ip->i_gl; in gfs2_meta_indirect_buffer()
Dops_fstype.c558 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, ji_gh); in gfs2_jindex_hold()
621 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_NOEXP | in check_journal_clean()
707 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, in init_journal()
887 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, in init_per_node()
895 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, in init_per_node()
Dexport.c115 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_get_name()
Drecovery.c36 struct gfs2_glock *gl = ip->i_gl; in gfs2_replay_read_block()
487 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, in gfs2_recover_func()
Dincore.h391 struct gfs2_glock *i_gl; /* Move into i_gh? */ member
Dglops.c507 struct gfs2_glock *j_gl = ip->i_gl; in freeze_go_xmote_bh()
Dtrace_gfs2.h452 __entry->dev = ip->i_gl->gl_name.ln_sbd->sd_vfs->s_dev;
Drgrp.c1033 struct gfs2_glock *gl = ip->i_gl; in gfs2_rindex_update()
2380 gfs2_trans_add_meta(ip->i_gl, dibh); in gfs2_alloc_blocks()
Dglock.c1515 struct gfs2_glock *gl = ip->i_gl; in gfs2_glock_finish_truncate()