/linux-3.4.99/fs/gfs2/ |
D | xattr.c | 121 error = gfs2_meta_read(ip->i_gl, ip->i_eattr, DIO_WAIT, &bh); in ea_foreach() 145 error = gfs2_meta_read(ip->i_gl, bn, DIO_WAIT, &eabh); in ea_foreach() 273 gfs2_trans_add_bh(ip->i_gl, bh, 1); in ea_dealloc_unstuffed() 312 gfs2_trans_add_bh(ip->i_gl, dibh, 1); in ea_dealloc_unstuffed() 435 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); in gfs2_listxattr() 478 error = gfs2_meta_read(ip->i_gl, be64_to_cpu(*dataptrs), 0, in ea_get_unstuffed() 621 *bhp = gfs2_meta_new(ip->i_gl, block); in ea_alloc_blk() 622 gfs2_trans_add_bh(ip->i_gl, *bhp, 1); in ea_alloc_blk() 683 bh = gfs2_meta_new(ip->i_gl, block); in ea_write() 684 gfs2_trans_add_bh(ip->i_gl, bh, 1); in ea_write() [all …]
|
D | inode.c | 158 error = gfs2_glock_get(sdp, no_addr, &gfs2_inode_glops, CREATE, &ip->i_gl); in gfs2_inode_lookup() 161 ip->i_gl->gl_object = ip; in gfs2_inode_lookup() 198 ip->i_gl->gl_object = NULL; in gfs2_inode_lookup() 199 gfs2_glock_put(ip->i_gl); in gfs2_inode_lookup() 302 if (gfs2_glock_is_locked_by_me(dip->i_gl) == NULL) { in gfs2_lookupi() 303 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); in gfs2_lookupi() 599 gfs2_trans_add_bh(ip->i_gl, dibh, 1); in link_dinode() 670 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_EXCLUSIVE, 0, ghs); in gfs2_create_inode() 782 struct gfs2_glock *gl = GFS2_I(inode)->i_gl; in gfs2_lookup() 822 gfs2_holder_init(dip->i_gl, LM_ST_EXCLUSIVE, 0, ghs); in gfs2_link() [all …]
|
D | super.c | 372 struct gfs2_glock *j_gl = ip->i_gl; in gfs2_make_fs_rw() 442 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE, in gfs2_statfs_init() 493 gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1); in gfs2_statfs_change() 521 gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1); in update_statfs() 532 gfs2_trans_add_bh(m_ip->i_gl, m_bh, 1); in update_statfs() 547 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE, in gfs2_statfs_sync() 619 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &lfcc->gh); in gfs2_lock_fs_check_clean() 753 struct address_space *metamapping = gfs2_glock2aspace(ip->i_gl); in gfs2_write_inode() 758 gfs2_log_flush(GFS2_SB(inode), ip->i_gl); in gfs2_write_inode() 796 if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { in gfs2_dirty_inode() [all …]
|
D | file.c | 65 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, in gfs2_llseek() 100 gfs2_holder_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); in gfs2_readdir() 164 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_get_flags() 229 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); in do_gfs2_set_flags() 262 gfs2_log_flush(sdp, ip->i_gl); in do_gfs2_set_flags() 276 gfs2_trans_add_bh(ip->i_gl, bh, 1); in do_gfs2_set_flags() 379 gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); in gfs2_page_mkwrite() 384 set_bit(GLF_DIRTY, &ip->i_gl->gl_flags); in gfs2_page_mkwrite() 495 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); in gfs2_mmap() 536 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, in gfs2_open() [all …]
|
D | dir.c | 95 bh = gfs2_meta_new(ip->i_gl, block); in gfs2_dir_get_new_buffer() 96 gfs2_trans_add_bh(ip->i_gl, bh, 1); in gfs2_dir_get_new_buffer() 109 error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, &bh); in gfs2_dir_get_existing_buffer() 130 gfs2_trans_add_bh(ip->i_gl, dibh, 1); in gfs2_dir_write_stuffed() 212 gfs2_trans_add_bh(ip->i_gl, bh, 1); in gfs2_dir_write_data() 234 gfs2_trans_add_bh(ip->i_gl, dibh, 1); in gfs2_dir_write_data() 304 bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); in gfs2_dir_read_data() 306 error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, &bh); in gfs2_dir_read_data() 650 gfs2_trans_add_bh(dip->i_gl, bh, 1); in dirent_del() 693 gfs2_trans_add_bh(ip->i_gl, bh, 1); in gfs2_init_dirent() [all …]
|
D | bmap.c | 96 gfs2_trans_add_bh(ip->i_gl, bh, 0); in gfs2_unstuffer_page() 156 gfs2_trans_add_bh(ip->i_gl, dibh, 1); in gfs2_unstuff_dinode() 471 gfs2_trans_add_bh(ip->i_gl, dibh, 1); in gfs2_bmap_alloc() 521 gfs2_indirect_init(mp, ip->i_gl, i, 0, bn++); in gfs2_bmap_alloc() 547 gfs2_trans_add_bh(ip->i_gl, mp->mp_bh[i-1], 1); in gfs2_bmap_alloc() 549 gfs2_indirect_init(mp, ip->i_gl, i, in gfs2_bmap_alloc() 559 gfs2_trans_add_bh(ip->i_gl, mp->mp_bh[end_of_metadata], 1); in gfs2_bmap_alloc() 796 gfs2_trans_add_bh(ip->i_gl, dibh, 1); in do_strip() 797 gfs2_trans_add_bh(ip->i_gl, bh, 1); in do_strip() 901 gfs2_metapath_ra(ip->i_gl, bh, top); in recursive_scan() [all …]
|
D | aops.c | 54 gfs2_trans_add_bh(ip->i_gl, bh, 0); in gfs2_page_add_databufs() 105 if (gfs2_assert_withdraw(sdp, gfs2_glock_is_held_excl(ip->i_gl))) in gfs2_writepage_common() 401 gfs2_log_flush(sdp, ip->i_gl); in gfs2_jdata_writepages() 499 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_readpage() 581 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_readpages() 623 gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &ip->i_gh); in gfs2_write_begin() 628 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, in gfs2_write_begin() 855 BUG_ON(gfs2_glock_is_locked_by_me(ip->i_gl) == NULL); in gfs2_write_end() 864 gfs2_trans_add_bh(ip->i_gl, dibh, 1); in gfs2_write_end() 925 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); in gfs2_bmap() [all …]
|
D | dentry.c | 64 had_lock = (gfs2_glock_is_locked_by_me(dip->i_gl) != NULL); in gfs2_drevalidate() 66 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); in gfs2_drevalidate()
|
D | meta_io.c | 358 bh = gfs2_getbuf(ip->i_gl, bstart, NO_CREATE); in gfs2_meta_wipe() 388 struct gfs2_glock *gl = ip->i_gl; in gfs2_meta_indirect_buffer() 395 gfs2_trans_add_bh(ip->i_gl, bh, 1); in gfs2_meta_indirect_buffer()
|
D | lops.c | 374 struct gfs2_glock *gl = ip->i_gl; in buf_lo_scan_elements() 423 gfs2_meta_sync(ip->i_gl); in buf_lo_after_scan() 429 gfs2_meta_sync(ip->i_gl); in buf_lo_after_scan() 745 struct gfs2_glock *gl = ip->i_gl; in databuf_lo_scan_elements() 796 gfs2_meta_sync(ip->i_gl); in databuf_lo_after_scan() 803 gfs2_meta_sync(ip->i_gl); in databuf_lo_after_scan()
|
D | quota.c | 323 error = gfs2_meta_read(ip->i_gl, bh_map.b_blocknr, DIO_WAIT, &bh); in bh_get() 585 gfs2_trans_add_bh(ip->i_gl, qd->qd_bh, 1); in do_qc() 721 gfs2_trans_add_bh(ip->i_gl, bh, 0); in gfs2_adjust_quota() 783 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); in do_sync() 841 gfs2_log_flush(ip->i_gl->gl_sbd, ip->i_gl); in do_sync() 892 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &i_gh); in do_glock() 1228 bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); in gfs2_quota_init() 1556 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); in gfs2_set_dqblk()
|
D | recovery.c | 36 struct gfs2_glock *gl = ip->i_gl; in gfs2_replay_read_block() 487 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, in gfs2_recover_func()
|
D | export.c | 118 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &gh); in gfs2_get_name()
|
D | trace_gfs2.h | 437 __entry->dev = ip->i_gl->gl_sbd->sd_vfs->s_dev; 474 __entry->dev = ip->i_gl->gl_sbd->sd_vfs->s_dev;
|
D | ops_fstype.c | 599 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, ji_gh); in gfs2_jindex_hold() 704 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, in init_journal() 868 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, in init_per_node() 876 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, in init_per_node()
|
D | incore.h | 329 struct gfs2_glock *i_gl; /* Move into i_gh? */ member
|
D | glops.c | 483 struct gfs2_glock *j_gl = ip->i_gl; in trans_go_xmote_bh()
|
D | rgrp.c | 694 struct gfs2_glock *gl = ip->i_gl; in gfs2_rindex_update() 1527 gfs2_trans_add_bh(ip->i_gl, dibh, 1); in gfs2_alloc_blocks()
|
D | glock.c | 1563 struct gfs2_glock *gl = ip->i_gl; in gfs2_glock_finish_truncate()
|