/linux-master/fs/gfs2/ |
H A D | inode.c | 136 &ip->i_gl); 167 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 174 gfs2_inode_already_deleted(ip->i_gl, no_formal_ino)) 185 set_bit(GLF_INSTANTIATE_NEEDED, &ip->i_gl->gl_flags); 192 glock_set_object(ip->i_gl, ip); 198 glock_clear_object(ip->i_gl, ip); 229 if (ip->i_gl) { 230 gfs2_glock_put(ip->i_gl); 231 ip->i_gl = NULL; 328 if (gfs2_glock_is_locked_by_me(dip->i_gl) [all...] |
H A D | acl.c | 70 if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { 71 int ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 129 if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { 130 ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
|
H A D | dentry.c | 63 had_lock = (gfs2_glock_is_locked_by_me(dip->i_gl) != NULL); 65 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh);
|
H A D | super.c | 132 struct gfs2_glock *j_gl = ip->i_gl; 190 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE, 229 gfs2_trans_add_meta(l_ip->i_gl, sdp->sd_sc_bh); 256 gfs2_trans_add_meta(l_ip->i_gl, sdp->sd_sc_bh); 257 gfs2_trans_add_meta(m_ip->i_gl, m_bh); 280 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE, 349 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &lfcc->gh); 447 struct address_space *metamapping = gfs2_glock2aspace(ip->i_gl); 453 gfs2_log_flush(GFS2_SB(inode), ip->i_gl, 496 if (unlikely(!ip->i_gl)) { [all...] |
H A D | xattr.c | 131 error = gfs2_meta_read(ip->i_gl, ip->i_eattr, DIO_WAIT, 0, &bh); 155 error = gfs2_meta_read(ip->i_gl, bn, DIO_WAIT, 0, &eabh); 278 gfs2_trans_add_meta(ip->i_gl, bh); 425 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh); 471 error = gfs2_meta_read(ip->i_gl, be64_to_cpu(*dataptrs), 0, 0, 504 gfs2_trans_add_meta(ip->i_gl, bh[x]); 613 if (!gfs2_glock_is_locked_by_me(ip->i_gl)) { 614 ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &gh); 646 *bhp = gfs2_meta_new(ip->i_gl, block); 647 gfs2_trans_add_meta(ip->i_gl, *bh [all...] |
H A D | util.c | 61 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_NOEXP | 123 struct gfs2_glock *i_gl; local 134 i_gl = ip->i_gl; 189 wait_on_bit(&i_gl->gl_flags, GLF_DEMOTE, TASK_UNINTERRUPTIBLE); 209 if (i_gl->gl_ops->go_free) { 210 set_bit(GLF_FREEING, &i_gl->gl_flags); 211 wait_on_bit(&i_gl->gl_flags, GLF_FREEING, TASK_UNINTERRUPTIBLE); 456 gfs2_dump_glock(NULL, ip->i_gl, 1);
|
H A D | file.c | 66 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, 112 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); 169 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); 228 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); 245 gfs2_log_flush(sdp, ip->i_gl, 264 gfs2_trans_add_meta(ip->i_gl, bh); 432 gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); 457 set_bit(GLF_DIRTY, &ip->i_gl->gl_flags); 558 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh); 598 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARE [all...] |
H A D | dir.c | 94 bh = gfs2_meta_new(ip->i_gl, block); 95 gfs2_trans_add_meta(ip->i_gl, bh); 108 error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, 0, &bh); 129 gfs2_trans_add_meta(ip->i_gl, dibh); 210 gfs2_trans_add_meta(ip->i_gl, bh); 232 gfs2_trans_add_meta(ip->i_gl, dibh); 301 bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); 303 error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, 0, &bh); 678 gfs2_trans_add_meta(dip->i_gl, bh); 717 gfs2_trans_add_meta(ip->i_gl, b [all...] |
H A D | meta_io.c | 437 if (!ip->i_gl) { 446 bh = gfs2_getbuf(ip->i_gl, bstart, NO_CREATE); 481 struct gfs2_glock *gl = ip->i_gl;
|
H A D | bmap.c | 81 gfs2_trans_add_data(ip->i_gl, bh); 127 gfs2_trans_add_meta(ip->i_gl, dibh); 680 gfs2_trans_add_meta(ip->i_gl, dibh); 724 gfs2_indirect_init(mp, ip->i_gl, i, 0, bn++); 750 gfs2_trans_add_meta(ip->i_gl, mp->mp_bh[i-1]); 752 gfs2_indirect_init(mp, ip->i_gl, i, 762 gfs2_trans_add_meta(ip->i_gl, mp->mp_bh[end_of_metadata]); 1183 set_bit(GLF_DIRTY, &ip->i_gl->gl_flags); 1384 gfs2_trans_add_meta(ip->i_gl, dibh); 1556 gfs2_trans_add_meta(ip->i_gl, b [all...] |
H A D | aops.c | 57 gfs2_trans_add_data(ip->i_gl, bh); 158 if (gfs2_assert_withdraw(sdp, ip->i_gl->gl_state == LM_ST_EXCLUSIVE)) 398 gfs2_log_flush(sdp, ip->i_gl, GFS2_LOG_HEAD_FLUSH_NORMAL | 596 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh);
|
H A D | lops.c | 784 struct gfs2_glock *gl = ip->i_gl; 839 gfs2_inode_metasync(ip->i_gl); 845 gfs2_inode_metasync(ip->i_gl); 1008 struct gfs2_glock *gl = ip->i_gl; 1059 gfs2_inode_metasync(ip->i_gl); 1066 gfs2_inode_metasync(ip->i_gl);
|
H A D | quota.c | 426 error = gfs2_meta_read(ip->i_gl, iomap.addr >> inode->i_blkbits, 721 gfs2_trans_add_meta(ip->i_gl, qd->qd_bh); 800 gfs2_trans_add_data(ip->i_gl, bh); 952 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh); 1010 gfs2_log_flush(ip->i_gl->gl_name.ln_sbd, ip->i_gl, 1070 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &i_gh); 1440 bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); 1728 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh);
|
H A D | recovery.c | 36 struct gfs2_glock *gl = ip->i_gl; 351 gfs2_inode_metasync(ip->i_gl); 445 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED,
|
H A D | export.c | 112 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &gh);
|
H A D | trace_gfs2.h | 455 __entry->dev = ip->i_gl->gl_name.ln_sbd->sd_vfs->s_dev; 491 __entry->dev = ip->i_gl->gl_name.ln_sbd->sd_vfs->s_dev; 523 __entry->dev = ip->i_gl->gl_name.ln_sbd->sd_vfs->s_dev;
|
H A D | ops_fstype.c | 580 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, ji_gh); 694 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_NOPID, 789 sdp->sd_jinode_gl = ip->i_gl; 790 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 962 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_NOPID,
|
H A D | incore.h | 391 struct gfs2_glock *i_gl; member in struct:gfs2_inode
|
H A D | glops.c | 598 struct gfs2_glock *j_gl = ip->i_gl;
|
H A D | glock.c | 673 if (gl == m_ip->i_gl) 987 gfs2_glock_poke(ip->i_gl);
|
H A D | rgrp.c | 1037 struct gfs2_glock *gl = ip->i_gl; 2458 gfs2_trans_add_meta(ip->i_gl, dibh);
|